DOE Office of Scientific and Technical Information (OSTI.GOV)
Hachtel, A. J.; Gillette, M. C.; Clements, E. R.
A novel home-built system for imaging cold atom samples is presented using a readily available astronomy camera which has the requisite sensitivity but no timing-control. We integrate the camera with LabVIEW achieving fast, low-jitter imaging with a convenient user-defined interface. We show that our system takes precisely timed millisecond exposures and offers significant improvements in terms of system jitter and readout time over previously reported home-built systems. Our system rivals current commercial “black box” systems in performance and user-friendliness.
Wang, Ximing; Liu, Brent J; Martinez, Clarisa; Zhang, Xuejun; Winstein, Carolee J
2015-01-01
Imaging based clinical trials can benefit from a solution to efficiently collect, analyze, and distribute multimedia data at various stages within the workflow. Currently, the data management needs of these trials are typically addressed with custom-built systems. However, software development of the custom- built systems for versatile workflows can be resource-consuming. To address these challenges, we present a system with a workflow engine for imaging based clinical trials. The system enables a project coordinator to build a data collection and management system specifically related to study protocol workflow without programming. Web Access to DICOM Objects (WADO) module with novel features is integrated to further facilitate imaging related study. The system was initially evaluated by an imaging based rehabilitation clinical trial. The evaluation shows that the cost of the development of system can be much reduced compared to the custom-built system. By providing a solution to customize a system and automate the workflow, the system will save on development time and reduce errors especially for imaging clinical trials. PMID:25870169
Imaging spectrometry of the Earth and other solar system bodies
NASA Technical Reports Server (NTRS)
Vane, Gregg
1993-01-01
Imaging spectrometry is a relatively new tool for remote sensing of the Earth and other bodies of the solar system. The technique dates back to the late 1970's and early 1980's. It is a natural extension of the earlier multi-spectral imagers developed for remote sensing that acquire images in a few, usually broad spectral bands. Imaging spectrometers combine aspects of classical spectrometers and imaging systems, making it possible to acquire literally hundreds of images of an object, each image in a separate, narrow spectral band. It is thus possible to perform spectroscopy on a pixel-by-pixel basis with the data acquired with an imaging spectrometer. Two imaging spectrometers have flown in space and several others are planned for future Earth and planetary missions. The French-built Phobos Infrared Spectrometer (ISM) was part of the payload of the Soviet Mars mission in 1988, and the JPL-built Near Infrared Mapping Spectrometer (NIMS) is currently en route to Jupiter aboard the Galileo spacecraft. Several airborne imaging spectrometers have been built in the past decade including the JPL-built Airborne Visible/Infrared Imaging Spectrometer (AVIRIS) which is the only such sensor that covers the full solar reflected portion of the spectrum in narrow, contiguous spectral bands. NASA plans two imaging spectrometers for its Earth Observing System, the Moderate and the High Resolution Imaging Spectrometers (MODIS and HIRIS). A brief overview of the applications of imaging spectrometry to Earth science will be presented to illustrate the value of the tool to remote sensing and indicate the types of measurements that are required. The system design for AVIRS and a planetary imaging spectrometer will be presented to illustrate the engineering considerations and challenges that must be met in building such instruments. Several key sensor technology areas will be discussed in which miniaturization and/or enhanced performance through micromachining and nanofabrication may allow smaller, more robust, and more capable imaging spectrometers to be built in the future.
NASA Astrophysics Data System (ADS)
Lefcourt, Alan M.; Kistler, Ross; Gadsden, S. Andrew
2016-05-01
The goal of this project was to construct a cart and a mounting system that would allow a hyperspectral laser-induced fluorescence imaging system (HLIFIS) to be used to detect fecal material in produce fields. Fecal contaminated produce is a recognized food safety risk. Previous research demonstrated the HLIFIS could detect fecal contamination in a laboratory setting. A cart was designed and built, and then tested to demonstrate that the cart was capable of moving at constant speeds or at precise intervals. A mounting system was designed and built to facilitate the critical alignment of the camera's imaging and the laser's illumination fields, and to allow the HLIFIS to be used in both field and laboratory settings without changing alignments. A hardened mount for the Powell lens that is used to produce the appropriate illumination profile was also designed, built, and tested.
A Multimedia Database Management System Supporting Contents Search in Media Data
1989-03-01
predicates: car (x), manufacturer (x, Horch ), year-built (x, 1922) to designate that the car was manufactured by Horch in the year 1922. The use of the name...manufacturer (imagel, x, " Horch "). year-built (imagel, x, 1922). query(I) :- car (I, Z), manufacturer (I, Z, " Horch "). ?- query(Image). yes. Image = image... Horch ) : imageJ 1, imageJ2, etc. In certain cases a predicate may generate entries in more than one list. For example, if the predicate of action
Initial clinical testing of a multi-spectral imaging system built on a smartphone platform
NASA Astrophysics Data System (ADS)
Mink, Jonah W.; Wexler, Shraga; Bolton, Frank J.; Hummel, Charles; Kahn, Bruce S.; Levitz, David
2016-03-01
Multi-spectral imaging systems are often expensive and bulky. An innovative multi-spectral imaging system was fitted onto a mobile colposcope, an imaging system built around a smartphone in order to image the uterine cervix from outside the body. The multi-spectral mobile colposcope (MSMC) acquires images at different wavelengths. This paper presents the clinical testing of MSMC imaging (technical validation of the MSMC system is described elsewhere 1 ). Patients who were referred to colposcopy following abnormal screening test (Pap or HPV DNA test) according to the standard of care were enrolled. Multi-spectral image sets of the cervix were acquired, consisting of images from the various wavelengths. Image acquisition took 1-2 sec. Areas suspected for dysplasia under white light imaging were biopsied, according to the standard of care. Biopsied sites were recorded on a clockface map of the cervix. Following the procedure, MSMC data was processed from the sites of biopsied sites. To date, the initial histopathological results are still outstanding. Qualitatively, structures in the cervical images were sharper at lower wavelengths than higher wavelengths. Patients tolerated imaging well. The result suggests MSMC holds promise for cervical imaging.
Wavelength Coded Image Transmission and Holographic Optical Elements.
1984-08-20
system has been designed and built for transmitting images of diffusely reflecting objects through optical fibers and displaying those images at a...passive components at the end of a fiber-optic designed to transmit high-resolution images of diffusely imaging system as described in this paper... designing a system for viewing diffusely reflecting The authors are with University of Minnesota. Electrical Engi- objects, one must consider that a
Developing an interactive teleradiology system for SARS diagnosis
NASA Astrophysics Data System (ADS)
Sun, Jianyong; Zhang, Jianguo; Zhuang, Jun; Chen, Xiaomeng; Yong, Yuanyuan; Tan, Yongqiang; Chen, Liu; Lian, Ping; Meng, Lili; Huang, H. K.
2004-04-01
Severe acute respiratory syndrome (SARS) is a respiratory illness that had been reported in Asia, North America, and Europe in last spring. Most of the China cases of SARS have occurred by infection in hospitals or among travelers. To protect the physicians, experts and nurses from the SARS during the diagnosis and treatment procedures, the infection control mechanisms were built in SARS hospitals. We built a Web-based interactive teleradiology system to assist the radiologists and physicians both in side and out side control area to make image diagnosis. The system consists of three major components: DICOM gateway (GW), Web-based image repository server (Server), and Web-based DICOM viewer (Viewer). This system was installed and integrated with CR, CT and the hospital information system (HIS) in Shanghai Xinhua hospital to provide image-based ePR functions for SARS consultation between the radiologists, physicians and experts inside and out side control area. The both users inside and out side the control area can use the system to process and manipulate the DICOM images interactively, and the system provide the remote control mechanism to synchronize their operations on images and display.
Designing Ratchets in Ultra-cold Atoms for the Advanced Undergraduate Laboratory
NASA Astrophysics Data System (ADS)
Hachtel, Andrew; Gillette, Matthew; Clements, Ethan; Zhong, Shan; Ducay, Rey; Bali, Samir
2014-05-01
We propose to perform ratchet experiments in cold Rubidium atoms using state-of-the-art home-built tapered amplifier and imaging systems. Our tapered amplifier system amplifies the output from home-built external cavity tunable diode lasers up to a factor 100 and costs less than 5,000, in contrast to commercial tapered amplifier systems, which cost upward of 20,000. We have developed an imaging system with LabVIEW integration, which allows for approximately 2 millisecond exposures and microsecond control of experimental parameters. Our imaging system also costs less than 5,000 in comparison to commercial options, which cost between 40-50,000. Progress toward implementation of a one-dimensional rocking ratchet is described. We gratefully acknowledge funding from the American Chemical Society Petroleum Research Fund and Miami University. We also acknowledge the Miami University Instrumentation Laboratory for their invaluable contributions.
Aerial imaging with manned aircraft for precision agriculture
USDA-ARS?s Scientific Manuscript database
Over the last two decades, numerous commercial and custom-built airborne imaging systems have been developed and deployed for diverse remote sensing applications, including precision agriculture. More recently, unmanned aircraft systems (UAS) have emerged as a versatile and cost-effective platform f...
Optical design and system characterization of an imaging microscope at 121.6 nm
NASA Astrophysics Data System (ADS)
Gao, Weichuan; Finan, Emily; Kim, Geon-Hee; Kim, Youngsik; Milster, Thomas D.
2018-03-01
We present the optical design and system characterization of an imaging microscope prototype at 121.6 nm. System engineering processes are demonstrated through the construction of a Schwarzschild microscope objective, including tolerance analysis, fabrication, alignment, and testing. Further improvements on the as-built system with a correction phase plate are proposed and analyzed. Finally, the microscope assembly and the imaging properties of the prototype are demonstrated.
Hyperspectral Imaging of Functional Patterns for Disease Assessment and Treatment Monitoring
DOE Office of Scientific and Technical Information (OSTI.GOV)
Demos, S; Hattery, D; Hassan, M
2003-12-05
We have designed and built a six-band multi-spectral NIR imaging system used in clinical testing on cancer patients. From our layered tissue model, we create blood volume and blood oxygenation images for patient treatment monitoring.
Imaging spectrometer using a liquid crystal tunable filter
NASA Astrophysics Data System (ADS)
Chrien, Thomas G.; Chovit, Christopher; Miller, Peter J.
1993-09-01
A demonstration imaging spectrometer using a liquid crystal tunable filter (LCTF) was built and tested on a hot air balloon platform. The LCTF is a tunable polarization interference or Lyot filter. The LCTF enables a small, light weight, low power, band sequential imaging spectrometer design. An overview of the prototype system is given along with a description of balloon experiment results. System model performance predictions are given for a future LCTF based imaging spectrometer design. System design considerations of LCTF imaging spectrometers are discussed.
An Insect Eye Inspired Miniaturized Multi-Camera System for Endoscopic Imaging.
Cogal, Omer; Leblebici, Yusuf
2017-02-01
In this work, we present a miniaturized high definition vision system inspired by insect eyes, with a distributed illumination method, which can work in dark environments for proximity imaging applications such as endoscopy. Our approach is based on modeling biological systems with off-the-shelf miniaturized cameras combined with digital circuit design for real time image processing. We built a 5 mm radius hemispherical compound eye, imaging a 180 ° ×180 ° degrees field of view while providing more than 1.1 megapixels (emulated ommatidias) as real-time video with an inter-ommatidial angle ∆ϕ = 0.5 ° at 18 mm radial distance. We made an FPGA implementation of the image processing system which is capable of generating 25 fps video with 1080 × 1080 pixel resolution at a 120 MHz processing clock frequency. When compared to similar size insect eye mimicking systems in literature, the system proposed in this paper features 1000 × resolution increase. To the best of our knowledge, this is the first time that a compound eye with built-in illumination idea is reported. We are offering our miniaturized imaging system for endoscopic applications like colonoscopy or laparoscopic surgery where there is a need for large field of view high definition imagery. For that purpose we tested our system inside a human colon model. We also present the resulting images and videos from the human colon model in this paper.
NASA Astrophysics Data System (ADS)
Wang, Ximing; Martinez, Clarisa; Wang, Jing; Liu, Ye; Liu, Brent
2014-03-01
Clinical trials usually have a demand to collect, track and analyze multimedia data according to the workflow. Currently, the clinical trial data management requirements are normally addressed with custom-built systems. Challenges occur in the workflow design within different trials. The traditional pre-defined custom-built system is usually limited to a specific clinical trial and normally requires time-consuming and resource-intensive software development. To provide a solution, we present a user customizable imaging informatics-based intelligent workflow engine system for managing stroke rehabilitation clinical trials with intelligent workflow. The intelligent workflow engine provides flexibility in building and tailoring the workflow in various stages of clinical trials. By providing a solution to tailor and automate the workflow, the system will save time and reduce errors for clinical trials. Although our system is designed for clinical trials for rehabilitation, it may be extended to other imaging based clinical trials as well.
Automated cart with VIS/NIR hyperspectral reflectance and fluorescence imaging capabilities
USDA-ARS?s Scientific Manuscript database
A system to take high-resolution VIS/NIR hyperspectral reflectance and fluorescence images in outdoor fields using ambient lighting or a pulsed laser (355 nm), respectively, for illumination was designed, built, and tested. Components of the system include a semi-autonomous cart, a gated-intensified...
Chosen results of field tests of synthetic aperture radar system installed on board UAV
NASA Astrophysics Data System (ADS)
Kaniewski, Piotr; Komorniczak, Wojciech; Lesnik, Czeslaw; Cyrek, Jacek; Serafin, Piotr; Labowski, Michal; Wajszczyk, Bronislaw
2017-04-01
The paper presents a synthetic information on a UAV-based radar terrain imaging system, its purpose, structure and working principle as well as terrain images obtained from flight experiments. A SAR technology demonstrator has been built as a result of a research project conducted by the Military University of Technology and WB Electronics S.A. under the name WATSAR. The developed system allows to obtain high resolution radar images, both in on-line and off-line modes, independently of the light conditions over the observed area. The software developed for the system allows to determine geographic coordinates of the imaged objects with high accuracy. Four LFM-CW radar sensors were built during the project: two for S band and two for Ku band, working with different signal bandwidths. Acquired signals were processed with the TDC algorithm, which allowed for a number of analyses in order to evaluate the performance of the system. The impact of the navigational corrections on a SAR image quality was assessed as well. The research methodology of the in-flight experiments of the system is presented in the paper. The projects results show that the developed system may be implemented as an aid to tactical C4ISR systems.
Technical Note: Experimental results from a prototype high-field inline MRI-linac
DOE Office of Scientific and Technical Information (OSTI.GOV)
Liney, G. P., E-mail: gary.liney@sswahs.nsw.gov.au
Purpose: The pursuit of real-time image guided radiotherapy using optimal tissue contrast has seen the development of several hybrid magnetic resonance imaging (MRI)-treatment systems, high field and low field, and inline and perpendicular configurations. As part of a new MRI-linac program, an MRI scanner was integrated with a linear accelerator to enable investigations of a coupled inline MRI-linac system. This work describes results from a prototype experimental system to demonstrate the feasibility of a high field inline MR-linac. Methods: The magnet is a 1.5 T MRI system (Sonata, Siemens Healthcare) was located in a purpose built radiofrequency (RF) cage enablingmore » shielding from and close proximity to a linear accelerator with inline (and future perpendicular) orientation. A portable linear accelerator (Linatron, Varian) was installed together with a multileaf collimator (Millennium, Varian) to provide dynamic field collimation and the whole assembly built onto a stainless-steel rail system. A series of MRI-linac experiments was performed to investigate (1) image quality with beam on measured using a macropodine (kangaroo) ex vivo phantom; (2) the noise as a function of beam state measured using a 6-channel surface coil array; and (3) electron contamination effects measured using Gafchromic film and an electronic portal imaging device (EPID). Results: (1) Image quality was unaffected by the radiation beam with the macropodine phantom image with the beam on being almost identical to the image with the beam off. (2) Noise measured with a surface RF coil produced a 25% elevation of background intensity when the radiation beam was on. (3) Film and EPID measurements demonstrated electron focusing occurring along the centerline of the magnet axis. Conclusions: A proof-of-concept high-field MRI-linac has been built and experimentally characterized. This system has allowed us to establish the efficacy of a high field inline MRI-linac and study a number of the technical challenges and solutions.« less
Polarimetry diagnostic on OMEGA EP using a 10-ps, 263-nm probe beam.
Davies, A; Haberberger, D; Boni, R; Ivancic, S; Brown, R; Froula, D H
2014-11-01
A polarimetry diagnostic was built and characterized for magnetic-field measurements in laser-plasma experiments on the OMEGA EP laser. This diagnostic was built into the existing 4ω (263-nm) probe system that employs a 10-ps laser pulse collected with an f/4 imaging system. The diagnostic measures the rotation of the probe beam's polarization. The polarimeter uses a Wollaston prism to split the probe beam into orthogonal polarization components. Spatially localized intensity variations between images indicate polarization rotation. Magnetic fields can be calculated by combining the polarimetry data with the measured plasma density profile obtained from angular filter refractometry.
Land cover classification of Landsat 8 satellite data based on Fuzzy Logic approach
NASA Astrophysics Data System (ADS)
Taufik, Afirah; Sakinah Syed Ahmad, Sharifah
2016-06-01
The aim of this paper is to propose a method to classify the land covers of a satellite image based on fuzzy rule-based system approach. The study uses bands in Landsat 8 and other indices, such as Normalized Difference Water Index (NDWI), Normalized difference built-up index (NDBI) and Normalized Difference Vegetation Index (NDVI) as input for the fuzzy inference system. The selected three indices represent our main three classes called water, built- up land, and vegetation. The combination of the original multispectral bands and selected indices provide more information about the image. The parameter selection of fuzzy membership is performed by using a supervised method known as ANFIS (Adaptive neuro fuzzy inference system) training. The fuzzy system is tested for the classification on the land cover image that covers Klang Valley area. The results showed that the fuzzy system approach is effective and can be explored and implemented for other areas of Landsat data.
Evaluation of a new breast-shaped compensation filter for a newly built breast imaging system
NASA Astrophysics Data System (ADS)
Cai, Weixing; Ning, Ruola; Zhang, Yan; Conover, David
2007-03-01
A new breast-shaped compensation filter has been designed and fabricated for breast imaging using our newly built breast imaging (CBCTBI) system, which is able to scan an uncompressed breast with pendant geometry. The shape of this compensation filter is designed based on an average-sized breast phantom. Unlike conventional bow-tie compensation filters, its cross-sectional profile varies along the chest wall-to-nipple direction for better compensation for the shape of a breast. Breast phantoms of three different sizes are used to evaluate the performance of this compensation filter. The reconstruction image quality was studied and compared to that obtained without the compensation filter in place. The uniformity of linear attenuation coefficient and the uniformity of noise distribution are significantly improved, and the contrast-to-noise ratios (CNR) of small lesions near the chest wall are increased as well. Multi-normal image method is used in the reconstruction process to correct compensation flood field and to reduce ring artifacts.
NASA Technical Reports Server (NTRS)
Masuoka, E.; Rose, J.; Quattromani, M.
1981-01-01
Recent developments related to microprocessor-based personal computers have made low-cost digital image processing systems a reality. Image analysis systems built around these microcomputers provide color image displays for images as large as 256 by 240 pixels in sixteen colors. Descriptive statistics can be computed for portions of an image, and supervised image classification can be obtained. The systems support Basic, Fortran, Pascal, and assembler language. A description is provided of a system which is representative of the new microprocessor-based image processing systems currently on the market. While small systems may never be truly independent of larger mainframes, because they lack 9-track tape drives, the independent processing power of the microcomputers will help alleviate some of the turn-around time problems associated with image analysis and display on the larger multiuser systems.
TH-AB-209-07: High Resolution X-Ray-Induced Acoustic Computed Tomography
DOE Office of Scientific and Technical Information (OSTI.GOV)
Xiang, L; Tang, S; Ahmad, M
Purpose: X-ray radiographic absorption imaging is an invaluable tool in medical diagnostics, biology and materials science. However, the use of conventional CT is limited by two factors: the detection sensitivity to weak absorption material and the radiation dose from CT scanning. The purpose of this study is to explore X-ray induced acoustic computed tomography (XACT), a new imaging modality, which combines X-ray absorption contrast and high ultrasonic resolution to address these challenges. Methods: First, theoretical models was built to analyze the XACT sensitivity to X-ray absorption and calculate the minimal radiation dose in XACT imaging. Then, an XACT system comprisedmore » of an ultrashort X-ray pulse, a low noise ultrasound detector and a signal acquisition system was built to evaluate the X-ray induced acoustic signal generation. A piece of chicken bone and a phantom with two golden fiducial markers were exposed to 270 kVp X-ray source with 60 ns exposure time, and the X-ray induced acoustic signal was received by a 2.25MHz ultrasound transducer in 200 positions. XACT images were reconstructed by a filtered back-projection algorithm. Results: The theoretical analysis shows that X-ray induced acoustic signals have 100% relative sensitivity to X-ray absorption, but not to X-ray scattering. Applying this innovative technology to breast imaging, we can reduce radiation dose by a factor of 50 compared with newly FDA approved breast CT. The reconstructed images of chicken bone and golden fiducial marker phantom reveal that the spatial resolution of the built XACT system is 350µm. Conclusion: In XACT, the imaging sensitivity to X-ray absorption is improved and the imaging dose is dramatically reduced by using ultrashort pulsed X-ray. Taking advantage of the high ultrasonic resolution, we can also perform 3D imaging with a single X-ray pulse. This new modality has the potential to revolutionize x-ray imaging applications in medicine and biology.« less
High Sensitivity SPECT for Small Animals and Plants
DOE Office of Scientific and Technical Information (OSTI.GOV)
Mitchell, Gregory S.
Imaging systems using single gamma-ray emitting radioisotopes typically implement collimators in order to form the images. However, a tradeoff in sensitivity is inherent in the use of collimators, and modern preclinical single-photon emission computed tomography (SPECT) systems detect a very small fraction of emitted gamma-rays (<0.3%). We have built a collimator-less system, which can reach sensitivity of 40% for 99mTc imaging, while still producing images of sufficient spatial resolution for certain applications in thin objects such as mice, small plants, and well plates used for in vitro experiments.
Remote media vision-based computer input device
NASA Astrophysics Data System (ADS)
Arabnia, Hamid R.; Chen, Ching-Yi
1991-11-01
In this paper, we introduce a vision-based computer input device which has been built at the University of Georgia. The user of this system gives commands to the computer without touching any physical device. The system receives input through a CCD camera; it is PC- based and is built on top of the DOS operating system. The major components of the input device are: a monitor, an image capturing board, a CCD camera, and some software (developed by use). These are interfaced with a standard PC running under the DOS operating system.
Image Analysis via Fuzzy-Reasoning Approach: Prototype Applications at NASA
NASA Technical Reports Server (NTRS)
Dominguez, Jesus A.; Klinko, Steven J.
2004-01-01
A set of imaging techniques based on Fuzzy Reasoning (FR) approach was built for NASA at Kennedy Space Center (KSC) to perform complex real-time visual-related safety prototype tasks, such as detection and tracking of moving Foreign Objects Debris (FOD) during the NASA Space Shuttle liftoff and visual anomaly detection on slidewires used in the emergency egress system for Space Shuttle at the launch pad. The system has also proved its prospective in enhancing X-ray images used to screen hard-covered items leading to a better visualization. The system capability was used as well during the imaging analysis of the Space Shuttle Columbia accident. These FR-based imaging techniques include novel proprietary adaptive image segmentation, image edge extraction, and image enhancement. Probabilistic Neural Network (PNN) scheme available from NeuroShell(TM) Classifier and optimized via Genetic Algorithm (GA) was also used along with this set of novel imaging techniques to add powerful learning and image classification capabilities. Prototype applications built using these techniques have received NASA Space Awards, including a Board Action Award, and are currently being filed for patents by NASA; they are being offered for commercialization through the Research Triangle Institute (RTI), an internationally recognized corporation in scientific research and technology development. Companies from different fields, including security, medical, text digitalization, and aerospace, are currently in the process of licensing these technologies from NASA.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Chan, James H. H.; Suyu, Sherry H.; Chiueh, Tzihong
Strong gravitationally lensed quasars provide powerful means to study galaxy evolution and cosmology. Current and upcoming imaging surveys will contain thousands of new lensed quasars, augmenting the existing sample by at least two orders of magnitude. To find such lens systems, we built a robot, Chitah, that hunts for lensed quasars by modeling the configuration of the multiple quasar images. Specifically, given an image of an object that might be a lensed quasar, Chitah first disentangles the light from the supposed lens galaxy and the light from the multiple quasar images based on color information. A simple rule is designed to categorize the given object as a potential four-image (quad) or two-image (double) lensed quasar system. The configuration of the identified quasar images is subsequently modeled to classify whether the object is a lensed quasar system. We test the performance of Chitah using simulated lens systems based on the Canada–France–Hawaii Telescope Legacy Survey. For bright quads with large image separations (with Einstein radiusmore » $${r}_{\\mathrm{ein}}\\gt 1\\buildrel{\\prime\\prime}\\over{.} 1$$) simulated using Gaussian point-spread functions, a high true-positive rate (TPR) of $$\\sim 90\\%$$ and a low false-positive rate of $$\\sim 3\\%$$ show that this is a promising approach to search for new lens systems. We obtain high TPR for lens systems with $${r}_{\\mathrm{ein}}\\gtrsim 0\\buildrel{\\prime\\prime}\\over{.} 5$$, so the performance of Chitah is set by the seeing. We further feed a known gravitational lens system, COSMOS 5921+0638, to Chitah, and demonstrate that Chitah is able to classify this real gravitational lens system successfully. Our newly built Chitah is omnivorous and can hunt in any ground-based imaging surveys.« less
Setoi, Ayana; Kose, Katsumi
2018-05-16
We developed ultrashort echo-time (UTE) imaging sequences with 3D Cones trajectories for a home-built compact MRI system using a 1.5T superconducting magnet and an unshielded gradient coil set. We achieved less than 7 min imaging time and obtained clear in vivo images of a human forearm with a TE of 0.4 ms. We concluded that UTE imaging using 3D Cones acquisition was successfully implemented in our 1.5T MRI system.
Solid state image sensing arrays
NASA Technical Reports Server (NTRS)
Sadasiv, G.
1972-01-01
The fabrication of a photodiode transistor image sensor array in silicon, and tests on individual elements of the array are described along with design for a scanning system for an image sensor array. The spectral response of p-n junctions was used as a technique for studying the optical-absorption edge in silicon. Heterojunction structures of Sb2S3- Si were fabricated and a system for measuring C-V curves on MOS structures was built.
Multi-spectral confocal microendoscope for in-vivo imaging
NASA Astrophysics Data System (ADS)
Rouse, Andrew Robert
The concept of in-vivo multi-spectral confocal microscopy is introduced. A slit-scanning multi-spectral confocal microendoscope (MCME) was built to demonstrate the technique. The MCME employs a flexible fiber-optic catheter coupled to a custom built slit-scan confocal microscope fitted with a custom built imaging spectrometer. The catheter consists of a fiber-optic imaging bundle linked to a miniature objective and focus assembly. The design and performance of the miniature objective and focus assembly are discussed. The 3mm diameter catheter may be used on its own or routed though the instrument channel of a commercial endoscope. The confocal nature of the system provides optical sectioning with 3mum lateral resolution and 30mum axial resolution. The prism based multi-spectral detection assembly is typically configured to collect 30 spectral samples over the visible chromatic range. The spectral sampling rate varies from 4nm/pixel at 490nm to 8nm/pixel at 660nm and the minimum resolvable wavelength difference varies from 7nm to 18nm over the same spectral range. Each of these characteristics are primarily dictated by the dispersive power of the prism. The MCME is designed to examine cellular structures during optical biopsy and to exploit the diagnostic information contained within the spectral domain. The primary applications for the system include diagnosis of disease in the gastro-intestinal tract and female reproductive system. Recent data from the grayscale imaging mode are presented. Preliminary multi-spectral results from phantoms, cell cultures, and excised human tissue are presented to demonstrate the potential of in-vivo multi-spectral imaging.
Names-to-Mars Chip for InSight Spacecraft
2015-12-17
The dime-size microchip in this close-up image carries 826,923 names that will go to Mars on NASA InSight lander. The image was taken in November 2015 inside a clean room at Lockheed Martin Space Systems, Denver, where the lander was built.
ERIC Educational Resources Information Center
Filinger, Ronald H.; Hall, Paul W.
Because large scale individualized learning systems place excessive demands on conventional means of producing audiovisual software, electronic image generation has been investigated as an alternative. A prototype, experimental device, Scanimate-500, was designed and built by the Computer Image Corporation. It uses photographic, television, and…
Model-based inspection of multipackage food products using a twin-beam x-ray system
NASA Astrophysics Data System (ADS)
Palmer, Stephen C.; Batchelor, Bruce G.
1998-10-01
A twin-orthogonal-fanbeam x-ray system has been built as part of a six-partner project funded by the Commission of the European Union. The images created by this system represent plan and side views of the object to be inspected. Using such a system, it is possible to locate a point-like feature that creates a significant shadow in both beams, in a 3D space. However, the real value of such a system lies in the fact that it is often possible to see a foreign body, such as a small piece of loose glass, within a jar using one beam, when the same contaminant is invisible to the other beam. Such a situation typically arises when the foreign body is obscured by the x-ray shadow of the neck-shoulder region of a jar. The x-ray system built by our colleagues in this consortium is being used to examine, simultaneously, six jars of semi-fluid savory sauce, held together by shrink-wrapping on a cardboard tray. The inspection algorithm consists of fitting multi-part models of the image intensity function to both the plan and side-view images. Once a model has been fitted, it is possible to use image comparison, in order to highlight any foreign bodies. The pre-processed plan and side-view images are analyzed and correlated together, so that in many cases, a foreign body whose view is obscured in one image can be detected in the other.
First Results from the Wide Angle Camera of the ROSETTA Mission .
NASA Astrophysics Data System (ADS)
Barbieri, C.; Fornasier, S.; Bertini, I.; Angrilli, F.; Bianchini, G. A.; Debei, S.; De Cecco, M.; Parzianello, G.; Zaccariotto, M.; Da Deppo, V.; Naletto, G.
This paper gives a brief description of the Wide Angle Camera (WAC), built by the Center of Studies and Activities for Space (CISAS) of the University of Padova for the ESA ROSETTA Mission, of data we have obtained about the new mission targets, and of the first results achieved after the launch in March 2004. The WAC is part of the OSIRIS imaging system, built under the PI-ship of Dr. U. Keller (Max-Planck-Institute for Solar System Studies) which comprises also a Narrow Angle Camera (NAC) built by the Laboratoire d'Astrophysique Spatiale (LAS) of Marseille. CISAS had also the responsibility to build the shutter and the front door mechanism for the NAC. The images show the excellent optical quality of the WAC, exceeding the specifications both in term of encircled energy (80% in one pixel over a FoV of 12×12 sq degree), limiting magnitude (fainter than the 13th in 30s exposure time through a wideband red filter) and amount of distortions.
Design and calibration of a vacuum compatible scanning tunneling microscope
NASA Technical Reports Server (NTRS)
Abel, Phillip B.
1990-01-01
A vacuum compatible scanning tunneling microscope was designed and built, capable of imaging solid surfaces with atomic resolution. The single piezoelectric tube design is compact, and makes use of sample mounting stubs standard to a commercially available surface analysis system. Image collection and display is computer controlled, allowing storage of images for further analysis. Calibration results from atomic scale images are presented.
Dates fruits classification using SVM
NASA Astrophysics Data System (ADS)
Alzu'bi, Reem; Anushya, A.; Hamed, Ebtisam; Al Sha'ar, Eng. Abdelnour; Vincy, B. S. Angela
2018-04-01
In this paper, we used SVM in classifying various types of dates using their images. Dates have interesting different characteristics that can be valuable to distinguish and determine a particular date type. These characteristics include shape, texture, and color. A system that achieves 100% accuracy was built to classify the dates which can be eatable and cannot be eatable. The built system helps the food industry and customer in classifying dates depending on specific quality measures giving best performance with specific type of dates.
ERIC Educational Resources Information Center
Mathematics Teacher, 2004
2004-01-01
Some inexpensive or free ways that enable to capture and use images in work are mentioned. The first tip demonstrates the methods of using some of the built-in capabilities of the Macintosh and Windows-based PC operating systems, and the second tip describes methods to capture and create images using SnagIt.
Image Processor Electronics (IPE): The High-Performance Computing System for NASA SWIFT Mission
NASA Technical Reports Server (NTRS)
Nguyen, Quang H.; Settles, Beverly A.
2003-01-01
Gamma Ray Bursts (GRBs) are believed to be the most powerful explosions that have occurred in the Universe since the Big Bang and are a mystery to the scientific community. Swift, a NASA mission that includes international participation, was designed and built in preparation for a 2003 launch to help to determine the origin of Gamma Ray Bursts. Locating the position in the sky where a burst originates requires intensive computing, because the duration of a GRB can range between a few milliseconds up to approximately a minute. The instrument data system must constantly accept multiple images representing large regions of the sky that are generated by sixteen gamma ray detectors operating in parallel. It then must process the received images very quickly in order to determine the existence of possible gamma ray bursts and their locations. The high-performance instrument data computing system that accomplishes this is called the Image Processor Electronics (IPE). The IPE was designed, built and tested by NASA Goddard Space Flight Center (GSFC) in order to meet these challenging requirements. The IPE is a small size, low power and high performing computing system for space applications. This paper addresses the system implementation and the system hardware architecture of the IPE. The paper concludes with the IPE system performance that was measured during end-to-end system testing.
NASA Astrophysics Data System (ADS)
Bognot, J. R.; Candido, C. G.; Blanco, A. C.; Montelibano, J. R. Y.
2018-05-01
Monitoring the progress of building's construction is critical in construction management. However, measuring the building construction's progress are still manual, time consuming, error prone, and impose tedious process of analysis leading to delays, additional costings and effort. The main goal of this research is to develop a methodology for building construction progress monitoring based on 3D as-built model of the building from unmanned aerial system (UAS) images, 4D as-planned model (with construction schedule integrated) and, GIS analysis. Monitoring was done by capturing videos of the building with a camera-equipped UAS. Still images were extracted, filtered, bundle-adjusted, and 3D as-built model was generated using open source photogrammetric software. The as-planned model was generated from digitized CAD drawings using GIS. The 3D as-built model was aligned with the 4D as-planned model of building formed from extrusion of building elements, and integration of the construction's planned schedule. The construction progress is visualized via color-coding the building elements in the 3D model. The developed methodology was conducted and applied from the data obtained from an actual construction site. Accuracy in detecting `built' or `not built' building elements ranges from 82-84 % and precision of 50-72 %. Quantified progress in terms of the number of building elements are 21.31% (November 2016), 26.84 % (January 2017) and 44.19 % (March 2017). The results can be used as an input for progress monitoring performance of construction projects and improving related decision-making process.
A design of optical modulation system with pixel-level modulation accuracy
NASA Astrophysics Data System (ADS)
Zheng, Shiwei; Qu, Xinghua; Feng, Wei; Liang, Baoqiu
2018-01-01
Vision measurement has been widely used in the field of dimensional measurement and surface metrology. However, traditional methods of vision measurement have many limits such as low dynamic range and poor reconfigurability. The optical modulation system before image formation has the advantage of high dynamic range, high accuracy and more flexibility, and the modulation accuracy is the key parameter which determines the accuracy and effectiveness of optical modulation system. In this paper, an optical modulation system with pixel level accuracy is designed and built based on multi-points reflective imaging theory and digital micromirror device (DMD). The system consisted of digital micromirror device, CCD camera and lens. Firstly we achieved accurate pixel-to-pixel correspondence between the DMD mirrors and the CCD pixels by moire fringe and an image processing of sampling and interpolation. Then we built three coordinate systems and calculated the mathematic relationship between the coordinate of digital micro-mirror and CCD pixels using a checkerboard pattern. A verification experiment proves that the correspondence error is less than 0.5 pixel. The results show that the modulation accuracy of system meets the requirements of modulation. Furthermore, the high reflecting edge of a metal circular piece can be detected using the system, which proves the effectiveness of the optical modulation system.
Color structured light imaging of skin
NASA Astrophysics Data System (ADS)
Yang, Bin; Lesicko, John; Moy, Austin; Reichenberg, Jason; Sacks, Michael; Tunnell, James W.
2016-05-01
We illustrate wide-field imaging of skin using a structured light (SL) approach that highlights the contrast from superficial tissue scattering. Setting the spatial frequency of the SL in a regime that limits the penetration depth effectively gates the image for photons that originate from the skin surface. Further, rendering the SL images in a color format provides an intuitive format for viewing skin pathologies. We demonstrate this approach in skin pathologies using a custom-built handheld SL imaging system.
1998-01-01
consisted of a videomicroscopy system and a tactile stimulator system. By using this setup, real-time images from the contact region as wvell as the... Videomicroscopy system . 4.3.2 Tactile stimulator svsteln . 4.3.3 Real-time imaging setup. 4.3.4 Active and passive touch experiments. 4.3.5...contact process is an important step. In this study, therefore, a videomicroscopy system was built’to visualize the contact re- gion of the fingerpad
The laboratory demonstration and signal processing of the inverse synthetic aperture imaging ladar
NASA Astrophysics Data System (ADS)
Gao, Si; Zhang, ZengHui; Xu, XianWen; Yu, WenXian
2017-10-01
This paper presents a coherent inverse synthetic-aperture imaging ladar(ISAL)system to obtain high resolution images. A balanced coherent optics system in laboratory is built with binary phase coded modulation transmit waveform which is different from conventional chirp. A whole digital signal processing solution is proposed including both quality phase gradient autofocus(QPGA) algorithm and cubic phase function(CPF) algorithm. Some high-resolution well-focused ISAL images of retro-reflecting targets are shown to validate the concepts. It is shown that high resolution images can be achieved and the influences from vibrations of platform involving targets and radar can be automatically compensated by the distinctive laboratory system and digital signal process.
Elliott, Jonathan T; Dsouza, Alisha V; Marra, Kayla; Pogue, Brian W; Roberts, David W; Paulsen, Keith D
2016-09-01
Fluorescence guided surgery has the potential to positively impact surgical oncology; current operating microscopes and stand-alone imaging systems are too insensitive or too cumbersome to maximally take advantage of new tumor-specific agents developed through the microdose pathway. To this end, a custom-built illumination and imaging module enabling picomolar-sensitive near-infrared fluorescence imaging on a commercial operating microscope is described. The limits of detection and system specifications are characterized, and in vivo efficacy of the system in detecting ABY-029 is evaluated in a rat orthotopic glioma model following microdose injections, showing the suitability of the device for microdose phase 0 clinical trials.
Dsouza, Alisha V.; Marra, Kayla; Pogue, Brian W.; Roberts, David W.; Paulsen, Keith D.
2016-01-01
Fluorescence guided surgery has the potential to positively impact surgical oncology; current operating microscopes and stand-alone imaging systems are too insensitive or too cumbersome to maximally take advantage of new tumor-specific agents developed through the microdose pathway. To this end, a custom-built illumination and imaging module enabling picomolar-sensitive near-infrared fluorescence imaging on a commercial operating microscope is described. The limits of detection and system specifications are characterized, and in vivo efficacy of the system in detecting ABY-029 is evaluated in a rat orthotopic glioma model following microdose injections, showing the suitability of the device for microdose phase 0 clinical trials. PMID:27699098
A novel stereotactic frame for real PET-guided biopsies: A preclinical proof-of-concept.
Cortes-Rodicio, J; Sanchez-Merino, G; Garcia-Fidalgo, M A; Tobalina-Larrea, I
2017-09-01
To design, build and test a stereotactic device that allows PET image-guided biopsies to be performed. An initial prototype consisting of four main pieces, one of which contains radioactive markers to make it visible in the PET images, was built using a 3D printer. Once the device is mounted, a spherical coordinate system is built with the entrance needle point in the skin as the origin of coordinates. Two in-house software programs, namely getCoord.ijm, which obtains the spherical coordinates of the tumour tissue to be biopsied, and getNeedle.ijm, which virtualizes the inner needle tip once the puncture has taken place, were written. This prototype was tested on an FDG-doped phantom to characterize both the accuracy of the system and the procedure time. Up to 11 complete biopsy procedures were conducted. The mean total procedure time was less than 20min, which is less than the procedure time of conventional standard CT-guided biopsies. The overall accuracy of the system was found to be 5.0±1.3mm, which outperforms the criterion used in routine clinical practice when targeting tumours with a diameter of 10mm. A stereotactic frame to conduct real PET image-guided biopsies has been designed and built. A proof-of-concept was performed to characterize the system. The procedure time and accuracy of the system were found to meet the current needs of physicians performing biopsies. Copyright © 2017 Associazione Italiana di Fisica Medica. Published by Elsevier Ltd. All rights reserved.
Real-time photoacoustic imaging of prostate brachytherapy seeds using a clinical ultrasound system.
Kuo, Nathanael; Kang, Hyun Jae; Song, Danny Y; Kang, Jin U; Boctor, Emad M
2012-06-01
Prostate brachytherapy is a popular prostate cancer treatment option that involves the permanent implantation of radioactive seeds into the prostate. However, contemporary brachytherapy procedure is limited by the lack of an imaging system that can provide real-time seed-position feedback. While many other imaging systems have been proposed, photoacoustic imaging has emerged as a potential ideal modality to address this need, since it could easily be incorporated into the current ultrasound system used in the operating room. We present such a photoacoustic imaging system built around a clinical ultrasound system to achieve the task of visualizing and localizing seeds. We performed several experiments to analyze the effects of various parameters on the appearance of brachytherapy seeds in photoacoustic images. We also imaged multiple seeds in an ex vivo dog prostate phantom to demonstrate the possibility of using this system in a clinical setting. Although still in its infancy, these initial results of a photoacoustic imaging system for the application of prostate brachytherapy seed localization are highly promising.
An automated image-collection system for crystallization experiments using SBS standard microplates.
Brostromer, Erik; Nan, Jie; Su, Xiao Dong
2007-02-01
As part of a structural genomics platform in a university laboratory, a low-cost in-house-developed automated imaging system for SBS microplate experiments has been designed and constructed. The imaging system can scan a microplate in 2-6 min for a 96-well plate depending on the plate layout and scanning options. A web-based crystallization database system has been developed, enabling users to follow their crystallization experiments from a web browser. As the system has been designed and built by students and crystallographers using commercially available parts, this report is aimed to serve as a do-it-yourself example for laboratory robotics.
NASA Astrophysics Data System (ADS)
Hopper, Darrel G.; Haralson, David G.; Simpson, Matthew A.; Longo, Sam J.
2002-08-01
Ultra-resolution visualization systems are achieved by the technique of tiling many direct or project-view displays. During the past fews years, several such systems have been built from commercial electronics components (displays, computers, image generators, networks, communication links, and software). Civil applications driving this development have independently determined that they require images at 10-100 megapixel (Mpx) resolution to enable state-of-the-art research, engineering, design, stock exchanges, flight simulators, business information and enterprise control centers, education, art and entertainment. Military applications also press the art of the possible to improve the productivity of warfighters and lower the cost of providing for the national defense. The environment in some 80% of defense applications can be addressed by ruggedization of commercial components. This paper reviews the status of ultra-resolution systems based on commercial components and describes a vision for their integration into advanced yet affordable military command centers, simulator/trainers, and, eventually, crew stations in air, land, sea and space systems.
Application Possibility of Smartphone as Payload for Photogrammetric Uav System
NASA Astrophysics Data System (ADS)
Yun, M. H.; Kim, J.; Seo, D.; Lee, J.; Choi, C.
2012-07-01
Smartphone can not only be operated under 3G network environment anytime and anyplace but also cost less than the existing photogrammetric UAV since it provides high-resolution image, 3D location and attitude data on a real-time basis from a variety of built-in sensors. This study is aimed to assess the possibility of smartphone as a payload for photogrammetric UAV system. Prior to such assessment, a smartphone-based photogrammetric UAV system application was developed, through which real-time image, location and attitude data was obtained using smartphone under both static and dynamic conditions. Subsequently the accuracy assessment on the location and attitude data obtained and sent by this system was conducted. The smartphone images were converted into ortho-images through image triangulation. The image triangulation was conducted in accordance with presence or absence of consideration of the interior orientation (IO) parameters determined by camera calibration. In case IO parameters were taken into account in the static experiment, the results from triangulation for any smartphone type were within 1.5 pixel (RMSE), which was improved at least by 35% compared to when IO parameters were not taken into account. On the contrary, the improvement effect of considering IO parameters on accuracy in triangulation for smartphone images in dynamic experiment was not significant compared to the static experiment. It was due to the significant impact of vibration and sudden attitude change of UAV on the actuator for automatic focus control within the camera built in smartphone under the dynamic condition. This cause appears to have a negative impact on the image-based DEM generation. Considering these study findings, it is suggested that smartphone is very feasible as a payload for UAV system. It is also expected that smartphone may be loaded onto existing UAV playing direct or indirect roles significantly.
NASA Technical Reports Server (NTRS)
2002-01-01
Nicknamed 'Dragon Lake,' this body of water is formed by the Bratskove Reservoir, built along the Angara river in southern Siberia, near the city of Bratsk. This image was acquired in winter, when the lake is frozen. This image was acquired by Landsat 7's Enhanced Thematic Mapper plus (ETM+) sensor on December 19, 1999. This is a natural color composite image made using blue, green, and red wavelengths. Image provided by the USGS EROS Data Center Satellite Systems Branch
The low intensity X-ray imaging scope /Lixiscope/
NASA Technical Reports Server (NTRS)
Yin, L. I.; Trombka, J. I.; Seltzer, S. M.; Webber, R. L.; Farr, M. R.; Rennie, J.
1978-01-01
A fully portable, small-format X-ray imaging system, Lixiscope (low intensity X-ray imaging scope) is described. In the prototype, which has been built to demonstrate the feasibility of the Lixiscope concept, only well-developed and available components have been used. Consideration is given to the principles of operation of the device, some of its performance characteristics as well as possible dental, medical and industrial applications.
NASA Astrophysics Data System (ADS)
Choi, Y.; Park, S.; Baik, S.; Jung, J.; Lee, S.; Yoo, J.
A small scale laboratory adaptive optics system using a Shack-Hartmann wave-front sensor (WFS) and a membrane deformable mirror (DM) has been built for robust image acquisition. In this study, an adaptive limited control technique is adopted to maintain the long-term correction stability of an adaptive optics system. To prevent the waste of dynamic correction range for correcting small residual wave-front distortions which are inefficient to correct, the built system tries to limit wave-front correction when a similar small difference wave-front pattern is repeatedly generated. Also, the effect of mechanical distortion in an adaptive optics system is studied and a pre-recognition method for the distortion is devised to prevent low-performance system operation. A confirmation process for a balanced work assignment among deformable mirror (DM) actuators is adopted for the pre-recognition. The corrected experimental results obtained by using a built small scale adaptive optics system are described in this paper.
Landsat Data Continuity Mission (LDCM) space to ground mission data architecture
Nelson, Jack L.; Ames, J.A.; Williams, J.; Patschke, R.; Mott, C.; Joseph, J.; Garon, H.; Mah, G.
2012-01-01
The Landsat Data Continuity Mission (LDCM) is a scientific endeavor to extend the longest continuous multi-spectral imaging record of Earth's land surface. The observatory consists of a spacecraft bus integrated with two imaging instruments; the Operational Land Imager (OLI), built by Ball Aerospace & Technologies Corporation in Boulder, Colorado, and the Thermal Infrared Sensor (TIRS), an in-house instrument built at the Goddard Space Flight Center (GSFC). Both instruments are integrated aboard a fine-pointing, fully redundant, spacecraft bus built by Orbital Sciences Corporation, Gilbert, Arizona. The mission is scheduled for launch in January 2013. This paper will describe the innovative end-to-end approach for efficiently managing high volumes of simultaneous realtime and playback of image and ancillary data from the instruments to the reception at the United States Geological Survey's (USGS) Landsat Ground Network (LGN) and International Cooperator (IC) ground stations. The core enabling capability lies within the spacecraft Command and Data Handling (C&DH) system and Radio Frequency (RF) communications system implementation. Each of these systems uniquely contribute to the efficient processing of high speed image data (up to 265Mbps) from each instrument, and provide virtually error free data delivery to the ground. Onboard methods include a combination of lossless data compression, Consultative Committee for Space Data Systems (CCSDS) data formatting, a file-based/managed Solid State Recorder (SSR), and Low Density Parity Check (LDPC) forward error correction. The 440 Mbps wideband X-Band downlink uses Class 1 CCSDS File Delivery Protocol (CFDP), and an earth coverage antenna to deliver an average of 400 scenes per day to a combination of LGN and IC ground stations. This paper will also describe the integrated capabilities and processes at the LGN ground stations for data reception using adaptive filtering, and the mission operations approach fro- the LDCM Mission Operations Center (MOC) to perform the CFDP accounting, file retransmissions, and management of the autonomous features of the SSR.
Architecture for biomedical multimedia information delivery on the World Wide Web
NASA Astrophysics Data System (ADS)
Long, L. Rodney; Goh, Gin-Hua; Neve, Leif; Thoma, George R.
1997-10-01
Research engineers at the National Library of Medicine are building a prototype system for the delivery of multimedia biomedical information on the World Wide Web. This paper discuses the architecture and design considerations for the system, which will be used initially to make images and text from the third National Health and Nutrition Examination Survey (NHANES) publicly available. We categorized our analysis as follows: (1) fundamental software tools: we analyzed trade-offs among use of conventional HTML/CGI, X Window Broadway, and Java; (2) image delivery: we examined the use of unconventional TCP transmission methods; (3) database manager and database design: we discuss the capabilities and planned use of the Informix object-relational database manager and the planned schema for the HNANES database; (4) storage requirements for our Sun server; (5) user interface considerations; (6) the compatibility of the system with other standard research and analysis tools; (7) image display: we discuss considerations for consistent image display for end users. Finally, we discuss the scalability of the system in terms of incorporating larger or more databases of similar data, and the extendibility of the system for supporting content-based retrieval of biomedical images. The system prototype is called the Web-based Medical Information Retrieval System. An early version was built as a Java applet and tested on Unix, PC, and Macintosh platforms. This prototype used the MiniSQL database manager to do text queries on a small database of records of participants in the second NHANES survey. The full records and associated x-ray images were retrievable and displayable on a standard Web browser. A second version has now been built, also a Java applet, using the MySQL database manager.
Design of a high-numerical-aperture digital micromirror device camera with high dynamic range.
Qiao, Yang; Xu, Xiping; Liu, Tao; Pan, Yue
2015-01-01
A high-NA imaging system with high dynamic range is presented based on a digital micromirror device (DMD). The DMD camera consists of an objective imaging system and a relay imaging system, connected by a DMD chip. With the introduction of a total internal reflection prism system, the objective imaging system is designed with a working F/# of 1.97, breaking through the F/2.45 limitation of conventional DMD projection lenses. As for the relay imaging system, an off-axis design that could correct off-axis aberrations of the tilt relay imaging system is developed. This structure has the advantage of increasing the NA of the imaging system while maintaining a compact size. Investigation revealed that the dynamic range of a DMD camera could be greatly increased, by 2.41 times. We built one prototype DMD camera with a working F/# of 1.23, and the field experiments proved the validity and reliability our work.
Small Interactive Image Processing System (SMIPS) system description
NASA Technical Reports Server (NTRS)
Moik, J. G.
1973-01-01
The Small Interactive Image Processing System (SMIPS) operates under control of the IBM-OS/MVT operating system and uses an IBM-2250 model 1 display unit as interactive graphic device. The input language in the form of character strings or attentions from keys and light pen is interpreted and causes processing of built-in image processing functions as well as execution of a variable number of application programs kept on a private disk file. A description of design considerations is given and characteristics, structure and logic flow of SMIPS are summarized. Data management and graphic programming techniques used for the interactive manipulation and display of digital pictures are also discussed.
Atmospheric Science Data Center
2013-04-16
... article title: Multi-layer Clouds Over the South Indian Ocean View Larger Image ... System-2 path 155. MISR was built and is managed by NASA's Jet Propulsion Laboratory, Pasadena, CA, for NASA's Science Mission ...
Optical correlators for recognition of human face thermal images
NASA Astrophysics Data System (ADS)
Bauer, Joanna; Podbielska, Halina; Suchwalko, Artur; Mazurkiewicz, Jacek
2005-09-01
In this paper, the application of the optical correlators for face thermograms recognition is described. The thermograms were colleted from 27 individuals. For each person 10 pictures in different conditions were recorded and the data base composed of 270 images was prepared. Two biometric systems based on joint transform correlator and 4f correlator were built. Each system was designed for realizing two various tasks: verification and identification. The recognition systems were tested and evaluated according to the Face Recognition Vendor Tests (FRVT).
Development of Targeting UAVs Using Electric Helicopters and Yamaha RMAX
2007-05-17
including the QNX real - time operating system . The video overlay board is useful to display the onboard camera’s image with important information such as... real - time operating system . Fully utilizing the built-in multi-processing architecture with inter-process synchronization and communication
Jin, Xin; Liu, Li; Chen, Yanqin; Dai, Qionghai
2017-05-01
This paper derives a mathematical point spread function (PSF) and a depth-invariant focal sweep point spread function (FSPSF) for plenoptic camera 2.0. Derivation of PSF is based on the Fresnel diffraction equation and image formation analysis of a self-built imaging system which is divided into two sub-systems to reflect the relay imaging properties of plenoptic camera 2.0. The variations in PSF, which are caused by changes of object's depth and sensor position variation, are analyzed. A mathematical model of FSPSF is further derived, which is verified to be depth-invariant. Experiments on the real imaging systems demonstrate the consistency between the proposed PSF and the actual imaging results.
NASA Astrophysics Data System (ADS)
Ruggeri, Marco; Maceo Heilman, Bianca M.; Yao, Yue; Chang, Yu-Cherng; Gonzalez, Alex; Rowaan, Cornelis; Mohamed, Ashik; Williams, Siobhan; Durkee, Heather A.; Silgado, Juan; Bernal, Andres; Arrieta-Quintero, Esdras; Ho, Arthur; Parel, Jean-Marie A.; Manns, Fabrice
2017-02-01
Age-related changes in the crystalline lens shape and refractive index gradient produce changes in dioptric power and high-order aberrations that influence the optics of the whole eye and contribute to a decrease in overall visual quality. Despite their key role, the changes in lens shape and refractive index gradient with age and accommodation and their effects on high-order aberrations are still not well understood. The goal of this project was to develop a combined laser ray tracing (LRT) and optical coherence tomography (OCT) system to measure high-order aberrations, shape and refractive index gradient in non-human primate and human lenses. A miniature motorized lens stretching system was built to enable imaging and aberrometry of the lens during simulated accommodation. A positioning system was also built to enable on- and off-axis OCT imaging and aberrometry for characterization of the peripheral defocus of the lens. We demonstrated the capability of the LRT-OCT system to produce OCT images and aberration measurements of crystalline lens with age and accommodation in vitro. In future work, the information acquired with the LRT-OCT system will be used to develop an accurate age-dependent lens model to predict the role of the lens in the development of refractive error and aberrations of the whole eye.
NASA Astrophysics Data System (ADS)
Wu, Jianping; Lu, Fei; Zou, Kai; Yan, Hong; Wan, Min; Kuang, Yan; Zhou, Yanqing
2018-03-01
An ultra-high angular velocity and minor-caliber high-precision stably control technology application for active-optics image-motion compensation, is put forward innovatively in this paper. The image blur problem due to several 100°/s high-velocity relative motion between imaging system and target is theoretically analyzed. The velocity match model of detection system and active optics compensation system is built, and active optics image motion compensation platform experiment parameters are designed. Several 100°/s high-velocity high-precision control optics compensation technology is studied and implemented. The relative motion velocity is up to 250°/s, and image motion amplitude is more than 20 pixel. After the active optics compensation, motion blur is less than one pixel. The bottleneck technology of ultra-high angular velocity and long exposure time in searching and infrared detection system is successfully broke through.
A multi-purpose electromagnetic actuator for magnetic resonance elastography.
Feng, Yuan; Zhu, Mo; Qiu, Suhao; Shen, Ping; Ma, Shengyuan; Zhao, Xuefeng; Hu, Chun-Hong; Guo, Liang
2018-04-19
An electromagnetic actuator was designed for magnetic resonance elastography (MRE). The actuator is unique in that it is simple, portable, and capable of brain, abdomen, and phantom imagings. A custom-built control unit was used for controlling the vibration frequency and synchronizing the trigger signals. An actuation unit was built and mounted on the specifically designed clamp and holders for different imaging applications. MRE experiments with respect to gel phantoms, brain, and liver showed that the actuator could produce stable and consistent mechanical waves. Estimated shear modulus using local frequency estimate method demonstrated that the measurement results were in line with that from MRE studies using different actuation systems. The relatively easy setup procedure and simple design indicated that the actuator system had the potential to be applied in many different clinical studies. Copyright © 2018 Elsevier Inc. All rights reserved.
NASA Astrophysics Data System (ADS)
Morgan, Christopher G.; Mitchell, A. C.; Murray, J. G.
1990-05-01
An imaging photon detector has been modified to incorporate fast timing electronics coupled to a custom built photon correlator interfaced to a RISC computer. Using excitation with intensity- muodulated light, fluorescence images can be readily obtained where contrast is determined by the decay time of emission, rather than by intensity. This technology is readily extended to multifrequency phase/demodulation fluorescence imaging or to differential polarised phase fluorometry. The potential use of the correlator for confocal imaging with a laser scanner is also briefly discussed.
Eom, Byeong Ho; Penanen, Konstantin; Hahn, Inseob
2014-09-01
Magnetic resonance imaging (MRI) at microtesla fields using superconducting quantum interference device (SQUID) detection has previously been demonstrated, and advantages have been noted. Although the ultralow-field SQUID MRI technique would not need the heavy superconducting magnet of conventional MRI systems, liquid helium required to cool the low-temperature detector still places a significant burden on its operation. We have built a prototype cryocooler-based SQUID MRI system that does not require a cryogen. The SQUID detector and the superconducting gradiometer were cooled down to 3.7 K and 4.3 K, respectively. We describe the prototype design, characterization, a phantom image, and areas of further improvements needed to bring the imaging performance to parity with conventional MRI systems.
A System for Open-Access 3He Human Lung Imaging at Very Low Field
RUSET, I.C.; TSAI, L.L.; MAIR, R.W.; PATZ, S.; HROVAT, M.I.; ROSEN, M.S.; MURADIAN, I.; NG, J.; TOPULOS, G.P.; BUTLER, J.P.; WALSWORTH, R.L.; HERSMAN, F.W.
2010-01-01
We describe a prototype system built to allow open-access very-low-field MRI of human lungs using laser-polarized 3He gas. The system employs an open four-coil electromagnet with an operational B0 field of 4 mT, and planar gradient coils that generate gradient fields up to 0.18 G/cm in the x and y direction and 0.41 G/cm in the z direction. This system was used to obtain 1H and 3He phantom images and supine and upright 3He images of human lungs. We include discussion on challenges unique to imaging at 50 –200 kHz, including noise filtering and compensation for narrow-bandwidth coils. PMID:20354575
NASA Astrophysics Data System (ADS)
Chen, Y.; Zhang, Y.; Gao, J.; Yuan, Y.; Lv, Z.
2018-04-01
Recently, built-up area detection from high-resolution satellite images (HRSI) has attracted increasing attention because HRSI can provide more detailed object information. In this paper, multi-resolution wavelet transform and local spatial autocorrelation statistic are introduced to model the spatial patterns of built-up areas. First, the input image is decomposed into high- and low-frequency subbands by wavelet transform at three levels. Then the high-frequency detail information in three directions (horizontal, vertical and diagonal) are extracted followed by a maximization operation to integrate the information in all directions. Afterward, a cross-scale operation is implemented to fuse different levels of information. Finally, local spatial autocorrelation statistic is introduced to enhance the saliency of built-up features and an adaptive threshold algorithm is used to achieve the detection of built-up areas. Experiments are conducted on ZY-3 and Quickbird panchromatic satellite images, and the results show that the proposed method is very effective for built-up area detection.
Simultaneous PET/MR imaging with a radio frequency-penetrable PET insert
Grant, Alexander M.; Lee, Brian J.; Chang, Chen-Ming; Levin, Craig S.
2017-01-01
Purpose A brain sized radio-frequency (RF)-penetrable PET insert has been designed for simultaneous operation with MRI systems. This system takes advantage of electro-optical coupling and battery power to electrically float the PET insert relative to the MRI ground, permitting RF signals to be transmitted through small gaps between the modules that form the PET ring. This design facilitates the use of the built-in body coil for RF transmission, and thus could be inserted into any existing MR site wishing to achieve simultaneous PET/MR imaging. The PET detectors employ non-magnetic silicon photomultipliers in conjunction with a compressed sensing signal multiplexing scheme, and optical fibers to transmit analog PET detector signals out of the MRI room for decoding, processing, and image reconstruction. Methods The PET insert was first constructed and tested in a laboratory benchtop setting, where tomographic images of a custom resolution phantom were successfully acquired. The PET insert was then placed within a 3T body MRI system, and tomographic resolution/contrast phantom images were acquired both with only the B0 field present, and under continuous pulsing from different MR imaging sequences. Results The resulting PET images have comparable contrast-to-noise ratios (CNR) under all MR pulsing conditions: the maximum percent CNR relative difference for each rod type among all four PET images acquired in the MRI system has a mean of 14.0±7.7%. MR images were successfully acquired through the RF-penetrable PET shielding using only the built-in MR body coil, suggesting that simultaneous imaging is possible without significant mutual interference. Conclusions These results show promise for this technology as an alternative to costly integrated PET/MR scanners; a PET insert that is compatible with any existing clinical MRI system could greatly increase the availability, accessibility, and dissemination of PET/MR. PMID:28102949
Earthshots: Satellite images of environmental change – Phnom Penh, Cambodia
,
2013-01-01
From 1975 to 1978, Cambodia was ruled by the Khmer Rouge regime, which sought among other things to build a vast system of irrigation canals. These images show an area around Cambodia's capital city of Phnom Penh where such waterworks were built. Many areas east of the Mekong River show a gridwork of canals by 1985.
PScan 1.0: flexible software framework for polygon based multiphoton microscopy
NASA Astrophysics Data System (ADS)
Li, Yongxiao; Lee, Woei Ming
2016-12-01
Multiphoton laser scanning microscopes exhibit highly localized nonlinear optical excitation and are powerful instruments for in-vivo deep tissue imaging. Customized multiphoton microscopy has a significantly superior performance for in-vivo imaging because of precise control over the scanning and detection system. To date, there have been several flexible software platforms catered to custom built microscopy systems i.e. ScanImage, HelioScan, MicroManager, that perform at imaging speeds of 30-100fps. In this paper, we describe a flexible software framework for high speed imaging systems capable of operating from 5 fps to 1600 fps. The software is based on the MATLAB image processing toolbox. It has the capability to communicate directly with a high performing imaging card (Matrox Solios eA/XA), thus retaining high speed acquisition. The program is also designed to communicate with LabVIEW and Fiji for instrument control and image processing. Pscan 1.0 can handle high imaging rates and contains sufficient flexibility for users to adapt to their high speed imaging systems.
Laser speckle contrast imaging using light field microscope approach
NASA Astrophysics Data System (ADS)
Ma, Xiaohui; Wang, Anting; Ma, Fenghua; Wang, Zi; Ming, Hai
2018-01-01
In this paper, a laser speckle contrast imaging (LSCI) system using light field (LF) microscope approach is proposed. As far as we known, it is first time to combine LSCI with LF. To verify this idea, a prototype consists of a modified LF microscope imaging system and an experimental device was built. A commercially used Lytro camera was modified for microscope imaging. Hollow glass tubes with different depth fixed in glass dish were used to simulate the vessels in brain and test the performance of the system. Compared with conventional LSCI, three new functions can be realized by using our system, which include refocusing, extending the depth of field (DOF) and gathering 3D information. Experiments show that the principle is feasible and the proposed system works well.
Low-cost, high-speed back-end processing system for high-frequency ultrasound B-mode imaging.
Chang, Jin Ho; Sun, Lei; Yen, Jesse T; Shung, K Kirk
2009-07-01
For real-time visualization of the mouse heart (6 to 13 beats per second), a back-end processing system involving high-speed signal processing functions to form and display images has been developed. This back-end system was designed with new signal processing algorithms to achieve a frame rate of more than 400 images per second. These algorithms were implemented in a simple and cost-effective manner with a single field-programmable gate array (FPGA) and software programs written in C++. The operating speed of the back-end system was investigated by recording the time required for transferring an image to a personal computer. Experimental results showed that the back-end system is capable of producing 433 images per second. To evaluate the imaging performance of the back-end system, a complete imaging system was built. This imaging system, which consisted of a recently reported high-speed mechanical sector scanner assembled with the back-end system, was tested by imaging a wire phantom, a pig eye (in vitro), and a mouse heart (in vivo). It was shown that this system is capable of providing high spatial resolution images with fast temporal resolution.
Low-Cost, High-Speed Back-End Processing System for High-Frequency Ultrasound B-Mode Imaging
Chang, Jin Ho; Sun, Lei; Yen, Jesse T.; Shung, K. Kirk
2009-01-01
For real-time visualization of the mouse heart (6 to 13 beats per second), a back-end processing system involving high-speed signal processing functions to form and display images has been developed. This back-end system was designed with new signal processing algorithms to achieve a frame rate of more than 400 images per second. These algorithms were implemented in a simple and cost-effective manner with a single field-programmable gate array (FPGA) and software programs written in C++. The operating speed of the back-end system was investigated by recording the time required for transferring an image to a personal computer. Experimental results showed that the back-end system is capable of producing 433 images per second. To evaluate the imaging performance of the back-end system, a complete imaging system was built. This imaging system, which consisted of a recently reported high-speed mechanical sector scanner assembled with the back-end system, was tested by imaging a wire phantom, a pig eye (in vitro), and a mouse heart (in vivo). It was shown that this system is capable of providing high spatial resolution images with fast temporal resolution. PMID:19574160
Biggest-Ever Heat Shield Prepared for Mars Spacecraft
2011-05-13
The heat shield for NASA Mars Science Laboratory is the largest ever built for a planetary mission. This image shows the heat shield being prepared at Lockheed Martin Space Systems, Denver, in April 2011.
Large Heat Shield for Mars Science Laboratory
2009-07-10
This image shows NASA Mars Science Laboratory heat shield, and a spacecraft worker at Lockheed Martin Space Systems, Denver. It is the largest heat shield ever built for descending through the atmosphere of any planet.
Utilization of a multimedia PACS workstation for surgical planning of epilepsy
NASA Astrophysics Data System (ADS)
Soo Hoo, Kent; Wong, Stephen T.; Hawkins, Randall A.; Knowlton, Robert C.; Laxer, Kenneth D.; Rowley, Howard A.
1997-05-01
Surgical treatment of temporal lobe epilepsy requires the localization of the epileptogenic zone for surgical resection. Currently, clinicians utilize electroencephalography, various neuroimaging modalities, and psychological tests together to determine the location of this zone. We investigate how a multimedia neuroimaging workstation built on top of the UCSF Picture Archiving and Communication System can be used to aid surgical planning of epilepsy and related brain diseases. This usage demonstrates the ability of the workstation to retrieve image and textural data from PACS and other image sources, register multimodality images, visualize and render 3D data sets, analyze images, generate new image and text data from the analysis, and organize all data in a relational database management system.
Rapid Protein Separations in Microfluidic Devices
NASA Technical Reports Server (NTRS)
Fan, Z. H.; Das, Champak; Xia, Zheng; Stoyanov, Alexander V.; Fredrickson, Carl K.
2004-01-01
This paper describes fabrication of glass and plastic microfluidic devices for protein separations. Although the long-term goal is to develop a microfluidic device for two-dimensional gel electrophoresis, this paper focuses on the first dimension-isoelectric focusing (IEF). A laser-induced fluorescence (LIF) imaging system has been built for imaging an entire channel in an IEF device. The whole-channel imaging eliminates the need to migrate focused protein bands, which is required if a single-point detector is used. Using the devices and the imaging system, we are able to perform IEF separations of proteins within minutes rather than hours in traditional bench-top instruments.
Chen, S C; Shao, C L; Liang, C K; Lin, S W; Huang, T H; Hsieh, M C; Yang, C H; Luo, C H; Wuo, C M
2004-01-01
In this paper, we present a text input system for the seriously disabled by using lips image recognition based on LabVIEW. This system can be divided into the software subsystem and the hardware subsystem. In the software subsystem, we adopted the technique of image processing to recognize the status of mouth-opened or mouth-closed depending the relative distance between the upper lip and the lower lip. In the hardware subsystem, parallel port built in PC is used to transmit the recognized result of mouth status to the Morse-code text input system. Integrating the software subsystem with the hardware subsystem, we implement a text input system by using lips image recognition programmed in LabVIEW language. We hope the system can help the seriously disabled to communicate with normal people more easily.
[Teleradiology using uncompressed DICOM format via exclusive fiber-optic system].
Okuda, Shigeo; Kuribayashi, Sachio; Hibi, Norihumi; Matsuura, Agato; Tani, Rie; Saga, Yasushi
2005-01-01
We developed a system for teleradiology using exclusive fiber optics for transferring images formatted in uncompressed DICOM. This system was built up with commercially available machines and software provided from various companies. We are now operating the system with five remote hospitals and have had one year of experience. The current system took advantage of the security and transfer efficiency of exclusive fiber optics. Uncompressed DICOM images were useful for the identification of cases and user-friendly for viewing. The reading room is located in our university hospital, and the location is convenient for consultation and discussion of cases.
Real-time biochemical sensor based on Raman scattering with CMOS contact imaging.
Muyun Cao; Yuhua Li; Yadid-Pecht, Orly
2015-08-01
This work presents a biochemical sensor based on Raman scattering with Complementary metal-oxide-semiconductor (CMOS) contact imaging. This biochemical optical sensor is designed for detecting the concentration of solutions. The system is built with a laser diode, an optical filter, a sample holder and a commercial CMOS sensor. The output of the system is analyzed by an image processing program. The system provides instant measurements with a resolution of 0.2 to 0.4 Mol. This low cost and easy-operated small scale system is useful in chemical, biomedical and environmental labs for quantitative bio-chemical concentration detection with results reported comparable to a highly cost commercial spectrometer.
Feasibility of employing a smartphone as the payload in a photogrammetric UAV system
NASA Astrophysics Data System (ADS)
Kim, Jinsoo; Lee, Seongkyu; Ahn, Hoyong; Seo, Dongju; Park, Soyoung; Choi, Chuluong
2013-05-01
Smartphones can be operated in a 3G network environment at any time or location, and they also cost less than existing photogrammetric UAV systems, providing high-resolution images and 3D location and attitude data from a variety of built-in sensors. This study aims to assess the feasibility of using a smartphone as the payload for a photogrammetric UAV system. To carry out the assessment, a smartphone-based photogrammetric UAV system was developed and utilized to obtain image, location, and attitude data under both static and dynamic conditions. The accuracy of the location and attitude data obtained and sent by this system was then evaluated. The smartphone images were converted into ortho-images via image triangulation, which was carried out both with and without consideration of the interior orientation (IO) parameters determined by camera calibration. In the static experiment, when the IO parameters were taken into account, the triangulation results were less than 1.28 pixels (RMSE) for all smartphone types, an improvement of at least 47% compared with the case when IO parameters were not taken into account. In the dynamic experiment, on the other hand, the accuracy of smartphone image triangulation was not significantly improved by considering IO parameters. This was because the electronic rolling shutter within the complementary metal-oxide semiconductor (CMOS) sensor built into the smartphone and the actuator for the voice coil motor (VCM)-type auto-focusing affected by the vibration and the speed of the UAV, which is likely to have a negative effect on image-based digital elevation model (DEM) generation. However, considering that these results were obtained using a single smartphone, this suggests that a smartphone is not only feasible as the payload for a photogrammetric UAV system but it may also play a useful role when installed in existing UAV systems.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Eom, Byeong Ho; Penanen, Konstantin; Hahn, Inseob, E-mail: ihahn@caltech.edu
2014-09-15
Magnetic resonance imaging (MRI) at microtesla fields using superconducting quantum interference device (SQUID) detection has previously been demonstrated, and advantages have been noted. Although the ultralow-field SQUID MRI technique would not need the heavy superconducting magnet of conventional MRI systems, liquid helium required to cool the low-temperature detector still places a significant burden on its operation. We have built a prototype cryocooler-based SQUID MRI system that does not require a cryogen. The SQUID detector and the superconducting gradiometer were cooled down to 3.7 K and 4.3 K, respectively. We describe the prototype design, characterization, a phantom image, and areas ofmore » further improvements needed to bring the imaging performance to parity with conventional MRI systems.« less
High throughput operando studies using Fourier transform infrared imaging and Raman spectroscopy.
Li, Guosheng; Hu, Dehong; Xia, Guanguang; White, J M; Zhang, Conrad
2008-07-01
A prototype high throughput operando (HTO) reactor designed and built for catalyst screening and characterization combines Fourier transform infrared (FT-IR) imaging and Raman spectroscopy in operando conditions. Using a focal plane array detector (HgCdTe focal plane array, 128x128 pixels, and 1610 Hz frame rate) for the FT-IR imaging system, the catalyst activity and selectivity of all parallel reaction channels can be simultaneously followed. Each image data set possesses 16 384 IR spectra with a spectral range of 800-4000 cm(-1) and with an 8 cm(-1) resolution. Depending on the signal-to-noise ratio, 2-20 s are needed to generate a full image of all reaction channels for a data set. Results on reactant conversion and product selectivity are obtained from FT-IR spectral analysis. Six novel Raman probes, one for each reaction channel, were specially designed and house built at Pacific Northwest National Laboratory, to simultaneously collect Raman spectra of the catalysts and possible reaction intermediates on the catalyst surface under operando conditions. As a model system, methanol partial oxidation reaction on silica-supported molybdenum oxide (MoO3SiO2) catalysts has been studied under different reaction conditions to demonstrate the performance of the HTO reactor.
Electronic readout system for the Belle II imaging Time-Of-Propagation detector
NASA Astrophysics Data System (ADS)
Kotchetkov, Dmitri
2017-07-01
The imaging Time-Of-Propagation (iTOP) detector, constructed for the Belle II experiment at the SuperKEKB e+e- collider, is an 8192-channel high precision Cherenkov particle identification detector with timing resolution below 50 ps. To acquire data from the iTOP, a novel front-end electronic readout system was designed, built, and integrated. Switched-capacitor array application-specific integrated circuits are used to sample analog signals. Triggering, digitization, readout, and data transfer are controlled by Xilinx Zynq-7000 system on a chip devices.
FRIDA: diffraction-limited imaging and integral-field spectroscopy for the GTC
NASA Astrophysics Data System (ADS)
Watson, Alan M.; Acosta-Pulido, José A.; Álvarez-Núñez, Luis C.; Bringas-Rico, Vicente; Cardiel, Nicolás.; Cardona, Salvador; Chapa, Oscar; Díaz García, José Javier; Eikenberry, Stephen S.; Espejo, Carlos; Flores-Meza, Rubén. A.; Fuentes-Fernández, Jorge; Gallego, Jesús; Garcés Medina, José Leonardo; Garzón López, Francisco; Hammersley, Peter; Keiman, Carolina; Lara, Gerardo; López, José Alberto; López, Pablo L.; Lucero, Diana; Moreno Arce, Heidy; Pascual Ramirez, Sergio; Patrón Recio, Jesús; Prieto, Almudena; Rodríguez, Alberto José; Marco de la Rosa, José; Sánchez, Beatriz; Uribe, Jorge A.; Váldez Berriozabal, Francisco
2016-08-01
FRIDA is a diffraction-limited imager and integral-field spectrometer that is being built for the adaptive-optics focus of the Gran Telescopio Canarias. In imaging mode FRIDA will provide scales of 0.010, 0.020 and 0.040 arcsec/pixel and in IFS mode spectral resolutions of 1500, 4000 and 30,000. FRIDA is starting systems integration and is scheduled to complete fully integrated system tests at the laboratory by the end of 2017 and to be delivered to GTC shortly thereafter. In this contribution we present a summary of its design, fabrication, current status and potential scientific applications.
NASA Astrophysics Data System (ADS)
Cusanno, F.; Argentieri, A.; Baiocchi, M.; Colilli, S.; Cisbani, E.; De Vincentis, G.; Fratoni, R.; Garibaldi, F.; Giuliani, F.; Gricia, M.; Lucentini, M.; Magliozzi, M. L.; Majewski, S.; Marano, G.; Musico, P.; Musumeci, M.; Santavenere, F.; Torrioli, S.; Tsui, B. M. W.; Vitelli, L.; Wang, Y.
2010-05-01
Cardiovascular diseases are the most common cause of death in western countries. Understanding the rupture of vulnerable atherosclerotic plaques and monitoring the effect of innovative therapies of heart failure is of fundamental importance. A flexible, high resolution, high sensitivity detector system for molecular imaging with radionuclides on small animal models has been designed for this aim. A prototype has been built using tungsten pinhole and LaBr3(Ce) scintillator coupled to Hamamatsu Flat Panel PMTs. Compact individual-channel readout has been designed, built and tested. Measurements with phantoms as well as pilot studies on mice have been performed, the results show that the myocardial perfusion in mice can be determined with sufficient precision. The detector will be improved replacing the Hamamatsu Flat Panel with Silicon Photomultipliers (SiPMs) to allow integration of the system with MRI scanners. Application of LaBr3(Ce) scintillator coupled to photosensor with high photon detection efficiency and excellent energy resolution will allow dual-label imaging to monitor simultaneously the cardiac perfusion and the molecular targets under investigation during the heart therapy.
NASA Astrophysics Data System (ADS)
Cota, Stephen A.; Lomheim, Terrence S.; Florio, Christopher J.; Harbold, Jeffrey M.; Muto, B. Michael; Schoolar, Richard B.; Wintz, Daniel T.; Keller, Robert A.
2011-10-01
In a previous paper in this series, we described how The Aerospace Corporation's Parameterized Image Chain Analysis & Simulation SOftware (PICASSO) tool may be used to model space and airborne imaging systems operating in the visible to near-infrared (VISNIR). PICASSO is a systems-level tool, representative of a class of such tools used throughout the remote sensing community. It is capable of modeling systems over a wide range of fidelity, anywhere from conceptual design level (where it can serve as an integral part of the systems engineering process) to as-built hardware (where it can serve as part of the verification process). In the present paper, we extend the discussion of PICASSO to the modeling of Thermal Infrared (TIR) remote sensing systems, presenting the equations and methods necessary to modeling in that regime.
Processing techniques for digital sonar images from GLORIA.
Chavez, P.S.
1986-01-01
Image processing techniques have been developed to handle data from one of the newest members of the remote sensing family of digital imaging systems. This paper discusses software to process data collected by the GLORIA (Geological Long Range Inclined Asdic) sonar imaging system, designed and built by the Institute of Oceanographic Sciences (IOS) in England, to correct for both geometric and radiometric distortions that exist in the original 'raw' data. Preprocessing algorithms that are GLORIA-specific include corrections for slant-range geometry, water column offset, aspect ratio distortion, changes in the ship's velocity, speckle noise, and shading problems caused by the power drop-off which occurs as a function of range.-from Author
Hybrid cryptosystem for image file using elgamal and double playfair cipher algorithm
NASA Astrophysics Data System (ADS)
Hardi, S. M.; Tarigan, J. T.; Safrina, N.
2018-03-01
In this paper, we present an implementation of an image file encryption using hybrid cryptography. We chose ElGamal algorithm to perform asymmetric encryption and Double Playfair for the symmetric encryption. Our objective is to show that these algorithms are capable to encrypt an image file with an acceptable running time and encrypted file size while maintaining the level of security. The application was built using C# programming language and ran as a stand alone desktop application under Windows Operating System. Our test shows that the system is capable to encrypt an image with a resolution of 500×500 to a size of 976 kilobytes with an acceptable running time.
Projecting 2D gene expression data into 3D and 4D space.
Gerth, Victor E; Katsuyama, Kaori; Snyder, Kevin A; Bowes, Jeff B; Kitayama, Atsushi; Ueno, Naoto; Vize, Peter D
2007-04-01
Video games typically generate virtual 3D objects by texture mapping an image onto a 3D polygonal frame. The feeling of movement is then achieved by mathematically simulating camera movement relative to the polygonal frame. We have built customized scripts that adapt video game authoring software to texture mapping images of gene expression data onto b-spline based embryo models. This approach, known as UV mapping, associates two-dimensional (U and V) coordinates within images to the three dimensions (X, Y, and Z) of a b-spline model. B-spline model frameworks were built either from confocal data or de novo extracted from 2D images, once again using video game authoring approaches. This system was then used to build 3D models of 182 genes expressed in developing Xenopus embryos and to implement these in a web-accessible database. Models can be viewed via simple Internet browsers and utilize openGL hardware acceleration via a Shockwave plugin. Not only does this database display static data in a dynamic and scalable manner, the UV mapping system also serves as a method to align different images to a common framework, an approach that may make high-throughput automated comparisons of gene expression patterns possible. Finally, video game systems also have elegant methods for handling movement, allowing biomechanical algorithms to drive the animation of models. With further development, these biomechanical techniques offer practical methods for generating virtual embryos that recapitulate morphogenesis.
NASA Technical Reports Server (NTRS)
Tanner, Alan B.; Wilson, William J.; Lambrigsten, Bjorn H.; Dinardo, Steven J.; Brown, Shannon T.; Kangaslahti, Pekka P.; Gaier, Todd C.; Ruf, C. S.; Gross, S. M.; Lim, B. H.;
2006-01-01
The design, error budget, and preliminary test results of a 50-56 GHz synthetic aperture radiometer demonstration system are presented. The instrument consists of a fixed 24-element array of correlation interferometers, and is capable of producing calibrated images with 0.8 degree spatial resolution within a 17 degree wide field of view. This system has been built to demonstrate performance and a design which can be scaled to a much larger geostationary earth imager. As a baseline, such a system would consist of about 300 elements, and would be capable of providing contiguous, full hemispheric images of the earth with 1 Kelvin of radiometric precision and 50 km spatial resolution.
Image fusion using sparse overcomplete feature dictionaries
Brumby, Steven P.; Bettencourt, Luis; Kenyon, Garrett T.; Chartrand, Rick; Wohlberg, Brendt
2015-10-06
Approaches for deciding what individuals in a population of visual system "neurons" are looking for using sparse overcomplete feature dictionaries are provided. A sparse overcomplete feature dictionary may be learned for an image dataset and a local sparse representation of the image dataset may be built using the learned feature dictionary. A local maximum pooling operation may be applied on the local sparse representation to produce a translation-tolerant representation of the image dataset. An object may then be classified and/or clustered within the translation-tolerant representation of the image dataset using a supervised classification algorithm and/or an unsupervised clustering algorithm.
Atmospheric Science Data Center
2013-04-19
... or frontal activity. This image is centered over the Indian Ocean (at about 38.9° South, 80.6° East), and was acquired on October ... System-2 path 134. MISR was built and is managed by NASA's Jet Propulsion Laboratory, Pasadena, CA, for NASA's Science Mission ...
Near real-time stereo vision system
NASA Technical Reports Server (NTRS)
Anderson, Charles H. (Inventor); Matthies, Larry H. (Inventor)
1993-01-01
The apparatus for a near real-time stereo vision system for use with a robotic vehicle is described. The system is comprised of two cameras mounted on three-axis rotation platforms, image-processing boards, a CPU, and specialized stereo vision algorithms. Bandpass-filtered image pyramids are computed, stereo matching is performed by least-squares correlation, and confidence ranges are estimated by means of Bayes' theorem. In particular, Laplacian image pyramids are built and disparity maps are produced from the 60 x 64 level of the pyramids at rates of up to 2 seconds per image pair. The first autonomous cross-country robotic traverses (of up to 100 meters) have been achieved using the stereo vision system of the present invention with all computing done onboard the vehicle. The overall approach disclosed herein provides a unifying paradigm for practical domain-independent stereo ranging.
Three-Dimensional Microwave Imaging for Indoor Environments
NASA Astrophysics Data System (ADS)
Scott, Simon
Microwave imaging involves the use of antenna arrays, operating at microwave and millimeter-wave frequencies, for capturing images of real-world objects. Typically, one or more antennas in the array illuminate the scene with a radio-frequency (RF) signal. Part of this signal reflects back to the other antennas, which record both the amplitude and phase of the reflected signal. These reflected RF signals are then processed to form an image of the scene. This work focuses on using planar antenna arrays, operating between 17 and 26 GHz, to capture three-dimensional images of people and other objects inside a room. Such an imaging system enables applications such as indoor positioning and tracking, health monitoring and hand gesture recognition. Microwave imaging techniques based on beamforming cannot be used for indoor imaging, as most objects lie within the array near-field. Therefore, the range-migration algorithm (RMA) is used instead, as it compensates for the curvature of the reflected wavefronts, hence enabling near-field imaging. It is also based on fast-Fourier transforms and is therefore computationally efficient. A number of novel RMA variants were developed to support a wider variety of antenna array configurations, as well as to generate 3-D velocity maps of objects moving around a room. The choice of antenna array configuration, microwave transceiver components and transmit power has a significant effect on both the energy consumed by the imaging system and the quality of the resulting images. A generic microwave imaging testbed was therefore built to characterize the effect of these antenna array parameters on image quality in the 20 GHz band. All variants of the RMA were compared and found to produce good quality three-dimensional images with transmit power levels as low as 1 muW. With an array size of 80x80 antennas, most of the imaging algorithms were able to image objects at 0.5 m range with 12.5 mm resolution, although some were only able to achieve 20 mm resolution. Increasing the size of the antenna array further results in a proportional improvement in image resolution and image SNR, until the resolution reaches the half-wavelength limit. While microwave imaging is not a new technology, it has seen little commercial success due to the cost and power consumption of the large number of antennas and radio transceivers required to build such a system. The cost and power consumption can be reduced by using low-power and low-cost components in both the transmit and receive RF chains, even if these components have poor noise figures. Alternatively, the cost and power consumption can be reduced by decreasing the number of antennas in the array, while keeping the aperture constant. This reduction in antenna count is achieved by randomly depopulating the array, resulting in a sparse antenna array. A novel compressive sensing algorithm, coupled with the wavelet transform, is used to process the samples collected by the sparse array and form a 3-D image of the scene. This algorithm works well for antenna arrays that are up to 96% sparse, equating to a 25 times reduction in the number of required antennas. For microwave imaging to be useful, it needs to capture images of the scene in real time. The architecture of a system capable of capturing real-time 3-D microwave images is therefore designed. The system consists of a modular antenna array, constructed by plugging RF daughtercards into a carrier board. Each daughtercard is a self-contained radio system, containing an antenna, RF transceiver baseband signal chain, and analog-to-digital converters. A small number of daughtercards have been built, and proven to be suitable for real-time microwave imaging. By arranging these daughtercards in different ways, any antenna array pattern can be built. This architecture allows real-time microwave imaging systems to be rapidly prototyped, while still being able to generate images at video frame rates.
Giaddui, Tawfik; Yu, Jialu; Manfredi, Denise; Linnemann, Nancy; Hunter, Joanne; O’Meara, Elizabeth; Galvin, James; Bialecki, Brian; Xiao, Ying
2016-01-01
Transmission of Imaging and Data (TRIAD) is a standard-based system built by the American College of Radiology (ACR) to provide seamless exchange of images and data for accreditation of clinical trials and registries. Scripts of structures’ names validation profiles created in TRIAD are used in the automated submission process. It is essential for users to understand the logistics of these scripts for successful submission of radiotherapy cases with less iteration. PMID:27053498
Natural image classification driven by human brain activity
NASA Astrophysics Data System (ADS)
Zhang, Dai; Peng, Hanyang; Wang, Jinqiao; Tang, Ming; Xue, Rong; Zuo, Zhentao
2016-03-01
Natural image classification has been a hot topic in computer vision and pattern recognition research field. Since the performance of an image classification system can be improved by feature selection, many image feature selection methods have been developed. However, the existing supervised feature selection methods are typically driven by the class label information that are identical for different samples from the same class, ignoring with-in class image variability and therefore degrading the feature selection performance. In this study, we propose a novel feature selection method, driven by human brain activity signals collected using fMRI technique when human subjects were viewing natural images of different categories. The fMRI signals associated with subjects viewing different images encode the human perception of natural images, and therefore may capture image variability within- and cross- categories. We then select image features with the guidance of fMRI signals from brain regions with active response to image viewing. Particularly, bag of words features based on GIST descriptor are extracted from natural images for classification, and a sparse regression base feature selection method is adapted to select image features that can best predict fMRI signals. Finally, a classification model is built on the select image features to classify images without fMRI signals. The validation experiments for classifying images from 4 categories of two subjects have demonstrated that our method could achieve much better classification performance than the classifiers built on image feature selected by traditional feature selection methods.
High frame-rate MR-guided near-infrared tomography system to monitor breast hemodynamics
NASA Astrophysics Data System (ADS)
Li, Zhiqiu; Jiang, Shudong; Krishnaswamy, Venkataramanan; Davis, Scott C.; Srinivasan, Subhadra; Paulsen, Keith D.; Pogue, Brian W.
2011-02-01
A near-infrared (NIR) tomography system with spectral-encoded sources at two wavelength bands was built to quantify the temporal contrast at 20 Hz bandwidth, while imaging breast tissue. The NIR system was integrated with a magnetic resonance (MR) machine through a custom breast coil interface, and both NIR data and MR images were acquired simultaneously. MR images provided breast tissue structural information for NIR reconstruction. Acquisition of finger pulse oximeter (PO) plethysmogram was synchronized with the NIR system in the experiment to offer a frequency-locked reference. The recovered absorption coefficients of the breast at two wavelengths showed identical temporal frequency as the PO output, proving this multi-modality design can recover the small pulsatile variation of absorption property in breast tissue related to the heartbeat. And it also showed the system's ability on novel contrast imaging of fast flow signals in deep tissue.
Improvements in Speed and Functionality of a 670-GHz Imaging Radar
NASA Technical Reports Server (NTRS)
Dengler, Robert J.; Cooper, Ken B.; Mehdi, Imran; Siegel, Peter H.; Tarsala, Jan A.; Bryllert, Thomas E.
2011-01-01
Significant improvements have been made in the instrument originally described in a prior NASA Tech Briefs article: Improved Speed and Functionality of a 580-GHz Imaging Radar (NPO-45156), Vol. 34, No. 7 (July 2010), p. 51. First, the wideband YIG oscillator has been replaced with a JPL-designed and built phase-locked, low-noise chirp source. Second, further refinements to the data acquisition and signal processing software have been performed by moving critical code sections to C code, and compiling those sections to Windows DLLs, which are then invoked from the main LabVIEW executive. This system is an active, single-pixel scanned imager operating at 670 GHz. The actual chirp signals for the RF and LO chains were generated by a pair of MITEQ 2.5 3.3 GHz chirp sources. Agilent benchtop synthesizers operating at fixed frequencies around 13 GHz were then used to up-convert the chirp sources to 15.5 16.3 GHz. The resulting signals were then multiplied 36 times by a combination of off-the-shelf millimeter- wave components, and JPL-built 200- GHz doublers and 300- and 600-GHz triplers. The power required to drive the submillimeter-wave multipliers was provided by JPL-built W-band amplifiers. The receive and transmit signal paths were combined using a thin, high-resistivity silicon wafer as a beam splitter. While the results at present are encouraging, the system still lacks sufficient speed to be usable for practical applications in a contraband detection. Ideally, an image acquisition speed of ten seconds, or a factor of 30 improvement, is desired. However, the system improvements to date have resulted in a factor of five increase in signal acquisition speed, as well as enhanced signal processing algorithms, permitting clearer imaging of contraband objects hidden underneath clothing. In particular, advances in three distinct areas have enabled these performance enhancements: base source phase noise reduction, chirp rate, and signal processing. Additionally, a second pixel was added, automatically reducing the imaging time by a factor of two. Although adding a second pixel to the system doubles the amount of submillimeter components required, some savings in microwave hardware can be realized by using a common low-noise source.
Table-driven image transformation engine algorithm
NASA Astrophysics Data System (ADS)
Shichman, Marc
1993-04-01
A high speed image transformation engine (ITE) was designed and a prototype built for use in a generic electronic light table and image perspective transformation application code. The ITE takes any linear transformation, breaks the transformation into two passes and resamples the image appropriately for each pass. The system performance is achieved by driving the engine with a set of look up tables computed at start up time for the calculation of pixel output contributions. Anti-aliasing is done automatically in the image resampling process. Operations such as multiplications and trigonometric functions are minimized. This algorithm can be used for texture mapping, image perspective transformation, electronic light table, and virtual reality.
Swami, Viren; Barron, David; Furnham, Adrian
2018-03-01
Five studies were conducted to understand the impact of nature exposure on body image. In three studies using different designs and outcome measures, British university students were exposed to photographs of natural or built environments. Results indicated that exposure to images of natural, but not built, environments resulted in improved state body image. In Study 4, British community participants went on a walk in a natural or built environment, with results indicating that the walk in a natural environment resulted in significantly higher state body appreciation, whereas the walk in a built environment resulted in significantly lower scores. In Study 5, British participants were recruited as they were entering a designed green space on their own volition. Results indicated that spending time in the green space led to improved state body appreciation. These results indicate that exposure to isomorphic or in-situ natural environments has positive effects on state body image. Copyright © 2017 Elsevier Ltd. All rights reserved.
Millimeter-wave imaging sensor data evaluation
NASA Technical Reports Server (NTRS)
Wilson, William J.; Ibbott, Anthony C.
1987-01-01
A passive 3-mm radiometer system with a mechanically scanned antenna was built for use on a small aircraft or an Unmanned Aerial Vehicle to produce real near-real-time, moderate-resolution (0.5) images of the ground. One of the main advantages of this passive imaging sensor is that it is able to provide surveillance information through dust, smoke, fog and clouds when visual and IR systems are unusable. It can also be used for a variety of remote sensing applications, such as measurements of surface moisture, surface temperature, vegetation extent and snow cover. It is also possible to detect reflective objects under vegetation cover.
Adapting smartphones for low-cost optical medical imaging
NASA Astrophysics Data System (ADS)
Pratavieira, Sebastião.; Vollet-Filho, José D.; Carbinatto, Fernanda M.; Blanco, Kate; Inada, Natalia M.; Bagnato, Vanderlei S.; Kurachi, Cristina
2015-06-01
Optical images have been used in several medical situations to improve diagnosis of lesions or to monitor treatments. However, most systems employ expensive scientific (CCD or CMOS) cameras and need computers to display and save the images, usually resulting in a high final cost for the system. Additionally, this sort of apparatus operation usually becomes more complex, requiring more and more specialized technical knowledge from the operator. Currently, the number of people using smartphone-like devices with built-in high quality cameras is increasing, which might allow using such devices as an efficient, lower cost, portable imaging system for medical applications. Thus, we aim to develop methods of adaptation of those devices to optical medical imaging techniques, such as fluorescence. Particularly, smartphones covers were adapted to connect a smartphone-like device to widefield fluorescence imaging systems. These systems were used to detect lesions in different tissues, such as cervix and mouth/throat mucosa, and to monitor ALA-induced protoporphyrin-IX formation for photodynamic treatment of Cervical Intraepithelial Neoplasia. This approach may contribute significantly to low-cost, portable and simple clinical optical imaging collection.
Spectrally-encoded color imaging
Kang, DongKyun; Yelin, Dvir; Bouma, Brett E.; Tearney, Guillermo J.
2010-01-01
Spectrally-encoded endoscopy (SEE) is a technique for ultraminiature endoscopy that encodes each spatial location on the sample with a different wavelength. One limitation of previous incarnations of SEE is that it inherently creates monochromatic images, since the spectral bandwidth is expended in the spatial encoding process. Here we present a spectrally-encoded imaging system that has color imaging capability. The new imaging system utilizes three distinct red, green, and blue spectral bands that are configured to illuminate the grating at different incident angles. By careful selection of the incident angles, the three spectral bands can be made to overlap on the sample. To demonstrate the method, a bench-top system was built, comprising a 2400-lpmm grating illuminated by three 525-μm-diameter beams with three different spectral bands. Each spectral band had a bandwidth of 75 nm, producing 189 resolvable points. A resolution target, color phantoms, and excised swine small intestine were imaged to validate the system's performance. The color SEE system showed qualitatively and quantitatively similar color imaging performance to that of a conventional digital camera. PMID:19688002
Polygon Patterned Ground on Mars and on Earth
NASA Technical Reports Server (NTRS)
2008-01-01
Some high-latitude areas on Mars (left) and Earth (right) exhibit similarly patterned ground where shallow fracturing has drawn polygons on the surface. This patterning may result from cycles of contraction and expansion. The left image shows ground within the targeted landing area NASA's Phoenix Mars Lander before the winter frost had entirely disappeared from the surface. The bright ice in shallow crevices accentuates the area's polygonal fracturing pattern. The polygons are a few meters (several feet) across. The image is a small portion of an exposure taken in March 2008 by the High Resolution Imaging Science Experiment (HiRISE) camera on NASA's Mars Reconnaissance Orbiter. The image on the right is an aerial view of similarly patterned ground in Antarctica. The Phoenix Mission is led by the University of Arizona on behalf of NASA. Project management of the mission is by NASA's Jet Propulsion Laboratory. Spacecraft development is by Lockheed Martin Space Systems. NASA's Jet Propulsion Laboratory, a division of the California Institute of Technology in Pasadena, manages the Mars Reconnaissance Orbiter for NASA's Science Mission Directorate, Washington. Lockheed Martin Space Systems, Denver, is the prime contractor for the project and built the spacecraft. The High Resolution Imaging Science Experiment is operated by the University of Arizona, Tucson, and the instrument was built by Ball Aerospace & Technologies Corp., Boulder, Colo.THz computed tomography system with zero-order Bessel beam
NASA Astrophysics Data System (ADS)
Niu, Liting; Wu, Qiao; Wang, Kejia; Liu, Jinsong; Yang, Zhengang
2018-01-01
Terahertz (THz) waves can penetrate many optically opaque dielectric materials such as plastics, ceramics and colorants. It is effective to reveal the internal structures of these materials. We have built a THz Computed Tomography (CT) system with 0.3 THz zero-order Bessel beam to improve the depth of focus of this imaging system for the non-diffraction property of Bessel beam. The THz CT system has been used to detect a paper cup with a metal rod inside. Finally, the acquired projection data have been processed by the filtered back-projection algorithm and the reconstructed image of the sample has been obtained.
Electron beam throughput from raster to imaging
NASA Astrophysics Data System (ADS)
Zywno, Marek
2016-12-01
Two architectures of electron beam tools are presented: single beam MEBES Exara designed and built by Etec Systems for mask writing, and the Reflected E-Beam Lithography tool (REBL), designed and built by KLA-Tencor under a DARPA Agreement No. HR0011-07-9-0007. Both tools have implemented technologies not used before to achieve their goals. The MEBES X, renamed Exara for marketing purposes, used an air bearing stage running in vacuum to achieve smooth continuous scanning. The REBL used 2 dimensional imaging to distribute charge to a 4k pixel swath to achieve writing times on the order of 1 wafer per hour, scalable to throughput approaching optical projection tools. Three stage architectures were designed for continuous scanning of wafers: linear maglev, rotary maglev, and dual linear maglev.
Automated alignment system for optical wireless communication systems using image recognition.
Brandl, Paul; Weiss, Alexander; Zimmermann, Horst
2014-07-01
In this Letter, we describe the realization of a tracked line-of-sight optical wireless communication system for indoor data distribution. We built a laser-based transmitter with adaptive focus and ray steering by a microelectromechanical systems mirror. To execute the alignment procedure, we used a CMOS image sensor at the transmitter side and developed an algorithm for image recognition to localize the receiver's position. The receiver is based on a self-developed optoelectronic integrated chip with low requirements on the receiver optics to make the system economically attractive. With this system, we were able to set up the communication link automatically without any back channel and to perform error-free (bit error rate <10⁻⁹) data transmission over a distance of 3.5 m with a data rate of 3 Gbit/s.
Detecting breast microcalcifications using super-resolution ultrasound imaging: a clinical study
NASA Astrophysics Data System (ADS)
Huang, Lianjie; Labyed, Yassin; Hanson, Kenneth; Sandoval, Daniel; Pohl, Jennifer; Williamson, Michael
2013-03-01
Imaging breast microcalcifications is crucial for early detection and diagnosis of breast cancer. It is challenging for current clinical ultrasound to image breast microcalcifications. However, new imaging techniques using data acquired with a synthetic-aperture ultrasound system have the potential to significantly improve ultrasound imaging. We recently developed a super-resolution ultrasound imaging method termed the phase-coherent multiple-signal classification (PC-MUSIC). This signal subspace method accounts for the phase response of transducer elements to improve image resolution. In this paper, we investigate the clinical feasibility of our super-resolution ultrasound imaging method for detecting breast microcalcifications. We use our custom-built, real-time synthetic-aperture ultrasound system to acquire breast ultrasound data for 40 patients whose mammograms show the presence of breast microcalcifications. We apply our super-resolution ultrasound imaging method to the patient data, and produce clear images of breast calcifications. Our super-resolution ultrasound PC-MUSIC imaging with synthetic-aperture ultrasound data can provide a new imaging modality for detecting breast microcalcifications in clinic without using ionizing radiation.
NASA Astrophysics Data System (ADS)
Michaelis, A.; Nemani, R. R.; Wang, W.; Votava, P.; Hashimoto, H.
2010-12-01
Given the increasing complexity of climate modeling and analysis tools, it is often difficult and expensive to build or recreate an exact replica of the software compute environment used in past experiments. With the recent development of new technologies for hardware virtualization, an opportunity exists to create full modeling, analysis and compute environments that are “archiveable”, transferable and may be easily shared amongst a scientific community or presented to a bureaucratic body if the need arises. By encapsulating and entire modeling and analysis environment in a virtual machine image, others may quickly gain access to the fully built system used in past experiments, potentially easing the task and reducing the costs of reproducing and verify past results produced by other researchers. Moreover, these virtual machine images may be used as a pedagogical tool for others that are interested in performing an academic exercise but don't yet possess the broad expertise required. We built two virtual machine images, one with the Community Earth System Model (CESM) and one with Weather Research Forecast Model (WRF), then ran several small experiments to assess the feasibility, performance overheads costs, reusability, and transferability. We present a list of the pros and cons as well as lessoned learned from utilizing virtualization technology in the climate and earth systems modeling domain.
On-Wafer Measurement of a Silicon-Based CMOS VCO at 324 GHz
NASA Technical Reports Server (NTRS)
Samoska, Lorene; Man Fung, King; Gaier, Todd; Huang, Daquan; Larocca, Tim; Chang, M. F.; Campbell, Richard; Andrews, Michael
2008-01-01
The world s first silicon-based complementary metal oxide/semiconductor (CMOS) integrated-circuit voltage-controlled oscillator (VCO) operating in a frequency range around 324 GHz has been built and tested. Concomitantly, equipment for measuring the performance of this oscillator has been built and tested. These accomplishments are intermediate steps in a continuing effort to develop low-power-consumption, low-phase-noise, electronically tunable signal generators as local oscillators for heterodyne receivers in submillimeter-wavelength (frequency > 300 GHz) scientific instruments and imaging systems. Submillimeter-wavelength imaging systems are of special interest for military and law-enforcement use because they could, potentially, be used to detect weapons hidden behind clothing and other opaque dielectric materials. In comparison with prior submillimeter- wavelength signal generators, CMOS VCOs offer significant potential advantages, including great reductions in power consumption, mass, size, and complexity. In addition, there is potential for on-chip integration of CMOS VCOs with other CMOS integrated circuitry, including phase-lock loops, analog- to-digital converters, and advanced microprocessors.
Design of the first optical system for real-time tomographic holography (RTTH)
NASA Astrophysics Data System (ADS)
Galeotti, John M.; Siegel, Mel; Rallison, Richard D.; Stetten, George
2008-08-01
The design of the first Real-Time-Tomographic-Holography (RTTH) optical system for augmented-reality applications is presented. RTTH places a viewpoint-independent real-time (RT) virtual image (VI) of an object into its actual location, enabling natural hand-eye coordination to guide invasive procedures, without requiring tracking or a head-mounted device. The VI is viewed through a narrow-band Holographic Optical Element (HOE) with built-in power that generates the largest possible near-field, in-situ VI from a small display chip without noticeable parallax error or obscuring direct view of the physical world. Rigidly fixed upon a medical-ultrasound probe, RTTH could show the scan in its actual location inside the patient, because the VI would move with the probe. We designed the image source along with the system-optics, allowing us to ignore both planer geometric distortions and field curvature, respectively compensated by using RT pre-processing software and attaching a custom-surfaced fiber-optic-faceplate (FOFP) to our image source. Focus in our fast, non-axial system was achieved by placing correcting lenses near the FOFP and custom-optically-fabricating our volume-phase HOE using a recording beam that was specially shaped by extra lenses. By simultaneously simulating and optimizing the system's playback performance across variations in both the total playback and HOE-recording optical systems, we derived and built a design that projects a 104x112 mm planar VI 1 m from the HOE using a laser-illuminated 19x16 mm LCD+FOFP image-source. The VI appeared fixed in space and well focused. Viewpoint-induced location errors were <3 mm, and unexpected first-order astigmatism produced 3 cm (3% of 1 m) ambiguity in depth, typically unnoticed by human observers.
Szarka, Mate; Guttman, Andras
2017-10-17
We present the application of a smartphone anatomy based technology in the field of liquid phase bioseparations, particularly in capillary electrophoresis. A simple capillary electrophoresis system was built with LED induced fluorescence detection and a credit card sized minicomputer to prove the concept of real time fluorescent imaging (zone adjustable time-lapse fluorescence image processor) and separation controller. The system was evaluated by analyzing under- and overloaded aminopyrenetrisulfonate (APTS)-labeled oligosaccharide samples. The open source software based image processing tool allowed undistorted signal modulation (reprocessing) if the signal was inappropriate for the actual detection system settings (too low or too high). The novel smart detection tool for fluorescently labeled biomolecules greatly expands dynamic range and enables retrospective correction for injections with unsuitable signal levels without the necessity to repeat the analysis.
Standoff passive video imaging at 350 GHz with 251 superconducting detectors
NASA Astrophysics Data System (ADS)
Becker, Daniel; Gentry, Cale; Smirnov, Ilya; Ade, Peter; Beall, James; Cho, Hsiao-Mei; Dicker, Simon; Duncan, William; Halpern, Mark; Hilton, Gene; Irwin, Kent; Li, Dale; Paulter, Nicholas; Reintsema, Carl; Schwall, Robert; Tucker, Carole
2014-06-01
Millimeter wavelength radiation holds promise for detection of security threats at a distance, including suicide bomb belts and maritime threats in poor weather. The high sensitivity of superconducting Transition Edge Sensor (TES) detectors makes them ideal for passive imaging of thermal signals at these wavelengths. We have built a 350 GHz video-rate imaging system using a large-format array of feedhorn-coupled TES bolometers. The system operates at a standoff distance of 16m to 28m with a spatial resolution of 1:4 cm (at 17m). It currently contains one 251-detector subarray, and will be expanded to contain four subarrays for a total of 1004 detectors. The system has been used to take video images which reveal the presence of weapons concealed beneath a shirt in an indoor setting. We present a summary of this work.
Review on short-wavelength infrared laser gated-viewing at Fraunhofer IOSB
NASA Astrophysics Data System (ADS)
Göhler, Benjamin; Lutzmann, Peter
2017-03-01
This paper reviews the work that has been done at Fraunhofer IOSB (and its predecessor institutes) in the past ten years in the area of laser gated-viewing (GV) in the short-wavelength infrared (SWIR) band. Experimental system demonstrators in various configurations have been built up to show the potential for different applications and to investigate specific topics. The wavelength of the pulsed illumination laser is 1.57 μm and lies in the invisible, retina-safe region allowing much higher pulse energies than for wavelengths in the visible or near-infrared band concerning eye safety. All systems built up, consist of gated Intevac LIVAR® cameras based on EBCCD/EBCMOS detectors sensitive in the SWIR band. This review comprises military and civilian applications in maritime and land domain-in particular vision enhancement in bad visibility, long-range applications, silhouette imaging, 3-D imaging by sliding gates and slope method, bistatic GV imaging, and looking through windows. In addition, theoretical studies that were conducted-e.g., estimating 3-D accuracy or modeling range performance-are presented. Finally, an outlook for future work in the area of SWIR laser GV at Fraunhofer IOSB is given.
Affordable Imaging Lab for Noninvasive Analysis of Biomass and Early Vigour in Cereal Crops
2018-01-01
Plant phenotyping by imaging allows automated analysis of plants for various morphological and physiological traits. In this work, we developed a low-cost RGB imaging phenotyping lab (LCP lab) for low-throughput imaging and analysis using affordable imaging equipment and freely available software. LCP lab comprising RGB imaging and analysis pipeline is set up and demonstrated with early vigour analysis in wheat. Using this lab, a few hundred pots can be photographed in a day and the pots are tracked with QR codes. The software pipeline for both imaging and analysis is built from freely available software. The LCP lab was evaluated for early vigour analysis of five wheat cultivars. A high coefficient of determination (R2 0.94) was obtained between the dry weight and the projected leaf area of 20-day-old wheat plants and R2 of 0.9 for the relative growth rate between 10 and 20 days of plant growth. Detailed description for setting up such a lab is provided together with custom scripts built for imaging and analysis. The LCP lab is an affordable alternative for analysis of cereal crops when access to a high-throughput phenotyping facility is unavailable or when the experiments require growing plants in highly controlled climate chambers. The protocols described in this work are useful for building affordable imaging system for small-scale research projects and for education. PMID:29850536
NASA Technical Reports Server (NTRS)
2008-01-01
[figure removed for brevity, see original site] Click on the image for movie of Zooming in on Landing Site This animation zooms in on the area on Mars where NASA's Phoenix Mars Lander will touchdown on May 25, 2008. The image was taken by the High Resolution Imaging Science Experiment (HiRISE) camera on NASA's Mars Reconnaissance Orbiter. The first shot shows the spacecraft's landing ellipse in green, the area where Phoenix has a high probability of landing. It then zooms in to show the region's arctic terrain. This polar landscape is relatively free of rocks, with only about 1 to 2 rocks 1.5 meters (4.9 feet) or larger in an area about as big as two football fields. NASA's Jet Propulsion Laboratory, a division of the California Institute of Technology in Pasadena, manages the Mars Reconnaissance Orbiter for NASA's Science Mission Directorate, Washington. Lockheed Martin Space Systems, Denver, is the prime contractor for the project and built the spacecraft. The High Resolution Imaging Science Experiment is operated by the University of Arizona, Tucson, and the instrument was built by Ball Aerospace & Technologies Corp., Boulder, Colo.Spirit's Tracks around 'Home Plate'
NASA Technical Reports Server (NTRS)
2006-01-01
[figure removed for brevity, see original site] Annotated Version This portion of an image acquired by the Mars Reconnaissance Orbiter's High Resolution Imaging Science Experiment camera shows the Spirit rover's winter campaign site. The rover is visible. So is the 'Low Ridge' feature where Spirit was parked with an 11-degree northerly tilt to maximize sunlight on the solar panels during the southern winter season. Tracks made by Spirit on the way to 'Home Plate' and to and from 'Tyrone,' an area of light-toned soils exposed by rover wheel motions, are also evident. The original image is catalogued as PSP_001513_1655_red and was taken Sept. 29, 2006. NASA's Jet Propulsion Laboratory, a division of the California Institute of Technology in Pasadena, manages the Mars Reconnaissance Orbiter for NASA's Science Mission Directorate, Washington. Lockheed Martin Space Systems, Denver, is the prime contractor for the project and built the spacecraft. The High Resolution Imaging Science Experiment is operated by the University of Arizona, Tucson, and the instrument was built by Ball Aerospace and Technology Corp., Boulder, Colo.Clinical experiences of collaborative imaging diagnosis in Shanghai district healthcare services
NASA Astrophysics Data System (ADS)
Zhang, Kai; Ling, Tonghui; Yang, Yuanyuan; Sun, Jianyong; Wang, Mingqing; Zhang, Jianguo
2016-03-01
To improve healthcare service quality with balancing healthcare resources between large and small hospitals, as well as reducing costs, each district health administration in Shanghai with more than 24 million citizens has built image-enabled electronic healthcare records (iEHR) system to share patient medical records and encourage patients to visit small hospitals for initial evaluations and preliminary diagnoses first, then go to large hospitals to have better specialists' services. We implemented solution for iEHR systems, based on the IHE XDS-I integration profile and installed the systems in more than 100 hospitals cross three districts in Shanghai and one city in Jiangsu Province in last few years. Here, we give operational results of these systems in these four districts and evaluated the performance of the systems in servicing the regional collaborative imaging diagnosis.
An implementation of wireless medical image transmission system on mobile devices.
Lee, SangBock; Lee, Taesoo; Jin, Gyehwan; Hong, Juhyun
2008-12-01
The advanced technology of computing system was followed by the rapid improvement of medical instrumentation and patient record management system. The typical examples are hospital information system (HIS) and picture archiving and communication system (PACS), which computerized the management procedure of medical records and images in hospital. Because these systems were built and used in hospitals, doctors out of hospital have problems to access them immediately on emergent cases. To solve these problems, this paper addressed the realization of system that could transmit the images acquired by medical imaging systems in hospital to the remote doctors' handheld PDA's using CDMA cellular phone network. The system consists of server and PDA. The server was developed to manage the accounts of doctors and patients and allocate the patient images to each doctor. The PDA was developed to display patient images through remote server connection. To authenticate the personal user, remote data access (RDA) method was used in PDA accessing the server database and file transfer protocol (FTP) was used to download patient images from the remove server. In laboratory experiments, it was calculated to take ninety seconds to transmit thirty images with 832 x 488 resolution and 24 bit depth and 0.37 Mb size. This result showed that the developed system has no problems for remote doctors to receive and review the patient images immediately on emergent cases.
Wideband radar for airborne minefield detection
NASA Astrophysics Data System (ADS)
Clark, William W.; Burns, Brian; Dorff, Gary; Plasky, Brian; Moussally, George; Soumekh, Mehrdad
2006-05-01
Ground Penetrating Radar (GPR) has been applied for several years to the problem of detecting both antipersonnel and anti-tank landmines. RDECOM CERDEC NVESD is developing an airborne wideband GPR sensor for the detection of minefields including surface and buried mines. In this paper, we describe the as-built system, data and image processing techniques to generate imagery, and current issues with this type of radar. Further, we will display images from a recent field test.
Gradiometer Using Middle Loops as Sensing Elements in a Low-Field SQUID MRI System
NASA Technical Reports Server (NTRS)
Penanen, Konstantin; Hahn, Inseob; Ho Eom, Byeong
2009-01-01
A new gradiometer scheme uses middle loops as sensing elements in lowfield superconducting quantum interference device (SQUID) magnetic resonance imaging (MRI). This design of a second order gradiometer increases its sensitivity and makes it more uniform, compared to the conventional side loop sensing scheme with a comparable matching SQUID. The space between the two middle loops becomes the imaging volume with the enclosing cryostat built accordingly.
Computer-based route-definition system for peripheral bronchoscopy.
Graham, Michael W; Gibbs, Jason D; Higgins, William E
2012-04-01
Multi-detector computed tomography (MDCT) scanners produce high-resolution images of the chest. Given a patient's MDCT scan, a physician can use an image-guided intervention system to first plan and later perform bronchoscopy to diagnostic sites situated deep in the lung periphery. An accurate definition of complete routes through the airway tree leading to the diagnostic sites, however, is vital for avoiding navigation errors during image-guided bronchoscopy. We present a system for the robust definition of complete airway routes suitable for image-guided bronchoscopy. The system incorporates both automatic and semiautomatic MDCT analysis methods for this purpose. Using an intuitive graphical user interface, the user invokes automatic analysis on a patient's MDCT scan to produce a series of preliminary routes. Next, the user visually inspects each route and quickly corrects the observed route defects using the built-in semiautomatic methods. Application of the system to a human study for the planning and guidance of peripheral bronchoscopy demonstrates the efficacy of the system.
Flavel, Richard J; Guppy, Chris N; Rabbi, Sheikh M R; Young, Iain M
2017-01-01
The objective of this study was to develop a flexible and free image processing and analysis solution, based on the Public Domain ImageJ platform, for the segmentation and analysis of complex biological plant root systems in soil from x-ray tomography 3D images. Contrasting root architectures from wheat, barley and chickpea root systems were grown in soil and scanned using a high resolution micro-tomography system. A macro (Root1) was developed that reliably identified with good to high accuracy complex root systems (10% overestimation for chickpea, 1% underestimation for wheat, 8% underestimation for barley) and provided analysis of root length and angle. In-built flexibility allowed the user interaction to (a) amend any aspect of the macro to account for specific user preferences, and (b) take account of computational limitations of the platform. The platform is free, flexible and accurate in analysing root system metrics.
Design and applications of a multimodality image data warehouse framework.
Wong, Stephen T C; Hoo, Kent Soo; Knowlton, Robert C; Laxer, Kenneth D; Cao, Xinhau; Hawkins, Randall A; Dillon, William P; Arenson, Ronald L
2002-01-01
A comprehensive data warehouse framework is needed, which encompasses imaging and non-imaging information in supporting disease management and research. The authors propose such a framework, describe general design principles and system architecture, and illustrate a multimodality neuroimaging data warehouse system implemented for clinical epilepsy research. The data warehouse system is built on top of a picture archiving and communication system (PACS) environment and applies an iterative object-oriented analysis and design (OOAD) approach and recognized data interface and design standards. The implementation is based on a Java CORBA (Common Object Request Broker Architecture) and Web-based architecture that separates the graphical user interface presentation, data warehouse business services, data staging area, and backend source systems into distinct software layers. To illustrate the practicality of the data warehouse system, the authors describe two distinct biomedical applications--namely, clinical diagnostic workup of multimodality neuroimaging cases and research data analysis and decision threshold on seizure foci lateralization. The image data warehouse framework can be modified and generalized for new application domains.
Design and Applications of a Multimodality Image Data Warehouse Framework
Wong, Stephen T.C.; Hoo, Kent Soo; Knowlton, Robert C.; Laxer, Kenneth D.; Cao, Xinhau; Hawkins, Randall A.; Dillon, William P.; Arenson, Ronald L.
2002-01-01
A comprehensive data warehouse framework is needed, which encompasses imaging and non-imaging information in supporting disease management and research. The authors propose such a framework, describe general design principles and system architecture, and illustrate a multimodality neuroimaging data warehouse system implemented for clinical epilepsy research. The data warehouse system is built on top of a picture archiving and communication system (PACS) environment and applies an iterative object-oriented analysis and design (OOAD) approach and recognized data interface and design standards. The implementation is based on a Java CORBA (Common Object Request Broker Architecture) and Web-based architecture that separates the graphical user interface presentation, data warehouse business services, data staging area, and backend source systems into distinct software layers. To illustrate the practicality of the data warehouse system, the authors describe two distinct biomedical applications—namely, clinical diagnostic workup of multimodality neuroimaging cases and research data analysis and decision threshold on seizure foci lateralization. The image data warehouse framework can be modified and generalized for new application domains. PMID:11971885
NASA Astrophysics Data System (ADS)
Boadi, J.; Sangwal, V.; MacNeil, S.; Matcher, S. J.
2015-03-01
The prevailing hypothesis for the existence and healing of the avascular corneal epithelium is that this layer of cells is continually produced by stem cells in the limbus and transported onto the cornea to mature into corneal epithelium. Limbal Stem Cell Deficiency (LSCD), in which the stem cell population is depleted, can lead to blindness. LSCD can be caused by chemical and thermal burns to the eye. A popular treatment, especially in emerging economies such as India, is the transplantation of limbal stem cells onto damaged limbus with hope of repopulating the region. Hence regenerating the corneal epithelium. In order to gain insights into the success rates of this treatment, new imaging technologies are needed in order to track the transplanted cells. Optical Coherence Tomography (OCT) is well known for its high resolution in vivo images of the retina. A custom OCT system has been built to image the corneal surface, to investigate the fate of transplanted limbal stem cells. We evaluate two methods to label and track transplanted cells: melanin labelling and magneto-labelling. To evaluate melanin labelling, stem cells are loaded with melanin and then transplanted onto a rabbit cornea denuded of its epithelium. The melanin displays strongly enhanced backscatter relative to normal cells. To evaluate magneto-labelling the stem cells are loaded with magnetic nanoparticles (20-30nm in size) and then imaged with a custom-built, magneto-motive OCT system.
NASA Technical Reports Server (NTRS)
2006-01-01
[figure removed for brevity, see original site] Annotated Version This portion of an image acquired by the Mars Reconnaissance Orbiter's High Resolution Imaging Science Experiment camera shows the Spirit rover's winter campaign site. Spirit was parked on a slope tilted 11 degrees to the north to maximize sunlight during the southern winter season. 'Tyrone' is an area where the rover's wheels disturbed light-toned soils. Remote sensing and in-situ analyses found the light-toned soil at Tyrone to be sulfate rich and hydrated. The original picture is catalogued as PSP_001513_1655_red and was taken on Sept. 29, 2006. NASA's Jet Propulsion Laboratory, a division of the California Institute of Technology in Pasadena, manages the Mars Reconnaissance Orbiter for NASA's Science Mission Directorate, Washington. Lockheed Martin Space Systems, Denver, is the prime contractor for the project and built the spacecraft. The High Resolution Imaging Science Experiment is operated by the University of Arizona, Tucson, and the instrument was built by Ball Aerospace and Technology Corp., Boulder, Colo.Improved microgrid arrangement for integrated imaging polarimeters.
LeMaster, Daniel A; Hirakawa, Keigo
2014-04-01
For almost 20 years, microgrid polarimetric imaging systems have been built using a 2×2 repeating pattern of polarization analyzers. In this Letter, we show that superior spatial resolution is achieved over this 2×2 case when the analyzers are arranged in a 2×4 repeating pattern. This unconventional result, in which a more distributed sampling pattern results in finer spatial resolution, is also achieved without affecting the conditioning of the polarimetric data-reduction matrix. Proof is provided theoretically and through Stokes image reconstruction of synthesized data.
Li, Yang; Ma, Jianguo; Martin, K Heath; Yu, Mingyue; Ma, Teng; Dayton, Paul A; Jiang, Xiaoning; Shung, K Kirk; Zhou, Qifa
2016-09-01
Superharmonic contrast-enhanced ultrasound imaging, also called acoustic angiography, has previously been used for the imaging of microvasculature. This approach excites microbubble contrast agents near their resonance frequency and receives echoes at nonoverlapping superharmonic bandwidths. No integrated system currently exists could fully support this application. To fulfill this need, an integrated dual-channel transmit/receive system for superharmonic imaging was designed, built, and characterized experimentally. The system was uniquely designed for superharmonic imaging and high-resolution B-mode imaging. A complete ultrasound system including a pulse generator, a data acquisition unit, and a signal processing unit were integrated into a single package. The system was controlled by a field-programmable gate array, on which multiple user-defined modes were implemented. A 6-, 35-MHz dual-frequency dual-element intravascular ultrasound transducer was designed and used for imaging. The system successfully obtained high-resolution B-mode images of coronary artery ex vivo with 45-dB dynamic range. The system was capable of acquiring in vitro superharmonic images of a vasa vasorum mimicking phantom with 30-dB contrast. It could detect a contrast agent filled tissue mimicking tube of 200 μm diameter. For the first time, high-resolution B-mode images and superharmonic images were obtained in an intravascular phantom, made possible by the dedicated integrated system proposed. The system greatly reduced the cost and complexity of the superharmonic imaging intended for preclinical study. Significant: The system showed promise for high-contrast intravascular microvascular imaging, which may have significant importance in assessment of the vasa vasorum associated with atherosclerotic plaques.
An information based approach to improving overhead imagery collection
NASA Astrophysics Data System (ADS)
Sourwine, Matthew J.; Hintz, Kenneth J.
2011-06-01
Recent growth in commercial imaging satellite development has resulted in a complex and diverse set of systems. To simplify this environment for both customer and vendor, an information based sensor management model was built to integrate tasking and scheduling systems. By establishing a relationship between image quality and information, tasking by NIIRS can be utilized to measure the customer's required information content. Focused on a reduction in uncertainty about a target of interest, the sensor manager finds the best sensors to complete the task given the active suite of imaging sensors' functions. This is done through determination of which satellite will meet customer information and timeliness requirements with low likelihood of interference at the highest rate of return.
Zhao, Ming; Li, Yu; Peng, Leilei
2014-05-05
We present a novel excitation-emission multiplexed fluorescence lifetime microscopy (FLIM) method that surpasses current FLIM techniques in multiplexing capability. The method employs Fourier multiplexing to simultaneously acquire confocal fluorescence lifetime images of multiple excitation wavelength and emission color combinations at 44,000 pixels/sec. The system is built with low-cost CW laser sources and standard PMTs with versatile spectral configuration, which can be implemented as an add-on to commercial confocal microscopes. The Fourier lifetime confocal method allows fast multiplexed FLIM imaging, which makes it possible to monitor multiple biological processes in live cells. The low cost and compatibility with commercial systems could also make multiplexed FLIM more accessible to biological research community.
Electronic magnification for astronomical camera tubes
NASA Technical Reports Server (NTRS)
Vine, J.; Hansen, J. R.; Pietrzyk, J. P.
1974-01-01
Definitions, test schemes, and analyses used to provide variable magnification in the image section of the television sensor for large space telescopes are outlined. Experimental results show a definite form of magnetic field distribution is necessary to achieve magnification in the range 3X to 4X. Coil systems to establish the required field shapes were built, and both image intensifiers and camera tubes were operated at high magnification. The experiments confirm that such operation is practical and can provide satisfactory image quality. The main problem with such a system was identified as heating of the photocathode due to concentration of coil power dissipation in that vicinity. Suggestions for overcoming this disadvantage are included.
Li, Feng
2015-07-01
This review paper is based on our research experience in the past 30 years. The importance of radiologists' role is discussed in the development or evaluation of new medical images and of computer-aided detection (CAD) schemes in chest radiology. The four main topics include (1) introducing what diseases can be included in a research database for different imaging techniques or CAD systems and what imaging database can be built by radiologists, (2) understanding how radiologists' subjective judgment can be combined with technical objective features to improve CAD performance, (3) sharing our experience in the design of successful observer performance studies, and (4) finally, discussing whether the new images and CAD systems can improve radiologists' diagnostic ability in chest radiology. In conclusion, advanced imaging techniques and detection/classification of CAD systems have a potential clinical impact on improvement of radiologists' diagnostic ability, for both the detection and the differential diagnosis of various lung diseases, in chest radiology.
Development of x-ray imaging technique for liquid screening at airport
NASA Astrophysics Data System (ADS)
Sulaiman, Nurhani binti; Srisatit, Somyot
2016-01-01
X-ray imaging technology is a viable option to recognize flammable liquids for the purposes of aviation security. In this study, an X-ray imaging technology was developed whereby, the image viewing system was built with the use of a digital camera coupled with a gadolinium oxysulfide (GOS) fluorescent screen. The camera was equipped with a software for remote control setting of the camera via a USB cable which allows the images to be captured. The image was analysed to determine the average grey level using a software designed by Microsoft Visual Basic 6.0. The data was obtained for various densities of liquid thickness of 4.5 cm, 6.0 cm and 7.5 cm respectively for X-ray energies ranging from 70 to 200 kVp. In order to verify the reliability of the constructed calibration data, the system was tested with a few types of unknown liquids. The developed system could be conveniently employed for security screening in order to discriminate between a threat and an innocuous liquid.
Usability of a real-time tracked augmented reality display system in musculoskeletal injections
NASA Astrophysics Data System (ADS)
Baum, Zachary; Ungi, Tamas; Lasso, Andras; Fichtinger, Gabor
2017-03-01
PURPOSE: Image-guided needle interventions are seldom performed with augmented reality guidance in clinical practice due to many workspace and usability restrictions. We propose a real-time optically tracked image overlay system to make image-guided musculoskeletal injections more efficient and assess its usability in a bed-side clinical environment. METHODS: An image overlay system consisting of an optically tracked viewbox, tablet computer, and semitransparent mirror allows users to navigate scanned patient volumetric images in real-time using software built on the open-source 3D Slicer application platform. A series of experiments were conducted to evaluate the latency and screen refresh rate of the system using different image resolutions. To assess the usability of the system and software, five medical professionals were asked to navigate patient images while using the overlay and completed a questionnaire to assess the system. RESULTS: In assessing the latency of the system with scanned images of varying size, screen refresh rates were approximately 5 FPS. The study showed that participants found using the image overlay system easy, and found the table-mounted system was significantly more usable and effective than the handheld system. CONCLUSION: It was determined that the system performs comparably with scanned images of varying size when assessing the latency of the system. During our usability study, participants preferred the table-mounted system over the handheld. The participants also felt that the system itself was simple to use and understand. With these results, the image overlay system shows promise for use in a clinical environment.
TH-AB-BRA-12: Experimental Results From the First High-Field Inline MRI-Linac
DOE Office of Scientific and Technical Information (OSTI.GOV)
Keall, P; Dong, B; Zhang, K
Purpose: The pursuit of real-time image guided radiotherapy using optimal tissue contrast has seen the development of several hybrid MRI-treatment systems, high field and low field, and inline and perpendicular configurations. As part of a new MRI-Linac program, an MRI scanner was integrated with a linear accelerator to enable investigations of a coupled inline MRI-Linac system. This work describes our experimental results from the first high-field inline MRI-Linac. Methods: A 1.5 Tesla magnet (Sonata, Siemens) was located in a purpose built RF cage enabling shielding from and close proximity to a linear accelerator with inline orientation. A portable linear acceleratormore » (Linatron, Varian) was installed together with a multi-leaf collimator (Millennium, Varian) to provide dynamic field collimation and the whole assembly built onto a stainless-steel rail system. A series of MRI-Linac experiments was performed to investigate: (1) image quality with beam on measured using a macropodine (kangaroo) ex vivo phantom; (2) the noise as a function of beam state measured using a 6-channel surface coil array and; (3) electron focusing measured using GafChromic film. Results: (1) The macropodine phantom image quality with the beam on was almost identical to that with the beam off. (2) Noise measured with a surface RF coil produced a 25% elevation of background noise when the radiation beam was on. (3) Film measurements demonstrated electron focusing occurring at the center of the radiation field. Conclusion: The first high-field MRI-Linac has been built and experimentally characterized. This system has allowed us to establish the efficacy of a high field in-line MRI-Linac and study a number of the technical challenges and solutions. Supported by the Australian National Health and Medical Research Council, the Australian Research Council, the Australian Cancer Research Foundation and the Health and Hospitals Fund.« less
Secrets of Hidden Valley on Mars
2015-10-08
An image taken at the Hidden Valley site, en-route to Mount Sharp, by NASA Curiosity rover. A variety of mudstone strata in the area indicate a lakebed deposit, with river- and stream-related deposits nearby. Decoding the history of how these sedimentary rocks were formed, and during what period of time, was a key component in the confirming of the role of water and sedimentation in the formation of the floor of Gale Crater and Mount Sharp. This image was taken by the Mast Camera (Mastcam) on Curiosity on the 703rd Martian day, or sol, of the mission. Malin Space Science Systems, San Diego, built and operates Curiosity's Mastcam. NASA's Jet Propulsion Laboratory, a division of the California Institute of Technology, Pasadena, built the rover and manages the project for NASA's Science Mission Directorate, Washington. http://photojournal.jpl.nasa.gov/catalog/PIA19840
Nanohole-array-based device for 2D snapshot multispectral imaging
Najiminaini, Mohamadreza; Vasefi, Fartash; Kaminska, Bozena; Carson, Jeffrey J. L.
2013-01-01
We present a two-dimensional (2D) snapshot multispectral imager that utilizes the optical transmission characteristics of nanohole arrays (NHAs) in a gold film to resolve a mixture of input colors into multiple spectral bands. The multispectral device consists of blocks of NHAs, wherein each NHA has a unique periodicity that results in transmission resonances and minima in the visible and near-infrared regions. The multispectral device was illuminated over a wide spectral range, and the transmission was spectrally unmixed using a least-squares estimation algorithm. A NHA-based multispectral imaging system was built and tested in both reflection and transmission modes. The NHA-based multispectral imager was capable of extracting 2D multispectral images representative of four independent bands within the spectral range of 662 nm to 832 nm for a variety of targets. The multispectral device can potentially be integrated into a variety of imaging sensor systems. PMID:24005065
Bae, Sam Y; Korniski, Ronald J; Shearn, Michael; Manohara, Harish M; Shahinian, Hrayr
2017-01-01
High-resolution three-dimensional (3-D) imaging (stereo imaging) by endoscopes in minimally invasive surgery, especially in space-constrained applications such as brain surgery, is one of the most desired capabilities. Such capability exists at larger than 4-mm overall diameters. We report the development of a stereo imaging endoscope of 4-mm maximum diameter, called Multiangle, Rear-Viewing Endoscopic Tool (MARVEL) that uses a single-lens system with complementary multibandpass filter (CMBF) technology to achieve 3-D imaging. In addition, the system is endowed with the capability to pan from side-to-side over an angle of [Formula: see text], which is another unique aspect of MARVEL for such a class of endoscopes. The design and construction of a single-lens, CMBF aperture camera with integrated illumination to generate 3-D images, and the actuation mechanism built into it is summarized.
Larson, Eric D; Lee, Won-Mean; Roubidoux, Marilyn A; Goodsitt, Mitchell M; Lashbrook, Chris; Davis, Cynthia E; Kripfgans, Oliver D; Carson, Paul L
2018-03-01
We analyzed the performance of a mammographically configured, automated breast ultrasound (McABUS) scanner combined with a digital breast tomosynthesis (DBT) system. The GE Invenia ultrasound system was modified for integration with GE DBT systems. Ultrasound and DBT imaging were performed in the same mammographic compression. Our small preliminary study included 13 cases, six of whom had contained invasive cancers. From analysis of these cases, current limitations and corresponding potential improvements of the system were determined. A registration analysis was performed to compare the ease of McABUS to DBT registration for this system with that of two systems designed previously. It was observed that in comparison to data from an earlier study, the McABUS-to-DBT registration alignment errors for both this system and a previously built combined system were smaller than those for a previously built standalone McABUS system. Copyright © 2018 World Federation for Ultrasound in Medicine and Biology. Published by Elsevier Inc. All rights reserved.
NASA Technical Reports Server (NTRS)
1993-01-01
Lixi, Inc. has built a thriving business on NASA-developed x-ray technology. The Low Intensity X-ray Imaging scope (LIXI) was designed to use less than one percent of radiation required by conventional x-ray devices. It is portable and can be used for a variety of industrial inspection systems as well as medical devices. A food processing plant uses the new LIXI Conveyor system to identify small bone fragments in chicken. The chicken packages on a conveyor belt enter an x-ray chamber and the image is displayed on a monitor. Defects measuring less than a millimeter can be detected. An important advantage of the system is its ability to inspect 100 percent of the product right on the production line.
Earthshots: Satellite images of environmental change – Dubai, United Arab Emirates
Adamson, Thomas
2016-01-01
In the first image of this series, desert fills much of the image. As Dubai expands, roads, buildings, and irrigated fields spread out over the desert. But the most prominent project in Dubai, and an impressive engineering feat, is the artificial islands built off its coast. The islands were built from sand dredged from the sea floor. Rock breakwaters protect them from erosion. These Landsat images show the rapid and impressive development of these islands.
Wong, Kevin S K; Jian, Yifan; Cua, Michelle; Bonora, Stefano; Zawadzki, Robert J; Sarunic, Marinko V
2015-02-01
Wavefront sensorless adaptive optics optical coherence tomography (WSAO-OCT) is a novel imaging technique for in vivo high-resolution depth-resolved imaging that mitigates some of the challenges encountered with the use of sensor-based adaptive optics designs. This technique replaces the Hartmann Shack wavefront sensor used to measure aberrations with a depth-resolved image-driven optimization algorithm, with the metric based on the OCT volumes acquired in real-time. The custom-built ultrahigh-speed GPU processing platform and fast modal optimization algorithm presented in this paper was essential in enabling real-time, in vivo imaging of human retinas with wavefront sensorless AO correction. WSAO-OCT is especially advantageous for developing a clinical high-resolution retinal imaging system as it enables the use of a compact, low-cost and robust lens-based adaptive optics design. In this report, we describe our WSAO-OCT system for imaging the human photoreceptor mosaic in vivo. We validated our system performance by imaging the retina at several eccentricities, and demonstrated the improvement in photoreceptor visibility with WSAO compensation.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Arinilhaq,; Widita, Rena
2014-09-30
Optical Coherence Tomography is often used in medical image acquisition to diagnose that change due easy to use and low price. Unfortunately, this type of examination produces a two-dimensional retinal image of the point of acquisition. Therefore, this study developed a method that combines and reconstruct 2-dimensional retinal images into three-dimensional images to display volumetric macular accurately. The system is built with three main stages: data acquisition, data extraction and 3-dimensional reconstruction. At data acquisition step, Optical Coherence Tomography produced six *.jpg images of each patient were further extracted with MATLAB 2010a software into six one-dimensional arrays. The six arraysmore » are combined into a 3-dimensional matrix using a kriging interpolation method with SURFER9 resulting 3-dimensional graphics of macula. Finally, system provides three-dimensional color graphs based on the data distribution normal macula. The reconstruction system which has been designed produces three-dimensional images with size of 481 × 481 × h (retinal thickness) pixels.« less
Coronagraphic imaging of circumstellar material around evolved massive stars
NASA Astrophysics Data System (ADS)
Lomax, Jamie R.; Levesque, Emily; Wisniewski, John
2018-01-01
While many astronomical subfields (e.g. the solar, exoplanet, and disk communities) have been using coronagraphy to combat contrast ratio problems for years, the use of coronagraphic imaging techniques to probe the circumstellar environments of massive stars has been surprisingly underutilized. While current extreme adaptive optics coronagraphic imaging systems (e.g. GPI on Gemini South, SPHERE at the VLT, and SCExAO at Subaru) were built for the sole purpose of detecting exoplanets, their ability to provide large contrast ratios and small inner working angles means they can detect gas, dust, and companions that are closer to the central star than ever before. In this poster we present pilot studies of evolved massive stars using several coronagraphic imaging systems and summarize potential science gains this technique might provide.
Grepstad, Jon Olav; Kaspar, Peter; Solgaard, Olav; Johansen, Ib-Rune; Sudbø, Aasmund S
2012-03-26
A sensor designed to detect bio-molecules is presented. The sensor exploits a planar 2D photonic crystal (PC) membrane with sub-micron thickness and through holes, to induce high optical fields that allow detection of nano-particles smaller than the diffraction limit of an optical microscope. We report on our design and fabrication of a PC membrane with a nano-particle trapped inside. We have also designed and built an imaging system where an optical microscope and a CCD camera are used to take images of the PC membrane. Results show how the trapped nano-particle appears as a bright spot in the image. In a first experimental realization of the imaging system, single particles with a radius of 75 nm can be detected.
NASA Astrophysics Data System (ADS)
FitzGerald, Jack G. M.
2015-02-01
The Rotating Scatter Mask (RSM) system is an inexpensive retrofit that provides imaging capabilities to scintillating detectors. Unlike traditional collimator systems that primarily absorb photons in order to form an image, this system primarily scatters the photons. Over a single rotation, there is a unique, smooth response curve for each defined source position. Testing was conducted using MCNPX simulations. Image reconstruction was performed using a chi-squared reconstruction technique. A simulated 100 uCi, Cs-137 source at 10 meters was detected after a single, 50-second rotation when a uniform terrestrial background was present. A Cs-137 extended source was also tested. The RSM field-of-view is 360 degrees azimuthally as well as 54 degrees above and 54 degrees below the horizontal plane. Since the RSM is built from polyethylene, the overall cost and weight of the system is low. The system was designed to search for lost or stolen radioactive material, also known as the orphan source problem.
Landsat-7 Enhanced Thematic Mapper plus radiometric calibration
Markham, B.L.; Boncyk, Wayne C.; Helder, D.L.; Barker, J.L.
1997-01-01
Landsat-7 is currently being built and tested for launch in 1998. The Enhanced Thematic Mapper Plus (ETM+) sensor for Landsat-7, a derivative of the highly successful Thematic Mapper (TM) sensors on Landsats 4 and 5, and the Landsat-7 ground system are being built to provide enhanced radiometric calibration performance. In addition, regular vicarious calibration campaigns are being planned to provide additional information for calibration of the ETM+ instrument. The primary upgrades to the instrument include the addition of two solar calibrators: the full aperture solar calibrator, a deployable diffuser, and the partial aperture solar calibrator, a passive device that allows the ETM+ to image the sun. The ground processing incorporates for the first time an off-line facility, the Image Assessment System (IAS), to perform calibration, evaluation and analysis. Within the IAS, processing capabilities include radiometric artifact characterization and correction, radiometric calibration from the multiple calibrator sources, inclusion of results from vicarious calibration and statistical trending of calibration data to improve calibration estimation. The Landsat Product Generation System, the portion of the ground system responsible for producing calibrated products, will incorporate the radiometric artifact correction algorithms and will use the calibration information generated by the IAS. This calibration information will also be supplied to ground processing systems throughout the world.
A biometric access personal optical storage device
NASA Astrophysics Data System (ADS)
Davies, David H.; Ray, Steve; Gurkowski, Mark; Lee, Lane
2007-01-01
A portable USB2.0 personal storage device that uses built-in encryption and allows data access through biometric scanning of a finger print is described. Biometric image derived templates are stored on the removable 32 mm write once (WO) media. The encrypted templates travel with the disc and allow access to the data providing the biometric feature (e.g. the finger itself) is present. The device also allows for export and import of the templates under secure key exchange protocols. The storage system is built around the small form factor optical engine that uses a tilt arm rotary actuator and front surface media.
2012-01-01
2002. [2] Max Blosser. Fundamental Modeling and Thermal Performance Issues for Metal- lic Thermal Protection System Concept. Journal of Spacecraft and...Directorate (AFRL/RX, Metallic Thermal Protection System Program, Universal Technology Corporation/AFRL Grant, ProgramManagers: Todd Warren and Reji John...retired Space Shuttle program were built with a Thermal Protection System (TPS) to withstand heating during atmo- spheric reentry. The partially
Multi-rate, real time image compression for images dominated by point sources
NASA Technical Reports Server (NTRS)
Huber, A. Kris; Budge, Scott E.; Harris, Richard W.
1993-01-01
An image compression system recently developed for compression of digital images dominated by point sources is presented. Encoding consists of minimum-mean removal, vector quantization, adaptive threshold truncation, and modified Huffman encoding. Simulations are presented showing that the peaks corresponding to point sources can be transmitted losslessly for low signal-to-noise ratios (SNR) and high point source densities while maintaining a reduced output bit rate. Encoding and decoding hardware has been built and tested which processes 552,960 12-bit pixels per second at compression rates of 10:1 and 4:1. Simulation results are presented for the 10:1 case only.
Multispectral Imaging Broadens Cellular Analysis
NASA Technical Reports Server (NTRS)
2007-01-01
Amnis Corporation, a Seattle-based biotechnology company, developed ImageStream to produce sensitive fluorescence images of cells in flow. The company responded to an SBIR solicitation from Ames Research Center, and proposed to evaluate several methods of extending the depth of field for its ImageStream system and implement the best as an upgrade to its commercial products. This would allow users to view whole cells at the same time, rather than just one section of each cell. Through Phase I and II SBIR contracts, Ames provided Amnis the funding the company needed to develop this extended functionality. For NASA, the resulting high-speed image flow cytometry process made its way into Medusa, a life-detection instrument built to collect, store, and analyze sample organisms from erupting hydrothermal vents, and has the potential to benefit space flight health monitoring. On the commercial end, Amnis has implemented the process in ImageStream, combining high-resolution microscopy and flow cytometry in a single instrument, giving researchers the power to conduct quantitative analyses of individual cells and cell populations at the same time, in the same experiment. ImageStream is also built for many other applications, including cell signaling and pathway analysis; classification and characterization of peripheral blood mononuclear cell populations; quantitative morphology; apoptosis (cell death) assays; gene expression analysis; analysis of cell conjugates; molecular distribution; and receptor mapping and distribution.
NASA Astrophysics Data System (ADS)
Hueso, R.; Mendikoa, I.; Sánchez-Lavega, A.; Pérez-Hoyos, S.; Rojas, J. F.; García-Melendo, E.
2015-10-01
PlanetCam UPV/EHU [1] is an astronomical instrument designed for high-resolution observations of Solar System planets. The main scientific themes are atmospheric dynamics and the vertical cloud structure of Jupiter and Saturn. The instrument uses a dichroic mirror to separate the light in two beams with spectral ranges from 380 nm to1 micron (visible channel) and from 1 to 1.7 microns (Short Wave InfraRed, SWIR channel) and two detectors working simultaneously with fast acquisition modes. High-resolution images are built using lucky imaging techniques [2]. Several hundred short exposed images are obtained and stored in fits files. Images are automatically reduced by a pipeline called PLAYLIST (written in IDL and requiring no interaction by the user)which selects the best frames and co-registers them using image correlation over several tie-points. The result is a high signal to noise ratio image that can be processed to show the faint structures in the data. PlanetCam is a visiting instrument mainly built for the 1.2 3 and 2.2m telescopes at Calar Alto Observatory in Spain but it has also been tested in the 1.5 m Telescope Carlos Sanchez in Tenerife and the 1.05 m Telescope at the Pic du Midi observatory.
NASA Astrophysics Data System (ADS)
Peng, F.; Cai, X.; Tan, W.
2017-09-01
Within-class spectral variation and between-class spectral confusion in remotely sensed imagery degrades the performance of built-up area detection when using planar texture, shape, and spectral features. Terrain slope and building height are often used to optimize the results, but extracted from auxiliary data (e.g. LIDAR data, DSM). Moreover, the auxiliary data must be acquired around the same time as image acquisition. Otherwise, built-up area detection accuracy is affected. Stereo imagery incorporates both planar and height information unlike single remotely sensed images. Stereo imagery acquired by many satellites (e.g. Worldview-4, Pleiades-HR, ALOS-PRISM, and ZY-3) can be used as data source of identifying built-up areas. A new method of identifying high-accuracy built-up areas from stereo imagery is achieved by using a combination of planar and height features. The digital surface model (DSM) and digital orthophoto map (DOM) are first generated from stereo images. Then, height values of above-ground objects (e.g. buildings) are calculated from the DSM, and used to obtain raw built-up areas. Other raw built-up areas are obtained from the DOM using Pantex and Gabor, respectively. Final high-accuracy built-up area results are achieved from these raw built-up areas using the decision level fusion. Experimental results show that accurate built-up areas can be achieved from stereo imagery. The height information used in the proposed method is derived from stereo imagery itself, with no need to require auxiliary height data (e.g. LIDAR data). The proposed method is suitable for spaceborne and airborne stereo pairs and triplets.
NASA Astrophysics Data System (ADS)
Solis-Najera, S.; Vazquez, F.; Hernandez, R.; Marrufo, O.; Rodriguez, A. O.
2016-12-01
A surface radio frequency coil was developed for small animal image acquisition in a pre-clinical magnetic resonance imaging system at 7 T. A flexible coil composed of two circular loops was developed to closely cover the object to be imaged. Electromagnetic numerical simulations were performed to evaluate its performance before the coil construction. An analytical expression of the mutual inductance for the two circular loops as a function of the separation between them was derived and used to validate the simulations. The RF coil is composed of two circular loops with a 5 cm external diameter and was tuned to 300 MHz and 50 Ohms matched. The angle between the loops was varied and the Q factor was obtained from the S11 simulations for each angle. B1 homogeneity was also evaluated using the electromagnetic simulations. The coil prototype was designed and built considering the numerical simulation results. To show the feasibility of the coil and its performance, saline-solution phantom images were acquired. A correlation of the simulations and imaging experimental results was conducted showing a concordance of 0.88 for the B1 field. The best coil performance was obtained at the 90° aperture angle. A more realistic phantom was also built using a formaldehyde-fixed rat phantom for ex vivo imaging experiments. All images showed a good image quality revealing clearly defined anatomical details of an ex vivo rat.
Image-Based Environmental Monitoring Sensor Application Using an Embedded Wireless Sensor Network
Paek, Jeongyeup; Hicks, John; Coe, Sharon; Govindan, Ramesh
2014-01-01
This article discusses the experiences from the development and deployment of two image-based environmental monitoring sensor applications using an embedded wireless sensor network. Our system uses low-power image sensors and the Tenet general purpose sensing system for tiered embedded wireless sensor networks. It leverages Tenet's built-in support for reliable delivery of high rate sensing data, scalability and its flexible scripting language, which enables mote-side image compression and the ease of deployment. Our first deployment of a pitfall trap monitoring application at the James San Jacinto Mountain Reserve provided us with insights and lessons learned into the deployment of and compression schemes for these embedded wireless imaging systems. Our three month-long deployment of a bird nest monitoring application resulted in over 100,000 images collected from a 19-camera node network deployed over an area of 0.05 square miles, despite highly variable environmental conditions. Our biologists found the on-line, near-real-time access to images to be useful for obtaining data on answering their biological questions. PMID:25171121
Distributed data collection for a database of radiological image interpretations
NASA Astrophysics Data System (ADS)
Long, L. Rodney; Ostchega, Yechiam; Goh, Gin-Hua; Thoma, George R.
1997-01-01
The National Library of Medicine, in collaboration with the National Center for Health Statistics and the National Institute for Arthritis and Musculoskeletal and Skin Diseases, has built a system for collecting radiological interpretations for a large set of x-ray images acquired as part of the data gathered in the second National Health and Nutrition Examination Survey. This system is capable of delivering across the Internet 5- and 10-megabyte x-ray images to Sun workstations equipped with X Window based 2048 X 2560 image displays, for the purpose of having these images interpreted for the degree of presence of particular osteoarthritic conditions in the cervical and lumbar spines. The collected interpretations can then be stored in a database at the National Library of Medicine, under control of the Illustra DBMS. This system is a client/server database application which integrates (1) distributed server processing of client requests, (2) a customized image transmission method for faster Internet data delivery, (3) distributed client workstations with high resolution displays, image processing functions and an on-line digital atlas, and (4) relational database management of the collected data.
Image-based environmental monitoring sensor application using an embedded wireless sensor network.
Paek, Jeongyeup; Hicks, John; Coe, Sharon; Govindan, Ramesh
2014-08-28
This article discusses the experiences from the development and deployment of two image-based environmental monitoring sensor applications using an embedded wireless sensor network. Our system uses low-power image sensors and the Tenet general purpose sensing system for tiered embedded wireless sensor networks. It leverages Tenet's built-in support for reliable delivery of high rate sensing data, scalability and its flexible scripting language, which enables mote-side image compression and the ease of deployment. Our first deployment of a pitfall trap monitoring application at the James San Cannot Mountain Reserve provided us with insights and lessons learned into the deployment of and compression schemes for these embedded wireless imaging systems. Our three month-long deployment of a bird nest monitoring application resulted in over 100,000 images collected from a 19-camera node network deployed over an area of 0.05 square miles, despite highly variable environmental conditions. Our biologists found the on-line, near-real-time access to images to be useful for obtaining data on answering their biological questions.
How Phoenix Creates Color Images (Animation)
NASA Technical Reports Server (NTRS)
2008-01-01
[figure removed for brevity, see original site] Click on image for animation This simple animation shows how a color image is made from images taken by Phoenix. The Surface Stereo Imager captures the same scene with three different filters. The images are sent to Earth in black and white and the color is added by mission scientists. By contrast, consumer digital cameras and cell phones have filters built in and do all of the color processing within the camera itself. The Phoenix Mission is led by the University of Arizona, Tucson, on behalf of NASA. Project management of the mission is by NASAaE(TM)s Jet Propulsion Laboratory, Pasadena, Calif. Spacecraft development is by Lockheed Martin Space Systems, Denver.A network-based training environment: a medical image processing paradigm.
Costaridou, L; Panayiotakis, G; Sakellaropoulos, P; Cavouras, D; Dimopoulos, J
1998-01-01
The capability of interactive multimedia and Internet technologies is investigated with respect to the implementation of a distance learning environment. The system is built according to a client-server architecture, based on the Internet infrastructure, composed of server nodes conceptually modelled as WWW sites. Sites are implemented by customization of available components. The environment integrates network-delivered interactive multimedia courses, network-based tutoring, SIG support, information databases of professional interest, as well as course and tutoring management. This capability has been demonstrated by means of an implemented system, validated with digital image processing content, specifically image enhancement. Image enhancement methods are theoretically described and applied to mammograms. Emphasis is given to the interactive presentation of the effects of algorithm parameters on images. The system end-user access depends on available bandwidth, so high-speed access can be achieved via LAN or local ISDN connections. Network based training offers new means of improved access and sharing of learning resources and expertise, as promising supplements in training.
NASA Astrophysics Data System (ADS)
Nurge, Mark A.
2007-05-01
An electrical capacitance volume tomography system has been created for use with a new image reconstruction algorithm capable of imaging high contrast dielectric distributions. The electrode geometry consists of two 4 × 4 parallel planes of copper conductors connected through custom built switch electronics to a commercially available capacitance to digital converter. Typical electrical capacitance tomography (ECT) systems rely solely on mutual capacitance readings to reconstruct images of dielectric distributions. This paper presents a method of reconstructing images of high contrast dielectric materials using only the self-capacitance measurements. By constraining the unknown dielectric material to one of two values, the inverse problem is no longer ill-determined. Resolution becomes limited only by the accuracy and resolution of the measurement circuitry. Images were reconstructed using this method with both synthetic and real data acquired using an aluminium structure inserted at different positions within the sensing region. Comparisons with standard two-dimensional ECT systems highlight the capabilities and limitations of the electronics and reconstruction algorithm.
Electrical capacitance volume tomography of high contrast dielectrics using a cuboid geometry
NASA Astrophysics Data System (ADS)
Nurge, Mark A.
An Electrical Capacitance Volume Tomography system has been created for use with a new image reconstruction algorithm capable of imaging high contrast dielectric distributions. The electrode geometry consists of two 4 x 4 parallel planes of copper conductors connected through custom built switch electronics to a commercially available capacitance to digital converter. Typical electrical capacitance tomography (ECT) systems rely solely on mutual capacitance readings to reconstruct images of dielectric distributions. This dissertation presents a method of reconstructing images of high contrast dielectric materials using only the self capacitance measurements. By constraining the unknown dielectric material to one of two values, the inverse problem is no longer ill-determined. Resolution becomes limited only by the accuracy and resolution of the measurement circuitry. Images were reconstructed using this method with both synthetic and real data acquired using an aluminum structure inserted at different positions within the sensing region. Comparisons with standard two dimensional ECT systems highlight the capabilities and limitations of the electronics and reconstruction algorithm.
Line-scanning Raman imaging spectroscopy for detection of fingerprints.
Deng, Sunan; Liu, Le; Liu, Zhiyi; Shen, Zhiyuan; Li, Guohua; He, Yonghong
2012-06-10
Fingerprints are the best form of personal identification for criminal investigation purposes. We present a line-scanning Raman imaging system and use it to detect fingerprints composed of β-carotene and fish oil on different substrates. Although the line-scanning Raman system has been used to map the distribution of materials such as polystyrene spheres and minerals within geological samples, this is the first time to our knowledge that the method is used in imaging fingerprints. Two Raman peaks of β-carotene (501.2, 510.3 nm) are detected and the results demonstrate that both peaks can generate excellent images with little difference between them. The system operates at a spectra resolution of about 0.4 nm and can detect β-carotene signals in petroleum ether solution with the limit of detection of 3.4×10(-9) mol/L. The results show that the line-scanning Raman imaging spectroscopy we have built has a high accuracy and can be used in the detection of latent fingerprints in the future.
SeeStar: an open-source, low-cost imaging system for subsea observations
NASA Astrophysics Data System (ADS)
Cazenave, F.; Kecy, C. D.; Haddock, S.
2016-02-01
Scientists and engineers at the Monterey Bay Aquarium Research Institute (MBARI) have collaborated to develop SeeStar, a modular, light weight, self-contained, low-cost subsea imaging system for short- to long-term monitoring of marine ecosystems. SeeStar is composed of separate camera, battery, and LED lighting modules. Two versions of the system exist: one rated to 300 meters depth, the other rated to 1500 meters. Users can download plans and instructions from an online repository and build the system using low-cost off-the-shelf components. The system utilizes an easily programmable Arduino based controller, and the widely distributed GoPro camera. The system can be deployed in a variety of scenarios taking still images and video and can be operated either autonomously or tethered on a range of platforms, including ROVs, AUVs, landers, piers, and moorings. Several Seestar systems have been built and used for scientific studies and engineering tests. The long-term goal of this project is to have a widely distributed marine imaging network across thousands of locations, to develop baselines of biological information.
Zhao, Ming; Li, Yu; Peng, Leilei
2014-01-01
We present a novel excitation-emission multiplexed fluorescence lifetime microscopy (FLIM) method that surpasses current FLIM techniques in multiplexing capability. The method employs Fourier multiplexing to simultaneously acquire confocal fluorescence lifetime images of multiple excitation wavelength and emission color combinations at 44,000 pixels/sec. The system is built with low-cost CW laser sources and standard PMTs with versatile spectral configuration, which can be implemented as an add-on to commercial confocal microscopes. The Fourier lifetime confocal method allows fast multiplexed FLIM imaging, which makes it possible to monitor multiple biological processes in live cells. The low cost and compatibility with commercial systems could also make multiplexed FLIM more accessible to biological research community. PMID:24921725
Image processing system design for microcantilever-based optical readout infrared arrays
NASA Astrophysics Data System (ADS)
Tong, Qiang; Dong, Liquan; Zhao, Yuejin; Gong, Cheng; Liu, Xiaohua; Yu, Xiaomei; Yang, Lei; Liu, Weiyu
2012-12-01
Compared with the traditional infrared imaging technology, the new type of optical-readout uncooled infrared imaging technology based on MEMS has many advantages, such as low cost, small size, producing simple. In addition, the theory proves that the technology's high thermal detection sensitivity. So it has a very broad application prospects in the field of high performance infrared detection. The paper mainly focuses on an image capturing and processing system in the new type of optical-readout uncooled infrared imaging technology based on MEMS. The image capturing and processing system consists of software and hardware. We build our image processing core hardware platform based on TI's high performance DSP chip which is the TMS320DM642, and then design our image capturing board based on the MT9P031. MT9P031 is Micron's company high frame rate, low power consumption CMOS chip. Last we use Intel's company network transceiver devices-LXT971A to design the network output board. The software system is built on the real-time operating system DSP/BIOS. We design our video capture driver program based on TI's class-mini driver and network output program based on the NDK kit for image capturing and processing and transmitting. The experiment shows that the system has the advantages of high capturing resolution and fast processing speed. The speed of the network transmission is up to 100Mbps.
Giaddui, Tawfik; Yu, Jialu; Manfredi, Denise; Linnemann, Nancy; Hunter, Joanne; O'Meara, Elizabeth; Galvin, James; Bialecki, Brian; Xiao, Ying
2016-01-01
Transmission of Imaging and Data (TRIAD) is a standard-based system built by the American College of Radiology to provide the seamless exchange of images and data for accreditation of clinical trials and registries. Scripts of structures' names validation profiles created in TRIAD are used in the automated submission process. It is essential for users to understand the logistics of these scripts for successful submission of radiation therapy cases with less iteration. Copyright © 2016 American Society for Radiation Oncology. Published by Elsevier Inc. All rights reserved.
A portfolio of products from the rapid terrain visualization interferometric SAR
NASA Astrophysics Data System (ADS)
Bickel, Douglas L.; Doerry, Armin W.
2007-04-01
The Rapid Terrain Visualization interferometric synthetic aperture radar was designed and built at Sandia National Laboratories as part of an Advanced Concept Technology Demonstration (ACTD) to "demonstrate the technologies and infrastructure to meet the Army requirement for rapid generation of digital topographic data to support emerging crisis or contingencies." This sensor was built by Sandia National Laboratories for the Joint Programs Sustainment and Development (JPSD) Project Office to provide highly accurate digital elevation models (DEMs) for military and civilian customers, both inside and outside of the United States. The sensor achieved better than HRTe Level IV position accuracy in near real-time. The system was flown on a deHavilland DHC-7 Army aircraft. This paper presents a collection of images and data products from the Rapid Terrain Visualization interferometric synthetic aperture radar. The imagery includes orthorectified images and DEMs from the RTV interferometric SAR radar.
Networked vision system using a Prolog controller
NASA Astrophysics Data System (ADS)
Batchelor, B. G.; Caton, S. J.; Chatburn, L. T.; Crowther, R. A.; Miller, J. W. V.
2005-11-01
Prolog offers a very different style of programming compared to conventional languages; it can define object properties and abstract relationships in a way that Java, C, C++, etc. find awkward. In an accompanying paper, the authors describe how a distributed web-based vision systems can be built using elements that may even be located on different continents. One particular system of this general type is described here. The top-level controller is a Prolog program, which operates one, or more, image processing engines. This type of function is natural to Prolog, since it is able to reason logically using symbolic (non-numeric) data. Although Prolog is not suitable for programming image processing functions directly, it is ideal for analysing the results derived by an image processor. This article describes the implementation of two systems, in which a Prolog program controls several image processing engines, a simple robot, a pneumatic pick-and-place arm), LED illumination modules and a various mains-powered devices.
Low cost light-sheet microscopy for whole brain imaging
NASA Astrophysics Data System (ADS)
Kumar, Manish; Nasenbeny, Jordan; Kozorovitskiy, Yevgenia
2018-02-01
Light-sheet microscopy has evolved as an indispensable tool in imaging biological samples. It can image 3D samples at fast speed, with high-resolution optical sectioning, and with reduced photobleaching effects. These properties make light-sheet microscopy ideal for imaging fluorophores in a variety of biological samples and organisms, e.g. zebrafish, drosophila, cleared mouse brains, etc. While most commercial turnkey light-sheet systems are expensive, the existing lower cost implementations, e.g. OpenSPIM, are focused on achieving high-resolution imaging of small samples or organisms like zebrafish. In this work, we substantially reduce the cost of light-sheet microscope system while targeting to image much larger samples, i.e. cleared mouse brains, at single-cell resolution. The expensive components of a lightsheet system - excitation laser, water-immersion objectives, and translation stage - are replaced with an incoherent laser diode, dry objectives, and a custom-built Arduino-controlled translation stage. A low-cost CUBIC protocol is used to clear fixed mouse brain samples. The open-source platforms of μManager and Fiji support image acquisition, processing, and visualization. Our system can easily be extended to multi-color light-sheet microscopy.
You can't measure what you can't see - detectors for microscopies
NASA Astrophysics Data System (ADS)
Denes, Peter
For centuries, the human eye has been the imaging detector of choice thanks to its high sensitivity, wide dynamic range, and direct connection to a built-in data recording and analysis system. The eye, however, is limited to visible light, which excludes microscopies with electrons and X-rays, and the built-in recording system stores archival information at very low rates. The former limitation has been overcome by ``indirect'' detectors, which convert probe particles to visible light, and the latter by a variety of recording techniques, from photographic film to semiconductor-based imagers. Semiconductor imagers have been used for decades as ``direct'' detectors in particle physics, and almost as long for hard X-rays. For soft X-ray microscopy, the challenge has been the small signal levels - plus getting the X-rays into the detector itself, given how quickly they are absorbed in inert layers. For electron microscopy, the challenge has been reconciling detector spatial resolution and pixel count with the large multiple scattering of electrons with energies used for microscopy. Further, a high recording rate (``movies'' rather than ``snapshots'') enables time-resolved studies, time-dependent corrections, shot-by-shot experiments and scanning techniques - at the expense of creating large data volumes. This talk will discuss solutions to these challenges, as well as an outlook towards future developments.
Design and Development of a New Multi-Projection X-Ray System for Chest Imaging
NASA Astrophysics Data System (ADS)
Chawla, Amarpreet S.; Boyce, Sarah; Washington, Lacey; McAdams, H. Page; Samei, Ehsan
2009-02-01
Overlapping anatomical structures may confound the detection of abnormal pathology, including lung nodules, in conventional single-projection chest radiography. To minimize this fundamental limiting factor, a dedicated digital multi-projection system for chest imaging was recently developed at the Radiology Department of Duke University. We are reporting the design of the multi-projection imaging system and its initial performance in an ongoing clinical trial. The system is capable of acquiring multiple full-field projections of the same patient along both the horizontal and vertical axes at variable speeds and acquisition frame rates. These images acquired in rapid succession from slightly different angles about the posterior-anterior (PA) orientation can be correlated to minimize the influence of overlying anatomy. The developed system has been tested for repeatability and motion blur artifacts to investigate its robustness for clinical trials. Excellent geometrical consistency was found in the tube motion, with positional errors for clinical settings within 1%. The effect of tube-motion on the image quality measured in terms of impact on the modulation transfer function (MTF) was found to be minimal. The system was deemed clinic-ready and a clinical trial was subsequently launched. The flexibility of image acquisition built into the system provides a unique opportunity to easily modify it for different clinical applications, including tomosynthesis, correlation imaging (CI), and stereoscopic imaging.
NASA Astrophysics Data System (ADS)
Liu, X.; Zhang, J. X.; Zhao, Z.; Ma, A. D.
2015-06-01
Synthetic aperture radar in the application of remote sensing technology is becoming more and more widely because of its all-time and all-weather operation, feature extraction research in high resolution SAR image has become a hot topic of concern. In particular, with the continuous improvement of airborne SAR image resolution, image texture information become more abundant. It's of great significance to classification and extraction. In this paper, a novel method for built-up areas extraction using both statistical and structural features is proposed according to the built-up texture features. First of all, statistical texture features and structural features are respectively extracted by classical method of gray level co-occurrence matrix and method of variogram function, and the direction information is considered in this process. Next, feature weights are calculated innovatively according to the Bhattacharyya distance. Then, all features are weighted fusion. At last, the fused image is classified with K-means classification method and the built-up areas are extracted after post classification process. The proposed method has been tested by domestic airborne P band polarization SAR images, at the same time, two groups of experiments based on the method of statistical texture and the method of structural texture were carried out respectively. On the basis of qualitative analysis, quantitative analysis based on the built-up area selected artificially is enforced, in the relatively simple experimentation area, detection rate is more than 90%, in the relatively complex experimentation area, detection rate is also higher than the other two methods. In the study-area, the results show that this method can effectively and accurately extract built-up areas in high resolution airborne SAR imagery.
NASA Astrophysics Data System (ADS)
Fujiwara, Yukihiro; Yoshii, Masakazu; Arai, Yasuhito; Adachi, Shuichi
Advanced safety vehicle(ASV)assists drivers’ manipulation to avoid trafic accidents. A variety of researches on automatic driving systems are necessary as an element of ASV. Among them, we focus on visual feedback approach in which the automatic driving system is realized by recognizing road trajectory using image information. The purpose of this paper is to examine the validity of this approach by experiments using a radio-controlled car. First, a practical image processing algorithm to recognize white lines on the road is proposed. Second, a model of the radio-controlled car is built by system identication experiments. Third, an automatic steering control system is designed based on H∞ control theory. Finally, the effectiveness of the designed control system is examined via traveling experiments.
Performance of the Gemini Planet Imager’s adaptive optics system
Poyneer, Lisa A.; Palmer, David W.; Macintosh, Bruce; ...
2016-01-07
The Gemini Planet Imager’s adaptive optics (AO) subsystem was designed specifically to facilitate high-contrast imaging. We give a definitive description of the system’s algorithms and technologies as built. Ultimately, the error budget indicates that for all targets and atmospheric conditions AO bandwidth error is the largest term.
Iotti, Bryan; Valazza, Alberto
2014-10-01
Picture Archiving and Communications Systems (PACS) are the most needed system in a modern hospital. As an integral part of the Digital Imaging and Communications in Medicine (DICOM) standard, they are charged with the responsibility for secure storage and accessibility of the diagnostic imaging data. These machines need to offer high performance, stability, and security while proving reliable and ergonomic in the day-to-day and long-term storage and retrieval of the data they safeguard. This paper reports the experience of the authors in developing and installing a compact and low-cost solution based on open-source technologies in the Veterinary Teaching Hospital for the University of Torino, Italy, during the course of the summer of 2012. The PACS server was built on low-cost x86-based hardware and uses an open source operating system derived from Oracle OpenSolaris (Oracle Corporation, Redwood City, CA, USA) to host the DCM4CHEE PACS DICOM server (DCM4CHEE, http://www.dcm4che.org ). This solution features very high data security and an ergonomic interface to provide easy access to a large amount of imaging data. The system has been in active use for almost 2 years now and has proven to be a scalable, cost-effective solution for practices ranging from small to very large, where the use of different hardware combinations allows scaling to the different deployments, while the use of paravirtualization allows increased security and easy migrations and upgrades.
Nanophotonic projection system.
Aflatouni, Firooz; Abiri, Behrooz; Rekhi, Angad; Hajimiri, Ali
2015-08-10
Low-power integrated projection technology can play a key role in development of low-cost mobile devices with built-in high-resolution projectors. Low-cost 3D imaging and holography systems are also among applications of such a technology. In this paper, an integrated projection system based on a two-dimensional optical phased array with fast beam steering capability is reported. Forward biased p-i-n phase modulators with 200MHz bandwidth are used per each array element for rapid phase control. An optimization algorithm is implemented to compensate for the phase dependent attenuation of the p-i-n modulators. Using rapid vector scanning technique, images were formed and recorded within a single snapshot of the IR camera.
NASA Astrophysics Data System (ADS)
Cao, Nan; Cao, Fengmei; Lin, Yabin; Bai, Tingzhu; Song, Shengyu
2015-04-01
For a new kind of retina-like senor camera and a traditional rectangular sensor camera, dual cameras acquisition and display system need to be built. We introduce the principle and the development of retina-like senor. Image coordinates transformation and interpolation based on sub-pixel interpolation need to be realized for our retina-like sensor's special pixels distribution. The hardware platform is composed of retina-like senor camera, rectangular sensor camera, image grabber and PC. Combined the MIL and OpenCV library, the software program is composed in VC++ on VS 2010. Experience results show that the system can realizes two cameras' acquisition and display.
Test Rover at JPL During Preparation for Mars Rover Low-Angle Selfie
2015-08-19
This view of a test rover at NASA's Jet Propulsion Laboratory, Pasadena, California, results from advance testing of arm positions and camera pointings for taking a low-angle self-portrait of NASA's Curiosity Mars rover. This rehearsal in California led to a dramatic Aug. 5, 2015, selfie of Curiosity, online at PIA19807. Curiosity's arm-mounted Mars Hand Lens Imager (MAHLI) camera took 92 of component images that were assembled into that mosaic. The rover team positioned the camera lower in relation to the rover body than for any previous full self-portrait of Curiosity. This practice version was taken at JPL's Mars Yard in July 2013, using the Vehicle System Test Bed (VSTB) rover, which has a test copy of MAHLI on its robotic arm. MAHLI was built by Malin Space Science Systems, San Diego. JPL, a division of the California Institute of Technology in Pasadena, manages the Mars Science Laboratory Project for the NASA Science Mission Directorate, Washington. JPL designed and built the project's Curiosity rover. http://photojournal.jpl.nasa.gov/catalog/PIA19810
Devising a Visual Inspection System for Canal Tunnels: Preliminary Studies
NASA Astrophysics Data System (ADS)
Albert, J.-L.; Charbonnier, P.; Chavant, P.; Foucher, P.; Muzet, V.; Prybyla, D.; Perrin, T.; Grussenmeyer, P.; Guillemin, S.; Koehl, M.
2013-07-01
In France, most tunnel canals were built during the 19th and 20th centuries. Maintaining them is not only a matter of heritage preservation but also a question of security. Inspecting tunnel canals is difficult and time consuming, which motivates the development of an image-based surveying system, as already exists for railway or road tunnels. However, while the imaging configuration is similar, referencing the data acquisition device is more difficult in the case of underground waterways, due to the drifts of the inspection barge. In this paper, we introduce the recording prototype we have designed and report the results of the test that were performed in an underground waterway, Niderviller's tunnel, to assess the feasibility of the system. In particular, we give details on the imaging system design. We also analyze the pros and cons of each location method, in terms of costs, practicability, computational burden and accuracy.
A compact 3 T all HTS cryogen-free MRI system
NASA Astrophysics Data System (ADS)
Parkinson, B. J.; Bouloukakis, K.; Slade, R. A.
2017-12-01
We have designed and built a passively shielded, cryogen-free 3 T 160 mm bore bismuth strontium calcium copper oxide HTS magnet with shielded gradient coils suitable for use in small animal imaging applications. The magnet is cooled to approximately 16 K using a two-stage cryocooler and is operated at 200 A. The magnet has been passively shimmed so as to achieve ±10 parts per million (ppm) homogeneity over a 60 mm diameter imaging volume. We have demonstrated that B 0 temporal stability is fit-for-purpose despite the magnet operating in the driven mode. The system has produced good quality spin-echo and gradient echo images. This compact HTS-MRI system is emerging as a true alternative to conventional low temperature superconductor based cryogen-free MRI systems, with much more efficient cryogenics since it operates entirely from a single phase alternating current electrical supply.
Face verification system for Android mobile devices using histogram based features
NASA Astrophysics Data System (ADS)
Sato, Sho; Kobayashi, Kazuhiro; Chen, Qiu
2016-07-01
This paper proposes a face verification system that runs on Android mobile devices. In this system, facial image is captured by a built-in camera on the Android device firstly, and then face detection is implemented using Haar-like features and AdaBoost learning algorithm. The proposed system verify the detected face using histogram based features, which are generated by binary Vector Quantization (VQ) histogram using DCT coefficients in low frequency domains, as well as Improved Local Binary Pattern (Improved LBP) histogram in spatial domain. Verification results with different type of histogram based features are first obtained separately and then combined by weighted averaging. We evaluate our proposed algorithm by using publicly available ORL database and facial images captured by an Android tablet.
A novel drill design for photoacoustic guided surgeries
NASA Astrophysics Data System (ADS)
Shubert, Joshua; Lediju Bell, Muyinatu A.
2018-02-01
Fluoroscopy is currently the standard approach for image guidance of surgical drilling procedures. In addition to the harmful radiation dose to the patient and surgeon, fluoroscopy fails to visualize critical structures such as blood vessels and nerves within the drill path. Photoacoustic imaging is a well-suited imaging method to visualize these structures and it does not require harmful ionizing radiation. However, there is currently no clinical system available to deliver light to occluded drill bit tips. To address this challenge, a prototype drill was designed, built, and tested using an internal light delivery system that allows laser energy to be transferred from a stationary laser source to the tip of a spinning drill bit. Photoacoustic images were successfully obtained with the drill bit submerged in water and with the drill tip inserted into a thoracic vertebra from a human cadaver.
Fast frame rate rodent cardiac x-ray imaging using scintillator lens coupled to CMOS camera
NASA Astrophysics Data System (ADS)
Swathi Lakshmi, B.; Sai Varsha, M. K. N.; Kumar, N. Ashwin; Dixit, Madhulika; Krishnamurthi, Ganapathy
2017-03-01
Micro-Computed Tomography (MCT) systems for small animal imaging plays a critical role for monitoring disease progression and therapy evaluation. In this work, an in-house built micro-CT system equipped with a X-ray scintillator lens coupled to a commercial CMOS camera was used to test the feasibility of its application to Digital Subtraction Angiography (DSA). Literature has reported such studies being done with clinical X-ray tubes that can be pulsed rapidly or with rotating gantry systems, thus increasing the cost and infrastructural requirements.The feasibility of DSA was evaluated by injected Iodinated contrast agent (ICA) through the tail vein of a mouse. Projection images of the heart were acquired pre and post contrast using the high frame rate X-ray detector and processing done to visualize transit of ICA through the heart.
A low cost imaging displacement measurement system for spacecraft thermal vacuum testing
NASA Technical Reports Server (NTRS)
Dempsey, Brian
2006-01-01
A low cost imaging displacement technique suitable for use in thermal vacuum testing was built and tested during thermal vacuum testing of the space infrared telescope facility (SIRTF, later renamed Spitzer infrared telescope facility). The problem was to measure the relative displacement of different portions of the spacecraft due to thermal expansion or contraction. Standard displacement measuring instrumentation could not be used because of the widely varying temperatures on the spacecraft and for fear of invalidating the thermal vacuum testing. The imaging system was conceived, designed, purchased, and installed in approximately 2 months at very low cost. The system performed beyond expectations proving that sub millimeter displacements could be measured from over 2 meters away. Using commercial optics it was possible to make displacement measurements down to 10 (mu)m. An automated image processing tool was used to process the data, which not only speeded up data reduction, but showed that velocities and accelerations could also be measured. Details of the design and capabilities of the system are discussed along with the results of the test on the observatory. Several images from the actual test are presented.
A Low Power, Parallel Wearable Multi-Sensor System for Human Activity Evaluation.
Li, Yuecheng; Jia, Wenyan; Yu, Tianjian; Luan, Bo; Mao, Zhi-Hong; Zhang, Hong; Sun, Mingui
2015-04-01
In this paper, the design of a low power heterogeneous wearable multi-sensor system, built with Zynq System-on-Chip (SoC), for human activity evaluation is presented. The powerful data processing capability and flexibility of this SoC represent significant improvements over our previous ARM based system designs. The new system captures and compresses multiple color images and sensor data simultaneously. Several strategies are adopted to minimize power consumption. Our wearable system provides a new tool for the evaluation of human activity, including diet, physical activity and lifestyle.
An information gathering system for medical image inspection
NASA Astrophysics Data System (ADS)
Lee, Young-Jin; Bajcsy, Peter
2005-04-01
We present an information gathering system for medical image inspection that consists of software tools for capturing computer-centric and human-centric information. Computer-centric information includes (1) static annotations, such as (a) image drawings enclosing any selected area, a set of areas with similar colors, a set of salient points, and (b) textual descriptions associated with either image drawings or links between pairs of image drawings, and (2) dynamic (or temporal) information, such as mouse movements, zoom level changes, image panning and frame selections from an image stack. Human-centric information is represented by video and audio signals that are acquired by computer-mounted cameras and microphones. The short-term goal of the presented system is to facilitate learning of medical novices from medical experts, while the long-term goal is to data mine all information about image inspection for assisting in making diagnoses. In this work, we built basic software functionality for gathering computer-centric and human-centric information of the aforementioned variables. Next, we developed the information playback capabilities of all gathered information for educational purposes. Finally, we prototyped text-based and image template-based search engines to retrieve information from recorded annotations, for example, (a) find all annotations containing the word "blood vessels", or (b) search for similar areas to a selected image area. The information gathering system for medical image inspection reported here has been tested with images from the Histology Atlas database.
Kiely, Daniel J; Stephanson, Kirk; Ross, Sue
2011-10-01
Low-cost laparoscopic box trainers built using home computers and webcams may provide residents with a useful tool for practice at home. This study set out to evaluate the image quality of low-cost laparoscopic box trainers compared with a commercially available model. Five low-cost laparoscopic box trainers including the components listed were compared in random order to one commercially available box trainer: A (high-definition USB 2.0 webcam, PC laptop), B (Firewire webcam, Mac laptop), C (high-definition USB 2.0 webcam, Mac laptop), D (standard USB webcam, PC desktop), E (Firewire webcam, PC desktop), and F (the TRLCD03 3-DMEd Standard Minimally Invasive Training System). Participants observed still image quality and performed a peg transfer task using each box trainer. Participants rated still image quality, image quality with motion, and whether the box trainer had sufficient image quality to be useful for training. Sixteen residents in obstetrics and gynecology took part in the study. The box trainers showing no statistically significant difference from the commercially available model were A, B, C, D, and E for still image quality; A for image quality with motion; and A and B for usefulness of the simulator based on image quality. The cost of the box trainers A-E is approximately $100 to $160 each, not including a computer or laparoscopic instruments. Laparoscopic box trainers built from a high-definition USB 2.0 webcam with a PC (box trainer A) or from a Firewire webcam with a Mac (box trainer B) provide image quality comparable with a commercial standard.
Pose-variant facial expression recognition using an embedded image system
NASA Astrophysics Data System (ADS)
Song, Kai-Tai; Han, Meng-Ju; Chang, Shuo-Hung
2008-12-01
In recent years, one of the most attractive research areas in human-robot interaction is automated facial expression recognition. Through recognizing the facial expression, a pet robot can interact with human in a more natural manner. In this study, we focus on the facial pose-variant problem. A novel method is proposed in this paper to recognize pose-variant facial expressions. After locating the face position in an image frame, the active appearance model (AAM) is applied to track facial features. Fourteen feature points are extracted to represent the variation of facial expressions. The distance between feature points are defined as the feature values. These feature values are sent to a support vector machine (SVM) for facial expression determination. The pose-variant facial expression is classified into happiness, neutral, sadness, surprise or anger. Furthermore, in order to evaluate the performance for practical applications, this study also built a low resolution database (160x120 pixels) using a CMOS image sensor. Experimental results show that the recognition rate is 84% with the self-built database.
Data on spatiotemporal urban sprawl of Dire Dawa City, Eastern Ethiopia.
Taffa, Chaltu; Mekonen, Teferi; Mulugeta, Messay; Tesfaye, Bechaye
2017-06-01
The data presented in this paper shows the spatiotemporal expansion of Dire Dawa City (eastern Ethiopia) and the ensuing land use land cover changes in its peri-urban areas between 1985 and 2015. The data were generated from satellite images of Thematic Mapper (TM), Enhanced Thematic Mapper-Plus (ETM+) and OLI (Operational Land Image) with path/raw value of 166/053 by using Arc GIS 10.1 software. The precision of the images was verified by geolocation data collected from ground control points by using Geographic Positioning System (GPS) receiver. Four LULC classes (built up area, vegetation, barren land and farmland) with their respective spatiotemporal dimensions were clearly identified in the analysis. Built up area had shown an overall annual increment of 15.8% (82 ha per year) from 517 ha in 1985 to 2976 ha in 2015. Expansion took place in all directions but it was more pronounced along the main road towards other nearby towns, recently established business/service areas and the Industrial Park. Barren land, farmland and vegetation areas showed speedy decline over the years.
Image quality assessment for video stream recognition systems
NASA Astrophysics Data System (ADS)
Chernov, Timofey S.; Razumnuy, Nikita P.; Kozharinov, Alexander S.; Nikolaev, Dmitry P.; Arlazarov, Vladimir V.
2018-04-01
Recognition and machine vision systems have long been widely used in many disciplines to automate various processes of life and industry. Input images of optical recognition systems can be subjected to a large number of different distortions, especially in uncontrolled or natural shooting conditions, which leads to unpredictable results of recognition systems, making it impossible to assess their reliability. For this reason, it is necessary to perform quality control of the input data of recognition systems, which is facilitated by modern progress in the field of image quality evaluation. In this paper, we investigate the approach to designing optical recognition systems with built-in input image quality estimation modules and feedback, for which the necessary definitions are introduced and a model for describing such systems is constructed. The efficiency of this approach is illustrated by the example of solving the problem of selecting the best frames for recognition in a video stream for a system with limited resources. Experimental results are presented for the system for identity documents recognition, showing a significant increase in the accuracy and speed of the system under simulated conditions of automatic camera focusing, leading to blurring of frames.
Development of x-ray imaging technique for liquid screening at airport
DOE Office of Scientific and Technical Information (OSTI.GOV)
Sulaiman, Nurhani binti, E-mail: nhani.sulaiman@gmail.com; Srisatit, Somyot, E-mail: somyot.s@chula.ac.th
2016-01-22
X-ray imaging technology is a viable option to recognize flammable liquids for the purposes of aviation security. In this study, an X-ray imaging technology was developed whereby, the image viewing system was built with the use of a digital camera coupled with a gadolinium oxysulfide (GOS) fluorescent screen. The camera was equipped with a software for remote control setting of the camera via a USB cable which allows the images to be captured. The image was analysed to determine the average grey level using a software designed by Microsoft Visual Basic 6.0. The data was obtained for various densities ofmore » liquid thickness of 4.5 cm, 6.0 cm and 7.5 cm respectively for X-ray energies ranging from 70 to 200 kVp. In order to verify the reliability of the constructed calibration data, the system was tested with a few types of unknown liquids. The developed system could be conveniently employed for security screening in order to discriminate between a threat and an innocuous liquid.« less
Zhao, Qiaole; Schelen, Ben; Schouten, Raymond; van den Oever, Rein; Leenen, René; van Kuijk, Harry; Peters, Inge; Polderdijk, Frank; Bosiers, Jan; Raspe, Marcel; Jalink, Kees; Geert Sander de Jong, Jan; van Geest, Bert; Stoop, Karel; Young, Ian Ted
2012-12-01
We have built an all-solid-state camera that is directly modulated at the pixel level for frequency-domain fluorescence lifetime imaging microscopy (FLIM) measurements. This novel camera eliminates the need for an image intensifier through the use of an application-specific charge coupled device design in a frequency-domain FLIM system. The first stage of evaluation for the camera has been carried out. Camera characteristics such as noise distribution, dark current influence, camera gain, sampling density, sensitivity, linearity of photometric response, and optical transfer function have been studied through experiments. We are able to do lifetime measurement using our modulated, electron-multiplied fluorescence lifetime imaging microscope (MEM-FLIM) camera for various objects, e.g., fluorescein solution, fixed green fluorescent protein (GFP) cells, and GFP-actin stained live cells. A detailed comparison of a conventional microchannel plate (MCP)-based FLIM system and the MEM-FLIM system is presented. The MEM-FLIM camera shows higher resolution and a better image quality. The MEM-FLIM camera provides a new opportunity for performing frequency-domain FLIM.
Volume estimation using food specific shape templates in mobile image-based dietary assessment
NASA Astrophysics Data System (ADS)
Chae, Junghoon; Woo, Insoo; Kim, SungYe; Maciejewski, Ross; Zhu, Fengqing; Delp, Edward J.; Boushey, Carol J.; Ebert, David S.
2011-03-01
As obesity concerns mount, dietary assessment methods for prevention and intervention are being developed. These methods include recording, cataloging and analyzing daily dietary records to monitor energy and nutrient intakes. Given the ubiquity of mobile devices with built-in cameras, one possible means of improving dietary assessment is through photographing foods and inputting these images into a system that can determine the nutrient content of foods in the images. One of the critical issues in such the image-based dietary assessment tool is the accurate and consistent estimation of food portion sizes. The objective of our study is to automatically estimate food volumes through the use of food specific shape templates. In our system, users capture food images using a mobile phone camera. Based on information (i.e., food name and code) determined through food segmentation and classification of the food images, our system choose a particular food template shape corresponding to each segmented food. Finally, our system reconstructs the three-dimensional properties of the food shape from a single image by extracting feature points in order to size the food shape template. By employing this template-based approach, our system automatically estimates food portion size, providing a consistent method for estimation food volume.
Real-time target tracking and locating system for UAV
NASA Astrophysics Data System (ADS)
Zhang, Chao; Tang, Linbo; Fu, Huiquan; Li, Maowen
2017-07-01
In order to achieve real-time target tracking and locating for UAV, a reliable processing system is built on the embedded platform. Firstly, the video image is acquired in real time by the photovoltaic system on the UAV. When the target information is known, KCF tracking algorithm is adopted to track the target. Then, the servo is controlled to rotate with the target, when the target is in the center of the image, the laser ranging module is opened to obtain the distance between the UAV and the target. Finally, to combine with UAV flight parameters obtained by BeiDou navigation system, through the target location algorithm to calculate the geodetic coordinates of the target. The results show that the system is stable for real-time tracking of targets and positioning.
Applying a visual language for image processing as a graphical teaching tool in medical imaging
NASA Astrophysics Data System (ADS)
Birchman, James J.; Tanimoto, Steven L.; Rowberg, Alan H.; Choi, Hyung-Sik; Kim, Yongmin
1992-05-01
Typical user interaction in image processing is with command line entries, pull-down menus, or text menu selections from a list, and as such is not generally graphical in nature. Although applying these interactive methods to construct more sophisticated algorithms from a series of simple image processing steps may be clear to engineers and programmers, it may not be clear to clinicians. A solution to this problem is to implement a visual programming language using visual representations to express image processing algorithms. Visual representations promote a more natural and rapid understanding of image processing algorithms by providing more visual insight into what the algorithms do than the interactive methods mentioned above can provide. Individuals accustomed to dealing with images will be more likely to understand an algorithm that is represented visually. This is especially true of referring physicians, such as surgeons in an intensive care unit. With the increasing acceptance of picture archiving and communications system (PACS) workstations and the trend toward increasing clinical use of image processing, referring physicians will need to learn more sophisticated concepts than simply image access and display. If the procedures that they perform commonly, such as window width and window level adjustment and image enhancement using unsharp masking, are depicted visually in an interactive environment, it will be easier for them to learn and apply these concepts. The software described in this paper is a visual programming language for imaging processing which has been implemented on the NeXT computer using NeXTstep user interface development tools and other tools in an object-oriented environment. The concept is based upon the description of a visual language titled `Visualization of Vision Algorithms' (VIVA). Iconic representations of simple image processing steps are placed into a workbench screen and connected together into a dataflow path by the user. As the user creates and edits a dataflow path, more complex algorithms can be built on the screen. Once the algorithm is built, it can be executed, its results can be reviewed, and operator parameters can be interactively adjusted until an optimized output is produced. The optimized algorithm can then be saved and added to the system as a new operator. This system has been evaluated as a graphical teaching tool for window width and window level adjustment, image enhancement using unsharp masking, and other techniques.
Accommodation-based liquid crystal adaptive optics system for large ocular aberration correction.
Mu, Quanquan; Cao, Zhaoliang; Li, Chao; Jiang, Baoguang; Hu, Lifa; Xuan, Li
2008-12-15
According to ocular aberration property and liquid crystal (LC) corrector characteristics, we calculated the minimum pixel demand of the LC corrector used for compensating large ocular aberrations. Then, an accommodation based optical configuration was introduced to reduce the demand. Based on this an adaptive optics (AO) retinal imaging system was built. Subjects with different defocus and astigmatism were tested to prove this. For myopia lower than 5D it performs well. When myopia is as large as 8D the accommodation error increased to nearly 3D, which requires the LC corrector to have 667 x 667 pixels to get a well-corrected image.
MyFreePACS: a free web-based radiology image storage and viewing tool.
de Regt, David; Weinberger, Ed
2004-08-01
We developed an easy-to-use method for central storage and subsequent viewing of radiology images for use on any PC equipped with Internet Explorer. We developed MyFreePACS, a program that uses a DICOM server to receive and store images and transmit them over the Web to the MyFreePACS Web client. The MyFreePACS Web client is a Web page that uses an ActiveX control for viewing and manipulating images. The client contains many of the tools found in modern image viewing stations including 3D localization and multiplanar reformation. The system is built entirely with free components and is freely available for download and installation from the Web at www.myfreepacs.com.
Development and bench testing of a multi-spectral imaging technology built on a smartphone platform
NASA Astrophysics Data System (ADS)
Bolton, Frank J.; Weiser, Reuven; Kass, Alex J.; Rose, Donny; Safir, Amit; Levitz, David
2016-03-01
Cervical cancer screening presents a great challenge for clinicians across the developing world. In many countries, cervical cancer screening is done by visualization with the naked eye. Simple brightfield white light imaging with photo documentation has been shown to make a significant impact on cervical cancer care. Adoption of smartphone based cervical imaging devices is increasing across Africa. However, advanced imaging technologies such as multispectral imaging systems, are seldom deployed in low resource settings, where they are needed most. To address this challenge, the optical system of a smartphone-based mobile colposcopy imaging system was refined, integrating components required for low cost, portable multi-spectral imaging of the cervix. This paper describes the refinement of the mobile colposcope to enable it to acquire images of the cervix at multiple illumination wavelengths, including modeling and laboratory testing. Wavelengths were selected to enable quantifying the main absorbers in tissue (oxyand deoxy-hemoglobin, and water), as well as scattering parameters that describe the size distribution of scatterers. The necessary hardware and software modifications are reviewed. Initial testing suggests the multi-spectral mobile device holds promise for use in low-resource settings.
The Wide Angle Camera of the ROSETTA Mission
NASA Astrophysics Data System (ADS)
Barbieri, C.; Fornasier, S.; Verani, S.; Bertini, I.; Lazzarin, M.; Rampazzi, F.; Cremonese, G.; Ragazzoni, R.; Marzari, F.; Angrilli, F.; Bianchini, G. A.; Debei, S.; Dececco, M.; Guizzo, G.; Parzianello, G.; Ramous, P.; Saggin, B.; Zaccariotto, M.; Da Deppo, V.; Naletto, G.; Nicolosi, G.; Pelizzo, M. G.; Tondello, G.; Brunello, P.; Peron, F.
This paper aims to give a brief description of the Wide Angle Camera (WAC), built by the Centro Servizi e AttivitàSpaziali (CISAS) of the University of Padova for the ESA ROSETTA Mission to comet 46P/Wirtanen and asteroids 4979 Otawara and 140 Siwa. The WAC is part of the OSIRIS imaging system, which comprises also a Narrow Angle Camera (NAC) built by the Laboratoire d'Astrophysique Spatiale (LAS) of Marseille. CISAS had also the responsibility to build the shutter and the front cover mechanism for the NAC. The flight model of the WAC was delivered in December 2001, and has been already integrated on ROSETTA.
Demonstration of Uncued Optical Surveillance of LEO
NASA Astrophysics Data System (ADS)
Zimmer, P.; Ackermann, M.; McGraw, J.
2014-09-01
J.T. McGraw and Associates, LLC, in collaboration with the University of New Mexico (UNM), has built and is operating two proof-of-concept wide-field imaging systems to test novel techniques for uncued surveillance of LEO. The imaging systems are built from off-the-shelf optics and detectors resulting in a 350mm aperture and a 6 square degree field of view. For streak detection, field of view is of critical importance because the maximum exposure time on the object is limited by its crossing time and measurements of apparent angular motion are better constrained with longer streaks. The current match of the detector to the optical system is optimized for detection of objects at altitudes above 450km, which for a circular orbit, corresponds to apparent motions of approximately 1 deg./sec. Using our GPU-accelerated detection scheme, the proof-of-concept systems have detected objects fainter than V=12.3, which approximately corresponds to a 24 cm object at 1000 km altitude at better than 6 sigma significance, from sites near and within Albuquerque, NM. This work demonstrates scalable optical systems designed for near real time detection of fast moving objects, which can be then handed off to other instruments capable of tracking and characterizing them. The two proof-of-concept systems, separated by ~30km, work together by taking simultaneous images of the same orbital volume to constrain the orbits of detected objects using parallax measurements. These detections are followed-up by photometric observations taken at UNM to independently assess the objects and the quality of the derived orbits. We believe this demonstrates the potential of small telescope arrays for detecting and cataloguing heretofore unknown LEO objects.
Dual-Modality PET/Ultrasound imaging of the Prostate
DOE Office of Scientific and Technical Information (OSTI.GOV)
Huber, Jennifer S.; Moses, William W.; Pouliot, Jean
2005-11-11
Functional imaging with positron emission tomography (PET)will detect malignant tumors in the prostate and/or prostate bed, as well as possibly help determine tumor ''aggressiveness''. However, the relative uptake in a prostate tumor can be so great that few other anatomical landmarks are visible in a PET image. Ultrasound imaging with a transrectal probe provides anatomical detail in the prostate region that can be co-registered with the sensitive functional information from the PET imaging. Imaging the prostate with both PET and transrectal ultrasound (TRUS) will help determine the location of any cancer within the prostate region. This dual-modality imaging should helpmore » provide better detection and treatment of prostate cancer. LBNL has built a high performance positron emission tomograph optimized to image the prostate.Compared to a standard whole-body PET camera, our prostate-optimized PET camera has the same sensitivity and resolution, less backgrounds and lower cost. We plan to develop the hardware and software tools needed for a validated dual PET/TRUS prostate imaging system. We also plan to develop dual prostate imaging with PET and external transabdominal ultrasound, in case the TRUS system is too uncomfortable for some patients. We present the design and intended clinical uses for these dual imaging systems.« less
NASA Astrophysics Data System (ADS)
Bramhe, V. S.; Ghosh, S. K.; Garg, P. K.
2018-04-01
With rapid globalization, the extent of built-up areas is continuously increasing. Extraction of features for classifying built-up areas that are more robust and abstract is a leading research topic from past many years. Although, various studies have been carried out where spatial information along with spectral features has been utilized to enhance the accuracy of classification. Still, these feature extraction techniques require a large number of user-specific parameters and generally application specific. On the other hand, recently introduced Deep Learning (DL) techniques requires less number of parameters to represent more abstract aspects of the data without any manual effort. Since, it is difficult to acquire high-resolution datasets for applications that require large scale monitoring of areas. Therefore, in this study Sentinel-2 image has been used for built-up areas extraction. In this work, pre-trained Convolutional Neural Networks (ConvNets) i.e. Inception v3 and VGGNet are employed for transfer learning. Since these networks are trained on generic images of ImageNet dataset which are having very different characteristics from satellite images. Therefore, weights of networks are fine-tuned using data derived from Sentinel-2 images. To compare the accuracies with existing shallow networks, two state of art classifiers i.e. Gaussian Support Vector Machine (SVM) and Back-Propagation Neural Network (BP-NN) are also implemented. Both SVM and BP-NN gives 84.31 % and 82.86 % overall accuracies respectively. Inception-v3 and VGGNet gives 89.43 % of overall accuracy using fine-tuned VGGNet and 92.10 % when using Inception-v3. The results indicate high accuracy of proposed fine-tuned ConvNets on a 4-channel Sentinel-2 dataset for built-up area extraction.
Automated eye blink detection and correction method for clinical MR eye imaging.
Wezel, Joep; Garpebring, Anders; Webb, Andrew G; van Osch, Matthias J P; Beenakker, Jan-Willem M
2017-07-01
To implement an on-line monitoring system to detect eye blinks during ocular MRI using field probes, and to reacquire corrupted k-space lines by means of an automatic feedback system integrated with the MR scanner. Six healthy subjects were scanned on a 7 Tesla MRI whole-body system using a custom-built receive coil. Subjects were asked to blink multiple times during the MR-scan. The local magnetic field changes were detected with an external fluorine-based field probe which was positioned close to the eye. The eye blink produces a field shift greater than a threshold level, this was communicated in real-time to the MR system which immediately reacquired the motion-corrupted k-space lines. The uncorrected images, using the original motion-corrupted data, showed severe artifacts, whereas the corrected images, using the reacquired data, provided an image quality similar to images acquired without blinks. Field probes can successfully detect eye blinks during MRI scans. By automatically reacquiring the eye blink-corrupted data, high quality MR-images of the eye can be acquired. Magn Reson Med 78:165-171, 2017. © 2016 International Society for Magnetic Resonance in Medicine. © 2016 International Society for Magnetic Resonance in Medicine.
New uses of position-sensitive photomultiplier tubes
NASA Astrophysics Data System (ADS)
Gordon, Jeffrey S.; Redus, Robert H.; Nagarkar, Vivek V.; Squillante, Michael R.
1992-12-01
Recent advances in photomultiplier tube technology have led to the availability of position sensitive photomultiplier tubes (PSPMTs). These tubes make it possible to build a new generation of imaging instruments for gamma rays and other types of ionizing radiation. We have investigated the use of these tubes for the construction of several prototype instruments. The first application investigated measures the quantity and distribution of radioactive compounds on filter papers used in microbiology research. The intent of this instrument is to replace film autoradiography with an electronic imaging system which can analyze samples 75 to 110 times faster than film. The second application involved the development of an intraoperative imaging probe to help surgeons identify cancerous tissue and ensure its complete removal. This instrument will replace a non-imaging probe now in use at many hospitals. A third prototype instrument under evaluation is an imaging nuclear survey system which obtains both a video and gamma ray image for the purpose of locating and quantifying radioactive materials. This system would be used at nuclear power plants and radioactive materials preparation facilities. A modification of this system could be built into robots used for inspecting and repairing power plants.
A restraint-free small animal SPECT imaging system with motion tracking
DOE Office of Scientific and Technical Information (OSTI.GOV)
Weisenberger, A.G.; Gleason, S.S.; Goddard, J.
2005-06-01
We report on an approach toward the development of a high-resolution single photon emission computed tomography (SPECT) system to image the biodistribution of radiolabeled tracers such as Tc-99m and I-125 in unrestrained/unanesthetized mice. An infrared (IR)-based position tracking apparatus has been developed and integrated into a SPECT gantry. The tracking system is designed to measure the spatial position of a mouse's head at a rate of 10-15 frames per second with submillimeter accuracy. The high-resolution, gamma imaging detectors are based on pixellated NaI(Tl) crystal scintillator arrays, position-sensitive photomultiplier tubes, and novel readout circuitry requiring fewer analog-digital converter (ADC) channels whilemore » retaining high spatial resolution. Two SPECT gamma camera detector heads based upon position-sensitive photomultiplier tubes have been built and installed onto the gantry. The IR landmark-based pose measurement and tracking system is under development to provide animal position data during a SPECT scan. The animal position and orientation data acquired by the tracking system will be used for motion correction during the tomographic image reconstruction.« less
NASA Astrophysics Data System (ADS)
Newswander, T.; Riesland, David W.; Miles, Duane; Reinhart, Lennon
2017-09-01
For space optical systems that image extended scenes such as earth-viewing systems, modulation transfer function (MTF) test data is directly applicable to system optical resolution. For many missions, it is the most direct metric for establishing the best focus of the instrument. Additionally, MTF test products can be combined to predict overall imaging performance. For fixed focus instruments, finding the best focus during ground testing is critical to achieving good imaging performance. The ground testing should account for the full-imaging system, operational parameters, and operational environment. Testing the full-imaging system removes uncertainty caused by breaking configurations and the combination of multiple subassembly test results. For earth viewing, the imaging system needs to be tested at infinite conjugate. Operational environment test conditions should include temperature and vacuum. Optical MTF testing in the presence of operational vibration and gravity release is less straightforward and may not be possible on the ground. Gravity effects are mitigated by testing in multiple orientations. Many space telescope systems are designed and built to have optimum performance in a gravity-free environment. These systems can have imaging performance that is dominated by aberration including astigmatism. This paper discusses how the slanted edge MTF test is applied to determine the best focus of a space optical telescope in ground testing accounting for gravity sag effects. Actual optical system test results and conclusions are presented.
A technique to measure strain distributions in single wood pulp fibers
Laurence Mott; Stephen M. Shaler; Leslie H. Groom
1996-01-01
Environmental scanning electron microscopy (ESEM) and digital image correlation (DIC) were used to measure microstrain distributions on the surface of wood pulp fibers. A loading stage incorporating a fiber gripping system was designed and built by the authors. Fitted to the tensile substage of an ESEM or a Polymer Laboratories MINIMAT tester, it provided a reliable...
Imaging Flash Lidar for Safe Landing on Solar System Bodies and Spacecraft Rendezvous and Docking
NASA Technical Reports Server (NTRS)
Amzajerdian, Farzin; Roback, Vincent E.; Bulyshev, Alexander E.; Brewster, Paul F.; Carrion, William A; Pierrottet, Diego F.; Hines, Glenn D.; Petway, Larry B.; Barnes, Bruce W.; Noe, Anna M.
2015-01-01
NASA has been pursuing flash lidar technology for autonomous, safe landing on solar system bodies and for automated rendezvous and docking. During the final stages of the landing from about 1 kilometer to 500 meters above the ground, the flash lidar can generate 3-Dimensional images of the terrain to identify hazardous features such as craters, rocks, and steep slopes. The onboard flight computer can then use the 3-D map of terrain to guide the vehicle to a safe location. As an automated rendezvous and docking sensor, the flash lidar can provide relative range, velocity, and bearing from an approaching spacecraft to another spacecraft or a space station. NASA Langley Research Center has developed and demonstrated a flash lidar sensor system capable of generating 16,000 pixels range images with 7 centimeters precision, at 20 Hertz frame rate, from a maximum slant range of 1800 m from the target area. This paper describes the lidar instrument and presents the results of recent flight tests onboard a rocket-propelled free-flyer vehicle (Morpheus) built by NASA Johnson Space Center. The flights were conducted at a simulated lunar terrain site, consisting of realistic hazard features and designated landing areas, built at NASA Kennedy Space Center specifically for this demonstration test. This paper also provides an overview of the plan for continued advancement of the flash lidar technology aimed at enhancing its performance to meet both landing and automated rendezvous and docking applications.
The impact of exposure to films of natural and built environments on state body appreciation.
Swami, Viren; Pickering, Mark; Barron, David; Patel, Shreepali
2018-06-12
Previous work has shown that exposure to images of nature results in elevated state body appreciation, but static images may lack ecological validity. Here, we examined the impact of exposure to short films of simulated, first-person walks in natural or built environments. Thirty-six university students completed a measure of state body appreciation before and after watching films of either a walk in a natural or a built environment created specifically for the present study. Two weeks later, they completed the same task but watched the other film type. Results indicated that exposure to the film of a natural environment resulted in significantly elevated state body appreciation (d = 0.66). There was no significant change in state body appreciation following exposure to the film of the built environment (d = 0.14). These findings suggest that exposure to films depicting the natural environment may promote immediate, moderate-sized improvements in state body image. Copyright © 2018 Elsevier Ltd. All rights reserved.
A Web-based telemedicine system for diabetic retinopathy screening using digital fundus photography.
Wei, Jack C; Valentino, Daniel J; Bell, Douglas S; Baker, Richard S
2006-02-01
The purpose was to design and implement a Web-based telemedicine system for diabetic retinopathy screening using digital fundus cameras and to make the software publicly available through Open Source release. The process of retinal imaging and case reviewing was modeled to optimize workflow and implement use of computer system. The Web-based system was built on Java Servlet and Java Server Pages (JSP) technologies. Apache Tomcat was chosen as the JSP engine, while MySQL was used as the main database and Laboratory of Neuro Imaging (LONI) Image Storage Architecture, from the LONI-UCLA, as the platform for image storage. For security, all data transmissions were carried over encrypted Internet connections such as Secure Socket Layer (SSL) and HyperText Transfer Protocol over SSL (HTTPS). User logins were required and access to patient data was logged for auditing. The system was deployed at Hubert H. Humphrey Comprehensive Health Center and Martin Luther King/Drew Medical Center of Los Angeles County Department of Health Services. Within 4 months, 1500 images of more than 650 patients were taken at Humphrey's Eye Clinic and successfully transferred to King/Drew's Department of Ophthalmology. This study demonstrates an effective architecture for remote diabetic retinopathy screening.
In vivo imaging and vibration measurement of Guinea pig cochlea
NASA Astrophysics Data System (ADS)
Choudhury, Niloy; Chen, Fangyi; Zheng, Jiefu; Nuttall, Alfred L.; Jacques, Steven L.
2008-02-01
An optical coherence tomography (OCT) system was built to acquire in vivo, both images and vibration measurements of the organ of Corti of the guinea pig. The organ of Corti was viewed through a ~500-μm diameter hole in the bony wall of the scala tympani of the first cochlear turn. In imaging mode, the image was acquired as reflectance R(x,z). In vibration mode, the basilar membrane (BM) or reticular lamina (RL) was selected based on the image. Under software control, the system would move the scanning mirrors to bring the sensing volume of the measurement to the desired tissue location. To address the gain stability problem of the homodyne OCT system, arising from the system moving in and out of the quadrature point and also to resolve the 180 degree ambiguity in the phase measurement using an interferometer, a vibration calibration method is developed by adding a vibrating source to the reference arm to monitor the operating point of the interferometric system. Amplitude gain and phase of various cochlear membranes was measured for different sound pressure level (SPL) varying from 65dB SPL to 93 dB SPL.
Wavelet optimization for content-based image retrieval in medical databases.
Quellec, G; Lamard, M; Cazuguel, G; Cochener, B; Roux, C
2010-04-01
We propose in this article a content-based image retrieval (CBIR) method for diagnosis aid in medical fields. In the proposed system, images are indexed in a generic fashion, without extracting domain-specific features: a signature is built for each image from its wavelet transform. These image signatures characterize the distribution of wavelet coefficients in each subband of the decomposition. A distance measure is then defined to compare two image signatures and thus retrieve the most similar images in a database when a query image is submitted by a physician. To retrieve relevant images from a medical database, the signatures and the distance measure must be related to the medical interpretation of images. As a consequence, we introduce several degrees of freedom in the system so that it can be tuned to any pathology and image modality. In particular, we propose to adapt the wavelet basis, within the lifting scheme framework, and to use a custom decomposition scheme. Weights are also introduced between subbands. All these parameters are tuned by an optimization procedure, using the medical grading of each image in the database to define a performance measure. The system is assessed on two medical image databases: one for diabetic retinopathy follow up and one for screening mammography, as well as a general purpose database. Results are promising: a mean precision of 56.50%, 70.91% and 96.10% is achieved for these three databases, when five images are returned by the system. Copyright 2009 Elsevier B.V. All rights reserved.
Baikejiang, Reheman; Zhang, Wei; Li, Changqing
2017-01-01
Diffuse optical tomography (DOT) has attracted attentions in the last two decades due to its intrinsic sensitivity in imaging chromophores of tissues such as hemoglobin, water, and lipid. However, DOT has not been clinically accepted yet due to its low spatial resolution caused by strong optical scattering in tissues. Structural guidance provided by an anatomical imaging modality enhances the DOT imaging substantially. Here, we propose a computed tomography (CT) guided multispectral DOT imaging system for breast cancer imaging. To validate its feasibility, we have built a prototype DOT imaging system which consists of a laser at the wavelength of 650 nm and an electron multiplying charge coupled device (EMCCD) camera. We have validated the CT guided DOT reconstruction algorithms with numerical simulations and phantom experiments, in which different imaging setup parameters, such as projection number of measurements and width of measurement patch, have been investigated. Our results indicate that an air-cooling EMCCD camera is good enough for the transmission mode DOT imaging. We have also found that measurements at six angular projections are sufficient for DOT to reconstruct the optical targets with 2 and 4 times absorption contrast when the CT guidance is applied. Finally, we have described our future research plan on integration of a multispectral DOT imaging system into a breast CT scanner.
Automated Detection of Microaneurysms Using Scale-Adapted Blob Analysis and Semi-Supervised Learning
DOE Office of Scientific and Technical Information (OSTI.GOV)
Adal, Kedir M.; Sidebe, Desire; Ali, Sharib
2014-01-07
Despite several attempts, automated detection of microaneurysm (MA) from digital fundus images still remains to be an open issue. This is due to the subtle nature of MAs against the surrounding tissues. In this paper, the microaneurysm detection problem is modeled as finding interest regions or blobs from an image and an automatic local-scale selection technique is presented. Several scale-adapted region descriptors are then introduced to characterize these blob regions. A semi-supervised based learning approach, which requires few manually annotated learning examples, is also proposed to train a classifier to detect true MAs. The developed system is built using onlymore » few manually labeled and a large number of unlabeled retinal color fundus images. The performance of the overall system is evaluated on Retinopathy Online Challenge (ROC) competition database. A competition performance measure (CPM) of 0.364 shows the competitiveness of the proposed system against state-of-the art techniques as well as the applicability of the proposed features to analyze fundus images.« less
Design and testing of a 750MHz CW-EPR digital console for small animal imaging.
Sato-Akaba, Hideo; Emoto, Miho C; Hirata, Hiroshi; Fujii, Hirotada G
2017-11-01
This paper describes the development of a digital console for three-dimensional (3D) continuous wave electron paramagnetic resonance (CW-EPR) imaging of a small animal to improve the signal-to-noise ratio and lower the cost of the EPR imaging system. A RF generation board, an RF acquisition board and a digital signal processing (DSP) & control board were built for the digital EPR detection. Direct sampling of the reflected RF signal from a resonator (approximately 750MHz), which contains the EPR signal, was carried out using a band-pass subsampling method. A direct automatic control system to reduce the reflection from the resonator was proposed and implemented in the digital EPR detection scheme. All DSP tasks were carried out in field programmable gate array ICs. In vivo 3D imaging of nitroxyl radicals in a mouse's head was successfully performed. Copyright © 2017 Elsevier Inc. All rights reserved.
Calcium neuroimaging in behaving zebrafish larvae using a turn-key light field camera
NASA Astrophysics Data System (ADS)
Cruz Perez, Carlos; Lauri, Antonella; Symvoulidis, Panagiotis; Cappetta, Michele; Erdmann, Arne; Westmeyer, Gil Gregor
2015-09-01
Reconstructing a three-dimensional scene from multiple simultaneously acquired perspectives (the light field) is an elegant scanless imaging concept that can exceed the temporal resolution of currently available scanning-based imaging methods for capturing fast cellular processes. We tested the performance of commercially available light field cameras on a fluorescent microscopy setup for monitoring calcium activity in the brain of awake and behaving reporter zebrafish larvae. The plenoptic imaging system could volumetrically resolve diverse neuronal response profiles throughout the zebrafish brain upon stimulation with an aversive odorant. Behavioral responses of the reporter fish could be captured simultaneously together with depth-resolved neuronal activity. Overall, our assessment showed that with some optimizations for fluorescence microscopy applications, commercial light field cameras have the potential of becoming an attractive alternative to custom-built systems to accelerate molecular imaging research on cellular dynamics.
Calcium neuroimaging in behaving zebrafish larvae using a turn-key light field camera.
Perez, Carlos Cruz; Lauri, Antonella; Symvoulidis, Panagiotis; Cappetta, Michele; Erdmann, Arne; Westmeyer, Gil Gregor
2015-09-01
Reconstructing a three-dimensional scene from multiple simultaneously acquired perspectives (the light field) is an elegant scanless imaging concept that can exceed the temporal resolution of currently available scanning-based imaging methods for capturing fast cellular processes. We tested the performance of commercially available light field cameras on a fluorescent microscopy setup for monitoring calcium activity in the brain of awake and behaving reporter zebrafish larvae. The plenoptic imaging system could volumetrically resolve diverse neuronal response profiles throughout the zebrafish brain upon stimulation with an aversive odorant. Behavioral responses of the reporter fish could be captured simultaneously together with depth-resolved neuronal activity. Overall, our assessment showed that with some optimizations for fluorescence microscopy applications, commercial light field cameras have the potential of becoming an attractive alternative to custom-built systems to accelerate molecular imaging research on cellular dynamics.
Design and testing of a 750 MHz CW-EPR digital console for small animal imaging
NASA Astrophysics Data System (ADS)
Sato-Akaba, Hideo; Emoto, Miho C.; Hirata, Hiroshi; Fujii, Hirotada G.
2017-11-01
This paper describes the development of a digital console for three-dimensional (3D) continuous wave electron paramagnetic resonance (CW-EPR) imaging of a small animal to improve the signal-to-noise ratio and lower the cost of the EPR imaging system. A RF generation board, an RF acquisition board and a digital signal processing (DSP) & control board were built for the digital EPR detection. Direct sampling of the reflected RF signal from a resonator (approximately 750 MHz), which contains the EPR signal, was carried out using a band-pass subsampling method. A direct automatic control system to reduce the reflection from the resonator was proposed and implemented in the digital EPR detection scheme. All DSP tasks were carried out in field programmable gate array ICs. In vivo 3D imaging of nitroxyl radicals in a mouse's head was successfully performed.
Migration of the digital interactive breast-imaging teaching file
NASA Astrophysics Data System (ADS)
Cao, Fei; Sickles, Edward A.; Huang, H. K.; Zhou, Xiaoqiang
1998-06-01
The digital breast imaging teaching file developed during the last two years in our laboratory has been used successfully at UCSF (University of California, San Francisco) as a routine teaching tool for training radiology residents and fellows in mammography. Building on this success, we have ported the teaching file from an old Pixar imaging/Sun SPARC 470 display system to our newly designed telemammography display workstation (Ultra SPARC 2 platform with two DOME Md5/SBX display boards). The old Pixar/Sun 470 system, although adequate for fast and high-resolution image display, is 4- year-old technology, expensive to maintain and difficult to upgrade. The new display workstation is more cost-effective and is also compatible with the digital image format from a full-field direct digital mammography system. The digital teaching file is built on a sophisticated computer-aided instruction (CAI) model, which simulates the management sequences used in imaging interpretation and work-up. Each user can be prompted to respond by making his/her own observations, assessments, and work-up decisions as well as the marking of image abnormalities. This effectively replaces the traditional 'show-and-tell' teaching file experience with an interactive, response-driven type of instruction.
Four-channel magnetic resonance imaging receiver using frequency domain multiplexing.
He, Wang; Qin, Xu; Jiejing, Ren; Gengying, Li
2007-01-01
An alternative technique that uses frequency domain multiplexing to acquire phased array magnetic resonance images is discussed in detail. The proposed method has advantages over traditional independent receiver chains in that it utilizes an analog-to-digital converter and a single-chip multicarrier receiver with high performance to reduce the size and cost of the phased array receiver system. A practical four-channel digital receiver using frequency domain multiplexing was implemented and verified on a home-built 0.3 T magnetic resonance imaging system. The experimental results confirmed that the cross talk between each channel was below -60 dB, the phase fluctuations were about 1 degrees , and there was no obvious signal-to-noise ratio degradation. It is demonstrated that the frequency domain multiplexing is a valuable and economical technique, particularly for array coil systems where the multichannel receiver is indispensable and dynamic range is not a critical problem.
NASA Technical Reports Server (NTRS)
Ostroff, A. J.; Romanczyk, K. C.
1973-01-01
One of the most significant problems associated with the development of large orbiting astronomical telescopes is that of maintaining the very precise pointing accuracy required. A proposed solution to this problem utilizes dual-level pointing control. The primary control system maintains the telescope structure attitude stabilized within the field of view to the desired accuracy. In order to demonstrate the feasibility of optically stabilizing the star images to the desired accuracy a regulating system has been designed and evaluated. The control system utilizes a digital star sensor and an optical star image motion compensator, both of which have been developed for this application. These components have been analyzed mathematically, analytical models have been developed, and hardware has been built and tested.
Qualification of a Quantitative Laryngeal Imaging System Using Videostroboscopy and Videokymography
Popolo, Peter S.; Titze, Ingo R.
2008-01-01
Objectives: We sought to determine whether full-cycle glottal width measurements could be obtained with a quantitative laryngeal imaging system using videostroboscopy, and whether glottal width and vocal fold length measurements were repeatable and reliable. Methods: Synthetic vocal folds were phonated on a laboratory bench, and dynamic images were obtained in repeated trials by use of videostroboscopy and videokymography (VKG) with an imaging system equipped with a 2-point laser projection device for measuring absolute dimensions. Video images were also obtained with an industrial videoscope system with a built-in laser measurement capability. Maximum glottal width and vocal fold length were compared among these 3 methods. Results: The average variation in maximum glottal width measurements between stroboscopic data and VKG data was 3.10%. The average variations in width measurements between the clinical system and the industrial system were 1.93% (stroboscopy) and 3.49% (VKG). The variations in vocal fold length were similarly small. The standard deviations across trials were 0.29 mm for width and 0.48 mm for length (stroboscopy), 0.18 mm for width (VKG), and 0.25 mm for width and 0.84 mm for length (industrial). Conclusions: For stable, periodic vibration, the full extent of the glottal width can be reliably measured with the quantitative videostroboscopy system. PMID:18646436
NASA Astrophysics Data System (ADS)
Damodaran, Vani; Rao, Suresh Ranga; Vasa, Nilesh J.
2016-08-01
In this paper, a study of in-house built optical coherence tomography (OCT) system with a wavelength of 840 nm for imaging of dental caries, progress in demineralisation and cavity restoration is presented. The caries when imaged with the 840 nm OCT system showed minute demineralisation in the order of 5 μm. The OCT system was also proposed to study the growth of lesion and this was demonstrated by artificially inducing caries with a demineralisation solution of pH 4.8. The progress of carious lesion to a depth of about 50-60 μm after 60 hours of demineralisation was clearly observed with the 840 nm OCT system. The tooth samples were subjected to accelerated demineralisation condition at pH of approximately 2.3 to study the adverse effects and the onset of cavity formation was clearly observed. The restoration of cavity was also studied by employing different restorative materials (filled and unfilled). In the case of restoration without filler material (unfilled), the restoration boundaries were clearly observed. Overall, results were comparable with that of the widely used 1310 nm OCT system. In the case of restoration with filler material, the 1310 nm OCT imaging displayed better imaging capacity due to lower scattering than 840 nm imaging.
NASA Astrophysics Data System (ADS)
Ye, Y.
2017-09-01
This paper presents a fast and robust method for the registration of multimodal remote sensing data (e.g., optical, LiDAR, SAR and map). The proposed method is based on the hypothesis that structural similarity between images is preserved across different modalities. In the definition of the proposed method, we first develop a pixel-wise feature descriptor named Dense Orientated Gradient Histogram (DOGH), which can be computed effectively at every pixel and is robust to non-linear intensity differences between images. Then a fast similarity metric based on DOGH is built in frequency domain using the Fast Fourier Transform (FFT) technique. Finally, a template matching scheme is applied to detect tie points between images. Experimental results on different types of multimodal remote sensing images show that the proposed similarity metric has the superior matching performance and computational efficiency than the state-of-the-art methods. Moreover, based on the proposed similarity metric, we also design a fast and robust automatic registration system for multimodal images. This system has been evaluated using a pair of very large SAR and optical images (more than 20000 × 20000 pixels). Experimental results show that our system outperforms the two popular commercial software systems (i.e. ENVI and ERDAS) in both registration accuracy and computational efficiency.
High-density fiber-optic DNA random microsphere array.
Ferguson, J A; Steemers, F J; Walt, D R
2000-11-15
A high-density fiber-optic DNA microarray sensor was developed to monitor multiple DNA sequences in parallel. Microarrays were prepared by randomly distributing DNA probe-functionalized 3.1-microm-diameter microspheres in an array of wells etched in a 500-microm-diameter optical imaging fiber. Registration of the microspheres was performed using an optical encoding scheme and a custom-built imaging system. Hybridization was visualized using fluorescent-labeled DNA targets with a detection limit of 10 fM. Hybridization times of seconds are required for nanomolar target concentrations, and analysis is performed in minutes.
NASA Technical Reports Server (NTRS)
2006-01-01
As winter turns to spring at the south polar ice cap of Mars, the rising sun reveals dark spots and fans emerging from the cold polar night. Using visual images (left) and temperature data (right) from the Thermal Emission Imaging system on NASA's Mars Odyssey orbiter, scientists have built a new model for the origin of the dark markings. Scientists propose the markings come from dark sand and dust strewn by high-speed jets of carbon-dioxide gas. These erupt from under a layer of carbon-dioxide ice that forms each Martian winter.Multispectral Imaging for Determination of Astaxanthin Concentration in Salmonids
Dissing, Bjørn S.; Nielsen, Michael E.; Ersbøll, Bjarne K.; Frosch, Stina
2011-01-01
Multispectral imaging has been evaluated for characterization of the concentration of a specific cartenoid pigment; astaxanthin. 59 fillets of rainbow trout, Oncorhynchus mykiss, were filleted and imaged using a rapid multispectral imaging device for quantitative analysis. The multispectral imaging device captures reflection properties in 19 distinct wavelength bands, prior to determination of the true concentration of astaxanthin. The samples ranged from 0.20 to 4.34 g per g fish. A PLSR model was calibrated to predict astaxanthin concentration from novel images, and showed good results with a RMSEP of 0.27. For comparison a similar model were built for normal color images, which yielded a RMSEP of 0.45. The acquisition speed of the multispectral imaging system and the accuracy of the PLSR model obtained suggest this method as a promising technique for rapid in-line estimation of astaxanthin concentration in rainbow trout fillets. PMID:21573000
New Mars Camera's First Image of Mars from Mapping Orbit (Full Frame)
NASA Technical Reports Server (NTRS)
2006-01-01
The high resolution camera on NASA's Mars Reconnaissance Orbiter captured its first image of Mars in the mapping orbit, demonstrating the full resolution capability, on Sept. 29, 2006. The High Resolution Imaging Science Experiment (HiRISE) acquired this first image at 8:16 AM (Pacific Time). With the spacecraft at an altitude of 280 kilometers (174 miles), the image scale is 25 centimeters per pixel (10 inches per pixel). If a person were located on this part of Mars, he or she would just barely be visible in this image. The image covers a small portion of the floor of Ius Chasma, one branch of the giant Valles Marineris system of canyons. The image illustrates a variety of processes that have shaped the Martian surface. There are bedrock exposures of layered materials, which could be sedimentary rocks deposited in water or from the air. Some of the bedrock has been faulted and folded, perhaps the result of large-scale forces in the crust or from a giant landslide. The image resolves rocks as small as small as 90 centimeters (3 feet) in diameter. It includes many dunes or ridges of windblown sand. This image (TRA_000823_1720) was taken by the High Resolution Imaging Science Experiment camera onboard the Mars Reconnaissance Orbiter spacecraft on Sept. 29, 2006. Shown here is the full image, centered at minus 7.8 degrees latitude, 279.5 degrees east longitude. The image is oriented such that north is to the top. The range to the target site was 297 kilometers (185.6 miles). At this distance the image scale is 25 centimeters (10 inches) per pixel (with one-by-one binning) so objects about 75 centimeters (30 inches) across are resolved. The image was taken at a local Mars time of 3:30 PM and the scene is illuminated from the west with a solar incidence angle of 59.7 degrees, thus the sun was about 30.3 degrees above the horizon. The season on Mars is northern winter, southern summer. [Photojournal note: Due to the large sizes of the high-resolution TIFF and JPEG files, some systems may experience extremely slow downlink time while viewing or downloading these images; some systems may be incapable of handling the download entirely.] NASA's Jet Propulsion Laboratory, a division of the California Institute of Technology in Pasadena, manages the Mars Reconnaissance Orbiter for NASA's Science Mission Directorate, Washington. Lockheed Martin Space Systems, Denver, is the prime contractor for the project and built the spacecraft. The HiRISE camera was built by Ball Aerospace & Technologies Corporation, Boulder, Colo., and is operated by the University of Arizona, Tucson.The 64 Mpixel wide field imager for the Wendelstein 2m telescope: design and calibration
NASA Astrophysics Data System (ADS)
Kosyra, Ralf; Gössl, Claus; Hopp, Ulrich; Lang-Bardl, Florian; Riffeser, Arno; Bender, Ralf; Seitz, Stella
2014-11-01
The Wendelstein Observatory of Ludwig Maximilians University of Munich has recently been upgraded with a modern 2m robotic telescope. One Nasmyth port of the telescope has been equipped with a wide-field corrector which preserves the excellent image quality (<0.8 " median seeing) of the site (Hopp et al. 2008) over a field of view of 0.7 degrees diameter. The available field is imaged by an optical imager (WWFI, the Wendelstein Wide Field Imager) built around a customized 2×2 mosaic of 4 k×4 k 15 μm e2v CCDs from Spectral Instruments. This paper provides an overview of the design and the WWFI's performance. We summarize the system mechanics (including a structural analysis), the electronics (and its electromagnetic interference (EMI) protection) and the control software. We discuss in detail detector system parameters, i.e. gain and readout noise, quantum efficiency as well as charge transfer efficiency (CTE) and persistent charges. First on sky tests yield overall good predictability of system throughput based on lab measurements.
Harrigan, Robert L; Yvernault, Benjamin C; Boyd, Brian D; Damon, Stephen M; Gibney, Kyla David; Conrad, Benjamin N; Phillips, Nicholas S; Rogers, Baxter P; Gao, Yurui; Landman, Bennett A
2016-01-01
The Vanderbilt University Institute for Imaging Science (VUIIS) Center for Computational Imaging (CCI) has developed a database built on XNAT housing over a quarter of a million scans. The database provides framework for (1) rapid prototyping, (2) large scale batch processing of images and (3) scalable project management. The system uses the web-based interfaces of XNAT and REDCap to allow for graphical interaction. A python middleware layer, the Distributed Automation for XNAT (DAX) package, distributes computation across the Vanderbilt Advanced Computing Center for Research and Education high performance computing center. All software are made available in open source for use in combining portable batch scripting (PBS) grids and XNAT servers. Copyright © 2015 Elsevier Inc. All rights reserved.
NASA Astrophysics Data System (ADS)
Peng, Xiao; Yang, Shaozhuang; Yu, Bin; Wang, Qi; Lin, Danying; Gao, Jian; Zhang, Peiqi; Ma, Yiqun; Qu, Junle; Niu, Hanben
2016-03-01
Optical Coherence Tomography (OCT) has been widely applied into microstructure imaging of tissues or blood vessels with a series of advantages, including non-destructiveness, real-time imaging, high resolution and high sensitivity. In this study, a Spectral Domain OCT (SD-OCT) system with higher sensitivity and signal-to-noise ratio (SNR) was built up, which was used to observe the blood vessel distribution and blood flow in the dorsal skin window chamber of the nude mouse tumor model. In order to obtain comparable data, the distribution images of blood vessels were collected from the same mouse before and after tumor injection. In conclusion, in vivo blood vessel distribution images of the tumor mouse model have been continuously obtained during around two weeks.
Qiu, Bensheng; El-Sharkawy, Abdel-Monem; Paliwal, Vaishali; Karmarkar, Parag; Gao, Fabao; Atalar, Ergin; Yang, Xiaoming
2005-07-01
Previous studies have confirmed the possibility of using an intravascular MR imaging guidewire (MRIG) as a heating source to enhance vascular gene transfection/expression. This motivated us to develop a new intravascular system that can perform MR imaging, radiofrequncy (RF) heating, and MR temperature monitoring simultaneously in an MR scanner. To validate this concept, a series of mathematical simulations of RF power loss along a 0.032-inch MRIG and RF energy spatial distribution were performed to determine the optimum RF heating frequency. Then, an RF generator/amplifier and a filter box were built. The possibility for simultaneous RF heating and MR thermal mapping of the system was confirmed in vitro using a phantom, and the obtained thermal mapping profile was compared with the simulated RF power distribution. Subsequently, the feasibility of simultaneous RF heating and temperature monitoring was successfully validated in vivo in the aorta of living rabbits. This MR imaging/RF heating system offers a potential tool for intravascular MR-mediated, RF-enhanced vascular gene therapy.
Quantitative phase imaging using grating-based quadrature phase interferometer
NASA Astrophysics Data System (ADS)
Wu, Jigang; Yaqoob, Zahid; Heng, Xin; Cui, Xiquan; Yang, Changhuei
2007-02-01
In this paper, we report the use of holographic gratings, which act as the free-space equivalent of the 3x3 fiber-optic coupler, to perform full field phase imaging. By recording two harmonically-related gratings in the same holographic plate, we are able to obtain nontrivial phase shift between different output ports of the gratings-based Mach-Zehnder interferometer. The phase difference can be adjusted by changing the relative phase of the recording beams when recording the hologram. We have built a Mach-Zehnder interferometer using harmonically-related holographic gratings with 600 and 1200 lines/mm spacing. Two CCD cameras at the output ports of the gratings-based Mach-Zehnder interferometer are used to record the full-field quadrature interferograms, which are subsequently processed to reconstruct the phase image. The imaging system has ~12X magnification with ~420μmx315μm field-of-view. To demonstrate the capability of our system, we have successfully performed phase imaging of a pure phase object and a paramecium caudatum.
Reflective all-sky thermal infrared cloud imager.
Redman, Brian J; Shaw, Joseph A; Nugent, Paul W; Clark, R Trevor; Piazzolla, Sabino
2018-04-30
A reflective all-sky imaging system has been built using a long-wave infrared microbolometer camera and a reflective metal sphere. This compact system was developed for measuring spatial and temporal patterns of clouds and their optical depth in support of applications including Earth-space optical communications. The camera is mounted to the side of the reflective sphere to leave the zenith sky unobstructed. The resulting geometric distortion is removed through an angular map derived from a combination of checkerboard-target imaging, geometric ray tracing, and sun-location-based alignment. A tape of high-emissivity material on the side of the reflector acts as a reference that is used to estimate and remove thermal emission from the metal sphere. Once a bias that is under continuing study was removed, sky radiance measurements from the all-sky imager in the 8-14 μm wavelength range agreed to within 0.91 W/(m 2 sr) of measurements from a previously calibrated, lens-based infrared cloud imager over its 110° field of view.
Vision, healing brush, and fiber bundles
NASA Astrophysics Data System (ADS)
Georgiev, Todor
2005-03-01
The Healing Brush is a tool introduced for the first time in Adobe Photoshop (2002) that removes defects in images by seamless cloning (gradient domain fusion). The Healing Brush algorithms are built on a new mathematical approach that uses Fibre Bundles and Connections to model the representation of images in the visual system. Our mathematical results are derived from first principles of human vision, related to adaptation transforms of von Kries type and Retinex theory. In this paper we present the new result of Healing in arbitrary color space. In addition to supporting image repair and seamless cloning, our approach also produces the exact solution to the problem of high dynamic range compression of17 and can be applied to other image processing algorithms.
Adaptable data management for systems biology investigations.
Boyle, John; Rovira, Hector; Cavnor, Chris; Burdick, David; Killcoyne, Sarah; Shmulevich, Ilya
2009-03-06
Within research each experiment is different, the focus changes and the data is generated from a continually evolving barrage of technologies. There is a continual introduction of new techniques whose usage ranges from in-house protocols through to high-throughput instrumentation. To support these requirements data management systems are needed that can be rapidly built and readily adapted for new usage. The adaptable data management system discussed is designed to support the seamless mining and analysis of biological experiment data that is commonly used in systems biology (e.g. ChIP-chip, gene expression, proteomics, imaging, flow cytometry). We use different content graphs to represent different views upon the data. These views are designed for different roles: equipment specific views are used to gather instrumentation information; data processing oriented views are provided to enable the rapid development of analysis applications; and research project specific views are used to organize information for individual research experiments. This management system allows for both the rapid introduction of new types of information and the evolution of the knowledge it represents. Data management is an important aspect of any research enterprise. It is the foundation on which most applications are built, and must be easily extended to serve new functionality for new scientific areas. We have found that adopting a three-tier architecture for data management, built around distributed standardized content repositories, allows us to rapidly develop new applications to support a diverse user community.
Time-resolved spectroscopic imaging reveals the fundamentals of cellular NADH fluorescence.
Li, Dong; Zheng, Wei; Qu, Jianan Y
2008-10-15
A time-resolved spectroscopic imaging system is built to study the fluorescence characteristics of nicotinamide adenine dinucleotide (NADH), an important metabolic coenzyme and endogenous fluorophore in cells. The system provides a unique approach to measure fluorescence signals in different cellular organelles and cytoplasm. The ratios of free over protein-bound NADH signals in cytosol and nucleus are slightly higher than those in mitochondria. The mitochondrial fluorescence contributes about 70% of overall cellular fluorescence and is not a completely dominant signal. Furthermore, NADH signals in mitochondria, cytosol, and the nucleus respond to the changes of cellular activity differently, suggesting that cytosolic and nuclear fluorescence may complicate the well-known relationship between mitochondrial fluorescence and cellular metabolism.
The orbiter PLB and Earth limb during STS-121
2006-07-15
S121-E-07909 (15 July 2006) --- Backdropped by the blackness of space and Earth's horizon, Space Shuttle Discovery's aft cargo bay, its vertical stabilizer and orbital maneuvering system (OMS) pods are seen in this image photographed by an STS-121 crewmember onboard the shuttle. The Italian-built Leonardo Multi-Purpose Logistics Module (MPLM) is visible in the cargo bay.
The orbiter PLB and Earth limb during STS-121
2006-07-15
S121-E-07904 (15 July 2006) --- Backdropped by the blackness of space and Earth's horizon, Space Shuttle Discovery's aft cargo bay, its vertical stabilizer and orbital maneuvering system (OMS) pods are seen in this image photographed by an STS-121 crewmember onboard the shuttle. The Italian-built Leonardo Multi-Purpose Logistics Module (MPLM) is visible in the cargo bay.
An all-reflective wide-angle flat-field telescope for space
NASA Technical Reports Server (NTRS)
Hallam, K. L.; Howell, B. J.; Wilson, M. E.
1984-01-01
An all-reflective wide-angle flat-field telescope (WAFFT) designed and built at Goddard Space Flight Center demonstrates the markedly improved wide-angle imaging capability which can be achieved with a design based on a recently announced class of unobscured 3-mirror optical systems. Astronomy and earth observation missions in space dictate the necessity or preference for wide-angle all-reflective systems which can provide UV through IR wavelength coverage and tolerate the space environment. An initial prototype unit has been designed to meet imaging requirements suitable for monitoring the ultraviolet sky from space. The unobscured f/4, 36 mm efl system achieves a full 20 x 30 deg field of view with resolution over a flat focal surface that is well matched for use with advanced ultraviolet image array detectors. Aspects of the design and fabrication approach, which have especially important bearing on the system solution, are reviewed; and test results are compared with the analytic performance predictions. Other possible applications of the WAFFT class of imaging system are briefly discussed. The exceptional wide-angle, high quality resolution, and very wide spectral coverage of the WAFFT-type optical system could make it a very important tool for future space research.
NASA Astrophysics Data System (ADS)
Zhang, Guozhi; Petrov, Dimitar; Marshall, Nicholas; Bosmans, Hilde
2017-03-01
Digital breast tomosynthesis (DBT) is a relatively new diagnostic imaging modality for women. Currently, various models of DBT systems are available on the market and the number of installations is rapidly increasing. EUREF, the European Reference Organization for Quality Assured Breast Screening and Diagnostic Services, has proposed a preliminary Guideline - protocol for the quality control of the physical and technical aspects of digital breast tomosynthesis systems, with an ultimate aim of providing limiting values guaranteeing proper performance for different applications of DBT. In this work, we introduce an adaptive toolkit developed in accordance with this guideline to facilitate the process of image quality evaluation in DBT performance test. This toolkit implements robust algorithms to quantify various technical parameters of DBT images and provides a convenient user interface in practice. Each test is built into a separate module with configurations set corresponding to the European guideline, which can be easily adapted to different settings and extended with additional tests. This toolkit largely improves the efficiency for image quality evaluation of DBT. It is also going to evolve with the development of protocols in quality control of DBT systems.
Feasibility study of a TIMEPIX detector for mammography applications
NASA Astrophysics Data System (ADS)
Ávila, Carlos A.; Mendoza, Luis M.; Roque, Gerardo A.; Loaiza, Leonardo; Racedo, Jorge; Rueda, Roberto
2017-11-01
We present a comparison study of two X-ray systems for mammography imaging. One is a SELENIA clinical system and the second is a TIMEPIX based system. The aim of the study is to determine the capability of a TIMEPIX detector for mammography applications. We first compare signal to noise ratio (SNR) of X-ray images of Al2O3 spheres with diameters of 0.16mm, 0.24mm and 0.32mm, of a commercial mammography accreditation phantom CIRS015, obtained with each system. Then, we make a similar comparison for a second phantom built with Hydroxyapatite crystals with different morphology and sizes ranging between 0.15mm and 0.83mm, which are embedded within the same block of PMMA of the CIRS015 phantom. Our study allows us to determine the minimum size of Al2O3 spheres on the order of 240μm, with 33% lower SNR for the TIMEPIX system as compared to the SELENIA system. When comparing the images of Hydroxyapatite crystals from both systems, the minimum size observed is about 300μm, with 23% lower SNR for TIMEPIX.
Gamma ray imager on the DIII-D tokamak
DOE Office of Scientific and Technical Information (OSTI.GOV)
Pace, D. C., E-mail: pacedc@fusion.gat.com; Taussig, D.; Eidietis, N. W.
2016-04-15
A gamma ray camera is built for the DIII-D tokamak [J. Luxon, Nucl. Fusion 42, 614 (2002)] that provides spatial localization and energy resolution of gamma flux by combining a lead pinhole camera with custom-built detectors and optimized viewing geometry. This diagnostic system is installed on the outer midplane of the tokamak such that its 123 collimated sightlines extend across the tokamak radius while also covering most of the vertical extent of the plasma volume. A set of 30 bismuth germanate detectors can be secured in any of the available sightlines, allowing for customizable coverage in experiments with runaway electronsmore » in the energy range of 1–60 MeV. Commissioning of the gamma ray imager includes the quantification of electromagnetic noise sources in the tokamak machine hall and a measurement of the energy spectrum of background gamma radiation. First measurements of gamma rays coming from the plasma provide a suitable testbed for implementing pulse height analysis that provides the energy of detected gamma photons.« less
Gamma ray imager on the DIII-D tokamak
Pace, D. C.; Cooper, C. M.; Taussig, D.; ...
2016-04-13
A gamma ray camera is built for the DIII-D tokamak [J. Luxon, Nucl. Fusion 42, 614 (2002)] that provides spatial localization and energy resolution of gamma flux by combining a lead pinhole camera with custom-built detectors and optimized viewing geometry. This diagnostic system is installed on the outer midplane of the tokamak such that its 123 collimated sightlines extend across the tokamak radius while also covering most of the vertical extent of the plasma volume. A set of 30 bismuth germanate detectors can be secured in any of the available sightlines, allowing for customizable coverage in experiments with runaway electronsmore » in the energy range of 1- 60 MeV. Commissioning of the gamma ray imager includes the quantification of electromagnetic noise sources in the tokamak machine hall and a measurement of the energy spectrum of background gamma radiation. In conclusion, first measurements of gamma rays coming from the plasma provide a suitable testbed for implementing pulse height analysis that provides the energy of detected gamma photons.« less
Mobile phone based mini-spectrometer for rapid screening of skin cancer
NASA Astrophysics Data System (ADS)
Das, Anshuman; Swedish, Tristan; Wahi, Akshat; Moufarrej, Mira; Noland, Marie; Gurry, Thomas; Aranda-Michel, Edgar; Aksel, Deniz; Wagh, Sneha; Sadashivaiah, Vijay; Zhang, Xu; Raskar, Ramesh
2015-06-01
We demonstrate a highly sensitive mobile phone based spectrometer that has potential to detect cancerous skin lesions in a rapid, non-invasive manner. Earlier reports of low cost spectrometers utilize the camera of the mobile phone to image the field after moving through a diffraction grating. These approaches are inherently limited by the closed nature of mobile phone image sensors and built in optical elements. The system presented uses a novel integrated grating and sensor that is compact, accurate and calibrated. Resolutions of about 10 nm can be achieved. Additionally, UV and visible LED excitation sources are built into the device. Data collection and analysis is simplified using the wireless interfaces and logical control on the smart phone. Furthermore, by utilizing an external sensor, the mobile phone camera can be used in conjunction with spectral measurements. We are exploring ways to use this device to measure endogenous fluorescence of skin in order to distinguish cancerous from non-cancerous lesions with a mobile phone based dermatoscope.
NASA Technical Reports Server (NTRS)
Gossmann, H.; Haberaecker, P. (Principal Investigator)
1980-01-01
The southwestern part of Central Europe between Basal and Frankfurt was used in a study to determine the accuracy with which a regionally bounded HCMM scene could be rectified with respect to a preassigned coordinate system. The scale to which excerpts from HCMM data can be sensibly enlarged and the question of how large natural structures must be in order to be identified in a satellite thermal image with the given resolution were also examined. Relief and forest and population distribution maps and a land use map derived from LANDSAT data were digitalized and adapted to a common reference system and then combined in a single multichannel data system. The control points for geometrical rectification were determined using the coordinates of the reference system. The multichannel scene was evaluated in several different manners such as the correlation of surface temperature and relief, surface temperature and land use, or surface temperature and built up areas.
Sun, Zhen-Jun; Ye, Bo; Sun, Yi; Zhang, Hong-Hai; Liu, Sheng
2014-07-01
This article describes a novel magnetically maneuverable capsule endoscope system with direction reference for image navigation. This direction reference was employed by utilizing a specific magnet configuration between a pair of external permanent magnets and a magnetic shell coated on the external capsule endoscope surface. A pair of customized Cartesian robots, each with only 4 degrees of freedom, was built to hold the external permanent magnets as their end-effectors. These robots, together with their external permanent magnets, were placed on two opposite sides of a "patient bed." Because of the optimized configuration based on magnetic analysis between the external permanent magnets and the magnetic shell, a simplified control strategy was proposed, and only two parameters, yaw step angle and moving step, were necessary for the employed robotic system. Step-by-step experiments demonstrated that the proposed system is capable of magnetically maneuvering the capsule endoscope while providing direction reference for image navigation. © IMechE 2014.
Development of Time-Series Human Settlement Mapping System Using Historical Landsat Archive
NASA Astrophysics Data System (ADS)
Miyazaki, H.; Nagai, M.; Shibasaki, R.
2016-06-01
Methodology of automated human settlement mapping is highly needed for utilization of historical satellite data archives for urgent issues of urban growth in global scale, such as disaster risk management, public health, food security, and urban management. As development of global data with spatial resolution of 10-100 m was achieved by some initiatives using ASTER, Landsat, and TerraSAR-X, next goal has targeted to development of time-series data which can contribute to studies urban development with background context of socioeconomy, disaster risk management, public health, transport and other development issues. We developed an automated algorithm to detect human settlement by classification of built-up and non-built-up in time-series Landsat images. A machine learning algorithm, Local and Global Consistency (LLGC), was applied with improvements for remote sensing data. The algorithm enables to use MCD12Q1, a MODIS-based global land cover map with 500-m resolution, as training data so that any manual process is not required for preparation of training data. In addition, we designed the method to composite multiple results of LLGC into a single output to reduce uncertainty. The LLGC results has a confidence value ranging 0.0 to 1.0 representing probability of built-up and non-built-up. The median value of the confidence for a certain period around a target time was expected to be a robust output of confidence to identify built-up or non-built-up areas against uncertainties in satellite data quality, such as cloud and haze contamination. Four scenes of Landsat data for each target years, 1990, 2000, 2005, and 2010, were chosen among the Landsat archive data with cloud contamination less than 20%.We developed a system with the algorithms on the Data Integration and Analysis System (DIAS) in the University of Tokyo and processed 5200 scenes of Landsat data for cities with more than one million people worldwide.
Video flow active control by means of adaptive shifted foveal geometries
NASA Astrophysics Data System (ADS)
Urdiales, Cristina; Rodriguez, Juan A.; Bandera, Antonio J.; Sandoval, Francisco
2000-10-01
This paper presents a control mechanism for video transmission that relies on transmitting non-uniform resolution images depending on the delay of the communication channel. These images are built in an active way to keep the areas of interest of the image at the highest resolution available. In order to shift the area of high resolution over the image and to achieve a data structure easy to process by using conventional algorithms, a shifted fovea multi resolution geometry of adaptive size is used. Besides, if delays are nevertheless too high, the different areas of resolution of the image can be transmitted at different rates. A functional system has been developed for corridor surveillance with static cameras. Tests with real video images have proven that the method allows an almost constant rate of images per second as long as the channel is not collapsed.
Construction of multi-functional open modulized Matlab simulation toolbox for imaging ladar system
NASA Astrophysics Data System (ADS)
Wu, Long; Zhao, Yuan; Tang, Meng; He, Jiang; Zhang, Yong
2011-06-01
Ladar system simulation is to simulate the ladar models using computer simulation technology in order to predict the performance of the ladar system. This paper presents the developments of laser imaging radar simulation for domestic and overseas studies and the studies of computer simulation on ladar system with different application requests. The LadarSim and FOI-LadarSIM simulation facilities of Utah State University and Swedish Defence Research Agency are introduced in details. This paper presents the low level of simulation scale, un-unified design and applications of domestic researches in imaging ladar system simulation, which are mostly to achieve simple function simulation based on ranging equations for ladar systems. Design of laser imaging radar simulation with open and modularized structure is proposed to design unified modules for ladar system, laser emitter, atmosphere models, target models, signal receiver, parameters setting and system controller. Unified Matlab toolbox and standard control modules have been built with regulated input and output of the functions, and the communication protocols between hardware modules. A simulation based on ICCD gain-modulated imaging ladar system for a space shuttle is made based on the toolbox. The simulation result shows that the models and parameter settings of the Matlab toolbox are able to simulate the actual detection process precisely. The unified control module and pre-defined parameter settings simplify the simulation of imaging ladar detection. Its open structures enable the toolbox to be modified for specialized requests. The modulization gives simulations flexibility.
View of the FGB prior to rendezvous and grapple
1998-12-06
S88-E-5044 (12-06-98) --- With Endeavour's astronauts waiting to mate the Russian-built Zarya control module with the U.S.-built Unity connecting module, an electronic still camera (ESC) was used to record this image of the approaching Zarya. Using the shuttle's 50-ft.-long Canadian-built robot arm, astronaut Nancy J. Currie, working from Endeavour's aft flight deck, plucked Zarya out of orbit at 5:47 p.m. (CST), Dec. 6. The craft had been orbiting Earth for a little over 16 days prior to the grapple and subsequent docking to Unity. This image was recorded at 22:53:55 GMT, Dec. 6.
View of the FGB prior to rendezvous and grapple
1998-12-06
S88-E-5045 (12-06-98) --- With Endeavour's astronauts waiting to mate the Russian-built Zarya control module with the U.S.-built Unity connecting module, an electronic still camera (ESC) was used to record this image of the approaching Zarya. Using the shuttle's 50-ft.-long Canadian-built robot arm, astronaut Nancy J. Currie, working from Endeavour's aft flight deck, plucked Zarya out of orbit at 5:47 p.m. (CST), Dec. 6. The craft had been orbiting Earth for a little over 16 days prior to the grapple and subsequent docking to Unity. This image was recorded at 23:02:06 GMT, Dec. 6.
NASA Astrophysics Data System (ADS)
Zhao, Huangxuan; Wang, Guangsong; Lin, Riqiang; Gong, Xiaojing; Song, Liang; Li, Tan; Wang, Wenjia; Zhang, Kunya; Qian, Xiuqing; Zhang, Haixia; Li, Lin; Liu, Zhicheng; Liu, Chengbo
2018-04-01
For the diagnosis and evaluation of ophthalmic diseases, imaging and quantitative characterization of vasculature in the iris are very important. The recently developed photoacoustic imaging, which is ultrasensitive in imaging endogenous hemoglobin molecules, provides a highly efficient label-free method for imaging blood vasculature in the iris. However, the development of advanced vascular quantification algorithms is still needed to enable accurate characterization of the underlying vasculature. We have developed a vascular information quantification algorithm by adopting a three-dimensional (3-D) Hessian matrix and applied for processing iris vasculature images obtained with a custom-built optical-resolution photoacoustic imaging system (OR-PAM). For the first time, we demonstrate in vivo 3-D vascular structures of a rat iris with a the label-free imaging method and also accurately extract quantitative vascular information, such as vessel diameter, vascular density, and vascular tortuosity. Our results indicate that the developed algorithm is capable of quantifying the vasculature in the 3-D photoacoustic images of the iris in-vivo, thus enhancing the diagnostic capability of the OR-PAM system for vascular-related ophthalmic diseases in vivo.
Handheld ultrasound array imaging device
NASA Astrophysics Data System (ADS)
Hwang, Juin-Jet; Quistgaard, Jens
1999-06-01
A handheld ultrasound imaging device, one that weighs less than five pounds, has been developed for diagnosing trauma in the combat battlefield as well as a variety of commercial mobile diagnostic applications. This handheld device consists of four component ASICs, each is designed using the state of the art microelectronics technologies. These ASICs are integrated with a convex array transducer to allow high quality imaging of soft tissues and blood flow in real time. The device is designed to be battery driven or ac powered with built-in image storage and cineloop playback capability. Design methodologies of a handheld device are fundamentally different to those of a cart-based system. As system architecture, signal and image processing algorithm as well as image control circuit and software in this device is deigned suitably for large-scale integration, the image performance of this device is designed to be adequate to the intent applications. To elongate the battery life, low power design rules and power management circuits are incorporated in the design of each component ASIC. The performance of the prototype device is currently being evaluated for various applications such as a primary image screening tool, fetal imaging in Obstetrics, foreign object detection and wound assessment for emergency care, etc.
Microscopie non-lineaire pour l'imagerie des cordes vocales
NASA Astrophysics Data System (ADS)
Deterre, Romain
The vocal cords are two folds of epithelial tissues located in the larynx and are involved in production of the human voice. Despite their apparent simplicity, their internal structure is complex. Each fold can be divided into several layers with different mechanical properties. The gold standard for studying their structure - histology - has the inconvenience of being very invasive. Non-linear microscopy is an optical imaging technique which allows images to be taken in depth within samples in a non invasive manner. It also offers intrinsic contrasts, allowing the identification of certain fibrous proteins - elastin and collagen - which are responsible for the mechanical properties of epithelious tissues. The main goal of this research project was to assess nonlinear microscopy's performances for vocal fold imaging. The study has been broken down in two separate tasks. The first one was to evaluate the nonlinear modalities contrast against histology. For that purpose, we chose to first take images of thin samples and compare them to the corresponding histological slides. The second task was to make tests to transcribe the results obtained to in vivo imaging. A custom-built nonlinear imaging system was used for these experiments. It was developed to allow acquisition of wide-field images. A C++ based software was developped to control the microscope and allow treatment and visualization of the images. After being built, the system was further tested to check its performances in comparison with the theoretical limit as described in the literature. Thin slices of vocal folds were obtained from the team of Pr Christopher J. Hartnick from Massachusetts Eye and Ear Infirmary, Harvard Medical School. Specialists from his team analysed the histological samples to extract structural data from the vocal folds. A good correlation was measured between histological and nonlinear data. A first step in evaluating the possibility for translating these results towards in vivo imaging was performed during this project. A swine's larynx was obtained, and vocal folds were extracted for imaging purposes. This experiment showed that it is indeed possible to localize various macrostructures of the tissues with nonlinear microscopy.
Li, Dongming; Sun, Changming; Yang, Jinhua; Liu, Huan; Peng, Jiaqi; Zhang, Lijuan
2017-04-06
An adaptive optics (AO) system provides real-time compensation for atmospheric turbulence. However, an AO image is usually of poor contrast because of the nature of the imaging process, meaning that the image contains information coming from both out-of-focus and in-focus planes of the object, which also brings about a loss in quality. In this paper, we present a robust multi-frame adaptive optics image restoration algorithm via maximum likelihood estimation. Our proposed algorithm uses a maximum likelihood method with image regularization as the basic principle, and constructs the joint log likelihood function for multi-frame AO images based on a Poisson distribution model. To begin with, a frame selection method based on image variance is applied to the observed multi-frame AO images to select images with better quality to improve the convergence of a blind deconvolution algorithm. Then, by combining the imaging conditions and the AO system properties, a point spread function estimation model is built. Finally, we develop our iterative solutions for AO image restoration addressing the joint deconvolution issue. We conduct a number of experiments to evaluate the performances of our proposed algorithm. Experimental results show that our algorithm produces accurate AO image restoration results and outperforms the current state-of-the-art blind deconvolution methods.
Li, Dongming; Sun, Changming; Yang, Jinhua; Liu, Huan; Peng, Jiaqi; Zhang, Lijuan
2017-01-01
An adaptive optics (AO) system provides real-time compensation for atmospheric turbulence. However, an AO image is usually of poor contrast because of the nature of the imaging process, meaning that the image contains information coming from both out-of-focus and in-focus planes of the object, which also brings about a loss in quality. In this paper, we present a robust multi-frame adaptive optics image restoration algorithm via maximum likelihood estimation. Our proposed algorithm uses a maximum likelihood method with image regularization as the basic principle, and constructs the joint log likelihood function for multi-frame AO images based on a Poisson distribution model. To begin with, a frame selection method based on image variance is applied to the observed multi-frame AO images to select images with better quality to improve the convergence of a blind deconvolution algorithm. Then, by combining the imaging conditions and the AO system properties, a point spread function estimation model is built. Finally, we develop our iterative solutions for AO image restoration addressing the joint deconvolution issue. We conduct a number of experiments to evaluate the performances of our proposed algorithm. Experimental results show that our algorithm produces accurate AO image restoration results and outperforms the current state-of-the-art blind deconvolution methods. PMID:28383503
NASA Technical Reports Server (NTRS)
2006-01-01
This full-frame image from the High Resolution Imaging Science Experiment camera on NASA's Mars Reconnaissance Orbiter shows faults and pits in Mars' north polar residual cap that have not been previously recognized. The faults and depressions between them are similar to features seen on Earth where the crust is being pulled apart. Such tectonic extension must have occurred very recently because the north polar residual cap is very young, as indicated by the paucity of impact craters on its surface. Alternatively, the faults and pits may be caused by collapse due to removal of material beneath the surface. The pits are aligned along the faults, either because material has drained into the subsurface along the faults or because gas has escaped from the subsurface through them. NASA's Jet Propulsion Laboratory, a division of the California Institute of Technology in Pasadena, manages the Mars Reconnaissance Orbiter for NASA's Science Mission Directorate, Washington. Lockheed Martin Space Systems, Denver, is the prime contractor for the project and built the spacecraft. The High Resolution Imaging Science Experiment is operated by the University of Arizona, Tucson, and the instrument was built by Ball Aerospace and Technology Corp., Boulder, Colo.ATTRIBUTES OF FORM IN THE BUILT ENVIRONMENT THAT INFLUENCE PERCEIVED WALKABILITY.
Oreskovic, Nicolas M; Charles, Pablina Roth Suzanne Lanyi; Shepherd, Dido Tsigaridi Kathrine; Nelson, Kerrie P; Bar, Moshe
2014-01-01
A recent focus of design and building regulations, including form-based codes and the Leadership in Energy and Environmental Design for Neighborhood Development rating system, has been on promoting pedestrian activity. This study assessed perceptions of walkability for residential and commercial streetscapes with different design attributes in order to inform form-based regulations and codes that aim to impact walkability. We scored 424 images on four design attributes purported to influence walkability: variation in building height, variation in building plane, presence of ground-floor windows, and presence of a street focal point. We then presented the images to 45 adults, who were asked to rate the images for walkability. The results showed that perceived walkability varied according to the degree to which a particular design attribute was present, with the presence of ground-floor windows and a street focal point most consistently associated with a space's perceived walkability. Understanding if and which design attributes are most related to walkability could allow planners and developers to focus on the most salient built-environment features influencing physical activity, as well as provide empirical scientific evidence for form-based regulations and zoning codes aimed at impacting walkabilit.
2015-07-23
A rock outcrop dubbed "Missoula," near Marias Pass on Mars, is seen in this image mosaic taken by the Mars Hand Lens Imager on NASA's Curiosity rover. Pale mudstone (bottom of outcrop) meets coarser sandstone (top) in this geological contact zone, which has piqued the interest of Mars scientists. White mineral veins that fill fractures in the lower rock unit abruptly end when they meet the upper rock unit. Such clues help scientists understand the possible timing of geological events. First, the fine sediment that now forms the lower unit would have hardened into rock. It then would have fractured, and groundwater would have deposited calcium sulfate minerals into the fractures. Next, the coarser sediment that forms the upper unit would have been deposited. The area pictured is about 16 inches (40 centimeters) across. The image was taken on the 1,031st Martian day, or sol, of the mission (July 1, 2015). MAHLI was built by Malin Space Science Systems, San Diego. NASA's Jet Propulsion Laboratory, a division of the California Institute of Technology in Pasadena, manages the Mars Science Laboratory Project for the NASA Science Mission Directorate, Washington. JPL designed and built the project's Curiosity rover. http://photojournal.jpl.nasa.gov/catalog/PIA19829
A novel snapshot polarimetric imager
NASA Astrophysics Data System (ADS)
Wong, Gerald; McMaster, Ciaran; Struthers, Robert; Gorman, Alistair; Sinclair, Peter; Lamb, Robert; Harvey, Andrew R.
2012-10-01
Polarimetric imaging (PI) is of increasing importance in determining additional scene information beyond that of conventional images. For very long-range surveillance, image quality is degraded due to turbulence. Furthermore, the high magnification required to create images with sufficient spatial resolution suitable for object recognition and identification require long focal length optical systems. These are incompatible with the size and weight restrictions for aircraft. Techniques which allow detection and recognition of an object at the single pixel level are therefore likely to provide advance warning of approaching threats or long-range object cueing. PI is a technique that has the potential to detect object signatures at the pixel level. Early attempts to develop PI used rotating polarisers (and spectral filters) which recorded sequential polarized images from which the complete Stokes matrix could be derived. This approach has built-in latency between frames and requires accurate registration of consecutive frames to analyze real-time video of moving objects. Alternatively, multiple optical systems and cameras have been demonstrated to remove latency, but this approach increases cost and bulk of the imaging system. In our investigation we present a simplified imaging system that divides an image into two orthogonal polarimetric components which are then simultaneously projected onto a single detector array. Thus polarimetric data is recorded without latency on a single snapshot. We further show that, for pixel-level objects, the data derived from only two orthogonal states (H and V) is sufficient to increase the probability of detection whilst reducing false alarms compared to conventional unpolarised imaging.
Height Measuring System On Video Using Otsu Method
NASA Astrophysics Data System (ADS)
Sandy, C. L. M.; Meiyanti, R.
2017-01-01
A measurement of height is comparing the value of the magnitude of an object with a standard measuring tool. The problems that exist in the measurement are still the use of a simple apparatus in which one of them is by using a meter. This method requires a relatively long time. To overcome these problems, this research aims to create software with image processing that is used for the measurement of height. And subsequent that image is tested, where the object captured by the video camera can be known so that the height of the object can be measured using the learning method of Otsu. The system was built using Delphi 7 of Vision Lab VCL 4.5 component. To increase the quality of work of the system in future research, the developed system can be combined with other methods.
Real-Time Interactive Facilities Associated With A 3-D Medical Workstation
NASA Astrophysics Data System (ADS)
Goldwasser, S. M.; Reynolds, R. A.; Talton, D.; Walsh, E.
1986-06-01
Biomedical workstations of the future will incorporate three-dimensional interactive capabilities which provide real-time response to most common operator requests. Such systems will find application in many areas of medicine including clinical diagnosis, surgical and radiation therapy planning, biomedical research based on functional imaging, and medical education. This paper considers the requirements of these future systems in terms of image quality, performance, and the interactive environment, and examines the relationship of workstation capabilities to specific medical applications. We describe a prototype physician's workstation that we have designed and built to meet many of these requirements (using conventional graphics technology in conjunction with a custom real-time 3-D processor), and give an account of the remaining issues and challenges that future designers of such systems will have to address.
Astronauts Sullivan and Leestma perform in-space simulation of refueling
1984-10-14
S84-43432 (11 Oct. 1984) --- Appearing small in the center background of this image, astronauts Kathryn D. Sullivan, left, and David C. Leestma, both 41-G mission specialists, perform an in-space simulation of refueling another spacecraft in orbit. Their station on the space shuttle Challenger is the orbital refueling system (ORS), positioned on the mission peculiar support structure (MPR ESS). The Large Format Camera (LFC) is left of the two mission specialists. In the left foreground is the antenna for the shuttle imaging radar (SIR-B) system onboard. The Canadian-built remote manipulator system (RMS) is positioned to allow close-up recording capability of the busy scene. A 50mm lens on a 70mm camera was used to photograph this scene. Photo credit: NASA
Submillimeter video imaging with a superconducting bolometer array
NASA Astrophysics Data System (ADS)
Becker, Daniel Thomas
Millimeter wavelength radiation holds promise for detection of security threats at a distance, including suicide bombers and maritime threats in poor weather. The high sensitivity of superconducting Transition Edge Sensor (TES) bolometers makes them ideal for passive imaging of thermal signals at millimeter and submillimeter wavelengths. I have built a 350 GHz video-rate imaging system using an array of feedhorn-coupled TES bolometers. The system operates at standoff distances of 16 m to 28 m with a measured spatial resolution of 1.4 cm (at 17 m). It currently contains one 251-detector sub-array, and can be expanded to contain four sub-arrays for a total of 1004 detectors. The system has been used to take video images that reveal the presence of weapons concealed beneath a shirt in an indoor setting. This dissertation describes the design, implementation and characterization of this system. It presents an overview of the challenges associated with standoff passive imaging and how these problems can be overcome through the use of large-format TES bolometer arrays. I describe the design of the system and cover the results of detector and optical characterization. I explain the procedure used to generate video images using the system, and present a noise analysis of those images. This analysis indicates that the Noise Equivalent Temperature Difference (NETD) of the video images is currently limited by artifacts of the scanning process. More sophisticated image processing algorithms can eliminate these artifacts and reduce the NETD to 100 mK, which is the target value for the most demanding passive imaging scenarios. I finish with an overview of future directions for this system.
CRISPR Primer Designer: Design primers for knockout and chromosome imaging CRISPR-Cas system.
Yan, Meng; Zhou, Shi-Rong; Xue, Hong-Wei
2015-07-01
The clustered regularly interspaced short palindromic repeats (CRISPR)-associated system enables biologists to edit genomes precisely and provides a powerful tool for perturbing endogenous gene regulation, modulation of epigenetic markers, and genome architecture. However, there are concerns about the specificity of the system, especially the usages of knocking out a gene. Previous designing tools either were mostly built-in websites or ran as command-line programs, and none of them ran locally and acquired a user-friendly interface. In addition, with the development of CRISPR-derived systems, such as chromosome imaging, there were still no tools helping users to generate specific end-user spacers. We herein present CRISPR Primer Designer for researchers to design primers for CRISPR applications. The program has a user-friendly interface, can analyze the BLAST results by using multiple parameters, score for each candidate spacer, and generate the primers when using a certain plasmid. In addition, CRISPR Primer Designer runs locally and can be used to search spacer clusters, and exports primers for the CRISPR-Cas system-based chromosome imaging system. © 2014 Institute of Botany, Chinese Academy of Sciences.
Development Of A Flash X-Ray Scanner For Stereoradiography And CT
NASA Astrophysics Data System (ADS)
Endorf, Robert J.; DiBianca, Frank A.; Fritsch, Daniel S.; Liu, Wen-Ching; Burns, Charles B.
1989-05-01
We are developing a flash x-ray scanner for stereoradiography and CT which will be able to produce a stereoradiograph in 30 to 70 ns and a complete CT scan in one microsecond. This type of imaging device will be valuable in studying high speed processes, high acceleration, and traumatic events. We have built a two channel flash x-ray system capable of producing stereo radiographs with stereo angles of from 15 to 165 degrees. The dynamic and static Miff 's for the flash x-ray system were measured and compared with similar MIT's measured for a conventional medical x-ray system. We have written and tested a stereo reconstruction algorithm to determine three dimensional space points from corresponding points in the two stereo images. To demonstrate the ability of the system to image traumatic events, a radiograph was obtained of a bone undergoing a fracture. The effects of accelerations of up to 600 g were examined on radiographs taken of human kidney tissue samples in a rapidly rotating centrifuge. Feasibility studies of CT reconstruction have been performed by making simulated Cr images of various phantoms for larger flash x-ray systems of from 8 to 29 flash x-ray tubes.
10-channel fiber array fabrication technique for parallel optical coherence tomography system
NASA Astrophysics Data System (ADS)
Arauz, Lina J.; Luo, Yuan; Castillo, Jose E.; Kostuk, Raymond K.; Barton, Jennifer
2007-02-01
Optical Coherence Tomography (OCT) shows great promise for low intrusive biomedical imaging applications. A parallel OCT system is a novel technique that replaces mechanical transverse scanning with electronic scanning. This will reduce the time required to acquire image data. In this system an array of small diameter fibers is required to obtain an image in the transverse direction. Each fiber in the array is configured in an interferometer and is used to image one pixel in the transverse direction. In this paper we describe a technique to package 15μm diameter fibers on a siliconsilica substrate to be used in a 2mm endoscopic probe tip. Single mode fibers are etched to reduce the cladding diameter from 125μm to 15μm. Etched fibers are placed into a 4mm by 150μm trench in a silicon-silica substrate and secured with UV glue. Active alignment was used to simplify the lay out of the fibers and minimize unwanted horizontal displacement of the fibers. A 10-channel fiber array was built, tested and later incorporated into a parallel optical coherence system. This paper describes the packaging, testing, and operation of the array in a parallel OCT system.
Focused ultrasound: concept for automated transcutaneous control of hemorrhage in austere settings.
Kucewicz, John C; Bailey, Michael R; Kaczkowski, Peter J; Carter, Stephen J
2009-04-01
High intensity focused ultrasound (HIFU) is being developed for a range of clinical applications. Of particular interest to NASA and the military is the use of HIFU for traumatic injuries because HIFU has the unique ability to transcutaneously stop bleeding. Automation of this technology would make possible its use in remote, austere settings by personnel not specialized in medical ultrasound. Here a system to automatically detect and target bleeding is tested and reported. The system uses Doppler ultrasound images from a clinical ultrasound scanner for bleeding detection and hardware for HIFU therapy. The system was tested using a moving string to simulate blood flow and targeting was visualized by Schlieren imaging to show the focusing of the HIFU acoustic waves. When instructed by the operator, a Doppler ultrasound image is acquired and processed to detect and localize the moving string, and the focus of the HIFU array is electronically adjusted to target the string. Precise and accurate targeting was verified in the Schlieren images. An automated system to detect and target simulated bleeding has been built and tested. The system could be combined with existing algorithms to detect, target, and treat clinical bleeding.
NASA Astrophysics Data System (ADS)
Venugopal, Vivek; Park, Minho; Ashitate, Yoshitomo; Neacsu, Florin; Kettenring, Frank; Frangioni, John V.; Gangadharan, Sidhu P.; Gioux, Sylvain
2013-12-01
We report the design, characterization, and validation of an optimized simultaneous color and near-infrared (NIR) fluorescence rigid endoscopic imaging system for minimally invasive surgery. This system is optimized for illumination and collection of NIR wavelengths allowing the simultaneous acquisition of both color and NIR fluorescence at frame rates higher than 6.8 fps with high sensitivity. The system employs a custom 10-mm diameter rigid endoscope optimized for NIR transmission. A dual-channel light source compatible with the constraints of an endoscope was built and includes a plasma source for white light illumination and NIR laser diodes for fluorescence excitation. A prism-based 2-CCD camera was customized for simultaneous color and NIR detection with a highly efficient filtration scheme for fluorescence imaging of both 700- and 800-nm emission dyes. The performance characterization studies indicate that the endoscope can efficiently detect fluorescence signal from both indocyanine green and methylene blue in dimethyl sulfoxide at the concentrations of 100 to 185 nM depending on the background optical properties. Finally, we performed the validation of this imaging system in vivo during a minimally invasive procedure for thoracic sentinel lymph node mapping in a porcine model.
NASA Astrophysics Data System (ADS)
Takashima, Ichiro; Kajiwara, Riichi; Murano, Kiyo; Iijima, Toshio; Morinaka, Yasuhiro; Komobuchi, Hiroyoshi
2001-04-01
We have designed and built a high-speed CCD imaging system for monitoring neural activity in an exposed animal cortex stained with a voltage-sensitive dye. Two types of custom-made CCD sensors were developed for this system. The type I chip has a resolution of 2664 (H) X 1200 (V) pixels and a wide imaging area of 28.1 X 13.8 mm, while the type II chip has 1776 X 1626 pixels and an active imaging area of 20.4 X 18.7 mm. The CCD arrays were constructed with multiple output amplifiers in order to accelerate the readout rate. The two chips were divided into either 24 (I) or 16 (II) distinct areas that were driven in parallel. The parallel CCD outputs were digitized by 12-bit A/D converters and then stored in the frame memory. The frame memory was constructed with synchronous DRAM modules, which provided a capacity of 128 MB per channel. On-chip and on-memory binning methods were incorporated into the system, e.g., this enabled us to capture 444 X 200 pixel-images for periods of 36 seconds at a rate of 500 frames/second. This system was successfully used to visualize neural activity in the cortices of rats, guinea pigs, and monkeys.
Initial experience with a nuclear medicine viewing workstation
NASA Astrophysics Data System (ADS)
Witt, Robert M.; Burt, Robert W.
1992-07-01
Graphical User Interfaced (GUI) workstations are now available from commercial vendors. We recently installed a GUI workstation in our nuclear medicine reading room for exclusive use of staff and resident physicians. The system is built upon a Macintosh platform and has been available as a DELTAmanager from MedImage and more recently as an ICON V from Siemens Medical Systems. The workstation provides only display functions and connects to our existing nuclear medicine imaging system via ethernet. The system has some processing capabilities to create oblique, sagittal and coronal views from transverse tomographic views. Hard copy output is via a screen save device and a thermal color printer. The DELTAmanager replaced a MicroDELTA workstation which had both process and view functions. The mouse activated GUI has made remarkable changes to physicians'' use of the nuclear medicine viewing system. Training time to view and review studies has been reduced from hours to about 30-minutes. Generation of oblique views and display of brain and heart tomographic studies has been reduced from about 30-minutes of technician''s time to about 5-minutes of physician''s time. Overall operator functionality has been increased so that resident physicians with little prior computer experience can access all images on the image server and display pertinent patient images when consulting with other staff.
Multidimensional custom-made non-linear microscope: from ex-vivo to in-vivo imaging
NASA Astrophysics Data System (ADS)
Cicchi, R.; Sacconi, L.; Jasaitis, A.; O'Connor, R. P.; Massi, D.; Sestini, S.; de Giorgi, V.; Lotti, T.; Pavone, F. S.
2008-09-01
We have built a custom-made multidimensional non-linear microscope equipped with a combination of several non-linear laser imaging techniques involving fluorescence lifetime, multispectral two-photon and second-harmonic generation imaging. The optical system was mounted on a vertical honeycomb breadboard in an upright configuration, using two galvo-mirrors relayed by two spherical mirrors as scanners. A double detection system working in non-descanning mode has allowed both photon counting and a proportional regime. This experimental setup offering high spatial (micrometric) and temporal (sub-nanosecond) resolution has been used to image both ex-vivo and in-vivo biological samples, including cells, tissues, and living animals. Multidimensional imaging was used to spectroscopically characterize human skin lesions, as malignant melanoma and naevi. Moreover, two-color detection of two photon excited fluorescence was applied to in-vivo imaging of living mice intact neocortex, as well as to induce neuronal microlesions by femtosecond laser burning. The presented applications demonstrate the capability of the instrument to be used in a wide range of biological and biomedical studies.
Looking Up at Mars Rover Curiosity in Buckskin Selfie
2015-08-19
This low-angle self-portrait of NASA's Curiosity Mars rover shows the vehicle at the site from which it reached down to drill into a rock target called "Buckskin" on lower Mount Sharp. The selfie combines several component images taken by Curiosity's Mars Hand Lens Imager (MAHLI) on Aug. 5, 2015, during the 1,065th Martian day, or sol, of the rover's work on Mars. For scale, the rover's wheels are 20 inches (50 centimeters) in diameter and about 16 inches (40 centimeters) wide. This view is a portion of a larger panorama available at PIA19807. A close look reveals a small rock stuck onto Curiosity's left middle wheel (on the right in this head-on view). The rock had been seen previously during periodic monitoring of wheel condition about three weeks earlier, in the MAHLI raw image at http://mars.nasa.gov/msl/multimedia/raw/?rawid=1046MH0002640000400290E01_DXXX&s=1046. MAHLI is mounted at the end of the rover's robotic arm. For this self-portrait, the rover team positioned the camera lower in relation to the rover body than for any previous full self-portrait of Curiosity. This yielded a view that includes the rover's "belly," as in a partial self-portrait (/catalog/PIA16137) taken about five weeks after Curiosity's August 2012 landing inside Mars' Gale Crater. The selfie at Buckskin does not include the rover's robotic arm beyond a portion of the upper arm held nearly vertical from the shoulder joint. With the wrist motions and turret rotations used in pointing the camera for the component images, the arm was positioned out of the shot in the frames or portions of frames used in this mosaic. This process was used previously in acquiring and assembling Curiosity self-portraits taken at sample-collection sites "Rocknest" (PIA16468), "John Klein" (PIA16937), "Windjana" (PIA18390) and "Mojave" (PIA19142). MAHLI was built by Malin Space Science Systems, San Diego. NASA's Jet Propulsion Laboratory, a division of the California Institute of Technology in Pasadena, manages the Mars Science Laboratory Project for the NASA Science Mission Directorate, Washington. JPL designed and built the project's Curiosity rover. MAHLI was built by Malin Space Science Systems, San Diego. NASA's Jet Propulsion Laboratory, a division of the California Institute of Technology in Pasadena, manages the Mars Science Laboratory Project for the NASA Science Mission Directorate, Washington. JPL designed and built the project's Curiosity rover. http://photojournal.jpl.nasa.gov/catalog/PIA19808
NASA Astrophysics Data System (ADS)
Lucey, Paul G.; Hinrichs, John L.; Akagi, Jason
2012-06-01
A prototype long wave infrared Fourier transform spectral imaging system using a wedged Fabry-Perot interferometer and a microbolometer array was designed and built. The instrument can be used at both short (cm) and long standoff ranges (infinity focus). Signal to noise ratios are in the several hundred range for 30 C targets. The sensor is compact, fitting in a volume about 12 x12 x 4 inches.
Orbiter Boom Sensor System and TPS tiles on orbiter Discovery as seen during EVA 3
2005-08-03
S114-E-6310 (3 August 2005) --- The Red Sea forms the backdrop for this view featuring a portion of thermal protection tiles on the Space Shuttle Discoverys underside and the Canadian-built remote manipulator system (RMS) robotic arm while docked to the international space station during the STS-114 mission. The image was photographed by astronaut Stephen K. Robinson (out of frame), mission specialist, during todays extravehicular activities (EVA).
Vertical Stabilizer and OMS pods from the aft FD window during STS-123 mission
2008-03-11
S123-E-005073 (11 Mar. 2008) --- This view out the aft windows on Endeavour's flight deck was one of a series of images recorded by the STS-123 crewmembers during their first full day in space. The end of the Canadian-built remote manipulator system's robot arm (right edge) along with the shuttle's vertical stabilizer and its two orbital maneuvering system (OMS) pods are visible. A heavily cloud-covered area of Earth fills the top half of the frame.
The Gemini Planet Imager Calibration Wavefront Sensor Instrument
NASA Technical Reports Server (NTRS)
Wallace, J. Kent; Burruss, Rick S.; Bartos, Randall D.; Trinh, Thang Q.; Pueyo, Laurent A.; Fregoso, Santos F.; Angione, John R.; Shelton, J. Chris
2010-01-01
The Gemini Planet Imager is an extreme adaptive optics system that will employ an apodized-pupil coronagraph to make direct detections of faint companions of nearby stars to a contrast level of the 10(exp -7) within a few lambda/D of the parent star. Such high contrasts from the ground require exquisite wavefront sensing and control both for the AO system as well as for the coronagraph. Un-sensed non-common path phase and amplitude errors after the wavefront sensor dichroic but before the coronagraph would lead to speckles which would ultimately limit the contrast. The calibration wavefront system for GPI will measure the complex wavefront at the system pupil before the apodizer and provide slow phase corrections to the AO system to mitigate errors that would cause a loss in contrast. The calibration wavefront sensor instrument for GPI has been built. We will describe the instrument and its performance.
Magnifying lens for 800 MeV proton radiography.
Merrill, F E; Campos, E; Espinoza, C; Hogan, G; Hollander, B; Lopez, J; Mariam, F G; Morley, D; Morris, C L; Murray, M; Saunders, A; Schwartz, C; Thompson, T N
2011-10-01
This article describes the design and performance of a magnifying magnetic-lens system designed, built, and commissioned at the Los Alamos National Laboratory (LANL) for 800 MeV flash proton radiography. The technique of flash proton radiography has been developed at LANL to study material properties under dynamic loading conditions through the analysis of time sequences of proton radiographs. The requirements of this growing experimental program have resulted in the need for improvements in spatial radiographic resolution. To meet these needs, a new magnetic lens system, consisting of four permanent magnet quadrupoles, has been developed. This new lens system was designed to reduce the second order chromatic aberrations, the dominant source of image blur in 800 MeV proton radiography, as well as magnifying the image to reduce the blur contribution from the detector and camera systems. The recently commissioned lens system performed as designed, providing nearly a factor of three improvement in radiographic resolution.
Magnifying lens for 800 MeV proton radiography
NASA Astrophysics Data System (ADS)
Merrill, F. E.; Campos, E.; Espinoza, C.; Hogan, G.; Hollander, B.; Lopez, J.; Mariam, F. G.; Morley, D.; Morris, C. L.; Murray, M.; Saunders, A.; Schwartz, C.; Thompson, T. N.
2011-10-01
This article describes the design and performance of a magnifying magnetic-lens system designed, built, and commissioned at the Los Alamos National Laboratory (LANL) for 800 MeV flash proton radiography. The technique of flash proton radiography has been developed at LANL to study material properties under dynamic loading conditions through the analysis of time sequences of proton radiographs. The requirements of this growing experimental program have resulted in the need for improvements in spatial radiographic resolution. To meet these needs, a new magnetic lens system, consisting of four permanent magnet quadrupoles, has been developed. This new lens system was designed to reduce the second order chromatic aberrations, the dominant source of image blur in 800 MeV proton radiography, as well as magnifying the image to reduce the blur contribution from the detector and camera systems. The recently commissioned lens system performed as designed, providing nearly a factor of three improvement in radiographic resolution.
Real-time stop sign detection and distance estimation using a single camera
NASA Astrophysics Data System (ADS)
Wang, Wenpeng; Su, Yuxuan; Cheng, Ming
2018-04-01
In modern world, the drastic development of driver assistance system has made driving a lot easier than before. In order to increase the safety onboard, a method was proposed to detect STOP sign and estimate distance using a single camera. In STOP sign detection, LBP-cascade classifier was applied to identify the sign in the image, and the principle of pinhole imaging was based for distance estimation. Road test was conducted using a detection system built with a CMOS camera and software developed by Python language with OpenCV library. Results shows that that the proposed system reach a detection accuracy of maximum of 97.6% at 10m, a minimum of 95.00% at 20m, and 5% max error in distance estimation. The results indicate that the system is effective and has the potential to be used in both autonomous driving and advanced driver assistance driving systems.
A Probabilistic Feature Map-Based Localization System Using a Monocular Camera.
Kim, Hyungjin; Lee, Donghwa; Oh, Taekjun; Choi, Hyun-Taek; Myung, Hyun
2015-08-31
Image-based localization is one of the most widely researched localization techniques in the robotics and computer vision communities. As enormous image data sets are provided through the Internet, many studies on estimating a location with a pre-built image-based 3D map have been conducted. Most research groups use numerous image data sets that contain sufficient features. In contrast, this paper focuses on image-based localization in the case of insufficient images and features. A more accurate localization method is proposed based on a probabilistic map using 3D-to-2D matching correspondences between a map and a query image. The probabilistic feature map is generated in advance by probabilistic modeling of the sensor system as well as the uncertainties of camera poses. Using the conventional PnP algorithm, an initial camera pose is estimated on the probabilistic feature map. The proposed algorithm is optimized from the initial pose by minimizing Mahalanobis distance errors between features from the query image and the map to improve accuracy. To verify that the localization accuracy is improved, the proposed algorithm is compared with the conventional algorithm in a simulation and realenvironments.
A Probabilistic Feature Map-Based Localization System Using a Monocular Camera
Kim, Hyungjin; Lee, Donghwa; Oh, Taekjun; Choi, Hyun-Taek; Myung, Hyun
2015-01-01
Image-based localization is one of the most widely researched localization techniques in the robotics and computer vision communities. As enormous image data sets are provided through the Internet, many studies on estimating a location with a pre-built image-based 3D map have been conducted. Most research groups use numerous image data sets that contain sufficient features. In contrast, this paper focuses on image-based localization in the case of insufficient images and features. A more accurate localization method is proposed based on a probabilistic map using 3D-to-2D matching correspondences between a map and a query image. The probabilistic feature map is generated in advance by probabilistic modeling of the sensor system as well as the uncertainties of camera poses. Using the conventional PnP algorithm, an initial camera pose is estimated on the probabilistic feature map. The proposed algorithm is optimized from the initial pose by minimizing Mahalanobis distance errors between features from the query image and the map to improve accuracy. To verify that the localization accuracy is improved, the proposed algorithm is compared with the conventional algorithm in a simulation and realenvironments. PMID:26404284
Kim, Dong-Keun; Yoo, Sun K; Kim, Sun H
2005-01-01
The instant transmission of radiological images may be important for making rapid clinical decisions about emergency patients. We have examined an instant image transfer system based on a personal digital assistant (PDA) phone with a built-in camera. Images displayed on a picture archiving and communication systems (PACS) monitor can be captured by the camera in the PDA phone directly. Images can then be transmitted from an emergency centre to a remote physician via a wireless high-bandwidth network (CDMA 1 x EVDO). We reviewed the radiological lesions in 10 normal and 10 abnormal cases produced by modalities such as computerized tomography (CT), magnetic resonance (MR) and digital angiography. The images were of 24-bit depth and 1,144 x 880, 1,120 x 840, 1,024 x 768, 800 x 600, 640 x 480 and 320 x 240 pixels. Three neurosurgeons found that for satisfactory remote consultation a minimum size of 640 x 480 pixels was required for CT and MR images and 1,024 x 768 pixels for angiography images. Although higher resolution produced higher clinical satisfaction, it also required more transmission time. At the limited bandwidth employed, higher resolutions could not be justified.
Image database for digital hand atlas
NASA Astrophysics Data System (ADS)
Cao, Fei; Huang, H. K.; Pietka, Ewa; Gilsanz, Vicente; Dey, Partha S.; Gertych, Arkadiusz; Pospiech-Kurkowska, Sywia
2003-05-01
Bone age assessment is a procedure frequently performed in pediatric patients to evaluate their growth disorder. A commonly used method is atlas matching by a visual comparison of a hand radiograph with a small reference set of old Greulich-Pyle atlas. We have developed a new digital hand atlas with a large set of clinically normal hand images of diverse ethnic groups. In this paper, we will present our system design and implementation of the digital atlas database to support the computer-aided atlas matching for bone age assessment. The system consists of a hand atlas image database, a computer-aided diagnostic (CAD) software module for image processing and atlas matching, and a Web user interface. Users can use a Web browser to push DICOM images, directly or indirectly from PACS, to the CAD server for a bone age assessment. Quantitative features on the examined image, which reflect the skeletal maturity, are then extracted and compared with patterns from the atlas image database to assess the bone age. The digital atlas method built on a large image database and current Internet technology provides an alternative to supplement or replace the traditional one for a quantitative, accurate and cost-effective assessment of bone age.
Ramanujan, V Krishnan; Ren, Songyang; Park, Sangyong; Farkas, Daniel L
2011-01-01
We report here a non-invasive multispectral imaging platform for monitoring spectral reflectance and fluorescence images from primary breast carcinoma and metastatic lymph nodes in preclinical rat model in vivo. The system is built around a monochromator light source and an acousto-optic tunable filter (AOTF) for spectral selection. Quantitative analysis of the measured reflectance profiles in the presence of a widely-used lymphazurin dye clearly demonstrates the capability of the proposed imaging platform to detect tumor-associated spectral signatures in the primary tumors as well as metastatic lymphatics. Tumor-associated changes in vascular oxygenation and interstitial fluid pressure are reasoned to be the physiological sources of the measured reflectance profiles. We also discuss the translational potential of our imaging platform in intra-operative clinical setting. PMID:21572915
Performance Characterization of an xy-Stage Applied to Micrometric Laser Direct Writing Lithography.
Jaramillo, Juan; Zarzycki, Artur; Galeano, July; Sandoz, Patrick
2017-01-31
This article concerns the characterization of the stability and performance of a motorized stage used in laser direct writing lithography. The system was built from commercial components and commanded by G-code. Measurements use a pseudo-periodic-pattern (PPP) observed by a camera and image processing is based on Fourier transform and phase measurement methods. The results report that the built system has a stability against vibrations determined by peak-valley deviations of 65 nm and 26 nm in the x and y directions, respectively, with a standard deviation of 10 nm in both directions. When the xy-stage is in movement, it works with a resolution of 0.36 μm, which is an acceptable value for most of research and development (R and D) microtechnology developments in which the typical feature size used is in the micrometer range.
Performance Characterization of an xy-Stage Applied to Micrometric Laser Direct Writing Lithography
Jaramillo, Juan; Zarzycki, Artur; Galeano, July; Sandoz, Patrick
2017-01-01
This article concerns the characterization of the stability and performance of a motorized stage used in laser direct writing lithography. The system was built from commercial components and commanded by G-code. Measurements use a pseudo-periodic-pattern (PPP) observed by a camera and image processing is based on Fourier transform and phase measurement methods. The results report that the built system has a stability against vibrations determined by peak-valley deviations of 65 nm and 26 nm in the x and y directions, respectively, with a standard deviation of 10 nm in both directions. When the xy-stage is in movement, it works with a resolution of 0.36 µm, which is an acceptable value for most of research and development (R and D) microtechnology developments in which the typical feature size used is in the micrometer range. PMID:28146126
A video wireless capsule endoscopy system powered wirelessly: design, analysis and experiment
NASA Astrophysics Data System (ADS)
Pan, Guobing; Xin, Wenhui; Yan, Guozheng; Chen, Jiaoliao
2011-06-01
Wireless capsule endoscopy (WCE), as a relatively new technology, has brought about a revolution in the diagnosis of gastrointestinal (GI) tract diseases. However, the existing WCE systems are not widely applied in clinic because of the low frame rate and low image resolution. A video WCE system based on a wireless power supply is developed in this paper. This WCE system consists of a video capsule endoscope (CE), a wireless power transmission device, a receiving box and an image processing station. Powered wirelessly, the video CE has the abilities of imaging the GI tract and transmitting the images wirelessly at a frame rate of 30 frames per second (f/s). A mathematical prototype was built to analyze the power transmission system, and some experiments were performed to test the capability of energy transferring. The results showed that the wireless electric power supply system had the ability to transfer more than 136 mW power, which was enough for the working of a video CE. In in vitro experiments, the video CE produced clear images of the small intestine of a pig with the resolution of 320 × 240, and transmitted NTSC format video outside the body. Because of the wireless power supply, the video WCE system with high frame rate and high resolution becomes feasible, and provides a novel solution for the diagnosis of the GI tract in clinic.
Arabic word recognizer for mobile applications
NASA Astrophysics Data System (ADS)
Khanna, Nitin; Abdollahian, Golnaz; Brame, Ben; Boutin, Mireille; Delp, Edward J.
2011-03-01
When traveling in a region where the local language is not written using a "Roman alphabet," translating written text (e.g., documents, road signs, or placards) is a particularly difficult problem since the text cannot be easily entered into a translation device or searched using a dictionary. To address this problem, we are developing the "Rosetta Phone," a handheld device (e.g., PDA or mobile telephone) capable of acquiring an image of the text, locating the region (word) of interest within the image, and producing both an audio and a visual English interpretation of the text. This paper presents a system targeted for interpreting words written in Arabic script. The goal of this work is to develop an autonomous, segmentation-free Arabic phrase recognizer, with computational complexity low enough to deploy on a mobile device. A prototype of the proposed system has been deployed on an iPhone with a suitable user interface. The system was tested on a number of noisy images, in addition to the images acquired from the iPhone's camera. It identifies Arabic words or phrases by extracting appropriate features and assigning "codewords" to each word or phrase. On a dictionary of 5,000 words, the system uniquely mapped (word-image to codeword) 99.9% of the words. The system has a 82% recognition accuracy on images of words captured using the iPhone's built-in camera.
A low-cost multimodal head-mounted display system for neuroendoscopic surgery.
Xu, Xinghua; Zheng, Yi; Yao, Shujing; Sun, Guochen; Xu, Bainan; Chen, Xiaolei
2018-01-01
With rapid advances in technology, wearable devices as head-mounted display (HMD) have been adopted for various uses in medical science, ranging from simply aiding in fitness to assisting surgery. We aimed to investigate the feasibility and practicability of a low-cost multimodal HMD system in neuroendoscopic surgery. A multimodal HMD system, mainly consisted of a HMD with two built-in displays, an action camera, and a laptop computer displaying reconstructed medical images, was developed to assist neuroendoscopic surgery. With this intensively integrated system, the neurosurgeon could freely switch between endoscopic image, three-dimensional (3D) reconstructed virtual endoscopy images, and surrounding environment images. Using a leap motion controller, the neurosurgeon could adjust or rotate the 3D virtual endoscopic images at a distance to better understand the positional relation between lesions and normal tissues at will. A total of 21 consecutive patients with ventricular system diseases underwent neuroendoscopic surgery with the aid of this system. All operations were accomplished successfully, and no system-related complications occurred. The HMD was comfortable to wear and easy to operate. Screen resolution of the HMD was high enough for the neurosurgeon to operate carefully. With the system, the neurosurgeon might get a better comprehension on lesions by freely switching among images of different modalities. The system had a steep learning curve, which meant a quick increment of skill with it. Compared with commercially available surgical assistant instruments, this system was relatively low-cost. The multimodal HMD system is feasible, practical, helpful, and relatively cost efficient in neuroendoscopic surgery.
Panretinal, high-resolution color photography of the mouse fundus.
Paques, Michel; Guyomard, Jean-Laurent; Simonutti, Manuel; Roux, Michel J; Picaud, Serge; Legargasson, Jean-François; Sahel, José-Alain
2007-06-01
To analyze high-resolution color photographs of the mouse fundus. A contact fundus camera based on topical endoscopy fundus imaging (TEFI) was built. Fundus photographs of C57 and Balb/c mice obtained by TEFI were qualitatively analyzed. High-resolution digital imaging of the fundus, including the ciliary body, was routinely obtained. The reflectance and contrast of retinal vessels varied significantly with the amount of incident and reflected light and, thus, with the degree of fundus pigmentation. The combination of chromatic and spherical aberration favored blue light imaging, in term of both field and contrast. TEFI is a small, low-cost system that allows high-resolution color fundus imaging and fluorescein angiography in conscious mice. Panretinal imaging is facilitated by the presence of the large rounded lens. TEFI significantly improves the quality of in vivo photography of retina and ciliary process of mice. Resolution is, however, affected by chromatic aberration, and should be improved by monochromatic imaging.
Kim, Taehoon; Visbal-Onufrak, Michelle A.; Konger, Raymond L.; Kim, Young L.
2017-01-01
Sensitive and accurate assessment of dermatologic inflammatory hyperemia in otherwise grossly normal-appearing skin conditions is beneficial to laypeople for monitoring their own skin health on a regular basis, to patients for looking for timely clinical examination, and to primary care physicians or dermatologists for delivering effective treatments. We propose that mathematical hyperspectral reconstruction from RGB images in a simple imaging setup can provide reliable visualization of hemoglobin content in a large skin area. Without relying on a complicated, expensive, and slow hyperspectral imaging system, we demonstrate the feasibility of determining heterogeneous or multifocal areas of inflammatory hyperemia associated with experimental photocarcinogenesis in mice. We envision that RGB-based reconstructed hyperspectral imaging of subclinical inflammatory hyperemic foci could potentially be integrated with the built-in camera (RGB sensor) of a smartphone to develop a simple imaging device that could offer affordable monitoring of dermatologic health. PMID:29188120
System Construction of the Stilbene Compact Neutron Scatter Camera
DOE Office of Scientific and Technical Information (OSTI.GOV)
Goldsmith, John E. M.; Gerling, Mark D.; Brennan, James S.
This report documents the construction of a stilbene-crystal-based compact neutron scatter camera. This system is essentially identical to the MINER (Mobile Imager of Neutrons for Emergency Responders) system previously built and deployed under DNN R&D funding,1 but with the liquid scintillator in the detection cells replaced by stilbene crystals. The availability of these two systems for side-by-side performance comparisons will enable us to unambiguously identify the performance enhancements provided by the stilbene crystals, which have only recently become commercially available in the large size required (3” diameter, 3” deep).
Modeling Of Object- And Scene-Prototypes With Hierarchically Structured Classes
NASA Astrophysics Data System (ADS)
Ren, Z.; Jensch, P.; Ameling, W.
1989-03-01
The success of knowledge-based image analysis methodology and implementation tools depends largely on an appropriately and efficiently built model wherein the domain-specific context information about and the inherent structure of the observed image scene have been encoded. For identifying an object in an application environment a computer vision system needs to know firstly the description of the object to be found in an image or in an image sequence, secondly the corresponding relationships between object descriptions within the image sequence. This paper presents models of image objects scenes by means of hierarchically structured classes. Using the topovisual formalism of graph and higraph, we are currently studying principally the relational aspect and data abstraction of the modeling in order to visualize the structural nature resident in image objects and scenes, and to formalize. their descriptions. The goal is to expose the structure of image scene and the correspondence of image objects in the low level image interpretation. process. The object-based system design approach has been applied to build the model base. We utilize the object-oriented programming language C + + for designing, testing and implementing the abstracted entity classes and the operation structures which have been modeled topovisually. The reference images used for modeling prototypes of objects and scenes are from industrial environments as'well as medical applications.
Automatic page layout using genetic algorithms for electronic albuming
NASA Astrophysics Data System (ADS)
Geigel, Joe; Loui, Alexander C. P.
2000-12-01
In this paper, we describe a flexible system for automatic page layout that makes use of genetic algorithms for albuming applications. The system is divided into two modules, a page creator module which is responsible for distributing images amongst various album pages, and an image placement module which positions images on individual pages. Final page layouts are specified in a textual form using XML for printing or viewing over the Internet. The system makes use of genetic algorithms, a class of search and optimization algorithms that are based on the concepts of biological evolution, for generating solutions with fitness based on graphic design preferences supplied by the user. The genetic page layout algorithm has been incorporated into a web-based prototype system for interactive page layout over the Internet. The prototype system is built using client-server architecture and is implemented in java. The system described in this paper has demonstrated the feasibility of using genetic algorithms for automated page layout in albuming and web-based imaging applications. We believe that the system adequately proves the validity of the concept, providing creative layouts in a reasonable number of iterations. By optimizing the layout parameters of the fitness function, we hope to further improve the quality of the final layout in terms of user preference and computation speed.
Using Purpose-Built Functions and Block Hashes to Enable Small Block and Sub-file Forensics
2010-01-01
JPEGs. We tested precarve using the nps-2009-canon2-gen6 (Garfinkel et al., 2009) disk image. The disk image was created with a 32 MB SD card and a...analysis of n-grams in the fragment. Fig. 1 e Usage of a 160 GB iPod reported by iTunes 8.2.1 (6) (top), as reported by the file system (bottom center), and...as computing with random sampling (bottom right). Note that iTunes usage actually in GiB, even though the program displays the “GB” label. Fig. 2 e
NASA Technical Reports Server (NTRS)
Lawson, R. Paul
2000-01-01
SPEC incorporated designed, built and operated a new instrument, called a pi-Nephelometer, on the NASA DC-8 for the SUCCESS field project. The pi-Nephelometer casts an image of a particle on a 400,000 pixel solid-state camera by freezing the motion of the particle using a 25 ns pulsed, high-power (60 W) laser diode. Unique optical imaging and particle detection systems precisely detect particles and define the depth-of-field so that at least one particle in the image is almost always in focus. A powerful image processing engine processes frames from the solid-state camera, identifies and records regions of interest (i.e. particle images) in real time. Images of ice crystals are displayed and recorded with 5 micron pixel resolution. In addition, a scattered light system simultaneously measures the scattering phase function of the imaged particle. The system consists of twenty-eight 1-mm optical fibers connected to microlenses bonded on the surface of avalanche photo diodes (APDs). Data collected with the pi-Nephelometer during the SUCCESS field project was reported in a special issue of Geophysical Research Letters. The pi-Nephelometer provided the basis for development of a commercial imaging probe, called the cloud particle imager (CPI), which has been installed on several research aircraft and used in More than a dozen field programs.
Du, Weiqi; Zhang, Gaofei; Ye, Liangchen
2016-01-01
Micromirror-based scanning displays have been the focus of a variety of applications. Lissajous scanning displays have advantages in terms of power consumption; however, the image quality is not good enough. The main reason for this is the varying size and the contrast ratio of pixels at different positions of the image. In this paper, the Lissajous scanning trajectory is analyzed and a new method based on the diamond pixel is introduced to Lissajous displays. The optical performance of micromirrors is discussed. A display system demonstrator is built, and tests of resolution and contrast ratio are conducted. The test results show that the new Lissajous scanning method can be used in displays by using diamond pixels and image quality remains stable at different positions. PMID:27187390
Du, Weiqi; Zhang, Gaofei; Ye, Liangchen
2016-05-11
Micromirror-based scanning displays have been the focus of a variety of applications. Lissajous scanning displays have advantages in terms of power consumption; however, the image quality is not good enough. The main reason for this is the varying size and the contrast ratio of pixels at different positions of the image. In this paper, the Lissajous scanning trajectory is analyzed and a new method based on the diamond pixel is introduced to Lissajous displays. The optical performance of micromirrors is discussed. A display system demonstrator is built, and tests of resolution and contrast ratio are conducted. The test results show that the new Lissajous scanning method can be used in displays by using diamond pixels and image quality remains stable at different positions.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Beiser, L.; Veligdan, J.
A Planar Optic Display (POD) is being built and tested for suitability as a high brightness replacement for the cathode ray tube, (CRT). The POD display technology utilizes a laminated optical waveguide structure which allows a projection type of display to be constructed in a thin (I to 2 inch) housing. Inherent in the optical waveguide is a black cladding matrix which gives the display a black appearance leading to very high contrast. A Digital Micromirror Device, (DMD) from Texas Instruments is used to create video images in conjunction with a 100 milliwatt green solid state laser. An anamorphic opticalmore » system is used to inject light into the POD to form a stigmatic image. In addition to the design of the POD screen, we discuss: image formation, image projection, and optical design constraints.« less
Park, Samuel Byeongjun; Kim, Jung-Gun; Lim, Ki-Woong; Yoon, Chae-Hyun; Kim, Dong-Jun; Kang, Han-Sung; Jo, Yung-Ho
2017-08-01
We developed an image-guided intervention robot system that can be operated in a magnetic resonance (MR) imaging gantry. The system incorporates a bendable needle intervention robot for breast cancer patients that overcomes the space limitations of the MR gantry. Most breast coil designs for breast MR imaging have side openings to allow manual localization. However, for many intervention procedures, the patient must be removed from the gantry. A robotic manipulation system with integrated image guidance software was developed. Our robotic manipulator was designed to be slim, so as to fit between the patient's side and the MR gantry wall. Only non-magnetic materials were used, and an electromagnetic shield was employed for cables and circuits. The image guidance software was built using open source libraries. In situ feasibility tests were performed in a 3-T MR system. One target point in the breast phantom was chosen by the clinician for each experiment, and our robot moved the needle close to the target point. Without image-guided feedback control, the needle end could not hit the target point (distance = 5 mm) in the first experiment. Using our robotic system, the needle hits the target lesion of the breast phantom at a distance of 2.3 mm from the same target point using image-guided feedback. The second experiment was performed using other target points, and the distance between the final needle end point and the target point was 0.8 mm. We successfully developed an MR-guided needle intervention robot for breast cancer patients. Further research will allow the expansion of these interventions.
Web-based visualization of very large scientific astronomy imagery
NASA Astrophysics Data System (ADS)
Bertin, E.; Pillay, R.; Marmo, C.
2015-04-01
Visualizing and navigating through large astronomy images from a remote location with current astronomy display tools can be a frustrating experience in terms of speed and ergonomics, especially on mobile devices. In this paper, we present a high performance, versatile and robust client-server system for remote visualization and analysis of extremely large scientific images. Applications of this work include survey image quality control, interactive data query and exploration, citizen science, as well as public outreach. The proposed software is entirely open source and is designed to be generic and applicable to a variety of datasets. It provides access to floating point data at terabyte scales, with the ability to precisely adjust image settings in real-time. The proposed clients are light-weight, platform-independent web applications built on standard HTML5 web technologies and compatible with both touch and mouse-based devices. We put the system to the test and assess the performance of the system and show that a single server can comfortably handle more than a hundred simultaneous users accessing full precision 32 bit astronomy data.
Blackboard architecture for medical image interpretation
NASA Astrophysics Data System (ADS)
Davis, Darryl N.; Taylor, Christopher J.
1991-06-01
There is a growing interest in using sophisticated knowledge-based systems for biomedical image interpretation. We present a principled attempt to use artificial intelligence methodologies in interpreting lateral skull x-ray images. Such radiographs are routinely used in cephalometric analysis to provide quantitative measurements useful to clinical orthodontists. Manual and interactive methods of analysis are known to be error prone and previous attempts to automate this analysis typically fail to capture the expertise and adaptability required to cope with the variability in biological structure and image quality. An integrated model-based system has been developed which makes use of a blackboard architecture and multiple knowledge sources. A model definition interface allows quantitative models, of feature appearance and location, to be built from examples as well as more qualitative modelling constructs. Visual task definition and blackboard control modules allow task-specific knowledge sources to act on information available to the blackboard in a hypothesise and test reasoning cycle. Further knowledge-based modules include object selection, location hypothesis, intelligent segmentation, and constraint propagation systems. Alternative solutions to given tasks are permitted.
Adal, Kedir M; Sidibé, Désiré; Ali, Sharib; Chaum, Edward; Karnowski, Thomas P; Mériaudeau, Fabrice
2014-04-01
Despite several attempts, automated detection of microaneurysm (MA) from digital fundus images still remains to be an open issue. This is due to the subtle nature of MAs against the surrounding tissues. In this paper, the microaneurysm detection problem is modeled as finding interest regions or blobs from an image and an automatic local-scale selection technique is presented. Several scale-adapted region descriptors are introduced to characterize these blob regions. A semi-supervised based learning approach, which requires few manually annotated learning examples, is also proposed to train a classifier which can detect true MAs. The developed system is built using only few manually labeled and a large number of unlabeled retinal color fundus images. The performance of the overall system is evaluated on Retinopathy Online Challenge (ROC) competition database. A competition performance measure (CPM) of 0.364 shows the competitiveness of the proposed system against state-of-the art techniques as well as the applicability of the proposed features to analyze fundus images. Copyright © 2013 Elsevier Ireland Ltd. All rights reserved.
CMOS Image Sensor with a Built-in Lane Detector.
Hsiao, Pei-Yung; Cheng, Hsien-Chein; Huang, Shih-Shinh; Fu, Li-Chen
2009-01-01
This work develops a new current-mode mixed signal Complementary Metal-Oxide-Semiconductor (CMOS) imager, which can capture images and simultaneously produce vehicle lane maps. The adopted lane detection algorithm, which was modified to be compatible with hardware requirements, can achieve a high recognition rate of up to approximately 96% under various weather conditions. Instead of a Personal Computer (PC) based system or embedded platform system equipped with expensive high performance chip of Reduced Instruction Set Computer (RISC) or Digital Signal Processor (DSP), the proposed imager, without extra Analog to Digital Converter (ADC) circuits to transform signals, is a compact, lower cost key-component chip. It is also an innovative component device that can be integrated into intelligent automotive lane departure systems. The chip size is 2,191.4 × 2,389.8 μm, and the package uses 40 pin Dual-In-Package (DIP). The pixel cell size is 18.45 × 21.8 μm and the core size of photodiode is 12.45 × 9.6 μm; the resulting fill factor is 29.7%.
Known Locations of Carbonate Rocks on Mars
NASA Technical Reports Server (NTRS)
2008-01-01
Green dots show the locations of orbital detections of carbonate-bearing rocks on Mars, determined by analysis of targeted observations by the Compact Reconnaissance Imaging Spectrometer for Mars (CRISM) acquired through January 2008. The spectrometer is on NASA's Mars Reconnaissance Orbiter. The base map is color-coded global topography (red is high, blue is low) overlain on mosaicked daytime thermal infrared images. The topography data are from the Mars Orbiter Laser Altimeter on NASA's Mars Global Surveyor. The thermal infrared imagery is from the Thermal Emission Imaging System camera on NASA's Mars Odyssey orbiter. The CRISM team, led by The Johns Hopkins University Applied Physics Laboratory, Laurel, Md., includes expertise from universities, government agencies and small businesses in the United States and abroad. Arizona State University, Tempe, operates the Thermal Emission Imaging System, which the university developed in collaboration with Raytheon Santa Barbara Remote Sensing. NASA's Jet Propulsion Laboratory, a division of the California Institute of Technology in Pasadena, manages the Mars Reconnaissance Orbiter and Mars Odyssey projects for the NASA Science Mission Directorate, Washington. Lockheed Martin Space Systems, Denver, built the orbiters.Cheng, Yufeng; Jin, Shuying; Wang, Mi; Zhu, Ying; Dong, Zhipeng
2017-06-20
The linear array push broom imaging mode is widely used for high resolution optical satellites (HROS). Using double-cameras attached by a high-rigidity support along with push broom imaging is one method to enlarge the field of view while ensuring high resolution. High accuracy image mosaicking is the key factor of the geometrical quality of complete stitched satellite imagery. This paper proposes a high accuracy image mosaicking approach based on the big virtual camera (BVC) in the double-camera system on the GaoFen2 optical remote sensing satellite (GF2). A big virtual camera can be built according to the rigorous imaging model of a single camera; then, each single image strip obtained by each TDI-CCD detector can be re-projected to the virtual detector of the big virtual camera coordinate system using forward-projection and backward-projection to obtain the corresponding single virtual image. After an on-orbit calibration and relative orientation, the complete final virtual image can be obtained by stitching the single virtual images together based on their coordinate information on the big virtual detector image plane. The paper subtly uses the concept of the big virtual camera to obtain a stitched image and the corresponding high accuracy rational function model (RFM) for concurrent post processing. Experiments verified that the proposed method can achieve seamless mosaicking while maintaining the geometric accuracy.
View of the FGB prior to rendezvous and grapple
1998-12-06
S88-E-5047 (12-06-98) --- With Endeavour's astronauts waiting to mate the Russian-built Zarya control module with the U.S.-built Unity connecting module, an electronic still camera (ESC) was used to record this image of the approaching Zarya. A portion of Unity is in the foreground. Using the shuttle's 50-ft.-long Canadian-built robot arm, astronaut Nancy J. Currie, working from Endeavour's aft flight deck, plucked Zarya out of orbit at 5:47 p.m. (CST), Dec. 6. The craft had been orbiting Earth for a little over 16 days prior to the grapple and subsequent docking to Unity. This image was recorded at 23:11:05 GMT, Dec. 6.
Training system for digital mammographic diagnoses of breast cancer
NASA Astrophysics Data System (ADS)
Thomaz, R. L.; Nirschl Crozara, M. G.; Patrocinio, A. C.
2013-03-01
As the technology evolves, the analog mammography systems are being replaced by digital systems. The digital system uses video monitors as the display of mammographic images instead of the previously used screen-film and negatoscope for analog images. The change in the way of visualizing mammographic images may require a different approach for training the health care professionals in diagnosing the breast cancer with digital mammography. Thus, this paper presents a computational approach to train the health care professionals providing a smooth transition between analog and digital technology also training to use the advantages of digital image processing tools to diagnose the breast cancer. This computational approach consists of a software where is possible to open, process and diagnose a full mammogram case from a database, which has the digital images of each of the mammographic views. The software communicates with a gold standard digital mammogram cases database. This database contains the digital images in Tagged Image File Format (TIFF) and the respective diagnoses according to BI-RADSTM, these files are read by software and shown to the user as needed. There are also some digital image processing tools that can be used to provide better visualization of each single image. The software was built based on a minimalist and a user-friendly interface concept that might help in the smooth transition. It also has an interface for inputting diagnoses from the professional being trained, providing a result feedback. This system has been already completed, but hasn't been applied to any professional training yet.
Low-power, low-cost urinalysis system with integrated dipstick evaluation and microscopic analysis.
Smith, Gennifer T; Li, Linkai; Zhu, Yue; Bowden, Audrey K
2018-06-21
We introduce a coupled dipstick and microscopy device for analyzing urine samples. The device is capable of accurately assessing urine dipstick results while simultaneously imaging the microscopic contents within the sample. We introduce a long working distance, cellphone-based microscope in combination with an oblique illumination scheme to accurately visualize and quantify particles within the urine sample. To facilitate accurate quantification, we couple the imaging set-up with a power-free filtration system. The proposed device is reusable, low-cost, and requires very little power. We show that results obtained with the proposed device and custom-built app are consistent with those obtained with the standard clinical protocol, suggesting the potential clinical utility of the device.
2005-09-06
This Tempel 1 image was built up from scaling images from NASA Deep Impact to 5 meters/pixel and aligned to fixed points. Each image at closer range replaced equivalent locations observed at a greater distance.
Lutterbey, G; Behrends, K; Falkenhausen, M V; Wattjes, M P; Morakkabati, N; Gieseke, J; Schild, H
2007-02-01
The purpose of this study was to compare the in-built body coil of the 3.0-Tesla (T) scanner with a dedicated surface coil of a 1.5 T system regarding knee imaging. We performed an intraindividual prospective clinical trial on 17 patients with knee pain using magnetic resonance imaging (MRI) at 1.5 and 3.0 T systems equipped with identical gradient systems. Proton-density-weighted turbo spin echo sequences with the same spatial resolution and comparable contrast parameters were used. A quantitative measurement of signal to noise ratio (SNR), relative contrast (RC) and contrast to noise ratio (CNR) between muscle and bone marrow was performed, followed by a qualitative assessment of anatomic/pathologic structures and the extent of artefacts. At 3.0 T, 30 lesions (91%) compared to 33 lesions at 1.5 T were detected. The SNR/CNR/RC were moderately reduced at 3.0 T versus 1.5 T (muscle 42 vs 47 and bone 83 vs 112/46 vs 69/0.33 vs 0.43). Motion artefacts from the pulsating popliteal artery were significantly increased at 3.0 T. A visible and measurable signal loss occurred at 3.0 T using the built-in body coil compared with the dedicated 1.5 T knee coil, but nearly all clinically important information could be obtained.
Convolutional Neural Network-Based Shadow Detection in Images Using Visible Light Camera Sensor.
Kim, Dong Seop; Arsalan, Muhammad; Park, Kang Ryoung
2018-03-23
Recent developments in intelligence surveillance camera systems have enabled more research on the detection, tracking, and recognition of humans. Such systems typically use visible light cameras and images, in which shadows make it difficult to detect and recognize the exact human area. Near-infrared (NIR) light cameras and thermal cameras are used to mitigate this problem. However, such instruments require a separate NIR illuminator, or are prohibitively expensive. Existing research on shadow detection in images captured by visible light cameras have utilized object and shadow color features for detection. Unfortunately, various environmental factors such as illumination change and brightness of background cause detection to be a difficult task. To overcome this problem, we propose a convolutional neural network-based shadow detection method. Experimental results with a database built from various outdoor surveillance camera environments, and from the context-aware vision using image-based active recognition (CAVIAR) open database, show that our method outperforms previous works.
Convolutional Neural Network-Based Shadow Detection in Images Using Visible Light Camera Sensor
Kim, Dong Seop; Arsalan, Muhammad; Park, Kang Ryoung
2018-01-01
Recent developments in intelligence surveillance camera systems have enabled more research on the detection, tracking, and recognition of humans. Such systems typically use visible light cameras and images, in which shadows make it difficult to detect and recognize the exact human area. Near-infrared (NIR) light cameras and thermal cameras are used to mitigate this problem. However, such instruments require a separate NIR illuminator, or are prohibitively expensive. Existing research on shadow detection in images captured by visible light cameras have utilized object and shadow color features for detection. Unfortunately, various environmental factors such as illumination change and brightness of background cause detection to be a difficult task. To overcome this problem, we propose a convolutional neural network-based shadow detection method. Experimental results with a database built from various outdoor surveillance camera environments, and from the context-aware vision using image-based active recognition (CAVIAR) open database, show that our method outperforms previous works. PMID:29570690
NVSIM: UNIX-based thermal imaging system simulator
NASA Astrophysics Data System (ADS)
Horger, John D.
1993-08-01
For several years the Night Vision and Electronic Sensors Directorate (NVESD) has been using an internally developed forward looking infrared (FLIR) simulation program. In response to interest in the simulation part of these projects by other organizations, NVESD has been working on a new version of the simulation, NVSIM, that will be made generally available to the FLIR using community. NVSIM uses basic FLIR specification data, high resolution thermal input imagery and spatial domain image processing techniques to produce simulated image outputs from a broad variety of FLIRs. It is being built around modular programming techniques to allow simpler addition of more sensor effects. The modularity also allows selective inclusion and exclusion of individual sensor effects at run time. The simulation has been written in the industry standard ANSI C programming language under the widely used UNIX operating system to make it easily portable to a wide variety of computer platforms.
Open-source platforms for navigated image-guided interventions.
Ungi, Tamas; Lasso, Andras; Fichtinger, Gabor
2016-10-01
Navigation technology is changing the clinical standards in medical interventions by making existing procedures more accurate, and new procedures possible. Navigation is based on preoperative or intraoperative imaging combined with 3-dimensional position tracking of interventional tools registered to the images. Research of navigation technology in medical interventions requires significant engineering efforts. The difficulty of developing such complex systems has been limiting the clinical translation of new methods and ideas. A key to the future success of this field is to provide researchers with platforms that allow rapid implementation of applications with minimal resources spent on reimplementing existing system features. A number of platforms have been already developed that can share data in real time through standard interfaces. Complete navigation systems can be built using these platforms using a layered software architecture. In this paper, we review the most popular platforms, and show an effective way to take advantage of them through an example surgical navigation application. Copyright © 2016 Elsevier B.V. All rights reserved.
FPGA Based High Speed Data Acquisition System for Electrical Impedance Tomography
Khan, S; Borsic, A; Manwaring, Preston; Hartov, Alexander; Halter, Ryan
2014-01-01
Electrical Impedance Tomography (EIT) systems are used to image tissue bio-impedance. EIT provides a number of features making it attractive for use as a medical imaging device including the ability to image fast physiological processes (>60 Hz), to meet a range of clinical imaging needs through varying electrode geometries and configurations, to impart only non-ionizing radiation to a patient, and to map the significant electrical property contrasts present between numerous benign and pathological tissues. To leverage these potential advantages for medical imaging, we developed a modular 32 channel data acquisition (DAQ) system using National Instruments’ PXI chassis, along with FPGA, ADC, Signal Generator and Timing and Synchronization modules. To achieve high frame rates, signal demodulation and spectral characteristics of higher order harmonics were computed using dedicated FFT-hardware built into the FPGA module. By offloading the computing onto FPGA, we were able to achieve a reduction in throughput required between the FPGA and PC by a factor of 32:1. A custom designed analog front end (AFE) was used to interface electrodes with our system. Our system is wideband, and capable of acquiring data for input signal frequencies ranging from 100 Hz to 12 MHz. The modular design of both the hardware and software will allow this system to be flexibly configured for the particular clinical application. PMID:24729790
NASA Astrophysics Data System (ADS)
Satoh, Hitoshi; Niki, Noboru; Mori, Kiyoshi; Eguchi, Kenji; Kaneko, Masahiro; Kakinuma, Ryutarou; Moriyama, Noriyuki; Ohmatsu, Hironobu; Masuda, Hideo; Machida, Suguru
2007-03-01
Multislice CT scanner advanced remarkably at the speed at which the chest CT images were acquired for mass screening. Mass screening based on multislice CT images requires a considerable number of images to be read. It is this time-consuming step that makes the use of helical CT for mass screening impractical at present. To overcome this problem, we have provided diagnostic assistance methods to medical screening specialists by developing a lung cancer screening algorithm that automatically detects suspected lung cancers in helical CT images and a coronary artery calcification screening algorithm that automatically detects suspected coronary artery calcification. Moreover, we have provided diagnostic assistance methods to medical screening specialists by using a lung cancer screening algorithm built into mobile helical CT scanner for the lung cancer mass screening done in the region without the hospital. We also have developed electronic medical recording system and prototype internet system for the community health in two or more regions by using the Virtual Private Network router and Biometric fingerprint authentication system and Biometric face authentication system for safety of medical information. Based on these diagnostic assistance methods, we have now developed a new computer-aided workstation and database that can display suspected lesions three-dimensionally in a short time. This paper describes basic studies that have been conducted to evaluate this new system.
Adaptable data management for systems biology investigations
Boyle, John; Rovira, Hector; Cavnor, Chris; Burdick, David; Killcoyne, Sarah; Shmulevich, Ilya
2009-01-01
Background Within research each experiment is different, the focus changes and the data is generated from a continually evolving barrage of technologies. There is a continual introduction of new techniques whose usage ranges from in-house protocols through to high-throughput instrumentation. To support these requirements data management systems are needed that can be rapidly built and readily adapted for new usage. Results The adaptable data management system discussed is designed to support the seamless mining and analysis of biological experiment data that is commonly used in systems biology (e.g. ChIP-chip, gene expression, proteomics, imaging, flow cytometry). We use different content graphs to represent different views upon the data. These views are designed for different roles: equipment specific views are used to gather instrumentation information; data processing oriented views are provided to enable the rapid development of analysis applications; and research project specific views are used to organize information for individual research experiments. This management system allows for both the rapid introduction of new types of information and the evolution of the knowledge it represents. Conclusion Data management is an important aspect of any research enterprise. It is the foundation on which most applications are built, and must be easily extended to serve new functionality for new scientific areas. We have found that adopting a three-tier architecture for data management, built around distributed standardized content repositories, allows us to rapidly develop new applications to support a diverse user community. PMID:19265554
NASA Astrophysics Data System (ADS)
Xu, Hanqiu; Huang, Shaolin; Zhang, Tiejun
2013-10-01
Worldwide urbanization has accelerated expansion of urban built-up lands and resulted in substantial negative impacts on the global environments. Precisely measuring the urban sprawl is becoming an increasing need. Among the satellite-based earth observation systems, the Landsat and ASTER data are most suitable for mesoscale measurements of urban changes. Nevertheless, to date the difference in the capability of mapping built-up land between the two sensors is not clear. Therefore, this study compared the performances of the Landsat-7 ETM+ and ASTER sensors for built-up land mapping in the coastal areas of southeastern China. The comparison was implemented on three date-coincident image pairs and achieved by using three approaches, including per-band-based, index-based, and classification-based comparisons. The index used is the Index-based Built-up Index (IBI), while the classification algorithm employed is the Support Vector Machine (SVM). Results show that in the study areas, ETM+ and ASTER have an overall similar performance in built-up land mapping but also differ in several aspects. The IBI values determined from ASTER were consistently higher than from ETM+ by up to 45.54% according to percentage difference. The ASTER also estimates more built-up land area than ETM+ by 5.9-6.3% estimated with the IBI-based approach or 3.9-6.1% with the SVM classification. The differences in the spectral response functions and spatial resolution between relative spectral bands of the two sensors are attributed to these different performances.
Time-domain Surveys and Data Shift: Case Study at the intermediate Palomar Transient Factory
NASA Astrophysics Data System (ADS)
Rebbapragada, Umaa; Bue, Brian; Wozniak, Przemyslaw R.
2015-01-01
Next generation time-domain surveys are susceptible to the problem of data shift that is caused by upgrades to data processing pipelines and instruments. Data shift degrades the performance of automated machine learning classifiers that vet detections and classify source types because fundamental assumptions are violated when classifiers are built in one data regime but are deployed on data from another. This issue is not currently discussed within the astronomical community, but will be increasingly pressing over the next decade with the advent of new time domain surveys.We look at the problem of data shift that was caused by a data pipeline upgrade when the intermediate Palomar Transient Factory (iPTF) succeeded the Palomar Transient Factory (PTF) in January 2013. iPTF relies upon machine-learned Real-Bogus classifiers to vet sources extracted from subtracted images on a scale of zero to one where zero indicates a bogus (image artifact) and one indicates a real astronomical transient, with the overwhelming majority of candidates are scored as bogus. An effective Real-Bogus system filters all but the most promising candidates, which are presented to human scanners who make decisions about triggering follow up assets.The Real-Bogus systems currently in operation at iPTF (RB4 and RB5) solve the data shift problem. The statistical models of RB4 and RB5 were built from the ground up using examples from iPTF alone, whereas an older system, RB2, was built using PTF data, but was deployed after iPTF launched. We discuss the machine learning assumptions that are violated when a system is trained on one domain (PTF) but deployed on another (iPTF) that experiences data shift. We provide illustrative examples of data parameters and statistics that experienced shift. Finally, we show results comparing the three systems in operation, demonstrating that systems that solve domain shift (RB4 and RB5) are superior to those that don't (RB2).Research described in this abstract was carried out at the Jet Propulsion Laboratory under contract with the National Aeronautics and Space Administration. US Government Support Acknowledged.
NASA Astrophysics Data System (ADS)
Saari, H.; Akujärvi, A.; Holmlund, C.; Ojanen, H.; Kaivosoja, J.; Nissinen, A.; Niemeläinen, O.
2017-10-01
The accurate determination of the quality parameters of crops requires a spectral range from 400 nm to 2500 nm (Kawamura et al., 2010, Thenkabail et al., 2002). Presently the hyperspectral imaging systems that cover this wavelength range consist of several separate hyperspectral imagers and the system weight is from 5 to 15 kg. In addition the cost of the Short Wave Infrared (SWIR) cameras is high ( 50 k€). VTT has previously developed compact hyperspectral imagers for drones and Cubesats for Visible and Very near Infrared (VNIR) spectral ranges (Saari et al., 2013, Mannila et al., 2013, Näsilä et al., 2016). Recently VTT has started to develop a hyperspectral imaging system that will enable imaging simultaneously in the Visible, VNIR, and SWIR spectral bands. The system can be operated from a drone, on a camera stand, or attached to a tractor. The targeted main applications of the DroneKnowledge hyperspectral system are grass, peas, and cereals. In this paper the characteristics of the built system are shortly described. The system was used for spectral measurements of wheat, several grass species and pea plants fixed to the camera mount in the test fields in Southern Finland and in the green house. The wheat, grass and pea field measurements were also carried out using the system mounted on the tractor. The work is part of the Finnish nationally funded DroneKnowledge - Towards knowledge based export of small UAS remote sensing technology project.
A web-based computer aided system for liver surgery planning: initial implementation on RayPlus
NASA Astrophysics Data System (ADS)
Luo, Ming; Yuan, Rong; Sun, Zhi; Li, Tianhong; Xie, Qingguo
2016-03-01
At present, computer aided systems for liver surgery design and risk evaluation are widely used in clinical all over the world. However, most systems are local applications that run on high-performance workstations, and the images have to processed offline. Compared with local applications, a web-based system is accessible anywhere and for a range of regardless of relative processing power or operating system. RayPlus (http://rayplus.life.hust.edu.cn), a B/S platform for medical image processing, was developed to give a jump start on web-based medical image processing. In this paper, we implement a computer aided system for liver surgery planning on the architecture of RayPlus. The system consists of a series of processing to CT images including filtering, segmentation, visualization and analyzing. Each processing is packaged into an executable program and runs on the server side. CT images in DICOM format are processed step by to interactive modeling on browser with zero-installation and server-side computing. The system supports users to semi-automatically segment the liver, intrahepatic vessel and tumor from the pre-processed images. Then, surface and volume models are built to analyze the vessel structure and the relative position between adjacent organs. The results show that the initial implementation meets satisfactorily its first-order objectives and provide an accurate 3D delineation of the liver anatomy. Vessel labeling and resection simulation are planned to add in the future. The system is available on Internet at the link mentioned above and an open username for testing is offered.
Wang, Chunliang; Ritter, Felix; Smedby, Orjan
2010-07-01
To enhance the functional expandability of a picture archiving and communication systems (PACS) workstation and to facilitate the integration of third-part image-processing modules, we propose a browser-server style method. In the proposed solution, the PACS workstation shows the front-end user interface defined in an XML file while the image processing software is running in the background as a server. Inter-process communication (IPC) techniques allow an efficient exchange of image data, parameters, and user input between the PACS workstation and stand-alone image-processing software. Using a predefined communication protocol, the PACS workstation developer or image processing software developer does not need detailed information about the other system, but will still be able to achieve seamless integration between the two systems and the IPC procedure is totally transparent to the final user. A browser-server style solution was built between OsiriX (PACS workstation software) and MeVisLab (Image-Processing Software). Ten example image-processing modules were easily added to OsiriX by converting existing MeVisLab image processing networks. Image data transfer using shared memory added <10ms of processing time while the other IPC methods cost 1-5 s in our experiments. The browser-server style communication based on IPC techniques is an appealing method that allows PACS workstation developers and image processing software developers to cooperate while focusing on different interests.
Spatial-scanning hyperspectral imaging probe for bio-imaging applications
NASA Astrophysics Data System (ADS)
Lim, Hoong-Ta; Murukeshan, Vadakke Matham
2016-03-01
The three common methods to perform hyperspectral imaging are the spatial-scanning, spectral-scanning, and snapshot methods. However, only the spectral-scanning and snapshot methods have been configured to a hyperspectral imaging probe as of today. This paper presents a spatial-scanning (pushbroom) hyperspectral imaging probe, which is realized by integrating a pushbroom hyperspectral imager with an imaging probe. The proposed hyperspectral imaging probe can also function as an endoscopic probe by integrating a custom fabricated image fiber bundle unit. The imaging probe is configured by incorporating a gradient-index lens at the end face of an image fiber bundle that consists of about 50 000 individual fiberlets. The necessary simulations, methodology, and detailed instrumentation aspects that are carried out are explained followed by assessing the developed probe's performance. Resolution test targets such as United States Air Force chart as well as bio-samples such as chicken breast tissue with blood clot are used as test samples for resolution analysis and for performance validation. This system is built on a pushbroom hyperspectral imaging system with a video camera and has the advantage of acquiring information from a large number of spectral bands with selectable region of interest. The advantages of this spatial-scanning hyperspectral imaging probe can be extended to test samples or tissues residing in regions that are difficult to access with potential diagnostic bio-imaging applications.
NASA Astrophysics Data System (ADS)
Boadi, Joseph; Byers, Robert A.; Fernandes, Jon; Mittar, Shweta; Hearnden, Vanessa; Lu, Zenghai; MacNeil, Sheila; Thornhill, Martin; Murdoch, Craig; Hunter, Keith D.; McKechnie, Alasdair; Matcher, Stephen J.
2016-02-01
OCT has demonstrated great potential to non-invasively detect oral epithelial cancers, potentially guiding biopsy and surgical resection. On non-ophthalmic tissues the preferred illumination wavelength is 1300 nm. Previous studies on skin have shown that useful image data can also be obtained at shorter wavelengths, with systems at 1060 nm and 820 nm offering reduced depth penetration but higher contrast. Here we apply a similar comparison to tissue engineered models of oral cancer and also to human biopsy samples, generally finding a similar trend. 1300 nm multi-beam OCT (Michelson Diagnostics EX1301) visualises stromal structures and surface keratin more clearly, providing useful image contrast down to around 1 mm. This system was compared with an ultra-high resolution home-built system operating at 890 nm (2.5 micron resolution vs 7.5 micron axial resolution for the EX1301). The UHR system reveals epithelial features more clearly, especially in the DOK pre-invasive cell line model and the biopsy samples. The relative effects of center wavelength vs axial resolution in generating the differential, wavelength-dependent contrast are assessed and the OCT biopsy images are compared with contemporary histology.
An automated system for whole microscopic image acquisition and analysis.
Bueno, Gloria; Déniz, Oscar; Fernández-Carrobles, María Del Milagro; Vállez, Noelia; Salido, Jesús
2014-09-01
The field of anatomic pathology has experienced major changes over the last decade. Virtual microscopy (VM) systems have allowed experts in pathology and other biomedical areas to work in a safer and more collaborative way. VMs are automated systems capable of digitizing microscopic samples that were traditionally examined one by one. The possibility of having digital copies reduces the risk of damaging original samples, and also makes it easier to distribute copies among other pathologists. This article describes the development of an automated high-resolution whole slide imaging (WSI) system tailored to the needs and problems encountered in digital imaging for pathology, from hardware control to the full digitization of samples. The system has been built with an additional digital monochromatic camera together with the color camera by default and LED transmitted illumination (RGB). Monochrome cameras are the preferred method of acquisition for fluorescence microscopy. The system is able to digitize correctly and form large high resolution microscope images for both brightfield and fluorescence. The quality of the digital images has been quantified using three metrics based on sharpness, contrast and focus. It has been proved on 150 tissue samples of brain autopsies, prostate biopsies and lung cytologies, at five magnifications: 2.5×, 10×, 20×, 40×, and 63×. The article is focused on the hardware set-up and the acquisition software, although results of the implemented image processing techniques included in the software and applied to the different tissue samples are also presented. © 2014 Wiley Periodicals, Inc.
Alternative techniques for high-resolution spectral estimation of spectrally encoded endoscopy
NASA Astrophysics Data System (ADS)
Mousavi, Mahta; Duan, Lian; Javidi, Tara; Ellerbee, Audrey K.
2015-09-01
Spectrally encoded endoscopy (SEE) is a minimally invasive optical imaging modality capable of fast confocal imaging of internal tissue structures. Modern SEE systems use coherent sources to image deep within the tissue and data are processed similar to optical coherence tomography (OCT); however, standard processing of SEE data via the Fast Fourier Transform (FFT) leads to degradation of the axial resolution as the bandwidth of the source shrinks, resulting in a well-known trade-off between speed and axial resolution. Recognizing the limitation of FFT as a general spectral estimation algorithm to only take into account samples collected by the detector, in this work we investigate alternative high-resolution spectral estimation algorithms that exploit information such as sparsity and the general region position of the bulk sample to improve the axial resolution of processed SEE data. We validate the performance of these algorithms using bothMATLAB simulations and analysis of experimental results generated from a home-built OCT system to simulate an SEE system with variable scan rates. Our results open a new door towards using non-FFT algorithms to generate higher quality (i.e., higher resolution) SEE images at correspondingly fast scan rates, resulting in systems that are more accurate and more comfortable for patients due to the reduced image time.
Software for minimalistic data management in large camera trap studies
Krishnappa, Yathin S.; Turner, Wendy C.
2014-01-01
The use of camera traps is now widespread and their importance in wildlife studies well understood. Camera trap studies can produce millions of photographs and there is a need for software to help manage photographs efficiently. In this paper, we describe a software system that was built to successfully manage a large behavioral camera trap study that produced more than a million photographs. We describe the software architecture and the design decisions that shaped the evolution of the program over the study’s three year period. The software system has the ability to automatically extract metadata from images, and add customized metadata to the images in a standardized format. The software system can be installed as a standalone application on popular operating systems. It is minimalistic, scalable and extendable so that it can be used by small teams or individual researchers for a broad variety of camera trap studies. PMID:25110471
Hipp, J Aaron; Manteiga, Alicia; Burgess, Amanda; Stylianou, Abby; Pless, Robert
2016-01-01
Active transportation opportunities and infrastructure are an important component of a community's design, livability, and health. Features of the built environment influence active transportation, but objective study of the natural experiment effects of built environment improvements on active transportation is challenging. The purpose of this study was to develop and present a novel method of active transportation research using webcams and crowdsourcing, and to determine if crosswalk enhancement was associated with changes in active transportation rates, including across a variety of weather conditions. The 20,529 publicly available webcam images from two street intersections in Washington, DC, USA were used to examine the impact of an improved crosswalk on active transportation. A crowdsource, Amazon Mechanical Turk, annotated image data. Temperature data were collected from the National Oceanic and Atmospheric Administration, and precipitation data were annotated from images by trained research assistants. Summary analyses demonstrated slight, bi-directional differences in the percent of images with pedestrians and bicyclists captured before and after the enhancement of the crosswalks. Chi-square analyses revealed these changes were not significant. In general, pedestrian presence increased in images captured during moderate temperatures compared to images captured during hot or cold temperatures. Chi-square analyses indicated the crosswalk improvement may have encouraged walking and biking in uncomfortable outdoor conditions (P < 0.5). The methods employed provide an objective, cost-effective alternative to traditional means of examining the effects of built environment changes on active transportation. The use of webcams to collect active transportation data has applications for community policymakers, planners, and health professionals. Future research will work to validate this method in a variety of settings as well as across different built environment and community policy initiatives.
Yu, Zhengyang; Zheng, Shusen; Chen, Huaiqing; Wang, Jianjun; Xiong, Qingwen; Jing, Wanjun; Zeng, Yu
2006-10-01
This research studies the process of dynamic concision and 3D reconstruction from medical body data using VRML and JavaScript language, focuses on how to realize the dynamic concision of 3D medical model built with VRML. The 2D medical digital images firstly are modified and manipulated by 2D image software. Then, based on these images, 3D mould is built with VRML and JavaScript language. After programming in JavaScript to control 3D model, the function of dynamic concision realized by Script node and sensor node in VRML. The 3D reconstruction and concision of body internal organs can be formed in high quality near to those got in traditional methods. By this way, with the function of dynamic concision, VRML browser can offer better windows of man-computer interaction in real time environment than before. 3D reconstruction and dynamic concision with VRML can be used to meet the requirement for the medical observation of 3D reconstruction and has a promising prospect in the fields of medical image.
TU-F-CAMPUS-I-05: Investigation of An EMCCD Detector with Variable Gain in a Micro-CT System
DOE Office of Scientific and Technical Information (OSTI.GOV)
Krishnakumar, S Bysani; Ionita, C; Rudin, S
Purpose: To investigate the performance of a newly built Electron Multiplying Charged Coupled Device (EMCCD) based Micro-CT system, with variable detector gain, using a phantom containing contrast agent of different concentrations. Methods: We built a micro- CT system with an EMCCD having 8 microns pixels and on-chip variable gain. We tested the system using a phantom containing five tubes filled with different iodine contrast solutions (30% to 70%). First, we scanned the phantom using various x-ray exposures values at 40 kVp and constant detector gain. Next, for the same tube currents, the detector gain was increased to maintain the airmore » value of the projection image constant. A standard FDK algorithm was used to reconstruct the data. Performance was analyzed by comparing the signal-to-noise ratio (SNR) measurements for increased gain with those for the low constant gain at each exposure. Results: The high detector gain reconstructed data SNR was always greater than the low gain data SNR for all x-ray settings and for all iodine features. The largest increases were observed for low contrast features, 30% iodine concentration, where the SNR improvement approached 2. Conclusion: One of the first implementations of an EMCCD based micro- CT system was presented and used to image a phantom with various iodine solution concentrations. The analysis of the reconstructed volumes showed a significant improvement of the SNR especially for low contrast features. The unique on-chip gain feature is a substantial benefit allowing the use of the system at very low x-ray exposures per frame.Partial support: NIH grant R01EB002873 and Toshiba Medical Systems Corp. Partial support: NIH grant R01EB002873 and Toshiba Medical Systems Corp.« less
International Space Station (ISS)
2000-09-01
This image of the International Space Station (ISS) was taken when Space Shuttle Atlantis (STS-106 mission) approached the ISS for docking. At the top is the Russian Progress supply ship that is linked with the Russian built Service Module or Zvezda. The Zvezda is cornected with the Russian built Functional Cargo Block (FGB) or Zarya. The U.S. built Node 1 or Unity module is seen at the bottom.
NASA Technical Reports Server (NTRS)
2002-01-01
This image is a mosaic of day and night infrared images of Melas Chasma taken by the camera system on NASA's Mars Odyssey spacecraft. The daytime temperature images are shown in black and white, superimposed on the martian topography. A single nighttime temperature image is superimposed in color. The daytime temperatures range from approximately -35 degrees Celsius (-31 degrees Fahrenheit) in black to -5 degrees Celsius (23 degrees Fahrenheit) in white. Overlapping landslides and individual layers in the walls of Melas Chasma can be seen in this image. The landslides flowed over 100 kilometers (62 miles) across the floor of Melas Chasma, producing deposits with ridges and grooves of alternating warm and cold materials that can still be seen. The temperature differences in the daytime images are due primarily to lighting effects, where sunlit slopes are warm (bright) and shadowed slopes are cool (dark). The nighttime temperature differences are due to differences in the abundance of rocky materials that retain their heat at night and stay relatively warm (red). Fine grained dust and sand (blue) cools off more rapidly at night. These images were acquired using the thermal infrared imaging system infrared Band 9, centered at 12.6 micrometers.Jet Propulsion Laboratory, a division of the California Institute of Technology in Pasadena, manages the 2001 Mars Odyssey mission for NASA's Office of Space Science in Washington, D.C. Investigators at Arizona State University in Tempe, the University of Arizona in Tucson and NASA's Johnson Space Center, Houston, operate the science instruments. Additional science partners are located at the Russian Aviation and Space Agency and at Los Alamos National Laboratories, New Mexico. Lockheed Martin Astronautics, Denver, is the prime contractor for the project, and developed and built the orbiter. Mission operations are conducted jointly from Lockheed Martin and from JPL. Aviation and Space Agency and at Los Alamos National Laboratories, New Mexico. Lockheed Martin Astronautics, Denver, is the prime contractor for the project, and developed and built the orbiter. Mission operations are conducted jointly from Lockheed Martin and from JPL.Fully automated laser ray tracing system to measure changes in the crystalline lens GRIN profile.
Qiu, Chen; Maceo Heilman, Bianca; Kaipio, Jari; Donaldson, Paul; Vaghefi, Ehsan
2017-11-01
Measuring the lens gradient refractive index (GRIN) accurately and reliably has proven an extremely challenging technical problem. A fully automated laser ray tracing (LRT) system was built to address this issue. The LRT system captures images of multiple laser projections before and after traversing through an ex vivo lens. These LRT images, combined with accurate measurements of the lens geometry, are used to calculate the lens GRIN profile. Mathematically, this is an ill-conditioned problem; hence, it is essential to apply biologically relevant constraints to produce a feasible solution. The lens GRIN measurements were compared with previously published data. Our GRIN retrieval algorithm produces fast and accurate measurements of the lens GRIN profile. Experiments to study the optics of physiologically perturbed lenses are the future direction of this research.
Fully automated laser ray tracing system to measure changes in the crystalline lens GRIN profile
Qiu, Chen; Maceo Heilman, Bianca; Kaipio, Jari; Donaldson, Paul; Vaghefi, Ehsan
2017-01-01
Measuring the lens gradient refractive index (GRIN) accurately and reliably has proven an extremely challenging technical problem. A fully automated laser ray tracing (LRT) system was built to address this issue. The LRT system captures images of multiple laser projections before and after traversing through an ex vivo lens. These LRT images, combined with accurate measurements of the lens geometry, are used to calculate the lens GRIN profile. Mathematically, this is an ill-conditioned problem; hence, it is essential to apply biologically relevant constraints to produce a feasible solution. The lens GRIN measurements were compared with previously published data. Our GRIN retrieval algorithm produces fast and accurate measurements of the lens GRIN profile. Experiments to study the optics of physiologically perturbed lenses are the future direction of this research. PMID:29188093
Interpolation strategies for reducing IFOV artifacts in microgrid polarimeter imagery.
Ratliff, Bradley M; LaCasse, Charles F; Tyo, J Scott
2009-05-25
Microgrid polarimeters are composed of an array of micro-polarizing elements overlaid upon an FPA sensor. In the past decade systems have been designed and built in all regions of the optical spectrum. These systems have rugged, compact designs and the ability to obtain a complete set of polarimetric measurements during a single image capture. However, these systems acquire the polarization measurements through spatial modulation and each measurement has a varying instantaneous field-of-view (IFOV). When these measurements are combined to estimate the polarization images, strong edge artifacts are present that severely degrade the estimated polarization imagery. These artifacts can be reduced when interpolation strategies are first applied to the intensity data prior to Stokes vector estimation. Here we formally study IFOV error and the performance of several bilinear interpolation strategies used for reducing it.
NASA Astrophysics Data System (ADS)
Zhang, Haichong K.; Lin, Melissa; Kim, Younsu; Paredes, Mateo; Kannan, Karun; Patel, Nisu; Moghekar, Abhay; Durr, Nicholas J.; Boctor, Emad M.
2017-03-01
Lumbar punctures (LPs) are interventional procedures used to collect cerebrospinal fluid (CSF), a bodily fluid needed to diagnose central nervous system disorders. Most lumbar punctures are performed blindly without imaging guidance. Because the target window is small, physicians can only accurately palpate the appropriate space about 30% of the time and perform a successful procedure after an average of three attempts. Although various forms of imaging based guidance systems have been developed to aid in this procedure, these systems complicate the procedure by including independent image modalities and requiring image-to-needle registration to guide the needle insertion. Here, we propose a simple and direct needle insertion platform utilizing a single ultrasound element within the needle through dynamic sensing and imaging. The needle-shaped ultrasound transducer can not only sense the distance between the tip and a potential obstacle such as bone, but also visually locate structures by combining transducer location tracking and back projection based tracked synthetic aperture beam-forming algorithm. The concept of the system was validated through simulation first, which revealed the tolerance to realistic error. Then, the initial prototype of the single element transducer was built into a 14G needle, and was mounted on a holster equipped with a rotation tracking encoder. We experimentally evaluated the system using a metal wire phantom mimicking high reflection bone structures and an actual spine bone phantom with both the controlled motion and freehand scanning. An ultrasound image corresponding to the model phantom structure was reconstructed using the beam-forming algorithm, and the resolution was improved compared to without beam-forming. These results demonstrated the proposed system has the potential to be used as an ultrasound imaging system for lumbar puncture procedures.
Reflective all-sky thermal infrared cloud imager
DOE Office of Scientific and Technical Information (OSTI.GOV)
Redman, Brian J.; Shaw, Joseph A.; Nugent, Paul W.
A reflective all-sky imaging system has been built using a long-wave infrared microbolometer camera and a reflective metal sphere. This compact system was developed for measuring spatial and temporal patterns of clouds and their optical depth in support of applications including Earth-space optical communications. The camera is mounted to the side of the reflective sphere to leave the zenith sky unobstructed. The resulting geometric distortion is removed through an angular map derived from a combination of checkerboard-target imaging, geometric ray tracing, and sun-location-based alignment. A tape of high-emissivity material on the side of the reflector acts as a reference thatmore » is used to estimate and remove thermal emission from the metal sphere. In conclusion, once a bias that is under continuing study was removed, sky radiance measurements from the all-sky imager in the 8-14 μm wavelength range agreed to within 0.91 W/(m 2 sr) of measurements from a previously calibrated, lens-based infrared cloud imager over its 110° field of view.« less
Reflective all-sky thermal infrared cloud imager
Redman, Brian J.; Shaw, Joseph A.; Nugent, Paul W.; ...
2018-04-17
A reflective all-sky imaging system has been built using a long-wave infrared microbolometer camera and a reflective metal sphere. This compact system was developed for measuring spatial and temporal patterns of clouds and their optical depth in support of applications including Earth-space optical communications. The camera is mounted to the side of the reflective sphere to leave the zenith sky unobstructed. The resulting geometric distortion is removed through an angular map derived from a combination of checkerboard-target imaging, geometric ray tracing, and sun-location-based alignment. A tape of high-emissivity material on the side of the reflector acts as a reference thatmore » is used to estimate and remove thermal emission from the metal sphere. In conclusion, once a bias that is under continuing study was removed, sky radiance measurements from the all-sky imager in the 8-14 μm wavelength range agreed to within 0.91 W/(m 2 sr) of measurements from a previously calibrated, lens-based infrared cloud imager over its 110° field of view.« less
NASA Astrophysics Data System (ADS)
Cui, Zhe; Wang, Anting; Ma, Qianli; Ming, Hai
2013-12-01
In this paper, the laser speckle pattern on human retina for a laser projection display is simulated. By introducing a specific eye model `Indiana Eye', the statistical properties of the laser speckle are numerical investigated. The results show that the aberrations of human eye (mostly spherical and chromatic) will decrease the speckle contrast felt by people. When the wavelength of the laser source is 550 nm (green), people will feel the strongest speck pattern and the weakest when the wavelength is 450 nm (blue). Myopia and hyperopia will decrease the speckle contrast by introducing large spherical aberrations. Although aberration is good for speckle reduction, but it will degrade the imaging capability of the eye. The results show that laser source (650 nm) will have the best image quality on the retina. At last, we compare the human eye with an aberration-free imaging system. Both the speckle contrast and the image quality appear different behavior in these two imaging systems. The results are useful when a standardized measurement procedure for speckle contrast needs to be built.
Assessing the Accuracy of Ortho-image using Photogrammetric Unmanned Aerial System
NASA Astrophysics Data System (ADS)
Jeong, H. H.; Park, J. W.; Kim, J. S.; Choi, C. U.
2016-06-01
Smart-camera can not only be operated under network environment anytime and any place but also cost less than the existing photogrammetric UAV since it provides high-resolution image, 3D location and attitude data on a real-time basis from a variety of built-in sensors. This study's proposed UAV photogrammetric method, low-cost UAV and smart camera were used. The elements of interior orientation were acquired through camera calibration. The image triangulation was conducted in accordance with presence or absence of consideration of the interior orientation (IO) parameters determined by camera calibration, The Digital Elevation Model (DEM) was constructed using the image data photographed at the target area and the results of the ground control point survey. This study also analyzes the proposed method's application possibility by comparing a Ortho-image the results of the ground control point survey. Considering these study findings, it is suggested that smartphone is very feasible as a payload for UAV system. It is also expected that smartphone may be loaded onto existing UAV playing direct or indirect roles significantly.
Scotti, F.; Soukhanovskii, V. A.
2015-12-09
A two-channel spectral imaging system based on a charge injection device radiation-hardened intensified camera was built for studies of plasma-surface interactions on divertor plasma facing components in the National Spherical Torus Experiment Upgrade (NSTX-U) tokamak. By means of commercially available mechanically referenced optical components, the two-wavelength setup images the light from the plasma, relayed by a fiber optic bundle, at two different wavelengths side-by-side on the same detector. Remotely controlled filter wheels are used for narrow band pass and neutral density filters on each optical path allowing for simultaneous imaging of emission at wavelengths differing in brightness up to 3more » orders of magnitude. Applications on NSTX-U will include the measurement of impurity influxes in the lower divertor strike point region and the imaging of plasma-material interaction on the head of the surface analysis probe MAPP (Material Analysis and Particle Probe). Furthermore, the diagnostic setup and initial results from its application on the lithium tokamak experiment are presented.« less
Henderson, Michael L; Dayhoff, Ruth E; Titton, Csaba P; Casertano, Andrew
2006-01-01
As part of its patient care mission, the U.S. Veterans Health Administration performs diagnostic imaging procedures at 141 medical centers and 850 outpatient clinics. VHA's VistA Imaging Package provides a full archival, display, and communications infrastructure and interfaces to radiology and other HIS modules as well as modalities and a worklist provider In addition, various medical center entities within VHA have elected to install commercial picture archiving and communications systems to enable image organization and interpretation. To evaluate interfaces between commercial PACS, the VistA hospital information system, and imaging modalities, VHA has built a fully constrained specification that is based on the Radiology Technical Framework (Rad-TF) Integrating the Healthcare Enterprise. The Health Level Seven normative conformance mechanism was applied to the IHE Rad-TF and agency requirements to arrive at a baseline set of message specifications. VHA provides a thorough implementation and testing process to promote the adoption of standards-based interoperability by all PACS vendors that want to interface with VistA Imaging.
NASA Astrophysics Data System (ADS)
Michaelis, H.; Behnke, T.; Bredthauer, R.; Holland, A.; Janesick, J.; Jaumann, R.; Keller, H. U.; Magrin, D.; Greggio, D.; Mottola, Stefano; Thomas, N.; Smith, P.
2017-11-01
When we talk about planetary exploration missions most people think spontaneously about fascinating images from other planets or close-up pictures of small planetary bodies such as asteroids and comets. Such images come in most cases from VIS/NIR- imaging- systems, simply called `cameras', which were typically built by institutes in collaboration with industry. Until now, they have nearly all been based on silicon CCD sensors, they have filter wheels and have often high power-consuming electronics. The question is, what are the challenges for future missions and what can be done to improve performance and scientific output. The exploration of Mars is ongoing. NASA and ESA are planning future missions to the outer planets like to the icy Jovian moons. Exploration of asteroids and comets are in focus of several recent and future missions. Furthermore, the detection and characterization of exo-planets will keep us busy for next generations. The paper is discussing the challenges and visions of imaging sensors for future planetary exploration missions. The focus of the talk is monolithic VIS/NIR- detectors.
Quiet echo planar imaging for functional and diffusion MRI
Price, Anthony N.; Cordero‐Grande, Lucilio; Malik, Shaihan; Ferrazzi, Giulio; Gaspar, Andreia; Hughes, Emer J.; Christiaens, Daan; McCabe, Laura; Schneider, Torben; Rutherford, Mary A.; Hajnal, Joseph V.
2017-01-01
Purpose To develop a purpose‐built quiet echo planar imaging capability for fetal functional and diffusion scans, for which acoustic considerations often compromise efficiency and resolution as well as angular/temporal coverage. Methods The gradient waveforms in multiband‐accelerated single‐shot echo planar imaging sequences have been redesigned to minimize spectral content. This includes a sinusoidal read‐out with a single fundamental frequency, a constant phase encoding gradient, overlapping smoothed CAIPIRINHA blips, and a novel strategy to merge the crushers in diffusion MRI. These changes are then tuned in conjunction with the gradient system frequency response function. Results Maintained image quality, SNR, and quantitative diffusion values while reducing acoustic noise up to 12 dB (A) is illustrated in two adult experiments. Fetal experiments in 10 subjects covering a range of parameters depict the adaptability and increased efficiency of quiet echo planar imaging. Conclusion Purpose‐built for highly efficient multiband fetal echo planar imaging studies, the presented framework reduces acoustic noise for all echo planar imaging‐based sequences. Full optimization by tuning to the gradient frequency response functions allows for a maximally time‐efficient scan within safe limits. This allows ambitious in‐utero studies such as functional brain imaging with high spatial/temporal resolution and diffusion scans with high angular/spatial resolution to be run in a highly efficient manner at acceptable sound levels. Magn Reson Med 79:1447–1459, 2018. © 2017 The Authors Magnetic Resonance in Medicine published by Wiley Periodicals, Inc. on behalf of International Society for Magnetic Resonance in Medicine. This is an open access article under the terms of the Creative Commons Attribution License, which permits use, distribution and reproduction in any medium, provided the original work is properly cited. PMID:28653363
Design and performance evaluation of a high resolution IRI-microPET preclinical scanner
NASA Astrophysics Data System (ADS)
Islami rad, S. Z.; Peyvandi, R. Gholipour; lehdarboni, M. Askari; Ghafari, A. A.
2015-05-01
PET for small animal, IRI-microPET, was designed and built at the NSTRI. The scanner is made of four detectors positioned on a rotating gantry at a distance 50 mm from the center. Each detector consists of a 10×10 crystal matrix of 2×2×10 mm3 directly coupled to a PS-PMT. A position encoding circuit for specific PS-PMT has been designed, built and tested with a PD-MFS-2MS/s-8/14 data acquisition board. After implementing reconstruction algorithms (FBP, MLEM and SART) on sinograms, images quality and system performance were evaluated by energy resolution, timing resolution, spatial resolution, scatter fraction, sensitivity, RMS contrast and SNR parameters. The energy spectra were obtained for the crystals with an energy window of 300-700 keV. The energy resolution in 511 keV averaged over all modules, detectors, and crystals, was 23.5%. A timing resolution of 2.4 ns FWHM obtained by coincidence timing spectrum was measured with crystal LYSO. The radial and tangential resolutions for 18F (1.15-mm inner diameter) at the center of the field of view were 1.81 mm and 1.90 mm, respectively. At a radial offset of 5 mm, the FWHM values were 1.96 and 2.06 mm. The system scatter fraction was 7.1% for the mouse phantom. The sensitivity was measured for different energy windows, leading to a sensitivity of 1.74% at the center of FOV. Also, images quality was evaluated by RMS contrast and SNR factors, and the results show that the reconstructed images by MLEM algorithm have the best RMS contrast, and SNR. The IRI-microPET presents high image resolution, low scatter fraction values and improved SNR for animal studies.
NASA Technical Reports Server (NTRS)
2005-01-01
[figure removed for brevity, see original site] Context image for PIA03026 Southern Clouds This image shows a system of clouds just off the margin of the South Polar cap. Taken during the summer season, these clouds contain both water-ice and dust. Image information: VIS instrument. Latitude 80.2S, Longitude 57.6E. 17 meter/pixel resolution. Note: this THEMIS visual image has not been radiometrically nor geometrically calibrated for this preliminary release. An empirical correction has been performed to remove instrumental effects. A linear shift has been applied in the cross-track and down-track direction to approximate spacecraft and planetary motion. Fully calibrated and geometrically projected images will be released through the Planetary Data System in accordance with Project policies at a later time. NASA's Jet Propulsion Laboratory manages the 2001 Mars Odyssey mission for NASA's Office of Space Science, Washington, D.C. The Thermal Emission Imaging System (THEMIS) was developed by Arizona State University, Tempe, in collaboration with Raytheon Santa Barbara Remote Sensing. The THEMIS investigation is led by Dr. Philip Christensen at Arizona State University. Lockheed Martin Astronautics, Denver, is the prime contractor for the Odyssey project, and developed and built the orbiter. Mission operations are conducted jointly from Lockheed Martin and from JPL, a division of the California Institute of Technology in Pasadena.Forensic Analysis of the Sony Playstation Portable
NASA Astrophysics Data System (ADS)
Conrad, Scott; Rodriguez, Carlos; Marberry, Chris; Craiger, Philip
The Sony PlayStation Portable (PSP) is a popular portable gaming device with features such as wireless Internet access and image, music and movie playback. As with most systems built around a processor and storage, the PSP can be used for purposes other than it was originally intended - legal as well as illegal. This paper discusses the features of the PSP browser and suggests best practices for extracting digital evidence.
ERIC Educational Resources Information Center
Simon, Jason
2013-01-01
The logo controversy was sparked by an article on the "San Jose Mercury News"' website that was promptly picked up by other news outlets and shared across social networks. Under the headline "University of California introduces a modern logo" sat a blurry, low-quality image of the new monogram next to the 145-year-old UC seal.…
Instrument for underwater high-angular resolution volume scattering function measurements
NASA Astrophysics Data System (ADS)
Dueweke, Paul W.; Bolstad, Jay; Leonard, Donald A.; Sweeney, Harold E.; Boyer, Philip A.; Winkler, Erik M.
1997-02-01
A prototype instrument for in situ measurements of the volume scattering function (VSF) and the beam attenuation of water has been built and tested in the EOO laboratory. The intended application of the instrument is the enhancement of Navy operational optical systems for finding and imaging underwater objects such as mines. A description of the apparatus that was built and preliminary laboratory data will be presented. The instrument measures the VSF, (beta) ((theta) ), near the optical axis in both the forward and back directions from approximately 0.2 degrees off axis to approximately 5 degrees in 0.1 degree steps and at side angles of 45 degrees, 90 degrees, and 135 degrees. A diode- pumped, frequency-doubled, Nd:YAG laser provides the 532 nm light. This is the most used wavelength for underwater optical systems. The forward and back scattered light is collected and focused to a plane where scattering angles in the water are mapped onto concentric rings. At this focal plane, a conical reflector compresses the annular optical data onto a line along the cone axis where it is read by a MOS linear image array providing over 500 separate angular measurements. The beam attenuation coefficient, c, is also measured by means of a unique dual path configuration.
A High-Resolution Minimicroscope System for Wireless Real-Time Monitoring.
Wang, Zongjie; Boddeda, Akash; Parker, Benjamin; Samanipour, Roya; Ghosh, Sanjoy; Menard, Frederic; Kim, Keekyoung
2018-07-01
Compact, cost-effective, and high-performance microscope that enables the real-time imaging of cells and lab-on-a-chip devices is highly demanded for cell biology and biomedical engineering. This paper aims to present the design and application of an inexpensive wireless minimicroscope with resolution up to 2592 × 1944 pixels and speed up to 90 f/s. The minimicroscope system was built on a commercial embedded system (Raspberry Pi). We modified a camera module and adopted an inverse dual lens system to obtain the clear field of view and appropriate magnification for tens of micrometer objects. The system was capable of capturing time-lapse images and transferring image data wirelessly. The entire system can be operated wirelessly and cordlessly in a conventional cell culturing incubator. The developed minimicroscope was used to monitor the attachment and proliferation of NIH-3T3 and HEK 293 cells inside an incubator for 50 h. In addition, the minimicroscope was used to monitor a droplet generation process in a microfluidic device. The high-quality images captured by the minimicroscope enabled us an automated analysis of experimental parameters. The successful applications prove the great potential of the developed minimicroscope for monitoring various biological samples and microfluidic devices. This paper presents the design of a high-resolution minimicroscope system that enables the wireless real-time imaging of cells inside the incubator. This system has been verified to be a useful tool to obtain high-quality images and videos for the automated quantitative analysis of biological samples and lab-on-a-chip devices in the long term.
Phoenix Telltale Movie with Clouds, Sol 103
NASA Technical Reports Server (NTRS)
2008-01-01
NASA's Phoenix Mars Lander's telltale catches a breeze as clouds move over the landing site on Sol 103 (Sept. 7, 2008), the 103rd Martian day since landing. Phoenix's Surface Stereo Imager took this series of images during daily telltale monitoring around 3 p.m. local solar time and captured the clouds moving over the landing site. Phoenix can measure wind speed and direction by imaging the telltale, which is about about 10 centimeters (4 inches) tall. The telltale was built by the University of Aarhus, Denmark. The Phoenix Mission is led by the University of Arizona, Tucson, on behalf of NASA. Project management of the mission is by NASA's Jet Propulsion Laboratory, Pasadena, Calif. Spacecraft development is by Lockheed Martin Space Systems, Denver.NASA Astrophysics Data System (ADS)
Wei, Qingyang; Wang, Shi; Ma, Tianyu; Wu, Jing; Liu, Hui; Xu, Tianpeng; Xia, Yan; Fan, Peng; Lyu, Zhenlei; Liu, Yaqiang
2015-06-01
PET, SPECT and CT imaging techniques are widely used in preclinical small animal imaging applications. In this paper, we present a compact small animal PET/SPECT/CT tri-modality system. A dual-functional, shared detector design is implemented which enables PET and SPECT imaging with a same LYSO ring detector. A multi-pinhole collimator is mounted on the system and inserted into the detector ring in SPECT imaging mode. A cone-beam CT consisting of a micro focus X-ray tube and a CMOS detector is implemented. The detailed design and the performance evaluations are reported in this paper. In PET imaging mode, the measured NEMA based spatial resolution is 2.12 mm (FWHM), and the sensitivity at the central field of view (CFOV) is 3.2%. The FOV size is 50 mm (∅)×100 mm (L). The SPECT has a spatial resolution of 1.32 mm (FWHM) and an average sensitivity of 0.031% at the center axial, and a 30 mm (∅)×90 mm (L) FOV. The CT spatial resolution is 8.32 lp/mm @10%MTF, and the contrast discrimination function value is 2.06% with 1.5 mm size cubic box object. In conclusion, a compact, tri-modality PET/SPECT/CT system was successfully built with low cost and high performance.
Maurice, P; Dhombres, F; Blondiaux, E; Friszer, S; Guilbaud, L; Lelong, N; Khoshnood, B; Charlet, J; Perrot, N; Jauniaux, E; Jurkovic, D; Jouannic, J-M
2017-05-01
We have developed a new knowledge base intelligent system for obstetrics and gynecology ultrasound imaging, based on an ontology and a reference image collection. This study evaluates the new system to support accurate annotations of ultrasound images. We have used the early ultrasound diagnosis of ectopic pregnancies as a model clinical issue. The ectopic pregnancy ontology was derived from medical texts (4260 ultrasound reports of ectopic pregnancy from a specialist center in the UK and 2795 Pubmed abstracts indexed with the MeSH term "Pregnancy, Ectopic") and the reference image collection was built on a selection from 106 publications. We conducted a retrospective analysis of the signs in 35 scans of ectopic pregnancy by six observers using the new system. The resulting ectopic pregnancy ontology consisted of 1395 terms, and 80 images were collected for the reference collection. The observers used the knowledge base intelligent system to provide a total of 1486 sign annotations. The precision, recall and F-measure for the annotations were 0.83, 0.62 and 0.71, respectively. The global proportion of agreement was 40.35% 95% CI [38.64-42.05]. The ontology-based intelligent system provides accurate annotations of ultrasound images and suggests that it may benefit non-expert operators. The precision rate is appropriate for accurate input of a computer-based clinical decision support and could be used to support medical imaging diagnosis of complex conditions in obstetrics and gynecology. Copyright © 2017. Published by Elsevier Masson SAS.
Nguyen, Quynh C; Sajjadi, Mehdi; McCullough, Matt; Pham, Minh; Nguyen, Thu T; Yu, Weijun; Meng, Hsien-Wen; Wen, Ming; Li, Feifei; Smith, Ken R; Brunisholz, Kim; Tasdizen, Tolga
2018-03-01
Neighbourhood quality has been connected with an array of health issues, but neighbourhood research has been limited by the lack of methods to characterise large geographical areas. This study uses innovative computer vision methods and a new big data source of street view images to automatically characterise neighbourhood built environments. A total of 430 000 images were obtained using Google's Street View Image API for Salt Lake City, Chicago and Charleston. Convolutional neural networks were used to create indicators of street greenness, crosswalks and building type. We implemented log Poisson regression models to estimate associations between built environment features and individual prevalence of obesity and diabetes in Salt Lake City, controlling for individual-level and zip code-level predisposing characteristics. Computer vision models had an accuracy of 86%-93% compared with manual annotations. Charleston had the highest percentage of green streets (79%), while Chicago had the highest percentage of crosswalks (23%) and commercial buildings/apartments (59%). Built environment characteristics were categorised into tertiles, with the highest tertile serving as the referent group. Individuals living in zip codes with the most green streets, crosswalks and commercial buildings/apartments had relative obesity prevalences that were 25%-28% lower and relative diabetes prevalences that were 12%-18% lower than individuals living in zip codes with the least abundance of these neighbourhood features. Neighbourhood conditions may influence chronic disease outcomes. Google Street View images represent an underused data resource for the construction of built environment features. © Article author(s) (or their employer(s) unless otherwise stated in the text of the article) 2018. All rights reserved. No commercial use is permitted unless otherwise expressly granted.
eHXI: A permanently installed, hard x-ray imager for the National Ignition Facility
Doppner, T.; Bachmann, B.; Albert, F.; ...
2016-06-14
We have designed and built a multi-pinhole imaging system for high energy x-rays (≥ 50 keV) that is permanently installed in the equatorial plane outside of the target chamber at the National Ignition Facility (NIF). It records absolutely-calibrated, time-integrated x-ray images with the same line-of-sight as the multi-channel, spatially integrating hard x-ray detector FFLEX [McDonald et al., Rev. Sci. Instrum. 75 (2004) 3753], having a side view of indirect-drive inertial confinement fusion (ICF) implosion targets. The equatorial hard x-ray imager (eHXI) has recorded images on the majority of ICF implosion experiments since May 2011. Lastly, eHXI provides valuable information onmore » hot electron distribution in hohlraum experiments, target alignment, potential hohlraum drive asymmetries and serves as a long term reference for the FFLEX diagnostics.« less
Network image data bank prototype: the RSI project (Resume de Sortie Images)
NASA Astrophysics Data System (ADS)
Abouchi, Nacer; Jourlin, Michel; Bohbot, Oriana; Faurie, Catherine; Grisel, Richard
1995-02-01
The Hospital Edouard Herriot in Lyon and 3M company, associated with the Electronic Department of Physics Chimics and Electronic Engineering School (CPE), decided in 1993 to begin a study on a project of image network. This project is composed of many practical applications to be checked one by one. The purpose of this paper is to discuss the context, which is kind of small picture archiving and communication system (PACS), to explain the methodology which has been used related to hardware and software implementation, and to give examples of the first results obtained. One of the main interests of the results is the possibility to obtain on the same support, 3M laser imager, a film including images from different modalities and abstract summing up the patient stay in the hospital. The framework used is built around Omnis7 and C++ language on a PC computer.
Endomicroscopy imaging of epithelial structures using tissue autofluorescence
NASA Astrophysics Data System (ADS)
Lin, Bevin; Urayama, Shiro; Saroufeem, Ramez M. G.; Matthews, Dennis L.; Demos, Stavros G.
2011-04-01
We explore autofluorescence endomicroscopy as a potential tool for real-time visualization of epithelial tissue microstructure and organization in a clinical setting. The design parameters are explored using two experimental systems--an Olympus Medical Systems Corp. stand-alone clinical prototype probe, and a custom built bench-top rigid fiber conduit prototype. Both systems entail ultraviolet excitation at 266 nm and/or 325 nm using compact laser sources. Preliminary results using ex vivo animal and human tissue specimens suggest that this technology can be translated toward in vivo application to address the need for real-time histology.
The Belle II imaging Time-of-Propagation (iTOP) detector
NASA Astrophysics Data System (ADS)
Fast, J.; Belle II Barrel Particle Identification Group
2017-12-01
High precision flavor physics measurements are an essential complement to the direct searches for new physics at the LHC ATLAS and CMS experiments. Such measurements will be performed using the upgraded Belle II detector that will take data at the SuperKEKB accelerator. With 40x the luminosity of KEKB, the detector systems must operate efficiently at much higher rates than the original Belle detector. A central element of the upgrade is the barrel particle identification system. Belle II has built and installed an imaging-Time-of-Propagation (iTOP) detector. The iTOP uses quartz optics as Cherenkov radiators. The photons are transported down the quartz bars via total internal reflection with a spherical mirror at the forward end to reflect photons to the backward end where they are imaged onto an array of segmented Micro-Channel Plate Photo-Multiplier Tubes (MCP-PMTs). The system is read out using giga-samples per second waveform sampling Application-Specific Integrated Circuits (ASICs). The combined timing and spatial distribution of the photons for each event are used to determine particle species. This paper provides an overview of the iTOP system.
2001-07-21
This ASTER sub-image covers a 12 x 12 km area in northern Shanxi Province, China, and was acquired January 9, 2001. The low sun angle, and light snow cover highlight a section of the Great Wall, visible as a black line running diagonally through the image from lower left to upper right. The Great Wall is over 2000 years old and was built over a period of 1000 years. Stretching 4500 miles from Korea to the Gobi Desert it was first built to protect China from marauders from the north. This image is located at 40.2 degrees north latitude and 112.8 degrees east longitude. http://photojournal.jpl.nasa.gov/catalog/PIA02669
Lee, R F; Giaquinto, R; Constantinides, C; Souza, S; Weiss, R G; Bottomley, P A
2000-02-01
Despite their proven gains in signal-to-noise ratio and field-of-view for routine clinical MRI, phased-array detection systems are currently unavailable for nuclei other than protons (1H). A broadband phased-array system was designed and built to convert the 1H transmitter signal to the non-1H frequency for excitation and to convert non-1H phased-array MRI signals to the 1H frequency for presentation to the narrowband 1H receivers of a clinical whole-body 1.5 T MRI system. With this system, the scanner operates at the 1H frequency, whereas phased-array MRI occurs at the frequency of the other nucleus. Pulse sequences were developed for direct phased-array sodium (23Na) and phosphorus (31P) MRI of high-energy phosphates using chemical selective imaging, thereby avoiding the complex processing and reconstruction required for phased-array magnetic resonance spectroscopy data. Flexible 4-channel 31P and 23Na phased-arrays were built and the entire system tested in phantom and human studies. The array produced a signal-to-noise ratio improvement of 20% relative to the best-positioned single coil, but gains of 300-400% were realized in many voxels located outside the effective field-of-view of the single coil. Cardiac phosphorus and sodium MRI were obtained in 6-13 min with 16 and 0.5 mL resolution, respectively. Lower resolution human cardiac 23Na MRI were obtained in as little as 4 sec. The system provides a practical approach to realizing the advantages of phased-arrays for nuclei other than 1H, and imaging metabolites directly.
PACS and teleradiology for on-call support of abdominal imaging
NASA Astrophysics Data System (ADS)
Horii, Steven C.; Garra, Brian S.; Mun, Seong K.; Zeman, Robert K.; Levine, Betty A.; Fielding, Robert
1991-07-01
One aspect of the Georgetown image management and communications system (IMACS or PACS) is a built-in capability to support teleradiology. Unlike many dedicated teleradiology systems, the support of this capability as a part of PACS means that any acquired images are remotely accessible, not just those specifically input for transmission. Over the past one and one-half years, two radiologists (SCH, BSG) in the abdominal imaging division of the department of radiology have been accumulating experience with teleradiology for on-call support of emergency abdominal imaging, chiefly in ultrasound. As of the time of this writing, use of the system during on-call (one of these attending radiologists primarily responsible) or back-up call (the attending responsible for the Fellow on primary call) has resulted in a marked reduction in the number of times one of them has to drive to the hospital at night or over the weekend. Approximately 80% of the time, use of the teleradiology system obviates having to go in to review a case. The remainder of the time, the radiologist has to perform a procedure (e.g., abscess drainage) or a scan (e.g., complex Doppler study) himself. This paper reviews the system used for teleradiology, how it is electronically and operationally integrated with the PACS, the clinical benefits and disadvantages of this use, and radiologist and referring physician acceptance.
MOC Image of Phobos with TES Temperature Overlay
NASA Technical Reports Server (NTRS)
1998-01-01
This image of Phobos, the inner and larger of the two moons of Mars, was taken by the Mars Global Surveyor on August 19, 1998. The Thermal Emission Spectrometer (TES) measured the brightness of thermal radiation at the same time the camera acquired this image. By analyzing the brightness, TES scientists could deduce the various fractions of the surface exposed to the Sun and their temperatures. This preliminary analysis shows that the surface temperature, dependent on slope and particle size, varies from a high of +25o F (-4o C) on the most illuminated slopes to -170o F (-112o C) in shadows. This large difference, and the fact that such differences can be found in close proximity, adds support to the notion that the surface of Phobos is covered by very small particles.
Malin Space Science Systems, Inc. and the California Institute of Technology built the MOC using spare hardware from the Mars Observer mission. MSSS operates the camera from its facilities in San Diego, CA. The Thermal Emission Spectrometer is operated by Arizona State University and was built by Raytheon Santa Barbara Remote Sensing. The Jet Propulsion Laboratory's Mars Surveyor Operations Project operates the Mars Global Surveyor spacecraft with its industrial partner, Lockheed Martin Astronautics, from facilities in Pasadena, CA and Denver, CO.NASA Technical Reports Server (NTRS)
Conrad, A. R.; Lupton, W. F.
1992-01-01
Each Keck instrument presents a consistent software view to the user interface programmer. The view consists of a small library of functions, which are identical for all instruments, and a large set of keywords, that vary from instrument to instrument. All knowledge of the underlying task structure is hidden from the application programmer by the keyword layer. Image capture software uses the same function library to collect data for the image header. Because the image capture software and the instrument control software are built on top of the same keyword layer, a given observation can be 'replayed' by extracting keyword-value pairs from the image header and passing them back to the control system. The keyword layer features non-blocking as well as blocking I/O. A non-blocking keyword write operation (such as setting a filter position) specifies a callback to be invoked when the operation is complete. A non-blocking keyword read operation specifies a callback to be invoked whenever the keyword changes state. The keyword-callback style meshes well with the widget-callback style commonly used in X window programs. The first keyword library was built for the two Keck optical instruments. More recently, keyword libraries have been developed for the infrared instruments and for telescope control. Although the underlying mechanisms used for inter-process communication by each of these systems vary widely (Lick MUSIC, Sun RPC, and direct socket I/O, respectively), a basic user interface has been written that can be used with any of these systems. Since the keyword libraries are bound to user interface programs dynamically at run time, only a single set of user interface executables is needed. For example, the same program, 'xshow', can be used to display continuously the telescope's position, the time left in an instrument's exposure, or both values simultaneously. Less generic tools that operate on specific keywords, for example an X display that controls optical instrument exposures, have also been written using the keyword layer.
Garden City Vein Complex on Lower Mount Sharp, Mars
2015-11-11
Prominent mineral veins at the "Garden City" site examined by NASA's Curiosity Mars rover vary in thickness and brightness, as seen in this image from Curiosity's Mast Camera (Mastcam). The image covers and area roughly 2 feet (60 centimeters) across. Types of vein material evident in the area include: 1) thin, dark-toned fracture filling material; 2) thick, dark-toned vein material in large fractures; 3) light-toned vein material, which was deposited last. Figure 1 includes annotations identifying each of those three major kinds and a scale bar indicating 10 centimeters (3.9 inches). Researchers used the Mastcam and other instruments on Curiosity in March and April 2015 to study the structure and composition of mineral veins at Garden City, for information about fluids that deposited minerals in fractured rock there. Malin Space Science Systems, San Diego, built and operates Curiosity's Mastcam. NASA's Jet Propulsion Laboratory, a division of the California Institute of Technology, Pasadena, built the rover and manages the project for NASA's Science Mission Directorate, Washington. http://photojournal.jpl.nasa.gov/catalog/PIA19922
Segments on Western Rim of Endeavour Crater, Mars
2017-04-19
This orbital image of the western rim of Mars' Endeavour Crater covers an area about 5 miles (8 kilometers) east-west by about 9 miles (14 kilometers) north-south and indicates the names of some of the raised segments of the rim. NASA's Mars Exploration Rover Opportunity arrived at Endeavour in 2011 after exploring smaller craters to the northwest during its first six years on Mars. It initially explored the "Cape York" segment, then headed south. It reached the northern end of "Cape Tribulation" in late 2014 and the southern tip of that segment in April 2017. A key destination in the "Cape Byron" segment is "Perseverance Valley," where the rover team plans to investigate whether the valley was carved by water, wind or a debris flow initiated by water. This image is from the Context Camera on NASA's Mars Reconnaissance Orbiter. Malin Space Science Systems, San Diego, California, built and operates that camera. NASA's Jet Propulsion Laboratory, a division of Caltech in Pasadena, California, built and operates Opportunity. https://photojournal.jpl.nasa.gov/catalog/PIA21490
Toward a workbench for rodent brain image data: systems architecture and design.
Moene, Ivar A; Subramaniam, Shankar; Darin, Dmitri; Leergaard, Trygve B; Bjaalie, Jan G
2007-01-01
We present a novel system for storing and manipulating microscopic images from sections through the brain and higher-level data extracted from such images. The system is designed and built on a three-tier paradigm and provides the research community with a web-based interface for facile use in neuroscience research. The Oracle relational database management system provides the ability to store a variety of objects relevant to the images and provides the framework for complex querying of data stored in the system. Further, the suite of applications intimately tied into the infrastructure in the application layer provide the user the ability not only to query and visualize the data, but also to perform analysis operations based on the tools embedded into the system. The presentation layer uses extant protocols of the modern web browser and this provides ease of use of the system. The present release, named Functional Anatomy of the Cerebro-Cerebellar System (FACCS), available through The Rodent Brain Workbench (http:// rbwb.org/), is targeted at the functional anatomy of the cerebro-cerebellar system in rats, and holds axonal tracing data from these projections. The system is extensible to other circuits and projections and to other categories of image data and provides a unique environment for analysis of rodent brain maps in the context of anatomical data. The FACCS application assumes standard animal brain atlas models and can be extended to future models. The system is available both for interactive use from a remote web-browser client as well as for download to a local server machine.
Progressive compressive imager
NASA Astrophysics Data System (ADS)
Evladov, Sergei; Levi, Ofer; Stern, Adrian
2012-06-01
We have designed and built a working automatic progressive sampling imaging system based on the vector sensor concept, which utilizes a unique sampling scheme of Radon projections. This sampling scheme makes it possible to progressively add information resulting in tradeoff between compression and the quality of reconstruction. The uniqueness of our sampling is that in any moment of the acquisition process the reconstruction can produce a reasonable version of the image. The advantage of the gradual addition of the samples is seen when the sparsity rate of the object is unknown, and thus the number of needed measurements. We have developed the iterative algorithm OSO (Ordered Sets Optimization) which employs our sampling scheme for creation of nearly uniform distributed sets of samples, which allows the reconstruction of Mega-Pixel images. We present the good quality reconstruction from compressed data ratios of 1:20.
The Real Time Correction of Stereoscopic Images: From the Serial to a Parallel Treatment
NASA Astrophysics Data System (ADS)
Irki, Zohir; Devy, Michel; Achour, Karim; Azzaz, Mohamed Salah
2008-06-01
The correction of the stereoscopic images is a task which consists in replacing acquired images by other images having the same properties but which are simpler to use in the other stages of stereovision. The use of the pre-calculated tables, built during an off line calibration step, made it possible to carry out the off line stereoscopic images rectification. An improvement of the built tables made it possible to carry out the real time rectification. In this paper, we describe an improvement of the real time correction approach so it can be exploited for a possible implementation on an FPGA component. This improvement holds in account the real time aspect of the correction and the available resources that can offer the FPGA Type Stratix 1S40F780C5.
A computational approach to real-time image processing for serial time-encoded amplified microscopy
NASA Astrophysics Data System (ADS)
Oikawa, Minoru; Hiyama, Daisuke; Hirayama, Ryuji; Hasegawa, Satoki; Endo, Yutaka; Sugie, Takahisa; Tsumura, Norimichi; Kuroshima, Mai; Maki, Masanori; Okada, Genki; Lei, Cheng; Ozeki, Yasuyuki; Goda, Keisuke; Shimobaba, Tomoyoshi
2016-03-01
High-speed imaging is an indispensable technique, particularly for identifying or analyzing fast-moving objects. The serial time-encoded amplified microscopy (STEAM) technique was proposed to enable us to capture images with a frame rate 1,000 times faster than using conventional methods such as CCD (charge-coupled device) cameras. The application of this high-speed STEAM imaging technique to a real-time system, such as flow cytometry for a cell-sorting system, requires successively processing a large number of captured images with high throughput in real time. We are now developing a high-speed flow cytometer system including a STEAM camera. In this paper, we describe our approach to processing these large amounts of image data in real time. We use an analog-to-digital converter that has up to 7.0G samples/s and 8-bit resolution for capturing the output voltage signal that involves grayscale images from the STEAM camera. Therefore the direct data output from the STEAM camera generates 7.0G byte/s continuously. We provided a field-programmable gate array (FPGA) device as a digital signal pre-processor for image reconstruction and finding objects in a microfluidic channel with high data rates in real time. We also utilized graphics processing unit (GPU) devices for accelerating the calculation speed of identification of the reconstructed images. We built our prototype system, which including a STEAM camera, a FPGA device and a GPU device, and evaluated its performance in real-time identification of small particles (beads), as virtual biological cells, owing through a microfluidic channel.
Flight Performance of the HEROES Solar Aspect System
NASA Astrophysics Data System (ADS)
Shih, Albert Y.; Christe, Steven; Rodriguez, Marcello; Gregory, Kyle; Cramer, Alexander; Edgerton, Melissa; Gaskin, Jessica; O'Connor, Brian; Sobey, Alexander
2014-06-01
Hard X-ray (HXR) observations of solar flares reveal the signatures of energetic electrons, and HXR images with high dynamic range and high sensitivity can distinguish between where electrons are accelerated and where they stop. Furthermore, high-sensitivity HXR measurements may be able to detect the presence of electron acceleration in the non-flaring corona. The High Energy Replicated Optics to Explore the Sun (HEROES) balloon mission added the capability of solar observations to an existing astrophysics balloon payload, HERO, which used grazing-incidence optics for direct HXR imaging. The HEROES Solar Aspect System (SAS) was developed and built to provide pointing knowledge during solar observations to better than the ~20 arcsec FWHM angular resolution of the HXR instrument. The SAS consists of two separate systems: the Pitch-Yaw aspect System (PYAS) and the Roll Aspect System (RAS). The PYAS compares the position of an optical image of the Sun relative to precise fiducials to determine the pitch and yaw pointing offsets from the desired solar target. The RAS images the Earth's horizon in opposite directions simultaneously to determine the roll of the gondola. HEROES launched in September 2013 from Fort Sumner, New Mexico, and had a successful one-day flight. We present the detailed analysis of the performance of the SAS for that flight.
Laser Based Color Film Recorder System With GaAs Microlaser
NASA Astrophysics Data System (ADS)
Difrancesco, David J.
1989-07-01
In 1984 Pixar's research and development group built and applied to the motion-picture arts at Lucasfilm's ILM facility a three color laser based film scanner/recorder system. The digital film printer is capable of reading and writing 35mm film formats on a variety of film stocks. The system has been used in award-winning special-effects work, and has been operated in a normal production environment since that time. The primary objective was to develop a full color high resolution system, free from scan artifacts, enabling traditionally photographed motion-picture film to be inter-cut with digital raster image photography. Its use is applied to the art of blue-screen traveling-matte cinematography for motion pic-ture special effects. The system was designed using the Pixar Image Computer and conventional gas laser technology as the illumination source. This paper will discuss recent experimental work in the application of GaAs microlaser technology to a digital film printing system of the future.
Development and optimization of hardware for delta relaxation enhanced MRI.
Harris, Chad T; Handler, William B; Araya, Yonathan; Martínez-Santiesteban, Francisco; Alford, Jamu K; Dalrymple, Brian; Van Sas, Frank; Chronik, Blaine A; Scholl, Timothy J
2014-10-01
Delta relaxation enhanced magnetic resonance (dreMR) imaging requires an auxiliary B0 electromagnet capable of shifting the main magnetic field within a clinical 1.5 Tesla (T) MR system. In this work, the main causes of interaction between an actively shielded, insertable resistive B0 electromagnet and a 1.5T superconducting system are systematically identified and mitigated. The effects of nonideal fabrication of the field-shifting magnet are taken into consideration through careful measurement during winding and improved accuracy in the design of the associated active shield. The shielding performance of the resultant electromagnet is compared against a previously built system in which the shield design was based on an ideal primary coil model. Hardware and software approaches implemented to eliminate residual image artifacts are presented in detail. The eddy currents produced by the newly constructed dreMR system are shown to have a significantly smaller "long-time-constant" component, consistent with the hypothesis that less energy is deposited into the cryostat of the MR system. With active compensation, the dreMR imaging system is capable of 0.22T field shifts within a clinical 1.5T MRI with no significant residual eddy-current fields. Copyright © 2013 Wiley Periodicals, Inc.
Simultaneous PET and Multispectral 3-Dimensional Fluorescence Optical Tomography Imaging System
Li, Changqing; Yang, Yongfeng; Mitchell, Gregory S.; Cherry, Simon R.
2015-01-01
Integrated PET and 3-dimensional (3D) fluorescence optical tomography (FOT) imaging has unique and attractive features for in vivo molecular imaging applications. We have designed, built, and evaluated a simultaneous PET and 3D FOT system. The design of the FOT system is compatible with many existing small-animal PET scanners. Methods The 3D FOT system comprises a novel conical mirror that is used to view the whole-body surface of a mouse with an electron-multiplying charge-coupled device camera when a collimated laser beam is projected on the mouse to stimulate fluorescence. The diffusion equation was used to model the propagation of optical photons inside the mouse body, and 3D fluorescence images were reconstructed iteratively from the fluorescence intensity measurements measured from the surface of the mouse. Insertion of the conical mirror into the gantry of a small-animal PET scanner allowed simultaneous PET and 3D FOT imaging. Results The mutual interactions between PET and 3D FOT were evaluated experimentally. PET has negligible effects on 3D FOT performance. The inserted conical mirror introduces a reduction in the sensitivity and noise-equivalent count rate of the PET system and increases the scatter fraction. PET–FOT phantom experiments were performed. An in vivo experiment using both PET and FOT was also performed. Conclusion Phantom and in vivo experiments demonstrate the feasibility of simultaneous PET and 3D FOT imaging. The first in vivo simultaneous PET–FOT results are reported. PMID:21810591
Pagoulatos, N; Edwards, W S; Haynor, D R; Kim, Y
1999-12-01
The use of stereotactic systems has been one of the main approaches for image-based guidance of the surgical tool within the brain. The main limitation of stereotactic systems is that they are based on preoperative images that might become outdated and invalid during the course of surgery. Ultrasound (US) is considered the most practical and cost-effective intraoperative imaging modality, but US images inherently have a low signal-to-noise ratio. Integrating intraoperative US with stereotactic systems has recently been attempted. In this paper, we present a new system for interactively registering two-dimensional US and three-dimensional magnetic resonance (MR) images. This registration is based on tracking the US probe with a dc magnetic position sensor. We have performed an extensive analysis of the errors of our system by using a custom-built phantom. The registration error between the MR and the position sensor space was found to have a mean value of 1.78 mm and a standard deviation of 0.18 mm. The registration error between US and MR space was dependent on the distance of the target point from the US probe face. For a 3.5-MHz phased one-dimensional array transducer and a depth of 6 cm, the mean value of the registration error was 2.00 mm and the standard deviation was 0.75 mm. The registered MR images were reconstructed using either zeroth-order or first-order interpolation. The ease of use and the interactive nature of our system (approximately 6.5 frames/s for 344 x 310 images and first-order interpolation on a Pentium II 450 MHz) demonstrates its potential to be used in the operating room.
Three-dimensional ophthalmic optical coherence tomography with a refraction correction algorithm
NASA Astrophysics Data System (ADS)
Zawadzki, Robert J.; Leisser, Christoph; Leitgeb, Rainer; Pircher, Michael; Fercher, Adolf F.
2003-10-01
We built an optical coherence tomography (OCT) system with a rapid scanning optical delay (RSOD) line, which allows probing full axial eye length. The system produces Three-dimensional (3D) data sets that are used to generate 3D tomograms of the model eye. The raw tomographic data were processed by an algorithm, which is based on Snell"s law to correct the interface positions. The Zernike polynomials representation of the interfaces allows quantitative wave aberration measurements. 3D images of our results are presented to illustrate the capabilities of the system and the algorithm performance. The system allows us to measure intra-ocular distances.
Report of AAPM Task Group 162: Software for planar image quality metrology.
Samei, Ehsan; Ikejimba, Lynda C; Harrawood, Brian P; Rong, John; Cunningham, Ian A; Flynn, Michael J
2018-02-01
The AAPM Task Group 162 aimed to provide a standardized approach for the assessment of image quality in planar imaging systems. This report offers a description of the approach as well as the details of the resultant software bundle to measure detective quantum efficiency (DQE) as well as its basis components and derivatives. The methodology and the associated software include the characterization of the noise power spectrum (NPS) from planar images acquired under specific acquisition conditions, modulation transfer function (MTF) using an edge test object, the DQE, and effective DQE (eDQE). First, a methodological framework is provided to highlight the theoretical basis of the work. Then, a step-by-step guide is included to assist in proper execution of each component of the code. Lastly, an evaluation of the method is included to validate its accuracy against model-based and experimental data. The code was built using a Macintosh OSX operating system. The software package contains all the source codes to permit an experienced user to build the suite on a Linux or other *nix type system. The package further includes manuals and sample images and scripts to demonstrate use of the software for new users. The results of the code are in close alignment with theoretical expectations and published results of experimental data. The methodology and the software package offered in AAPM TG162 can be used as baseline for characterization of inherent image quality attributes of planar imaging systems. © 2017 American Association of Physicists in Medicine.
NASA Technical Reports Server (NTRS)
2005-01-01
[figure removed for brevity, see original site] Context image for PIA03025 Channeled Winds This low resolution VIS image shows a large portion of etched terrain near the south pole of Mars. Image information: VIS instrument. Latitude 10S, Longitude 37.2E. 18 meter/pixel resolution. Note: this THEMIS visual image has not been radiometrically nor geometrically calibrated for this preliminary release. An empirical correction has been performed to remove instrumental effects. A linear shift has been applied in the cross-track and down-track direction to approximate spacecraft and planetary motion. Fully calibrated and geometrically projected images will be released through the Planetary Data System in accordance with Project policies at a later time. NASA's Jet Propulsion Laboratory manages the 2001 Mars Odyssey mission for NASA's Office of Space Science, Washington, D.C. The Thermal Emission Imaging System (THEMIS) was developed by Arizona State University, Tempe, in collaboration with Raytheon Santa Barbara Remote Sensing. The THEMIS investigation is led by Dr. Philip Christensen at Arizona State University. Lockheed Martin Astronautics, Denver, is the prime contractor for the Odyssey project, and developed and built the orbiter. Mission operations are conducted jointly from Lockheed Martin and from JPL, a division of the California Institute of Technology in Pasadena.NASA Technical Reports Server (NTRS)
2005-01-01
[figure removed for brevity, see original site] Context image for PIA03638 Polar Textures This image illustrates the variety of textures that appear in the south polar region during late summer. Image information: VIS instrument. Latitude 80.5S, Longitude 57.9E. 17 meter/pixel resolution. Note: this THEMIS visual image has not been radiometrically nor geometrically calibrated for this preliminary release. An empirical correction has been performed to remove instrumental effects. A linear shift has been applied in the cross-track and down-track direction to approximate spacecraft and planetary motion. Fully calibrated and geometrically projected images will be released through the Planetary Data System in accordance with Project policies at a later time. NASA's Jet Propulsion Laboratory manages the 2001 Mars Odyssey mission for NASA's Office of Space Science, Washington, D.C. The Thermal Emission Imaging System (THEMIS) was developed by Arizona State University, Tempe, in collaboration with Raytheon Santa Barbara Remote Sensing. The THEMIS investigation is led by Dr. Philip Christensen at Arizona State University. Lockheed Martin Astronautics, Denver, is the prime contractor for the Odyssey project, and developed and built the orbiter. Mission operations are conducted jointly from Lockheed Martin and from JPL, a division of the California Institute of Technology in Pasadena.Chen, Chi-Jim; Pai, Tun-Wen; Cheng, Mox
2015-01-01
A sweeping fingerprint sensor converts fingerprints on a row by row basis through image reconstruction techniques. However, a built fingerprint image might appear to be truncated and distorted when the finger was swept across a fingerprint sensor at a non-linear speed. If the truncated fingerprint images were enrolled as reference targets and collected by any automated fingerprint identification system (AFIS), successful prediction rates for fingerprint matching applications would be decreased significantly. In this paper, a novel and effective methodology with low time computational complexity was developed for detecting truncated fingerprints in a real time manner. Several filtering rules were implemented to validate existences of truncated fingerprints. In addition, a machine learning method of supported vector machine (SVM), based on the principle of structural risk minimization, was applied to reject pseudo truncated fingerprints containing similar characteristics of truncated ones. The experimental result has shown that an accuracy rate of 90.7% was achieved by successfully identifying truncated fingerprint images from testing images before AFIS enrollment procedures. The proposed effective and efficient methodology can be extensively applied to all existing fingerprint matching systems as a preliminary quality control prior to construction of fingerprint templates. PMID:25835186
Ruggeri, Marco; Major, James C.; McKeown, Craig; Knighton, Robert W.; Puliafito, Carmen A.
2010-01-01
Purpose. To reveal three-dimensional (3-D) information about the retinal structures of birds of prey in vivo. Methods. An ultra-high resolution spectral-domain optical coherence tomography (SD-OCT) system was built for in vivo imaging of retinas of birds of prey. The calibrated imaging depth and axial resolution of the system were 3.1 mm and 2.8 μm (in tissue), respectively. 3-D segmentation was performed for calculation of the retinal nerve fiber layer (RNFL) map. Results. High-resolution OCT images were obtained of the retinas of four species of birds of prey: two diurnal hawks (Buteo platypterus and Buteo brachyurus) and two nocturnal owls (Bubo virginianus and Strix varia). These images showed the detailed retinal anatomy, including the retinal layers and the structure of the deep and shallow foveae. The calculated thickness map showed the RNFL distribution. Traumatic injury to one bird's retina was also successfully imaged. Conclusions. Ultra-high resolution SD-OCT provides unprecedented high-quality 2-D and 3-D in vivo visualization of the retinal structures of birds of prey. SD-OCT is a powerful imaging tool for vision research in birds of prey. PMID:20554605
KENIS: a high-performance thermal imager developed using the OSPREY IR detector
NASA Astrophysics Data System (ADS)
Goss, Tristan M.; Baker, Ian M.
2000-07-01
`KENIS', a complete, high performance, compact and lightweight thermal imager, is built around the `OSPREY' infrared detector from BAE systems Infrared Ltd. The `OSPREY' detector uses a 384 X 288 element CMT array with a 20 micrometers pixel size and cooled to 120 K. The relatively small pixel size results in very compact cryogenics and optics, and the relatively high operating temperature provides fast start-up time, low power consumption and long operating life. Requiring single input supply voltage and consuming less than 30 watts of power, the thermal imager generates both analogue and digital format outputs. The `KENIS' lens assembly features a near diffraction limited dual field-of-view optical system that has been designed to be athermalized and switches between fields in less than one second. The `OSPREY' detector produces near background limited performance with few defects and has special, pixel level circuitry to eliminate crosstalk and blooming effects. This, together with signal processing based on an effective two-point fixed pattern noise correction algorithm, results in high quality imagery and a thermal imager that is suitable for most traditional thermal imaging applications. This paper describes the rationale used in the development of the `KENIS' thermal imager, and highlights the potential performance benefits to the user's system, primarily gained by selecting the `OSPREY' infra-red detector within the core of the thermal imager.
NASA Technical Reports Server (NTRS)
2005-01-01
[figure removed for brevity, see original site] Context image for PIA03285 Ganges Features This image shows part of Ganges Chasma. Several landslides occur at the top of the image, while dunes and canyon floor deposits are visible at the bottom of the image. Image information: VIS instrument. Latitude -6.8N, Longitude 312.2E. 17 meter/pixel resolution. Note: this THEMIS visual image has not been radiometrically nor geometrically calibrated for this preliminary release. An empirical correction has been performed to remove instrumental effects. A linear shift has been applied in the cross-track and down-track direction to approximate spacecraft and planetary motion. Fully calibrated and geometrically projected images will be released through the Planetary Data System in accordance with Project policies at a later time. NASA's Jet Propulsion Laboratory manages the 2001 Mars Odyssey mission for NASA's Office of Space Science, Washington, D.C. The Thermal Emission Imaging System (THEMIS) was developed by Arizona State University, Tempe, in collaboration with Raytheon Santa Barbara Remote Sensing. The THEMIS investigation is led by Dr. Philip Christensen at Arizona State University. Lockheed Martin Astronautics, Denver, is the prime contractor for the Odyssey project, and developed and built the orbiter. Mission operations are conducted jointly from Lockheed Martin and from JPL, a division of the California Institute of Technology in Pasadena.The application of digital image plane holography technology to identify Chinese herbal medicine
NASA Astrophysics Data System (ADS)
Wang, Huaying; Guo, Zhongjia; Liao, Wei; Zhang, Zhihui
2012-03-01
In this paper, the imaging technology of digital image plane holography to identify the Chinese herbal medicine is studied. The optical experiment system of digital image plane holography which is the special case of pre-magnification digital holography was built. In the record system, one is an object light by using plane waves which illuminates the object, and the other one is recording hologram by using spherical light wave as reference light. There is a Micro objective lens behind the object. The second phase factor which caus ed by the Micro objective lens can be eliminated by choosing the proper position of the reference point source when digital image plane holography is recorded by spherical light. In this experiment, we use the Lygodium cells and Onion cells as the object. The experiment results with Lygodium cells and Onion cells show that digital image plane holography avoid the process of finding recording distance by using auto-focusing approach, and the phase information of the object can be reconstructed more accurately. The digital image plane holography is applied to the microscopic imaging of cells more effectively, and it is suit to apply for the identify of Chinese Herbal Medicine. And it promotes the application of digital holographic in practice.
ART-Ada design project, phase 2
NASA Technical Reports Server (NTRS)
Lee, S. Daniel; Allen, Bradley P.
1990-01-01
Interest in deploying expert systems in Ada has increased. An Ada based expert system tool is described called ART-Ada, which was built to support research into the language and methodological issues of expert systems in Ada. ART-Ada allows applications of an existing expert system tool called ART-IM (Automated Reasoning Tool for Information Management) to be deployed in various Ada environments. ART-IM, a C-based expert system tool, is used to generate Ada source code which is compiled and linked with an Ada based inference engine to produce an Ada executable image. ART-Ada is being used to implement several expert systems for NASA's Space Station Freedom Program and the U.S. Air Force.
An Automatic Image Processing Workflow for Daily Magnetic Resonance Imaging Quality Assurance.
Peltonen, Juha I; Mäkelä, Teemu; Sofiev, Alexey; Salli, Eero
2017-04-01
The performance of magnetic resonance imaging (MRI) equipment is typically monitored with a quality assurance (QA) program. The QA program includes various tests performed at regular intervals. Users may execute specific tests, e.g., daily, weekly, or monthly. The exact interval of these measurements varies according to the department policies, machine setup and usage, manufacturer's recommendations, and available resources. In our experience, a single image acquired before the first patient of the day offers a low effort and effective system check. When this daily QA check is repeated with identical imaging parameters and phantom setup, the data can be used to derive various time series of the scanner performance. However, daily QA with manual processing can quickly become laborious in a multi-scanner environment. Fully automated image analysis and results output can positively impact the QA process by decreasing reaction time, improving repeatability, and by offering novel performance evaluation methods. In this study, we have developed a daily MRI QA workflow that can measure multiple scanner performance parameters with minimal manual labor required. The daily QA system is built around a phantom image taken by the radiographers at the beginning of day. The image is acquired with a consistent phantom setup and standardized imaging parameters. Recorded parameters are processed into graphs available to everyone involved in the MRI QA process via a web-based interface. The presented automatic MRI QA system provides an efficient tool for following the short- and long-term stability of MRI scanners.
Documet, Jorge; Le, Anh; Liu, Brent; Chiu, John; Huang, H K
2010-05-01
This paper presents the concept of bridging the gap between diagnostic images and image-assisted surgical treatment through the development of a one-stop multimedia electronic patient record (ePR) system that manages and distributes the real-time multimodality imaging and informatics data that assists the surgeon during all clinical phases of the operation from planning Intra-Op to post-care follow-up. We present the concept of this multimedia ePR for surgery by first focusing on image-assisted minimally invasive spinal surgery as a clinical application. Three clinical phases of minimally invasive spinal surgery workflow in Pre-Op, Intra-Op, and Post-Op are discussed. The ePR architecture was developed based on the three-phased workflow, which includes the Pre-Op, Intra-Op, and Post-Op modules and four components comprising of the input integration unit, fault-tolerant gateway server, fault-tolerant ePR server, and the visualization and display. A prototype was built and deployed to a minimally invasive spinal surgery clinical site with user training and support for daily use. A step-by-step approach was introduced to develop a multimedia ePR system for imaging-assisted minimally invasive spinal surgery that includes images, clinical forms, waveforms, and textual data for planning the surgery, two real-time imaging techniques (digital fluoroscopic, DF) and endoscope video images (Endo), and more than half a dozen live vital signs of the patient during surgery. Clinical implementation experiences and challenges were also discussed.
A Planar Two-Dimensional Superconducting Bolometer Array for the Green Bank Telescope
NASA Technical Reports Server (NTRS)
Benford, Dominic; Staguhn, Johannes G.; Chervenak, James A.; Chen, Tina C.; Moseley, S. Harvey; Wollack, Edward J.; Devlin, Mark J.; Dicker, Simon R.; Supanich, Mark
2004-01-01
In order to provide high sensitivity rapid imaging at 3.3mm (90GHz) for the Green Bank Telescope - the world's largest steerable aperture - a camera is being built by the University of Pennsylvania, NASA/GSFC, and NRAO. The heart of this camera is an 8x8 close-packed, Nyquist-sampled detector array. We have designed and are fabricating a functional superconducting bolometer array system using a monolithic planar architecture. Read out by SQUID multiplexers, the superconducting transition edge sensors will provide fast, linear, sensitive response for high performance imaging. This will provide the first ever superconducting bolometer array on a facility instrument.
Zooglider - an Autonomous Vehicle for Optical and Acoustic Sensing of Marine Zooplankton
NASA Astrophysics Data System (ADS)
Ohman, M. D.; Davis, R. E.; Sherman, J. T.; Grindley, K.; Whitmore, B. M.
2016-02-01
We will present results from early sea trials of the Zooglider, an autonomous zooplankton glider designed and built by the Instrument Development Group at Scripps. The Zooglider is built upon a modified Spray glider and includes a low power camera with telecentric lens and a custom dual frequency sonar (200/1000 kHz). The imaging system quantifies zooplankton as they flow through a sampling tunnel within a well-defined sampling volume. The maximum operating depth is 500 m. Other sensors include a pumped CTD and Chl-a fluorometer. The Zooglider permits in situ measurements of mesozooplankton distributions and three dimensional orientation in relation to other biotic and physical properties of the ocean water column. Zooglider development is supported by the Gordon and Betty Moore Foundation.
Subsidence Evaluation of High-Speed Railway in Shenyang Based on Time-Series Insar
NASA Astrophysics Data System (ADS)
Zhang, Yun; Wei, Lianhuan; Li, Jiayu; Liu, Shanjun; Mao, Yachun; Wu, Lixin
2018-04-01
More and more high-speed railway are under construction in China. The slow settlement along high-speed railway tracks and newly-built stations would lead to inhomogeneous deformation of local area, and the accumulation may be a threat to the safe operation of high-speed rail system. In this paper, surface deformation of the newly-built high-speed railway station as well as the railway lines in Shenyang region will be retrieved by time series InSAR analysis using multi-orbit COSMO-SkyMed images. This paper focuses on the non-uniform subsidence caused by the changing of local environment along the railway. The accuracy of the settlement results can be verified by cross validation of the results obtained from two different orbits during the same period.
EPICS-based control and data acquisition for the APS slope profiler (Conference Presentation)
NASA Astrophysics Data System (ADS)
Sullivan, Joseph; Assoufid, Lahsen; Qian, Jun; Jemian, Peter R.; Mooney, Tim; Rivers, Mark L.; Goetze, Kurt; Sluiter, Ronald L.; Lang, Keenan
2016-09-01
The motion control, data acquisition and analysis system for APS Slope Measuring Profiler was implemented using the Experimental Physics and Industrial Control System (EPICS). EPICS was designed as a framework with software tools and applications that provide a software infrastructure used in building distributed control systems to operate devices such as particle accelerators, large experiments and major telescopes. EPICS was chosen to implement the APS Slope Measuring Profiler because it is also applicable to single purpose systems. The control and data handling capability available in the EPICS framework provides the basic functionality needed for high precision X-ray mirror measurement. Those built in capabilities include hardware integration of high-performance motion control systems (3-axis gantry and tip-tilt stages), mirror measurement devices (autocollimator, laser spot camera) and temperature sensors. Scanning the mirror and taking measurements was accomplished with an EPICS feature (the sscan record) which synchronizes motor positioning with measurement triggers and data storage. Various mirror scanning modes were automatically configured using EPICS built-in scripting. EPICS tools also provide low-level image processing (areaDetector). Operation screens were created using EPICS-aware GUI screen development tools.
NASA Astrophysics Data System (ADS)
Sun, Yi; You, Sixian; Tu, Haohua; Spillman, Darold R.; Marjanovic, Marina; Chaney, Eric J.; Liu, George Z.; Ray, Partha S.; Higham, Anna; Boppart, Stephen A.
2017-02-01
Label-free multi-photon imaging has been a powerful tool for studying tissue microstructures and biochemical distributions, particularly for investigating tumors and their microenvironments. However, it remains challenging for traditional bench-top multi-photon microscope systems to conduct ex vivo tumor tissue imaging in the operating room due to their bulky setups and laser sources. In this study, we designed, built, and clinically demonstrated a portable multi-modal nonlinear label-free microscope system that combined four modalities, including two- and three- photon fluorescence for studying the distributions of FAD and NADH, and second and third harmonic generation, respectively, for collagen fiber structures and the distribution of micro-vesicles found in tumors and the microenvironment. Optical realignments and switching between modalities were motorized for more rapid and efficient imaging and for a light-tight enclosure, reducing ambient light noise to only 5% within the brightly lit operating room. Using up to 20 mW of laser power after a 20x objective, this system can acquire multi-modal sets of images over 600 μm × 600 μm at an acquisition rate of 60 seconds using galvo-mirror scanning. This portable microscope system was demonstrated in the operating room for imaging fresh, resected, unstained breast tissue specimens, and for assessing tumor margins and the tumor microenvironment. This real-time label-free nonlinear imaging system has the potential to uniquely characterize breast cancer margins and the microenvironment of tumors to intraoperatively identify structural, functional, and molecular changes that could indicate the aggressiveness of the tumor.
Qian, Zeng-Hui; Feng, Xu; Li, Yang; Tang, Ke
2018-01-01
Studying the three-dimensional (3D) anatomy of the cavernous sinus is essential for treating lesions in this region with skull base surgeries. Cadaver dissection is a conventional method that has insurmountable flaws with regard to understanding spatial anatomy. The authors' research aimed to build an image model of the cavernous sinus region in a virtual reality system to precisely, individually and objectively elucidate the complete and local stereo-anatomy. Computed tomography and magnetic resonance imaging scans were performed on 5 adult cadaver heads. Latex mixed with contrast agent was injected into the arterial system and then into the venous system. Computed tomography scans were performed again following the 2 injections. Magnetic resonance imaging scans were performed again after the cranial nerves were exposed. Image data were input into a virtual reality system to establish a model of the cavernous sinus. Observation results of the image models were compared with those of the cadaver heads. Visualization of the cavernous sinus region models built using the virtual reality system was good for all the cadavers. High resolutions were achieved for the images of different tissues. The observed results were consistent with those of the cadaver head. The spatial architecture and modality of the cavernous sinus were clearly displayed in the 3D model by rotating the model and conveniently changing its transparency. A 3D virtual reality model of the cavernous sinus region is helpful for globally and objectively understanding anatomy. The observation procedure was accurate, convenient, noninvasive, and time and specimen saving.
ERIC Educational Resources Information Center
Dillon, Robert
2012-01-01
Images can be powerful tools for change, but without compelling images of the future of education, everyone will be forced to use the images of the past. The problem is that schools can't be built on old images: they must reflect current best practices that infuse technology, relationships, background knowledge, culturally responsive texts, and…
Wang, Baoju; Zhan, Qiuqiang; Zhao, Yuxiang; Wu, Ruitao; Liu, Jing; He, Sailing
2016-01-25
Further development of multiphoton microscopic imaging is confronted with a number of limitations, including high-cost, high complexity and relatively low spatial resolution due to the long excitation wavelength. To overcome these problems, for the first time, we propose visible-to-visible four-photon ultrahigh resolution microscopic imaging by using a common cost-effective 730-nm laser diode to excite the prepared Nd(3+)-sensitized upconversion nanoparticles (Nd(3+)-UCNPs). An ordinary multiphoton scanning microscope system was built using a visible CW diode laser and the lateral imaging resolution as high as 161-nm was achieved via the four-photon upconversion process. The demonstrated large saturation excitation power for Nd(3+)-UCNPs would be more practical and facilitate the four-photon imaging in the application. A sample with fine structure was imaged to demonstrate the advantages of visible-to-visible four-photon ultrahigh resolution microscopic imaging with 730-nm diode laser excited nanocrystals. Combining the uniqueness of UCNPs, the proposed visible-to-visible four-photon imaging would be highly promising and attractive in the field of multiphoton imaging.
Computational polarization difference underwater imaging based on image fusion
NASA Astrophysics Data System (ADS)
Han, Hongwei; Zhang, Xiaohui; Guan, Feng
2016-01-01
Polarization difference imaging can improve the quality of images acquired underwater, whether the background and veiling light are unpolarized or partial polarized. Computational polarization difference imaging technique which replaces the mechanical rotation of polarization analyzer and shortens the time spent to select the optimum orthogonal ǁ and ⊥axes is the improvement of the conventional PDI. But it originally gets the output image by setting the weight coefficient manually to an identical constant for all pixels. In this paper, a kind of algorithm is proposed to combine the Q and U parameters of the Stokes vector through pixel-level image fusion theory based on non-subsample contourlet transform. The experimental system built by the green LED array with polarizer to illuminate a piece of flat target merged in water and the CCD with polarization analyzer to obtain target image under different angle is used to verify the effect of the proposed algorithm. The results showed that the output processed by our algorithm could show more details of the flat target and had higher contrast compared to original computational polarization difference imaging.
The CCD/Transit Instrument (CTI) data-analysis system
NASA Technical Reports Server (NTRS)
Cawson, M. G. M.; Mcgraw, J. T.; Keane, M. J.
1995-01-01
The automated software system for archiving, analyzing, and interrogating data from the CCD/Transit Instrument (CTI) is described. The CTI collects up to 450 Mbytes of image-data each clear night in the form of a narrow strip of sky observed in two colors. The large data-volumes and the scientific aims of the project make it imperative that the data are analyzed within the 24-hour period following the observations. To this end a fully automatic and self evaluating software system has been developed. The data are collected from the telescope in real-time and then transported to Tucson for analysis. Verification is performed by visual inspection of random subsets of the data and obvious cosmic rays are detected and removed before permanent archival is made to the optical disc. The analysis phase is performed by a pair of linked algorithms, one operating on the absolute pixel-values and the other on the spatial derivative of the data. In this way both isolated and merged images are reliably detected in a single pass. In order to isolate the latter algorithm from the effects of noise spikes a 3x3 Hanning filter is applied to the raw data before the analysis is run. The algorithms reduce the input pixel-data to a database of measured parameters for each image which has been found. A contrast filter is applied in order to assign a detection-probability to each image and then x-y calibration and intensity calibration are performed using known reference stars in the strip. These are added to as necessary by secondary standards boot-strapped from the CTI data itself. The final stages involve merging the new data into the CTI Master-list and History-list and the automatic comparison of each new detection with a set of pre-defined templates in parameter-space to find interesting objects such as supernovae, quasars and variable stars. Each stage of the processing from verification to interesting image selection is performed under a data-logging system which both controls the pipe-lining of data through the system and records key performance monitor parameters which are built into the software. Furthermore, the data from each stage are stored in databases to facilitate evaluation, and all stages offer the facility to enter keyword-indexed free-format text into the data-logging system. In this way a large measure of certification is built into the system to provide the necessary confidence in the end results.
Design and operation of SUCHI: the space ultra-compact hyperspectral imager for a small satellite
NASA Astrophysics Data System (ADS)
Crites, S. T.; Lucey, P. G.; Wright, R.; Chan, J.; Garbeil, H.; Horton, K. A.; Imai, A.; Pilger, E. J.; Wood, M.; Yoneshige, Lance
2014-06-01
The primary payload on the University of Hawaii-built `HiakaSat' micro-satellite will be the Space Ultra Compact Hyperspectral Imager (SUCHI). SUCHI is a low-mass (<9kg), low-volume (10x10x36 cm3) long wave infrared hyperspectral imager designed and built at the University of Hawaii. SUCHI is based on a variable-gap Fabry-Perot interferometer employed as a Fourier transform spectrometer with images collected by a commercial 320x256 microbolometer array. The microbolometer camera and vacuum-sensitive electronics are contained within a sealed vessel at 1 atm. SUCHI will collect spectral radiance data from 8 to 14 microns and demonstrate the potential of this instrument for geological studies from orbit (e.g. mapping of major rock-forming minerals) and volcanic hazard observation and assessment (e.g. quantification of volcanic sulfur dioxide pollution and lava flow cooling rates). The sensor has been integrated with the satellite which will launch on the Office of Responsive Space ORS-4 mission scheduled for 2014. The primary mission will last 6 months, with extended operations anticipated for approximately 2 years. A follow-on mission has been proposed to perform imaging of Earth's surface in the 3-5 micron range with a field of view of 5 km with 5.25 m sampling (from a 350 km orbit). The 19-kg proposed instrument will be a prototype sensor for a constellation of small satellites for Earth imaging. The integrated satellite properties will be incorporated into the Hawaii Space Flight Laboratory's constellation maintenance software environment COSMOS (Comprehensive Openarchitecture Space Mission Operations System) to ease future implementation of the instrument as part of a constellation.
Satellite Imagery Via Personal Computer
NASA Technical Reports Server (NTRS)
1989-01-01
Automatic Picture Transmission (APT) was incorporated by NASA in the Tiros 8 weather satellite. APT included an advanced satellite camera that immediately transmitted a picture as well as low cost receiving equipment. When an advanced scanning radiometer was later introduced, ground station display equipment would not readily adjust to the new format until GSFC developed an APT Digital Scan Converter that made them compatible. A NASA Technical Note by Goddard's Vermillion and Kamoski described how to build a converter. In 1979, Electro-Services, using this technology, built the first microcomputer weather imaging system in the U.S. The company changed its name to Satellite Data Systems, Inc. and now manufactures the WeatherFax facsimile display graphics system which converts a personal computer into a weather satellite image acquisition and display workstation. Hardware, antennas, receivers, etc. are also offered. Customers include U.S. Weather Service, schools, military, etc.
Visible and Extended Near-Infrared Multispectral Imaging for Skin Cancer Diagnosis
Rey-Barroso, Laura; Burgos-Fernández, Francisco J.; Delpueyo, Xana; Ares, Miguel; Malvehy, Josep; Puig, Susana
2018-01-01
With the goal of diagnosing skin cancer in an early and noninvasive way, an extended near infrared multispectral imaging system based on an InGaAs sensor with sensitivity from 995 nm to 1613 nm was built to evaluate deeper skin layers thanks to the higher penetration of photons at these wavelengths. The outcomes of this device were combined with those of a previously developed multispectral system that works in the visible and near infrared range (414 nm–995 nm). Both provide spectral and spatial information from skin lesions. A classification method to discriminate between melanomas and nevi was developed based on the analysis of first-order statistics descriptors, principal component analysis, and support vector machine tools. The system provided a sensitivity of 78.6% and a specificity of 84.6%, the latter one being improved with respect to that offered by silicon sensors. PMID:29734747
Matsumoto, Atsushi; Miyazaki, Naoyuki; Takagi, Junichi; Iwasaki, Kenji
2017-03-23
In this study, we develop an approach termed "2D hybrid analysis" for building atomic models by image matching from electron microscopy (EM) images of biological molecules. The key advantage is that it is applicable to flexible molecules, which are difficult to analyze by 3DEM approach. In the proposed approach, first, a lot of atomic models with different conformations are built by computer simulation. Then, simulated EM images are built from each atomic model. Finally, they are compared with the experimental EM image. Two kinds of models are used as simulated EM images: the negative stain model and the simple projection model. Although the former is more realistic, the latter is adopted to perform faster computations. The use of the negative stain model enables decomposition of the averaged EM images into multiple projection images, each of which originated from a different conformation or orientation. We apply this approach to the EM images of integrin to obtain the distribution of the conformations, from which the pathway of the conformational change of the protein is deduced.
NASA Astrophysics Data System (ADS)
Trokielewicz, Mateusz; Bartuzi, Ewelina; Michowska, Katarzyna; Andrzejewska, Antonina; Selegrat, Monika
2015-09-01
In the age of modern, hyperconnected society that increasingly relies on mobile devices and solutions, implementing a reliable and accurate biometric system employing iris recognition presents new challenges. Typical biometric systems employing iris analysis require expensive and complicated hardware. We therefore explore an alternative way using visible spectrum iris imaging. This paper aims at answering several questions related to applying iris biometrics for images obtained in the visible spectrum using smartphone camera. Can irides be successfully and effortlessly imaged using a smartphone's built-in camera? Can existing iris recognition methods perform well when presented with such images? The main advantage of using near-infrared (NIR) illumination in dedicated iris recognition cameras is good performance almost independent of the iris color and pigmentation. Are the images obtained from smartphone's camera of sufficient quality even for the dark irides? We present experiments incorporating simple image preprocessing to find the best visibility of iris texture, followed by a performance study to assess whether iris recognition methods originally aimed at NIR iris images perform well with visible light images. To our best knowledge this is the first comprehensive analysis of iris recognition performance using a database of high-quality images collected in visible light using the smartphones flashlight together with the application of commercial off-the-shelf (COTS) iris recognition methods.
NASA Astrophysics Data System (ADS)
Sameen, Maher Ibrahim; Pradhan, Biswajeet
2016-06-01
In this study, we propose a novel built-up spectral index which was developed by using particle-swarm-optimization (PSO) technique for Worldview-2 images. PSO was used to select the relevant bands from the eight (8) spectral bands of Worldview-2 image and then were used for index development. Multiobiective optimization was used to minimize the number of selected spectral bands and to maximize the classification accuracy. The results showed that the most important and relevant spectral bands among the eight (8) bands for built-up area extraction are band4 (yellow) and band7 (NIR1). Using those relevant spectral bands, the final spectral index was form ulated by developing a normalized band ratio. The validation of the classification result using the proposed spectral index showed that our novel spectral index performs well compared to the existing WV -BI index. The accuracy assessment showed that the new proposed spectral index could extract built-up areas from Worldview-2 image with an area under curve (AUC) of (0.76) indicating the effectiveness of the developed spectral index. Further improvement could be done by using several datasets during the index development process to ensure the transferability of the index to other datasets and study areas.
A GaAs pixel detectors-based digital mammographic system: Performances and imaging tests results
NASA Astrophysics Data System (ADS)
Annovazzi, A.; Amendolia, S. R.; Bigongiari, A.; Bisogni, M. G.; Catarsi, F.; Cesqui, F.; Cetronio, A.; Colombo, F.; Delogu, P.; Fantacci, M. E.; Gilberti, A.; Lanzieri, C.; Lavagna, S.; Novelli, M.; Passuello, G.; Paternoster, G.; Pieracci, M.; Poletti, M.; Quattrocchi, M.; Rosso, V.; Stefanini, A.; Testa, A.; Venturelli, L.
2007-06-01
The prototype presented in this paper is based on GaAs pixel detectors read-out by the PCC/MEDIPIX I circuit. The active area of a sensor is about 1 cm 2 therefore to cover the typical irradiation field used in mammography (18×24 cm 2), 18 GaAs detection units have been organized in two staggered rows of nine chips each and moved by a stepper motor in the orthogonal direction. The system is integrated in a mammographic equipment which comprehends the X-ray tube, the bias and data acquisition systems and the PC-based control system. The prototype has been developed in the framework of the Integrated Mammographic Imaging (IMI) project, an industrial research activity aiming to develop innovative instrumentation for morphologic and functional imaging. The project has been supported by the Italian Ministry of Education, University and Research (MIUR) and by five Italian High Tech companies, Alenia Marconi Systems (AMS), CAEN, Gilardoni, LABEN and Poli.Hi.Tech., in collaboration with the universities of Ferrara, Roma "La Sapienza", Pisa and the Istituto Nazionale di Fisica Nucleare (INFN). In this paper, we report on the electrical characterization and the first imaging test results of the digital mammographic system. To assess the imaging capability of such a detector we have built a phantom, which simulates the breast tissue with malignancies. The radiographs of the phantom, obtained by delivering an entrance dose of 4.8 mGy, have shown particulars with a measured contrast below 1%.
3D ultrasound image guidance system used in RF uterine adenoma and uterine bleeding ablation system
NASA Astrophysics Data System (ADS)
Ding, Mingyue; Luo, Xiaoan; Cai, Chao; Zhou, Chengping; Fenster, Aaron
2006-03-01
Uterine adenoma and uterine bleeding are the two most prevalent diseases in Chinese women. Many women lose their fertility from these diseases. Currently, a minimally invasive ablation system using an RF button electrode is being used in Chinese hospitals to destroy tumor cells or stop bleeding. In this paper, we report on a 3D US guidance system developed to avoid accidents or death of the patient by inaccurate localization of the tumor position during treatment. A 3D US imaging system using a rotational scanning approach of an abdominal probe was built. In order to reduce the distortion produced when the rotational axis is not collinear with the central beam of the probe, a new 3D reconstruction algorithm is used. Then, a fast 3D needle segmentation algorithm is used to find the electrode. Finally, the tip of electrode is determined along the segmented 3D needle and the whole electrode is displayed. Experiments with a water phantom demonstrated the feasibility of our approach.
4D light-field sensing system for people counting
NASA Astrophysics Data System (ADS)
Hou, Guangqi; Zhang, Chi; Wang, Yunlong; Sun, Zhenan
2016-03-01
Counting the number of people is still an important task in social security applications, and a few methods based on video surveillance have been proposed in recent years. In this paper, we design a novel optical sensing system to directly acquire the depth map of the scene from one light-field camera. The light-field sensing system can count the number of people crossing the passageway, and record the direction and intensity of rays at a snapshot without any assistant light devices. Depth maps are extracted from the raw light-ray sensing data. Our smart sensing system is equipped with a passive imaging sensor, which is able to naturally discern the depth difference between the head and shoulders for each person. Then a human model is built. Through detecting the human model from light-field images, the number of people passing the scene can be counted rapidly. We verify the feasibility of the sensing system as well as the accuracy by capturing real-world scenes passing single and multiple people under natural illumination.
High-performance floating-point image computing workstation for medical applications
NASA Astrophysics Data System (ADS)
Mills, Karl S.; Wong, Gilman K.; Kim, Yongmin
1990-07-01
The medical imaging field relies increasingly on imaging and graphics techniques in diverse applications with needs similar to (or more stringent than) those of the military, industrial and scientific communities. However, most image processing and graphics systems available for use in medical imaging today are either expensive, specialized, or in most cases both. High performance imaging and graphics workstations which can provide real-time results for a number of applications, while maintaining affordability and flexibility, can facilitate the application of digital image computing techniques in many different areas. This paper describes the hardware and software architecture of a medium-cost floating-point image processing and display subsystem for the NeXT computer, and its applications as a medical imaging workstation. Medical imaging applications of the workstation include use in a Picture Archiving and Communications System (PACS), in multimodal image processing and 3-D graphics workstation for a broad range of imaging modalities, and as an electronic alternator utilizing its multiple monitor display capability and large and fast frame buffer. The subsystem provides a 2048 x 2048 x 32-bit frame buffer (16 Mbytes of image storage) and supports both 8-bit gray scale and 32-bit true color images. When used to display 8-bit gray scale images, up to four different 256-color palettes may be used for each of four 2K x 2K x 8-bit image frames. Three of these image frames can be used simultaneously to provide pixel selectable region of interest display. A 1280 x 1024 pixel screen with 1: 1 aspect ratio can be windowed into the frame buffer for display of any portion of the processed image or images. In addition, the system provides hardware support for integer zoom and an 82-color cursor. This subsystem is implemented on an add-in board occupying a single slot in the NeXT computer. Up to three boards may be added to the NeXT for multiple display capability (e.g., three 1280 x 1024 monitors, each with a 16-Mbyte frame buffer). Each add-in board provides an expansion connector to which an optional image computing coprocessor board may be added. Each coprocessor board supports up to four processors for a peak performance of 160 MFLOPS. The coprocessors can execute programs from external high-speed microcode memory as well as built-in internal microcode routines. The internal microcode routines provide support for 2-D and 3-D graphics operations, matrix and vector arithmetic, and image processing in integer, IEEE single-precision floating point, or IEEE double-precision floating point. In addition to providing a library of C functions which links the NeXT computer to the add-in board and supports its various operational modes, algorithms and medical imaging application programs are being developed and implemented for image display and enhancement. As an extension to the built-in algorithms of the coprocessors, 2-D Fast Fourier Transform (FF1), 2-D Inverse FFF, convolution, warping and other algorithms (e.g., Discrete Cosine Transform) which exploit the parallel architecture of the coprocessor board are being implemented.
The Design of Optical Sensor for the Pinhole/Occulter Facility
NASA Technical Reports Server (NTRS)
Greene, Michael E.
1990-01-01
Three optical sight sensor systems were designed, built and tested. Two optical lines of sight sensor system are capable of measuring the absolute pointing angle to the sun. The system is for use with the Pinhole/Occulter Facility (P/OF), a solar hard x ray experiment to be flown from Space Shuttle or Space Station. The sensor consists of a pinhole camera with two pairs of perpendicularly mounted linear photodiode arrays to detect the intensity distribution of the solar image produced by the pinhole, track and hold circuitry for data reduction, an analog to digital converter, and a microcomputer. The deflection of the image center is calculated from these data using an approximation for the solar image. A second system consists of a pinhole camera with a pair of perpendicularly mounted linear photodiode arrays, amplification circuitry, threshold detection circuitry, and a microcomputer board. The deflection of the image is calculated by knowing the position of each pixel of the photodiode array and merely counting the pixel numbers until threshold is surpassed. A third optical sensor system is capable of measuring the internal vibration of the P/OF between the mask and base. The system consists of a white light source, a mirror and a pair of perpendicularly mounted linear photodiode arrays to detect the intensity distribution of the solar image produced by the mirror, amplification circuitry, threshold detection circuitry, and a microcomputer board. The deflection of the image and hence the vibration of the structure is calculated by knowing the position of each pixel of the photodiode array and merely counting the pixel numbers until threshold is surpassed.
Affordable CZT SPECT with dose-time minimization (Conference Presentation)
NASA Astrophysics Data System (ADS)
Hugg, James W.; Harris, Brian W.; Radley, Ian
2017-03-01
PURPOSE Pixelated CdZnTe (CZT) detector arrays are used in molecular imaging applications that can enable precision medicine, including small-animal SPECT, cardiac SPECT, molecular breast imaging (MBI), and general purpose SPECT. The interplay of gamma camera, collimator, gantry motion, and image reconstruction determines image quality and dose-time-FOV tradeoffs. Both dose and exam time can be minimized without compromising diagnostic content. METHODS Integration of pixelated CZT detectors with advanced ASICs and readout electronics improves system performance. Because historically CZT was expensive, the first clinical applications were limited to small FOV. Radiation doses were initially high and exam times long. Advances have significantly improved efficiency of CZT-based molecular imaging systems and the cost has steadily declined. We have built a general purpose SPECT system using our 40 cm x 53 cm CZT gamma camera with 2 mm pixel pitch and characterized system performance. RESULTS Compared to NaI scintillator gamma cameras: intrinsic spatial resolution improved from 3.8 mm to 2.0 mm; energy resolution improved from 9.8% to <4 % at 140 keV; maximum count rate is <1.5 times higher; non-detection camera edges are reduced 3-fold. Scattered photons are greatly reduced in the photopeak energy window; image contrast is improved; and the optimal FOV is increased to the entire camera area. CONCLUSION Continual improvements in CZT detector arrays for molecular imaging, coupled with optimal collimator and image reconstruction, result in minimized dose and exam time. With CZT cost improving, affordable whole-body CZT general purpose SPECT is expected to enable precision medicine applications.
Image/video understanding systems based on network-symbolic models
NASA Astrophysics Data System (ADS)
Kuvich, Gary
2004-03-01
Vision is a part of a larger information system that converts visual information into knowledge structures. These structures drive vision process, resolve ambiguity and uncertainty via feedback projections, and provide image understanding that is an interpretation of visual information in terms of such knowledge models. Computer simulation models are built on the basis of graphs/networks. The ability of human brain to emulate similar graph/network models is found. Symbols, predicates and grammars naturally emerge in such networks, and logic is simply a way of restructuring such models. Brain analyzes an image as a graph-type relational structure created via multilevel hierarchical compression of visual information. Primary areas provide active fusion of image features on a spatial grid-like structure, where nodes are cortical columns. Spatial logic and topology naturally present in such structures. Mid-level vision processes like perceptual grouping, separation of figure from ground, are special kinds of network transformations. They convert primary image structure into the set of more abstract ones, which represent objects and visual scene, making them easy for analysis by higher-level knowledge structures. Higher-level vision phenomena are results of such analysis. Composition of network-symbolic models combines learning, classification, and analogy together with higher-level model-based reasoning into a single framework, and it works similar to frames and agents. Computational intelligence methods transform images into model-based knowledge representation. Based on such principles, an Image/Video Understanding system can convert images into the knowledge models, and resolve uncertainty and ambiguity. This allows creating intelligent computer vision systems for design and manufacturing.
Intravascular optical imaging of high-risk plaques in vivo by targeting macrophage mannose receptors
NASA Astrophysics Data System (ADS)
Kim, Ji Bak; Park, Kyeongsoon; Ryu, Jiheun; Lee, Jae Joong; Lee, Min Woo; Cho, Han Saem; Nam, Hyeong Soo; Park, Ok Kyu; Song, Joon Woo; Kim, Tae Shik; Oh, Dong Joo; Gweon, Daegab; Oh, Wang-Yuhl; Yoo, Hongki; Kim, Jin Won
2016-03-01
Macrophages mediate atheroma expansion and disruption, and denote high-risk arterial plaques. Therefore, they are substantially gaining importance as a diagnostic imaging target for the detection of rupture-prone plaques. Here, we developed an injectable near-infrared fluorescence (NIRF) probe by chemically conjugating thiolated glycol chitosan with cholesteryl chloroformate, NIRF dye (cyanine 5.5 or 7), and maleimide-polyethylene glycol-mannose as mannose receptor binding ligands to specifically target a subset of macrophages abundant in high-risk plaques. This probe showed high affinity to mannose receptors, low toxicity, and allowed the direct visualization of plaque macrophages in murine carotid atheroma. After the scale-up of the MMR-NIRF probe, the administration of the probe facilitated in vivo intravascular imaging of plaque inflammation in coronary-sized vessels of atheromatous rabbits using a custom-built dual-modal optical coherence tomography (OCT)-NIRF catheter-based imaging system. This novel imaging approach represents a potential imaging strategy enabling the identification of high-risk plaques in vivo and holds promise for future clinical implications.
Lindsey, Brooks D; Light, Edward D; Nicoletto, Heather A; Bennett, Ellen R; Laskowitz, Daniel T; Smith, Stephen W
2011-06-01
Because stroke remains an important and time-sensitive health concern in developed nations, we present a system capable of fusing 3-D transcranial ultrasound volumes acquired from two sides of the head. This system uses custom sparse array transducers built on flexible multilayer circuits that can be positioned for simultaneous imaging through both temporal acoustic windows, allowing for potential registration of multiple real-time 3-D scans of cerebral vasculature. We examine hardware considerations for new matrix arrays-transducer design and interconnects-in this application. Specifically, it is proposed that SNR may be increased by reducing the length of probe cables. This claim is evaluated as part of the presented system through simulation, experimental data, and in vivo imaging. Ultimately, gains in SNR of 7 dB are realized by replacing a standard probe cable with a much shorter flex interconnect; higher gains may be possible using ribbon-based probe cables. In vivo images are presented, showing cerebral arteries with and without the use of microbubble contrast agent; they have been registered and fused using a simple algorithm which maximizes normalized cross-correlation.
Lindsey, Brooks D.; Light, Edward D.; Nicoletto, Heather A.; Bennett, Ellen R.; Laskowitz, Daniel T.; Smith, Stephen W.
2012-01-01
Because stroke remains an important and time-sensitive health concern in developed nations, we present a system capable of fusing 3-D transcranial ultrasound volumes acquired from two sides of the head. This system uses custom sparse array transducers built on flexible multilayer circuits that can be positioned for simultaneous imaging through both temporal acoustic windows, allowing for potential registration of multiple real-time 3-D scans of cerebral vasculature. We examine hardware considerations for new matrix arrays—transducer design and interconnects—in this application. Specifically, it is proposed that SNR may be increased by reducing the length of probe cables. This claim is evaluated as part of the presented system through simulation, experimental data, and in vivo imaging. Ultimately, gains in SNR of 7 dB are realized by replacing a standard probe cable with a much shorter flex interconnect; higher gains may be possible using ribbon-based probe cables. In vivo images are presented, showing cerebral arteries with and without the use of microbubble contrast agent; they have been registered and fused using a simple algorithm which maximizes normalized cross-correlation. PMID:21693401
NASA Astrophysics Data System (ADS)
Hsu, Jen-Feng; Dhingra, Shonali; D'Urso, Brian
2017-01-01
Mirror galvanometer systems (galvos) are commonly employed in research and commercial applications in areas involving laser imaging, laser machining, laser-light shows, and others. Here, we present a robust, moderate-speed, and cost-efficient home-built galvo system. The mechanical part of this design consists of one mirror, which is tilted around two axes with multiple surface transducers. We demonstrate the ability of this galvo by scanning the mirror using a computer, via a custom driver circuit. The performance of the galvo, including scan range, noise, linearity, and scan speed, is characterized. As an application, we show that this galvo system can be used in a confocal scanning microscopy system.
SkySat-1: very high-resolution imagery from a small satellite
NASA Astrophysics Data System (ADS)
Murthy, Kiran; Shearn, Michael; Smiley, Byron D.; Chau, Alexandra H.; Levine, Josh; Robinson, M. Dirk
2014-10-01
This paper presents details of the SkySat-1 mission, which is the first microsatellite-class commercial earth- observation system to generate sub-meter resolution panchromatic imagery, in addition to sub-meter resolution 4-band pan-sharpened imagery. SkySat-1 was built and launched for an order of magnitude lower cost than similarly performing missions. The low-cost design enables the deployment of a large imaging constellation that can provide imagery with both high temporal resolution and high spatial resolution. One key enabler of the SkySat-1 mission was simplifying the spacecraft design and instead relying on ground- based image processing to achieve high-performance at the system level. The imaging instrument consists of a custom-designed high-quality optical telescope and commercially-available high frame rate CMOS image sen- sors. While each individually captured raw image frame shows moderate quality, ground-based image processing algorithms improve the raw data by combining data from multiple frames to boost image signal-to-noise ratio (SNR) and decrease the ground sample distance (GSD) in a process Skybox calls "digital TDI". Careful qual-ity assessment and tuning of the spacecraft, payload, and algorithms was necessary to generate high-quality panchromatic, multispectral, and pan-sharpened imagery. Furthermore, the framing sensor configuration en- abled the first commercial High-Definition full-frame rate panchromatic video to be captured from space, with approximately 1 meter ground sample distance. Details of the SkySat-1 imaging instrument and ground-based image processing system are presented, as well as an overview of the work involved with calibrating and validating the system. Examples of raw and processed imagery are shown, and the raw imagery is compared to pre-launch simulated imagery used to tune the image processing algorithms.
Image reconstruction and system modeling techniques for virtual-pinhole PET insert systems
Keesing, Daniel B; Mathews, Aswin; Komarov, Sergey; Wu, Heyu; Song, Tae Yong; O'Sullivan, Joseph A; Tai, Yuan-Chuan
2012-01-01
Virtual-pinhole PET (VP-PET) imaging is a new technology in which one or more high-resolution detector modules are integrated into a conventional PET scanner with lower-resolution detectors. It can locally enhance the spatial resolution and contrast recovery near the add-on detectors, and depending on the configuration, may also increase the sensitivity of the system. This novel scanner geometry makes the reconstruction problem more challenging compared to the reconstruction of data from a standalone PET scanner, as new techniques are needed to model and account for the non-standard acquisition. In this paper, we present a general framework for fully 3D modeling of an arbitrary VP-PET insert system. The model components are incorporated into a statistical reconstruction algorithm to estimate an image from the multi-resolution data. For validation, we apply the proposed model and reconstruction approach to one of our custom-built VP-PET systems – a half-ring insert device integrated into a clinical PET/CT scanner. Details regarding the most important implementation issues are provided. We show that the proposed data model is consistent with the measured data, and that our approach can lead to reconstructions with improved spatial resolution and lesion detectability. PMID:22490983
NASA Technical Reports Server (NTRS)
2006-01-01
[figure removed for brevity, see original site] Context image for PIA02171 Cloud Front These clouds formed in the south polar region. The faintness of the cloud system likely indicates that these are mainly ice clouds, with relatively little dust content. Image information: VIS instrument. Latitude -86.7N, Longitude 212.3E. 17 meter/pixel resolution. Note: this THEMIS visual image has not been radiometrically nor geometrically calibrated for this preliminary release. An empirical correction has been performed to remove instrumental effects. A linear shift has been applied in the cross-track and down-track direction to approximate spacecraft and planetary motion. Fully calibrated and geometrically projected images will be released through the Planetary Data System in accordance with Project policies at a later time. NASA's Jet Propulsion Laboratory manages the 2001 Mars Odyssey mission for NASA's Office of Space Science, Washington, D.C. The Thermal Emission Imaging System (THEMIS) was developed by Arizona State University, Tempe, in collaboration with Raytheon Santa Barbara Remote Sensing. The THEMIS investigation is led by Dr. Philip Christensen at Arizona State University. Lockheed Martin Astronautics, Denver, is the prime contractor for the Odyssey project, and developed and built the orbiter. Mission operations are conducted jointly from Lockheed Martin and from JPL, a division of the California Institute of Technology in Pasadena.Aryanto, K Y E; Broekema, A; Langenhuysen, R G A; Oudkerk, M; van Ooijen, P M A
2015-05-01
To develop and test a fast and easy rule-based web-environment with optional de-identification of imaging data to facilitate data distribution within a hospital environment. A web interface was built using Hypertext Preprocessor (PHP), an open source scripting language for web development, and Java with SQL Server to handle the database. The system allows for the selection of patient data and for de-identifying these when necessary. Using the services provided by the RSNA Clinical Trial Processor (CTP), the selected images were pushed to the appropriate services using a protocol based on the module created for the associated task. Five pipelines, each performing a different task, were set up in the server. In a 75 month period, more than 2,000,000 images are transferred and de-identified in a proper manner while 20,000,000 images are moved from one node to another without de-identification. While maintaining a high level of security and stability, the proposed system is easy to setup, it integrate well with our clinical and research practice and it provides a fast and accurate vendor-neutral process of transferring, de-identifying, and storing DICOM images. Its ability to run different de-identification processes in parallel pipelines is a major advantage in both clinical and research setting.
Multi-microphone adaptive array augmented with visual cueing.
Gibson, Paul L; Hedin, Dan S; Davies-Venn, Evelyn E; Nelson, Peggy; Kramer, Kevin
2012-01-01
We present the development of an audiovisual array that enables hearing aid users to converse with multiple speakers in reverberant environments with significant speech babble noise where their hearing aids do not function well. The system concept consists of a smartphone, a smartphone accessory, and a smartphone software application. The smartphone accessory concept is a multi-microphone audiovisual array in a form factor that allows attachment to the back of the smartphone. The accessory will also contain a lower power radio by which it can transmit audio signals to compatible hearing aids. The smartphone software application concept will use the smartphone's built in camera to acquire images and perform real-time face detection using the built-in face detection support of the smartphone. The audiovisual beamforming algorithm uses the location of talking targets to improve the signal to noise ratio and consequently improve the user's speech intelligibility. Since the proposed array system leverages a handheld consumer electronic device, it will be portable and low cost. A PC based experimental system was developed to demonstrate the feasibility of an audiovisual multi-microphone array and these results are presented.
NASA Technical Reports Server (NTRS)
1983-01-01
Grumman OV-1C in flight. This OV-1C Mohawk, serial #67-15932, was used in a joint NASA/US Army Aviation Engineering Flight Activity (USAAEFA) program to study a stall-speed warning system in the early 1980s. NASA designed and built an automated stall-speed warning system which presented both airspeed and stall speed to the pilot. Visual indication of impending stall would be displayed to the pilot as a cursor or pointer located on a conventional airspeed indicator. In addition, an aural warning at predetermined stall margins was presented to the pilot through a voice synthesizer. The Mohawk was developed by Grumman Aircraft as a photo observation and reconnaissance aircraft for the US Marines and the US Army. The OV-1 entered production in October 1959 and served the US Army in Europe, Korea, the Viet Nam War, Central and South America, Alaska, and during Desert Shield/Desert Storm in the Middle East. The Mohawk was retired from service in September 1996. 133 OV-1Cs were built, the 'C' designating the model which used an IR (infrared) imaging system to provide reconnaissance.
NASA Astrophysics Data System (ADS)
McMackin, Lenore; Herman, Matthew A.; Weston, Tyler
2016-02-01
We present the design of a multi-spectral imager built using the architecture of the single-pixel camera. The architecture is enabled by the novel sampling theory of compressive sensing implemented optically using the Texas Instruments DLP™ micro-mirror array. The array not only implements spatial modulation necessary for compressive imaging but also provides unique diffractive spectral features that result in a multi-spectral, high-spatial resolution imager design. The new camera design provides multi-spectral imagery in a wavelength range that extends from the visible to the shortwave infrared without reduction in spatial resolution. In addition to the compressive imaging spectrometer design, we present a diffractive model of the architecture that allows us to predict a variety of detailed functional spatial and spectral design features. We present modeling results, architectural design and experimental results that prove the concept.
Omega-3 chicken egg detection system using a mobile-based image processing segmentation method
NASA Astrophysics Data System (ADS)
Nurhayati, Oky Dwi; Kurniawan Teguh, M.; Cintya Amalia, P.
2017-02-01
An Omega-3 chicken egg is a chicken egg produced through food engineering technology. It is produced by hen fed with high omega-3 fatty acids. So, it has fifteen times nutrient content of omega-3 higher than Leghorn's. Visually, its shell has the same shape and colour as Leghorn's. Each egg can be distinguished by breaking the egg's shell and testing the egg yolk's nutrient content in a laboratory. But, those methods were proven not effective and efficient. Observing this problem, the purpose of this research is to make an application to detect the type of omega-3 chicken egg by using a mobile-based computer vision. This application was built in OpenCV computer vision library to support Android Operating System. This experiment required some chicken egg images taken using an egg candling box. We used 60 omega-3 chicken and Leghorn eggs as samples. Then, using an Android smartphone, image acquisition of the egg was obtained. After that, we applied several steps using image processing methods such as Grab Cut, convert RGB image to eight bit grayscale, median filter, P-Tile segmentation, and morphology technique in this research. The next steps were feature extraction which was used to extract feature values via mean, variance, skewness, and kurtosis from each image. Finally, using digital image measurement, some chicken egg images were classified. The result showed that omega-3 chicken egg and Leghorn egg had different values. This system is able to provide accurate reading around of 91%.
NASA Technical Reports Server (NTRS)
2007-01-01
Lockheed Martin Space Systems technicians Jim Young (left) and Jack Farmerie (right) work on the science deck of NASA's Phoenix Mars Lander. The spacecraft was built in a 100,000-class clean room near Denver under NASA's planetary protection practices to keep organics from being taken to Mars. The lander's robotic arm, built by the Jet Propulsion Laboratory, Pasadena, is seen at the top of the picture. The color and grey dots will be used to calibrate the spacecraft's Surface Stereoscopic Imager camera once the spacecraft has landed on the red planet. The Phoenix mission is led by Principal Investigator Peter H. Smith of the University of Arizona, Tucson, with project management at NASA's Jet Propulsion Laboratory and development partnership with Lockheed Martin Space Systems. International contributions for Phoenix are provided by the Canadian Space Agency, the University of Neuchatel (Switzerland), the University of Copenhagen, and the Max Planck Institute in Germany. JPL is a division of the California Institute of Technology in Pasadena.NASA Astrophysics Data System (ADS)
McFee, John E.; Russell, Kevin L.; Chesney, Robert H.; Faust, Anthony A.; Das, Yogadhish
2006-05-01
The Improved Landmine Detection System (ILDS) is intended to meet Canadian military mine clearance requirements in rear area combat situations and peacekeeping on roads and tracks. The system consists of two teleoperated vehicles and a command vehicle. The teleoperated protection vehicle precedes, clearing antipersonnel mines and magnetic and tilt rod-fuzed antitank mines. It consists of an armoured personnel carrier with a forward looking infrared imager, a finger plow or roller and a magnetic signature duplicator. The teleoperated detection vehicle follows to detect antitank mines. The purpose-built vehicle carries forward looking infrared and visible imagers, a 3 m wide, down-looking sensitive electromagnetic induction detector array and a 3 m wide down-looking ground probing radar, which scan the ground in front of the vehicle. Sensor information is combined using navigation sensors and custom navigation, registration, spatial correspondence and data fusion algorithms. Suspicious targets are then confirmed by a thermal neutron activation detector. The prototype, designed and built by Defence R&D Canada, was completed in October 1997. General Dynamics Canada delivered four production units, based on the prototype concept and technologies, to the Canadian Forces (CF) in 2002. ILDS was deployed in Afghanistan in 2003, making the system the first militarily fielded, teleoperated, multi-sensor vehicle-mounted mine detector and the first with a fielded confirmation sensor. Performance of the prototype in Canadian and independent US trials is summarized and recent results from the production version of the confirmation sensor are discussed. CF operations with ILDS in Afghanistan are described.
Li, Hui; Cui, Quan; Zhang, Zhihong; Luo, Qingming
2015-01-01
Background The nonlinear optical microscopy has become the current state-of-the-art for intravital imaging. Due to its advantages of high resolution, superior tissue penetration, lower photodamage and photobleaching, as well as intrinsic z-sectioning ability, this technology has been widely applied in immunoimaging for a decade. However, in terms of monitoring immune events in native physiological environment, the conventional nonlinear optical microscope system has to be optimized for live animal imaging. Generally speaking, three crucial capabilities are desired, including high-speed, large-area and multicolor imaging. Among numerous high-speed scanning mechanisms used in nonlinear optical imaging, polygon scanning is not only linearly but also dispersion-freely with high stability and tunable rotation speed, which can overcome disadvantages of multifocal scanning, resonant scanner and acousto-optical deflector (AOD). However, low frame rate, lacking large-area or multicolor imaging ability make current polygonbased nonlinear optical microscopes unable to meet the requirements of immune event monitoring. Methods We built up a polygon-based nonlinear optical microscope system which was custom optimized for immunoimaging with high-speed, large-are and multicolor imaging abilities. Results Firstly, we validated the imaging performance of the system by standard methods. Then, to demonstrate the ability to monitor immune events, migration of immunocytes observed by the system based on typical immunological models such as lymph node, footpad and dorsal skinfold chamber are shown. Finally, we take an outlook for the possible advance of related technologies such as sample stabilization and optical clearing for more stable and deeper intravital immunoimaging. Conclusions This study will be helpful for optimizing nonlinear optical microscope to obtain more comprehensive and accurate information of immune events. PMID:25694951
NASA Technical Reports Server (NTRS)
Udomkesmalee, Suraphol; Padgett, Curtis; Zhu, David; Lung, Gerald; Howard, Ayanna
2000-01-01
A three-dimensional microelectronic device (3DANN-R) capable of performing general image convolution at the speed of 1012 operations/second (ops) in a volume of less than 1.5 cubic centimeter has been successfully built under the BMDO/JPL VIGILANTE program. 3DANN-R was developed in partnership with Irvine Sensors Corp., Costa Mesa, California. 3DANN-R is a sugar-cube-sized, low power image convolution engine that in its core computation circuitry is capable of performing 64 image convolutions with large (64x64) windows at video frame rates. This paper explores potential applications of 3DANN-R such as target recognition, SAR and hyperspectral data processing, and general machine vision using real data and discuss technical challenges for providing deployable systems for BMDO surveillance and interceptor programs.
Wang, Pu; Ma, Teng; Slipchenko, Mikhail N.; Liang, Shanshan; Hui, Jie; Shung, K. Kirk; Roy, Sukesh; Sturek, Michael; Zhou, Qifa; Chen, Zhongping; Cheng, Ji-Xin
2014-01-01
Lipid deposition inside the arterial wall is a key indicator of plaque vulnerability. An intravascular photoacoustic (IVPA) catheter is considered a promising device for quantifying the amount of lipid inside the arterial wall. Thus far, IVPA systems suffered from slow imaging speed (~50 s per frame) due to the lack of a suitable laser source for high-speed excitation of molecular overtone vibrations. Here, we report an improvement in IVPA imaging speed by two orders of magnitude, to 1.0 s per frame, enabled by a custom-built, 2-kHz master oscillator power amplifier (MOPA)-pumped, barium nitrite [Ba(NO3)2] Raman laser. This advancement narrows the gap in translating the IVPA technology to the clinical setting. PMID:25366991
Phoenix's Laser Beam in Action on Mars
NASA Technical Reports Server (NTRS)
2008-01-01
[figure removed for brevity, see original site] Click on image to view the animation The Surface Stereo Imager camera aboard NASA's Phoenix Mars Lander acquired a series of images of the laser beam in the Martian night sky. Bright spots in the beam are reflections from ice crystals in the low level ice-fog. The brighter area at the top of the beam is due to enhanced scattering of the laser light in a cloud. The Canadian-built lidar instrument emits pulses of laser light and records what is scattered back. The Phoenix Mission is led by the University of Arizona, Tucson, on behalf of NASA. Project management of the mission is by NASA's Jet Propulsion Laboratory, Pasadena, Calif. Spacecraft development is by Lockheed Martin Space Systems, Denver.Unconventional imaging with contained granular media
NASA Astrophysics Data System (ADS)
Quadrelli, Marco B.; Basinger, Scott; Sidick, Erkin
2017-09-01
Typically, the cost of a space-borne imaging system is driven by the size and mass of the primary aperture. The solution that we propose uses a method to construct an imaging system in space in which the nonlinear optical properties of a cloud of micron-sized particles, shaped into a specific surface by electromagnetic means, and allows one to form a very large and lightweight aperture of an optical system, hence reducing overall mass and cost. Recent work at JPL has investigated the feasibility of a granular imaging system, concluding that such a system could be built and controlled in orbit. We conducted experiments and simulation of the optical response of a granular lens. In all cases, the optical response, measured by the Modulation Transfer Function, of hexagonal reflectors was closely comparable to that of a conventional spherical mirror. We conducted some further analyses by evaluating the sensitivity to fill factor and grain shape, and found a marked sensitivity to fill factor but no sensitivity to grain shape. We have also found that at fill factors as low as 30%, the reflection from a granular lens is still excellent. Furthermore, we replaced the monolithic primary mirror in an existing integrated model of an optical system (WFIRST Coronagraph) with a granular lens, and found that the granular lens that can be useful for exoplanet detection provides excellent contrast levels. We will present our testbed and simulation results in this paper.
Towards eye-safe standoff Raman imaging systems
NASA Astrophysics Data System (ADS)
Glimtoft, Martin; Bââth, Petra; Saari, Heikki; Mäkynen, Jussi; Näsilä, Antti; Östmark, Henric
2014-05-01
Standoff Raman imaging systems have shown the ability to detect single explosives particles. However, in many cases, the laser intensities needed restrict the applications where they can be safely used. A new generation imaging Raman system has been developed based on a 355 nm UV laser that, in addition to eye safety, allows discrete and invisible measurements. Non-dangerous exposure levels for the eye are several orders of magnitude higher in UVA than in the visible range that previously has been used. The UV Raman system has been built based on an UV Fabry-Perot Interferometer (UV-FPI) developed by VTT. The design allows for precise selection of Raman shifts in combination with high out-of-band blocking. The stable operation of the UV-FPI module under varying environmental conditions is arranged by controlling the temperature of the module and using a closed loop control of the FPI air gap based on capacitive measurement. The system presented consists of a 3rd harmonics Nd:YAG laser with 1.5 W average output at 1000 Hz, a 200 mm Schmidt-Cassegrain telescope, UV-FPI filter and an ICCD camera for signal gating and detection. The design principal leads to a Raman spectrum in each image pixel. The system is designed for field use and easy manoeuvring. Preliminary results show that in measurements of <60 s on 10 m distance, single AN particles of <300 μm diameter can be identified.
Implementation of a General Real-Time Visual Anomaly Detection System Via Soft Computing
NASA Technical Reports Server (NTRS)
Dominguez, Jesus A.; Klinko, Steve; Ferrell, Bob; Steinrock, Todd (Technical Monitor)
2001-01-01
The intelligent visual system detects anomalies or defects in real time under normal lighting operating conditions. The application is basically a learning machine that integrates fuzzy logic (FL), artificial neural network (ANN), and generic algorithm (GA) schemes to process the image, run the learning process, and finally detect the anomalies or defects. The system acquires the image, performs segmentation to separate the object being tested from the background, preprocesses the image using fuzzy reasoning, performs the final segmentation using fuzzy reasoning techniques to retrieve regions with potential anomalies or defects, and finally retrieves them using a learning model built via ANN and GA techniques. FL provides a powerful framework for knowledge representation and overcomes uncertainty and vagueness typically found in image analysis. ANN provides learning capabilities, and GA leads to robust learning results. An application prototype currently runs on a regular PC under Windows NT, and preliminary work has been performed to build an embedded version with multiple image processors. The application prototype is being tested at the Kennedy Space Center (KSC), Florida, to visually detect anomalies along slide basket cables utilized by the astronauts to evacuate the NASA Shuttle launch pad in an emergency. The potential applications of this anomaly detection system in an open environment are quite wide. Another current, potentially viable application at NASA is in detecting anomalies of the NASA Space Shuttle Orbiter's radiator panels.
A non-contact time-domain scanning brain imaging system: first in-vivo results
NASA Astrophysics Data System (ADS)
Mazurenka, M.; Di Sieno, L.; Boso, G.; Contini, D.; Pifferi, A.; Dalla Mora, A.; Tosi, A.; Wabnitz, H.; Macdonald, R.
2013-06-01
We present results of first in-vivo tests of an optical non-contact scanning imaging system, intended to study oxidative metabolism related processes in biological tissue by means of time-resolved near-infrared spectroscopy. Our method is a novel realization of the short source-detector separation approach and based on a fast-gated single-photon avalanche diode to detect late photons only. The scanning system is built in quasi-confocal configuration and utilizes polarizationsensitive detection. It scans an area of 4×4 cm2, recording images with 32×32 pixels, thus creating a high density of source-detector pairs. To test the system we performed a range of in vivo measurements of hemodynamic changes in several types of biological tissues, i.e. skin (Valsalva maneuver), muscle (venous and arterial occlusions) and brain (motor and cognitive tasks). Task-related changes in hemoglobin concentrations were clearly detected in skin and muscle. The brain activation shows weaker, but yet detectable changes. These changes were localized in pixels near the motor cortex area (C3). However, it was found that even very short hair substantially impairs the measurement. Thus the applicability of the scanner is limited to hairless parts of body. The results of our first in-vivo tests prove the feasibility of non-contact scanning imaging as a first step towards development of a prototype for biological tissue imaging for various medical applications.
Coprates Chasma Landslides in IR
NASA Technical Reports Server (NTRS)
2005-01-01
[figure removed for brevity, see original site] Today's daytime IR image is of a portion of Coprates Chasma, part of Valles Marineris. As with yesterday's image, this image shows multiple large landslides. Image information: IR instrument. Latitude -8.2, Longitude 300.2 East (59.8 West). 100 meter/pixel resolution. Note: this THEMIS visual image has not been radiometrically nor geometrically calibrated for this preliminary release. An empirical correction has been performed to remove instrumental effects. A linear shift has been applied in the cross-track and down-track direction to approximate spacecraft and planetary motion. Fully calibrated and geometrically projected images will be released through the Planetary Data System in accordance with Project policies at a later time. NASA's Jet Propulsion Laboratory manages the 2001 Mars Odyssey mission for NASA's Office of Space Science, Washington, D.C. The Thermal Emission Imaging System (THEMIS) was developed by Arizona State University, Tempe, in collaboration with Raytheon Santa Barbara Remote Sensing. The THEMIS investigation is led by Dr. Philip Christensen at Arizona State University. Lockheed Martin Astronautics, Denver, is the prime contractor for the Odyssey project, and developed and built the orbiter. Mission operations are conducted jointly from Lockheed Martin and from JPL, a division of the California Institute of Technology in Pasadena.Total body photography for skin cancer screening.
Dengel, Lynn T; Petroni, Gina R; Judge, Joshua; Chen, David; Acton, Scott T; Schroen, Anneke T; Slingluff, Craig L
2015-11-01
Total body photography may aid in melanoma screening but is not widely applied due to time and cost. We hypothesized that a near-simultaneous automated skin photo-acquisition system would be acceptable to patients and could rapidly obtain total body images that enable visualization of pigmented skin lesions. From February to May 2009, a study of 20 volunteers was performed at the University of Virginia to test a prototype 16-camera imaging booth built by the research team and to guide development of special purpose software. For each participant, images were obtained before and after marking 10 lesions (five "easy" and five "difficult"), and images were evaluated to estimate visualization rates. Imaging logistical challenges were scored by the operator, and participant opinion was assessed by questionnaire. Average time for image capture was three minutes (range 2-5). All 55 "easy" lesions were visualized (sensitivity 100%, 90% CI 95-100%), and 54/55 "difficult" lesions were visualized (sensitivity 98%, 90% CI 92-100%). Operators and patients graded the imaging process favorably, with challenges identified regarding lighting and positioning. Rapid-acquisition automated skin photography is feasible with a low-cost system, with excellent lesion visualization and participant acceptance. These data provide a basis for employing this method in clinical melanoma screening. © 2014 The International Society of Dermatology.
Sub-word image clustering in Farsi printed books
NASA Astrophysics Data System (ADS)
Soheili, Mohammad Reza; Kabir, Ehsanollah; Stricker, Didier
2015-02-01
Most OCR systems are designed for the recognition of a single page. In case of unfamiliar font faces, low quality papers and degraded prints, the performance of these products drops sharply. However, an OCR system can use redundancy of word occurrences in large documents to improve recognition results. In this paper, we propose a sub-word image clustering method for the applications dealing with large printed documents. We assume that the whole document is printed by a unique unknown font with low quality print. Our proposed method finds clusters of equivalent sub-word images with an incremental algorithm. Due to the low print quality, we propose an image matching algorithm for measuring the distance between two sub-word images, based on Hamming distance and the ratio of the area to the perimeter of the connected components. We built a ground-truth dataset of more than 111000 sub-word images to evaluate our method. All of these images were extracted from an old Farsi book. We cluster all of these sub-words, including isolated letters and even punctuation marks. Then all centers of created clusters are labeled manually. We show that all sub-words of the book can be recognized with more than 99.7% accuracy by assigning the label of each cluster center to all of its members.
NASA Technical Reports Server (NTRS)
2005-01-01
Welcome to another brief interval of THEMIS Images as Art. For two weeks, we will be showcasing images for their aesthetic value rather than their science content. Portions of these images resemble things in our everyday lives, from animals to letters of the alphabet. We hope you enjoy our fanciful look at Mars! X marks the spot! Note: this THEMIS visual image has not been radiometrically nor geometrically calibrated for this preliminary release. An empirical correction has been performed to remove instrumental effects. A linear shift has been applied in the cross-track and down-track direction to approximate spacecraft and planetary motion. Fully calibrated and geometrically projected images will be released through the Planetary Data System in accordance with Project policies at a later time. NASA's Jet Propulsion Laboratory manages the 2001 Mars Odyssey mission for NASA's Office of Space Science, Washington, D.C. The Thermal Emission Imaging System (THEMIS) was developed by Arizona State University, Tempe, in collaboration with Raytheon Santa Barbara Remote Sensing. The THEMIS investigation is led by Dr. Philip Christensen at Arizona State University. Lockheed Martin Astronautics, Denver, is the prime contractor for the Odyssey project, and developed and built the orbiter. Mission operations are conducted jointly from Lockheed Martin and from JPL, a division of the California Institute of Technology in Pasadena.NASA Technical Reports Server (NTRS)
2005-01-01
Welcome to another brief interval of THEMIS Images as Art. For two weeks, we will be showcasing images for their aesthetic value rather than their science content. Portions of these images resemble things in our everyday lives, from animals to letters of the alphabet. We hope you enjoy our fanciful look at Mars! Martian Unicorns? Note: this THEMIS visual image has not been radiometrically nor geometrically calibrated for this preliminary release. An empirical correction has been performed to remove instrumental effects. A linear shift has been applied in the cross-track and down-track direction to approximate spacecraft and planetary motion. Fully calibrated and geometrically projected images will be released through the Planetary Data System in accordance with Project policies at a later time. NASA's Jet Propulsion Laboratory manages the 2001 Mars Odyssey mission for NASA's Office of Space Science, Washington, D.C. The Thermal Emission Imaging System (THEMIS) was developed by Arizona State University, Tempe, in collaboration with Raytheon Santa Barbara Remote Sensing. The THEMIS investigation is led by Dr. Philip Christensen at Arizona State University. Lockheed Martin Astronautics, Denver, is the prime contractor for the Odyssey project, and developed and built the orbiter. Mission operations are conducted jointly from Lockheed Martin and from JPL, a division of the California Institute of Technology in Pasadena.NASA Technical Reports Server (NTRS)
2005-01-01
Welcome to another brief interval of THEMIS Images as Art. For two weeks, we will be showcasing images for their aesthetic value rather than their science content. Portions of these images resemble things in our everyday lives, from animals to letters of the alphabet. We hope you enjoy our fanciful look at Mars! Mars has a kiss for you today! Note: this THEMIS visual image has not been radiometrically nor geometrically calibrated for this preliminary release. An empirical correction has been performed to remove instrumental effects. A linear shift has been applied in the cross-track and down-track direction to approximate spacecraft and planetary motion. Fully calibrated and geometrically projected images will be released through the Planetary Data System in accordance with Project policies at a later time. NASA's Jet Propulsion Laboratory manages the 2001 Mars Odyssey mission for NASA's Office of Space Science, Washington, D.C. The Thermal Emission Imaging System (THEMIS) was developed by Arizona State University, Tempe, in collaboration with Raytheon Santa Barbara Remote Sensing. The THEMIS investigation is led by Dr. Philip Christensen at Arizona State University. Lockheed Martin Astronautics, Denver, is the prime contractor for the Odyssey project, and developed and built the orbiter. Mission operations are conducted jointly from Lockheed Martin and from JPL, a division of the California Institute of Technology in Pasadena.Performance Enhancement of the RatCAP Awake Rate Brain PET System
DOE Office of Scientific and Technical Information (OSTI.GOV)
Vaska, P.; Vaska, P.; Woody, C.
The first full prototype of the RatCAP PET system, designed to image the brain of a rat while conscious, has been completed. Initial results demonstrated excellent spatial resolution, 1.8 mm FWHM with filtered backprojection and <1.5 mm FWHM with a Monte Carlo based MLEM method. However, noise equivalent countrate studies indicated the need for better timing to mitigate the effect of randoms. Thus, the front-end ASIC has been redesigned to minimize time walk, an accurate coincidence time alignment method has been implemented, and a variance reduction technique for the randoms is being developed. To maximize the quantitative capabilities required formore » neuroscience, corrections are being implemented and validated for positron range and photon noncollinearity, scatter (including outside the field of view), attenuation, randoms, and detector efficiency (deadtime is negligible). In addition, a more robust and compact PCI-based optical data acquisition system has been built to replace the original VME-based system while retaining the linux-based data processing and image reconstruction codes. Finally, a number of new animal imaging experiments have been carried out to demonstrate the performance of the RatCAP in real imaging situations, including an F-18 fluoride bone scan, a C-11 raclopride scan, and a dynamic C-11 methamphetamine scan.« less
NASA Astrophysics Data System (ADS)
Mu, Wei; Qi, Jin; Lu, Hong; Schabath, Matthew; Balagurunathan, Yoganand; Tunali, Ilke; Gillies, Robert James
2018-02-01
Purpose: Investigate the ability of using complementary information provided by the fusion of PET/CT images to predict immunotherapy response in non-small cell lung cancer (NSCLC) patients. Materials and methods: We collected 64 patients diagnosed with primary NSCLC treated with anti PD-1 checkpoint blockade. Using PET/CT images, fused images were created following multiple methodologies, resulting in up to 7 different images for the tumor region. Quantitative image features were extracted from the primary image (PET/CT) and the fused images, which included 195 from primary images and 1235 features from the fusion images. Three clinical characteristics were also analyzed. We then used support vector machine (SVM) classification models to identify discriminant features that predict immunotherapy response at baseline. Results: A SVM built with 87 fusion features and 13 primary PET/CT features on validation dataset had an accuracy and area under the ROC curve (AUROC) of 87.5% and 0.82, respectively, compared to a model built with 113 original PET/CT features on validation dataset 78.12% and 0.68. Conclusion: The fusion features shows better ability to predict immunotherapy response prediction compared to individual image features.
Three-pass protocol scheme for bitmap image security by using vernam cipher algorithm
NASA Astrophysics Data System (ADS)
Rachmawati, D.; Budiman, M. A.; Aulya, L.
2018-02-01
Confidentiality, integrity, and efficiency are the crucial aspects of data security. Among the other digital data, image data is too prone to abuse of operation like duplication, modification, etc. There are some data security techniques, one of them is cryptography. The security of Vernam Cipher cryptography algorithm is very dependent on the key exchange process. If the key is leaked, security of this algorithm will collapse. Therefore, a method that minimizes key leakage during the exchange of messages is required. The method which is used, is known as Three-Pass Protocol. This protocol enables message delivery process without the key exchange. Therefore, the sending messages process can reach the receiver safely without fear of key leakage. The system is built by using Java programming language. The materials which are used for system testing are image in size 200×200 pixel, 300×300 pixel, 500×500 pixel, 800×800 pixel and 1000×1000 pixel. The result of experiments showed that Vernam Cipher algorithm in Three-Pass Protocol scheme could restore the original image.
Le, T Hoang Ngan; Luu, Khoa; Savvides, Marios
2013-08-01
Robust facial hair detection and segmentation is a highly valued soft biometric attribute for carrying out forensic facial analysis. In this paper, we propose a novel and fully automatic system, called SparCLeS, for beard/moustache detection and segmentation in challenging facial images. SparCLeS uses the multiscale self-quotient (MSQ) algorithm to preprocess facial images and deal with illumination variation. Histogram of oriented gradients (HOG) features are extracted from the preprocessed images and a dynamic sparse classifier is built using these features to classify a facial region as either containing skin or facial hair. A level set based approach, which makes use of the advantages of both global and local information, is then used to segment the regions of a face containing facial hair. Experimental results demonstrate the effectiveness of our proposed system in detecting and segmenting facial hair regions in images drawn from three databases, i.e., the NIST Multiple Biometric Grand Challenge (MBGC) still face database, the NIST Color Facial Recognition Technology FERET database, and the Labeled Faces in the Wild (LFW) database.
Multifunctional microcontrollable interface module
NASA Astrophysics Data System (ADS)
Spitzer, Mark B.; Zavracky, Paul M.; Rensing, Noa M.; Crawford, J.; Hockman, Angela H.; Aquilino, P. D.; Girolamo, Henry J.
2001-08-01
This paper reports the development of a complete eyeglass- mounted computer interface system including display, camera and audio subsystems. The display system provides an SVGA image with a 20 degree horizontal field of view. The camera system has been optimized for face recognition and provides a 19 degree horizontal field of view. A microphone and built-in pre-amp optimized for voice recognition and a speaker on an articulated arm are included for audio. An important feature of the system is a high degree of adjustability and reconfigurability. The system has been developed for testing by the Military Police, in a complete system comprising the eyeglass-mounted interface, a wearable computer, and an RF link. Details of the design, construction, and performance of the eyeglass-based system are discussed.
NASA Astrophysics Data System (ADS)
Mindur, B.; Alimov, S.; Fiutowski, T.; Schulz, C.; Wilpert, T.
2014-12-01
A two-dimensional (2D) position sensitive detector for neutron scattering applications based on low-pressure gas amplification and micro-strip technology was built and tested with an innovative readout electronics and data acquisition system. This detector contains a thin solid neutron converter and was developed for time- and thus wavelength-resolved neutron detection in single-event counting mode, which improves the image contrast in comparison with integrating detectors. The prototype detector of a Micro-Strip Gas Chamber (MSGC) was built with a solid natGd/CsI thermal neutron converter for spatial resolutions of about 100 μm and counting rates up to 107 neutrons/s. For attaining very high spatial resolutions and counting rates via micro-strip readout with centre-of-gravity evaluation of the signal amplitude distributions, a fast, channel-wise, self-triggering ASIC was developed. The front-end chips (MSGCROCs), which are very first signal processing components, are read out into powerful ADC-FPGA boards for on-line data processing and thereafter via Gigabit Ethernet link into the data receiving PC. The workstation PC is controlled by a modular, high performance dedicated software suite. Such a fast and accurate system is crucial for efficient radiography/tomography, diffraction or imaging applications based on high flux thermal neutron beam. In this paper a brief description of the detector concept with its operation principles, readout electronics requirements and design together with the signals processing stages performed in hardware and software are presented. In more detail the neutron test beam conditions and measurement results are reported. The focus of this paper is on the system integration, two dimensional spatial resolution, the time resolution of the readout system and the imaging capabilities of the overall setup. The detection efficiency of the detector prototype is estimated as well.
A motion deblurring method with long/short exposure image pairs
NASA Astrophysics Data System (ADS)
Cui, Guangmang; Hua, Weiping; Zhao, Jufeng; Gong, Xiaoli; Zhu, Liyao
2018-01-01
In this paper, a motion deblurring method with long/short exposure image pairs is presented. The long/short exposure image pairs are captured for the same scene under different exposure time. The image pairs are treated as the input of the deblurring method and more information could be used to obtain a deblurring result with high image quality. Firstly, the luminance equalization process is carried out to the short exposure image. And the blur kernel is estimated with the image pair under the maximum a posteriori (MAP) framework using conjugate gradient algorithm. Then a L0 image smoothing based denoising method is applied to the luminance equalized image. And the final deblurring result is obtained with the gain controlled residual image deconvolution process with the edge map as the gain map. Furthermore, a real experimental optical system is built to capture the image pair in order to demonstrate the effectiveness of the proposed deblurring framework. The long/short image pairs are obtained under different exposure time and camera gain control. Experimental results show that the proposed method could provide a superior deblurring result in both subjective and objective assessment compared with other deblurring approaches.
Calibration and testing of a Raman hyperspectral imaging system to reveal powdered food adulteration
Lohumi, Santosh; Lee, Hoonsoo; Kim, Moon S.; Qin, Jianwei; Kandpal, Lalit Mohan; Bae, Hyungjin; Rahman, Anisur
2018-01-01
The potential adulteration of foodstuffs has led to increasing concern regarding food safety and security, in particular for powdered food products where cheap ground materials or hazardous chemicals can be added to increase the quantity of powder or to obtain the desired aesthetic quality. Due to the resulting potential health threat to consumers, the development of a fast, label-free, and non-invasive technique for the detection of adulteration over a wide range of food products is necessary. We therefore report the development of a rapid Raman hyperspectral imaging technique for the detection of food adulteration and for authenticity analysis. The Raman hyperspectral imaging system comprises of a custom designed laser illumination system, sensing module, and a software interface. Laser illumination system generates a 785 nm laser line of high power, and the Gaussian like intensity distribution of laser beam is shaped by incorporating an engineered diffuser. The sensing module utilize Rayleigh filters, imaging spectrometer, and detector for collection of the Raman scattering signals along the laser line. A custom-built software to acquire Raman hyperspectral images which also facilitate the real time visualization of Raman chemical images of scanned samples. The developed system was employed for the simultaneous detection of Sudan dye and Congo red dye adulteration in paprika powder, and benzoyl peroxide and alloxan monohydrate adulteration in wheat flour at six different concentrations (w/w) from 0.05 to 1%. The collected Raman imaging data of the adulterated samples were analyzed to visualize and detect the adulterant concentrations by generating a binary image for each individual adulterant material. The results obtained based on the Raman chemical images of adulterants showed a strong correlation (R>0.98) between added and pixel based calculated concentration of adulterant materials. This developed Raman imaging system thus, can be considered as a powerful analytical technique for the quality and authenticity analysis of food products. PMID:29708973
Lohumi, Santosh; Lee, Hoonsoo; Kim, Moon S; Qin, Jianwei; Kandpal, Lalit Mohan; Bae, Hyungjin; Rahman, Anisur; Cho, Byoung-Kwan
2018-01-01
The potential adulteration of foodstuffs has led to increasing concern regarding food safety and security, in particular for powdered food products where cheap ground materials or hazardous chemicals can be added to increase the quantity of powder or to obtain the desired aesthetic quality. Due to the resulting potential health threat to consumers, the development of a fast, label-free, and non-invasive technique for the detection of adulteration over a wide range of food products is necessary. We therefore report the development of a rapid Raman hyperspectral imaging technique for the detection of food adulteration and for authenticity analysis. The Raman hyperspectral imaging system comprises of a custom designed laser illumination system, sensing module, and a software interface. Laser illumination system generates a 785 nm laser line of high power, and the Gaussian like intensity distribution of laser beam is shaped by incorporating an engineered diffuser. The sensing module utilize Rayleigh filters, imaging spectrometer, and detector for collection of the Raman scattering signals along the laser line. A custom-built software to acquire Raman hyperspectral images which also facilitate the real time visualization of Raman chemical images of scanned samples. The developed system was employed for the simultaneous detection of Sudan dye and Congo red dye adulteration in paprika powder, and benzoyl peroxide and alloxan monohydrate adulteration in wheat flour at six different concentrations (w/w) from 0.05 to 1%. The collected Raman imaging data of the adulterated samples were analyzed to visualize and detect the adulterant concentrations by generating a binary image for each individual adulterant material. The results obtained based on the Raman chemical images of adulterants showed a strong correlation (R>0.98) between added and pixel based calculated concentration of adulterant materials. This developed Raman imaging system thus, can be considered as a powerful analytical technique for the quality and authenticity analysis of food products.
Ex vivo brain tumor analysis using spectroscopic optical coherence tomography
NASA Astrophysics Data System (ADS)
Lenz, Marcel; Krug, Robin; Welp, Hubert; Schmieder, Kirsten; Hofmann, Martin R.
2016-03-01
A big challenge during neurosurgeries is to distinguish between healthy tissue and cancerous tissue, but currently a suitable non-invasive real time imaging modality is not available. Optical Coherence Tomography (OCT) is a potential technique for such a modality. OCT has a penetration depth of 1-2 mm and a resolution of 1-15 μm which is sufficient to illustrate structural differences between healthy tissue and brain tumor. Therefore, we investigated gray and white matter of healthy central nervous system and meningioma samples with a Spectral Domain OCT System (Thorlabs Callisto). Additional OCT images were generated after paraffin embedding and after the samples were cut into 10 μm thin slices for histological investigation with a bright field microscope. All samples were stained with Hematoxylin and Eosin. In all cases B-scans and 3D images were made. Furthermore, a camera image of the investigated area was made by the built-in video camera of our OCT system. For orientation, the backsides of all samples were marked with blue ink. The structural differences between healthy tissue and meningioma samples were most pronounced directly after removal. After paraffin embedding these differences diminished. A correlation between OCT en face images and microscopy images can be seen. In order to increase contrast, post processing algorithms were applied. Hence we employed Spectroscopic OCT, pattern recognition algorithms and machine learning algorithms such as k-means Clustering and Principal Component Analysis.
Riza, Nabeel A; La Torre, Juan Pablo; Amin, M Junaid
2016-06-13
Proposed and experimentally demonstrated is the CAOS-CMOS camera design that combines the coded access optical sensor (CAOS) imager platform with the CMOS multi-pixel optical sensor. The unique CAOS-CMOS camera engages the classic CMOS sensor light staring mode with the time-frequency-space agile pixel CAOS imager mode within one programmable optical unit to realize a high dynamic range imager for extreme light contrast conditions. The experimentally demonstrated CAOS-CMOS camera is built using a digital micromirror device, a silicon point-photo-detector with a variable gain amplifier, and a silicon CMOS sensor with a maximum rated 51.3 dB dynamic range. White light imaging of three different brightness simultaneously viewed targets, that is not possible by the CMOS sensor, is achieved by the CAOS-CMOS camera demonstrating an 82.06 dB dynamic range. Applications for the camera include industrial machine vision, welding, laser analysis, automotive, night vision, surveillance and multispectral military systems.
Active illuminated space object imaging and tracking simulation
NASA Astrophysics Data System (ADS)
Yue, Yufang; Xie, Xiaogang; Luo, Wen; Zhang, Feizhou; An, Jianzhu
2016-10-01
Optical earth imaging simulation of a space target in orbit and it's extraction in laser illumination condition were discussed. Based on the orbit and corresponding attitude of a satellite, its 3D imaging rendering was built. General simulation platform was researched, which was adaptive to variable 3D satellite models and relative position relationships between satellite and earth detector system. Unified parallel projection technology was proposed in this paper. Furthermore, we denoted that random optical distribution in laser-illuminated condition was a challenge for object discrimination. Great randomicity of laser active illuminating speckles was the primary factor. The conjunction effects of multi-frame accumulation process and some tracking methods such as Meanshift tracking, contour poid, and filter deconvolution were simulated. Comparison of results illustrates that the union of multi-frame accumulation and contour poid was recommendable for laser active illuminated images, which had capacities of high tracking precise and stability for multiple object attitudes.
Femtowatt incoherent image conversion from mid-infrared light to near-infrared light
NASA Astrophysics Data System (ADS)
Huang, Nan; Liu, Hongjun; Wang, Zhaolu; Han, Jing; Zhang, Shuan
2017-03-01
We report on the experimental conversion imaging of an incoherent continuous-wave dim source from mid-infrared light to near-infrared light with a lowest input power of 31 femtowatt (fW). Incoherent mid-infrared images of light emission from a heat lamp bulb with an adjustable power supply at window wavelengths ranging from 2.9 µm to 3.5 µm are used for upconversion. The sum-frequency generation is realized in a laser cavity with the resonant wavelength of 1064 nm pumped by an LD at 806 nm built around a periodically poled lithium niobate (PPLN) crystal. The converted infrared image in the wavelength range ~785 nm with a resolution of about 120 × 70 is low-noise detected using a silicon-based camera. By optimizing the system parameters, the upconversion quantum efficiency is predicted to be 28% for correctly polarized, on-axis and phase-matching light.
Visual Communications And Image Processing
NASA Astrophysics Data System (ADS)
Hsing, T. Russell; Tzou, Kou-Hu
1989-07-01
This special issue on Visual Communications and Image Processing contains 14 papers that cover a wide spectrum in this fast growing area. For the past few decades, researchers and scientists have devoted their efforts to these fields. Through this long-lasting devotion, we witness today the growing popularity of low-bit-rate video as a convenient tool for visual communication. We also see the integration of high-quality video into broadband digital networks. Today, with more sophisticated processing, clearer and sharper pictures are being restored from blurring and noise. Also, thanks to the advances in digital image processing, even a PC-based system can be built to recognize highly complicated Chinese characters at the speed of 300 characters per minute. This special issue can be viewed as a milestone of visual communications and image processing on its journey to eternity. It presents some overviews on advanced topics as well as some new development in specific subjects.
A multiphoton laser scanning microscope setup for transcranial in vivo brain imaging on mice
NASA Astrophysics Data System (ADS)
Nase, Gabriele; Helm, P. Johannes; Reppen, Trond; Ottersen, Ole Petter
2005-12-01
We describe a multiphoton laser scanning microscope setup for transcranial in vivo brain imaging in mice. The modular system is based on a modified industrial standard Confocal Scanning Laser Microscope (CSLM) and is assembled mainly from commercially available components. A special multifunctional stage, which is optimized for both laser scanning microscopic observation and preparative animal surgery, has been developed and built. The detection unit includes a highly efficient photomultiplier tube installed in a Peltier-cooled thermal box shielding the detector from changes in room temperature and from distortions caused by external electromagnetic fields. The images are recorded using a 12-bit analog-to-digital converter. Depending on the characteristics of the staining, individual nerve cells can be imaged down to at least 100μm below the intact cranium and down to at least 200μm below the opened cranium.
Forman, Bruce H.; Eccles, Randy; Piggins, Judith; Raila, Wayne; Estey, Greg; Barnett, G. Octo
1990-01-01
We have developed a visually oriented, computer-controlled learning environment designed for use by students of gross anatomy. The goals of this module are to reinforce the concepts of organ relationships and topography by using computed axial tomographic (CAT) images accessed from a videodisc integrated with color graphics and to introduce students to cross-sectional radiographic anatomy. We chose to build the program around CAT scan images because they not only provide excellent structural detail but also offer an anatomic orientation (transverse) that complements that used in the dissection laboratory (basically a layer-by-layer, anterior-to-posterior, or coronal approach). Our system, built using a Microsoft Windows-386 based authoring environment which we designed and implemented, integrates text, video images, and graphics into a single screen display. The program allows both user browsing of information, facilitated by hypertext links, and didactic sessions including mini-quizzes for self-assessment.
NASA Astrophysics Data System (ADS)
Weng, Jiawen; Clark, David C.; Kim, Myung K.
2016-05-01
A numerical reconstruction method based on compressive sensing (CS) for self-interference incoherent digital holography (SIDH) is proposed to achieve sectional imaging by single-shot in-line self-interference incoherent hologram. The sensing operator is built up based on the physical mechanism of SIDH according to CS theory, and a recovery algorithm is employed for image restoration. Numerical simulation and experimental studies employing LEDs as discrete point-sources and resolution targets as extended sources are performed to demonstrate the feasibility and validity of the method. The intensity distribution and the axial resolution along the propagation direction of SIDH by angular spectrum method (ASM) and by CS are discussed. The analysis result shows that compared to ASM the reconstruction by CS can improve the axial resolution of SIDH, and achieve sectional imaging. The proposed method may be useful to 3D analysis of dynamic systems.
Simultaneous in vivo positron emission tomography and magnetic resonance imaging.
Catana, Ciprian; Procissi, Daniel; Wu, Yibao; Judenhofer, Martin S; Qi, Jinyi; Pichler, Bernd J; Jacobs, Russell E; Cherry, Simon R
2008-03-11
Positron emission tomography (PET) and magnetic resonance imaging (MRI) are widely used in vivo imaging technologies with both clinical and biomedical research applications. The strengths of MRI include high-resolution, high-contrast morphologic imaging of soft tissues; the ability to image physiologic parameters such as diffusion and changes in oxygenation level resulting from neuronal stimulation; and the measurement of metabolites using chemical shift imaging. PET images the distribution of biologically targeted radiotracers with high sensitivity, but images generally lack anatomic context and are of lower spatial resolution. Integration of these technologies permits the acquisition of temporally correlated data showing the distribution of PET radiotracers and MRI contrast agents or MR-detectable metabolites, with registration to the underlying anatomy. An MRI-compatible PET scanner has been built for biomedical research applications that allows data from both modalities to be acquired simultaneously. Experiments demonstrate no effect of the MRI system on the spatial resolution of the PET system and <10% reduction in the fraction of radioactive decay events detected by the PET scanner inside the MRI. The signal-to-noise ratio and uniformity of the MR images, with the exception of one particular pulse sequence, were little affected by the presence of the PET scanner. In vivo simultaneous PET and MRI studies were performed in mice. Proof-of-principle in vivo MR spectroscopy and functional MRI experiments were also demonstrated with the combined scanner.
NASA Astrophysics Data System (ADS)
Jang, Sun-Joo; Park, Taejin; Shin, Inho; Park, Hyun Sang; Shin, Paul; Oh, Wang-Yuhl
2016-02-01
Optical coherence tomography (OCT) is a useful imaging method for in vivo tissue imaging with deep penetration and high spatial resolution. However, imaging of the beating mouse heart is still challenging due to limited temporal resolution or penetration depth. Here, we demonstrate a multifunctional OCT system for a beating mouse heart, providing various types of visual information about heart pathophysiology with high spatiotemporal resolution and deep tissue imaging. Angiographic imaging and polarization-sensitive (PS) imaging were implemented with the electrocardiogram (ECG)-triggered beam scanning scheme on the high-speed OCT platform (A-line rate: 240 kHz). Depth-resolved local birefringence and the local orientation of the mouse myocardial fiber were visualized from the PS-OCT. ECG-triggered angiographic OCT (AOCT) with the custom-built motion stabilization imaging window provided myocardial vasculature of a beating mouse heart. Mice underwent coronary artery ligation to derive myocardial infarction (MI) and were imaged with the multifunctional OCT system at multiple time points. AOCT and PS-OCT visualize change of functionality of coronary vessels and myocardium respectively at different phases (acute and chronic) of MI in an ischemic mouse heart. Taken together, the integrated imaging of PS-OCT and AOCT would play an important role in study of MI providing multi-dimensional information of the ischemic mouse heart in vivo.
Fast Fiber-Coupled Imaging Devices
DOE Office of Scientific and Technical Information (OSTI.GOV)
Brockington, Samuel; Case, Andrew; Witherspoon, Franklin Douglas
HyperV Technologies Corp. has successfully designed, built and experimentally demonstrated a full scale 1024 pixel 100 MegaFrames/s fiber coupled camera with 12 or 14 bits, and record lengths of 32K frames, exceeding our original performance objectives. This high-pixel-count, fiber optically-coupled, imaging diagnostic can be used for investigating fast, bright plasma events. In Phase 1 of this effort, a 100 pixel fiber-coupled fast streak camera for imaging plasma jet profiles was constructed and successfully demonstrated. The resulting response from outside plasma physics researchers emphasized development of increased pixel performance as a higher priority over increasing pixel count. In this Phase 2more » effort, HyperV therefore focused on increasing the sample rate and bit-depth of the photodiode pixel designed in Phase 1, while still maintaining a long record length and holding the cost per channel to levels which allowed up to 1024 pixels to be constructed. Cost per channel was 53.31 dollars, very close to our original target of $50 per channel. The system consists of an imaging "camera head" coupled to a photodiode bank with an array of optical fibers. The output of these fast photodiodes is then digitized at 100 Megaframes per second and stored in record lengths of 32,768 samples with bit depths of 12 to 14 bits per pixel. Longer record lengths are possible with additional memory. A prototype imaging system with up to 1024 pixels was designed and constructed and used to successfully take movies of very fast moving plasma jets as a demonstration of the camera performance capabilities. Some faulty electrical components on the 64 circuit boards resulted in only 1008 functional channels out of 1024 on this first generation prototype system. We experimentally observed backlit high speed fan blades in initial camera testing and then followed that with full movies and streak images of free flowing high speed plasma jets (at 30-50 km/s). Jet structure and jet collisions onto metal pillars in the path of the plasma jets were recorded in a single shot. This new fast imaging system is an attractive alternative to conventional fast framing cameras for applications and experiments where imaging events using existing techniques are inefficient or impossible. The development of HyperV's new diagnostic was split into two tracks: a next generation camera track, in which HyperV built, tested, and demonstrated a prototype 1024 channel camera at its own facility, and a second plasma community beta test track, where selected plasma physics programs received small systems of a few test pixels to evaluate the expected performance of a full scale camera on their experiments. These evaluations were performed as part of an unfunded collaboration with researchers at Los Alamos National Laboratory and the University of California at Davis. Results from the prototype 1024-pixel camera are discussed, as well as results from the collaborations with test pixel system deployment sites.« less
Beaudette, Kathy; Baac, Hyoung Won; Madore, Wendy-Julie; Villiger, Martin; Godbout, Nicolas; Bouma, Brett E; Boudoux, Caroline
2015-04-01
Double-clad fiber (DCF) is herein used in conjunction with a double-clad fiber coupler (DCFC) to enable simultaneous and co-registered optical coherence tomography (OCT) and laser tissue coagulation. The DCF allows a single channel fiber-optic probe to be shared: i.e. the core propagating the OCT signal while the inner cladding delivers the coagulation laser light. We herein present a novel DCFC designed and built to combine both signals within a DCF (>90% of single-mode transmission; >65% multimode coupling). Potential OCT imaging degradation mechanisms are also investigated and solutions to mitigate them are presented. The combined DCFC-based system was used to induce coagulation of an ex vivo swine esophagus allowing a real-time assessment of thermal dynamic processes. We therefore demonstrate a DCFC-based system combining OCT imaging with laser coagulation through a single fiber, thus enabling both modalities to be performed simultaneously and in a co-registered manner. Such a system enables endoscopic image-guided laser marking of superficial epithelial tissues or laser thermal therapy of epithelial lesions in pathologies such as Barrett's esophagus.
Beaudette, Kathy; Baac, Hyoung Won; Madore, Wendy-Julie; Villiger, Martin; Godbout, Nicolas; Bouma, Brett E.; Boudoux, Caroline
2015-01-01
Double-clad fiber (DCF) is herein used in conjunction with a double-clad fiber coupler (DCFC) to enable simultaneous and co-registered optical coherence tomography (OCT) and laser tissue coagulation. The DCF allows a single channel fiber-optic probe to be shared: i.e. the core propagating the OCT signal while the inner cladding delivers the coagulation laser light. We herein present a novel DCFC designed and built to combine both signals within a DCF (>90% of single-mode transmission; >65% multimode coupling). Potential OCT imaging degradation mechanisms are also investigated and solutions to mitigate them are presented. The combined DCFC-based system was used to induce coagulation of an ex vivo swine esophagus allowing a real-time assessment of thermal dynamic processes. We therefore demonstrate a DCFC-based system combining OCT imaging with laser coagulation through a single fiber, thus enabling both modalities to be performed simultaneously and in a co-registered manner. Such a system enables endoscopic image-guided laser marking of superficial epithelial tissues or laser thermal therapy of epithelial lesions in pathologies such as Barrett’s esophagus. PMID:25909013
Watching elderly and disabled person's physical condition by remotely controlled monorail robot
NASA Astrophysics Data System (ADS)
Nagasaka, Yasunori; Matsumoto, Yoshinori; Fukaya, Yasutoshi; Takahashi, Tomoichi; Takeshita, Toru
2001-10-01
We are developing a nursing system using robots and cameras. The cameras are mounted on a remote controlled monorail robot which moves inside a room and watches the elderly. It is necessary to pay attention to the elderly at home or nursing homes all time. This requires staffs to pay attention to them at every time. The purpose of our system is to help those staffs. This study intends to improve such situation. A host computer controls a monorail robot to go in front of the elderly using the images taken by cameras on the ceiling. A CCD camera is mounted on the monorail robot to take pictures of their facial expression or movements. The robot sends the images to a host computer that checks them whether something unusual happens or not. We propose a simple calibration method for positioning the monorail robots to track the moves of the elderly for keeping their faces at center of camera view. We built a small experiment system, and evaluated our camera calibration method and image processing algorithm.
Identification of Alfalfa Leaf Diseases Using Image Recognition Technology
Qin, Feng; Liu, Dongxia; Sun, Bingda; Ruan, Liu; Ma, Zhanhong; Wang, Haiguang
2016-01-01
Common leaf spot (caused by Pseudopeziza medicaginis), rust (caused by Uromyces striatus), Leptosphaerulina leaf spot (caused by Leptosphaerulina briosiana) and Cercospora leaf spot (caused by Cercospora medicaginis) are the four common types of alfalfa leaf diseases. Timely and accurate diagnoses of these diseases are critical for disease management, alfalfa quality control and the healthy development of the alfalfa industry. In this study, the identification and diagnosis of the four types of alfalfa leaf diseases were investigated using pattern recognition algorithms based on image-processing technology. A sub-image with one or multiple typical lesions was obtained by artificial cutting from each acquired digital disease image. Then the sub-images were segmented using twelve lesion segmentation methods integrated with clustering algorithms (including K_means clustering, fuzzy C-means clustering and K_median clustering) and supervised classification algorithms (including logistic regression analysis, Naive Bayes algorithm, classification and regression tree, and linear discriminant analysis). After a comprehensive comparison, the segmentation method integrating the K_median clustering algorithm and linear discriminant analysis was chosen to obtain lesion images. After the lesion segmentation using this method, a total of 129 texture, color and shape features were extracted from the lesion images. Based on the features selected using three methods (ReliefF, 1R and correlation-based feature selection), disease recognition models were built using three supervised learning methods, including the random forest, support vector machine (SVM) and K-nearest neighbor methods. A comparison of the recognition results of the models was conducted. The results showed that when the ReliefF method was used for feature selection, the SVM model built with the most important 45 features (selected from a total of 129 features) was the optimal model. For this SVM model, the recognition accuracies of the training set and the testing set were 97.64% and 94.74%, respectively. Semi-supervised models for disease recognition were built based on the 45 effective features that were used for building the optimal SVM model. For the optimal semi-supervised models built with three ratios of labeled to unlabeled samples in the training set, the recognition accuracies of the training set and the testing set were both approximately 80%. The results indicated that image recognition of the four alfalfa leaf diseases can be implemented with high accuracy. This study provides a feasible solution for lesion image segmentation and image recognition of alfalfa leaf disease. PMID:27977767
Identification of Alfalfa Leaf Diseases Using Image Recognition Technology.
Qin, Feng; Liu, Dongxia; Sun, Bingda; Ruan, Liu; Ma, Zhanhong; Wang, Haiguang
2016-01-01
Common leaf spot (caused by Pseudopeziza medicaginis), rust (caused by Uromyces striatus), Leptosphaerulina leaf spot (caused by Leptosphaerulina briosiana) and Cercospora leaf spot (caused by Cercospora medicaginis) are the four common types of alfalfa leaf diseases. Timely and accurate diagnoses of these diseases are critical for disease management, alfalfa quality control and the healthy development of the alfalfa industry. In this study, the identification and diagnosis of the four types of alfalfa leaf diseases were investigated using pattern recognition algorithms based on image-processing technology. A sub-image with one or multiple typical lesions was obtained by artificial cutting from each acquired digital disease image. Then the sub-images were segmented using twelve lesion segmentation methods integrated with clustering algorithms (including K_means clustering, fuzzy C-means clustering and K_median clustering) and supervised classification algorithms (including logistic regression analysis, Naive Bayes algorithm, classification and regression tree, and linear discriminant analysis). After a comprehensive comparison, the segmentation method integrating the K_median clustering algorithm and linear discriminant analysis was chosen to obtain lesion images. After the lesion segmentation using this method, a total of 129 texture, color and shape features were extracted from the lesion images. Based on the features selected using three methods (ReliefF, 1R and correlation-based feature selection), disease recognition models were built using three supervised learning methods, including the random forest, support vector machine (SVM) and K-nearest neighbor methods. A comparison of the recognition results of the models was conducted. The results showed that when the ReliefF method was used for feature selection, the SVM model built with the most important 45 features (selected from a total of 129 features) was the optimal model. For this SVM model, the recognition accuracies of the training set and the testing set were 97.64% and 94.74%, respectively. Semi-supervised models for disease recognition were built based on the 45 effective features that were used for building the optimal SVM model. For the optimal semi-supervised models built with three ratios of labeled to unlabeled samples in the training set, the recognition accuracies of the training set and the testing set were both approximately 80%. The results indicated that image recognition of the four alfalfa leaf diseases can be implemented with high accuracy. This study provides a feasible solution for lesion image segmentation and image recognition of alfalfa leaf disease.
Development of the ARICH monitor system for the Belle II experiment
NASA Astrophysics Data System (ADS)
Hataya, K.; Adachi, I.; Dolenec, R.; Iori, S.; Iwata, S.; Kakuno, H.; Kataura, R.; Kawai, H.; Kindo, H.; Kobayashi, T.; Korpar, S.; Križan, P.; Kumita, T.; Mrvar, M.; Nishida, S.; Ogawa, K.; Ogawa, S.; Pestotnik, R.; Šantelj, L.; Sumiyoshi, T.; Tabata, M.; Yonenaga, M.; Yusa, Y.
2017-12-01
The Belle II detector is under construction at KEK in Japan. In the forward endcap region of the Belle II detector, particle identification (PID) is performed by the Aerogel Ring Imaging Cherenkov (ARICH) counter composed of aerogel tiles and 144-channel Hybrid Avalanche Photo-Detectors (HAPDs). The photon detection efficiency of the photosensor is important for a stable operation of the ARICH. To examine the performance of the HAPDs periodically, a monitor system using scattered photons injected by optical fibers is being developed. In this paper, we report the test using the prototype monitor system and the tests with a partially built ARICH detector.
Full-Frame Reference for Test Photo of Moon
NASA Technical Reports Server (NTRS)
2005-01-01
This pair of views shows how little of the full image frame was taken up by the Moon in test images taken Sept. 8, 2005, by the High Resolution Imaging Science Experiment (HiRISE) camera on NASA's Mars Reconnaissance Orbiter. The Mars-bound camera imaged Earth's Moon from a distance of about 10 million kilometers (6 million miles) away -- 26 times the distance between Earth and the Moon -- as part of an activity to test and calibrate the camera. The images are very significant because they show that the Mars Reconnaissance Orbiter spacecraft and this camera can properly operate together to collect very high-resolution images of Mars. The target must move through the camera's telescope view in just the right direction and speed to acquire a proper image. The day's test images also demonstrate that the focus mechanism works properly with the telescope to produce sharp images. Out of the 20,000-pixel-by-6,000-pixel full frame, the Moon's diameter is about 340 pixels, if the full Moon could be seen. The illuminated crescent is about 60 pixels wide, and the resolution is about 10 kilometers (6 miles) per pixel. At Mars, the entire image region will be filled with high-resolution information. The Mars Reconnaissance Orbiter, launched on Aug. 12, 2005, is on course to reach Mars on March 10, 2006. After gradually adjusting the shape of its orbit for half a year, it will begin its primary science phase in November 2006. From the mission's planned science orbit about 300 kilometers (186 miles) above the surface of Mars, the high resolution camera will be able to discern features as small as one meter or yard across. The Mars Reconnaissance Orbiter mission is managed by NASA's Jet Propulsion Laboratory, a division of the California Institute of Technology, Pasadena, for the NASA Science Mission Directorate. Lockheed Martin Space Systems, Denver, prime contractor for the project, built the spacecraft. Ball Aerospace & Technologies Corp., Boulder, Colo., built the High Resolution Imaging Science Experiment instrument for the University of Arizona, Tucson, to provide to the mission. The HiRISE Operations Center at the University of Arizona processes images from the camera.Effect of contact lens on optical coherence tomography imaging of rodent retina.
Liu, Xiaojing; Wang, Chia-Hao; Dai, Cuixia; Camesa, Adam; Zhang, Hao F; Jiao, Shuliang
2013-12-01
To evaluate the effect of powerless contact lens on improving the quality of optical coherence tomography imaging of rodent retina. A spectral-domain optical coherence tomography (SD-OCT) system was built for in vivo imaging of rodent retina. The calibrated depth resolution of the system was 3 µm in tissue. A commercial powerless contact lens for rat eye was tested in the experiments. For each rat eye, the retina was imaged in vivo sequentially first without wearing contact lens and then with wearing contact lens. The lateral resolution and signal-to-noise ratio of the OCT images with and without contact lens were compared to evaluate the improvement of image quality. The fundus images generated from the measured 3D OCT datasets with contact lens showed sharper retinal blood vessels than those without contact lens. The contrast of the retinal blood vessels was also significantly enhanced in the OCT fundus images with contact lens. As high as 10 dB improvements in SNR was observed for OCT images with contact lens compared to the images of the same retinal area without contact lens. We have demonstrated that the use of powerless contact lens on rat eye can significantly improve OCT image quality of rodent retina, which is a benefit in addition to preventing cataract formation. We believe the improvement in image quality is the result of partial compensation of the optical aberrations of the rodent eye by the contact lens.
SU-E-J-234: Application of a Breathing Motion Model to ViewRay Cine MR Images
DOE Office of Scientific and Technical Information (OSTI.GOV)
O’Connell, D. P.; Thomas, D. H.; Dou, T. H.
2015-06-15
Purpose: A respiratory motion model previously used to generate breathing-gated CT images was used with cine MR images. Accuracy and predictive ability of the in-plane models were evaluated. Methods: Sagittalplane cine MR images of a patient undergoing treatment on a ViewRay MRI/radiotherapy system were acquired before and during treatment. Images were acquired at 4 frames/second with 3.5 × 3.5 mm resolution and a slice thickness of 5 mm. The first cine frame was deformably registered to following frames. Superior/inferior component of the tumor centroid position was used as a breathing surrogate. Deformation vectors and surrogate measurements were used to determinemore » motion model parameters. Model error was evaluated and subsequent treatment cines were predicted from breathing surrogate data. A simulated CT cine was created by generating breathing-gated volumetric images at 0.25 second intervals along the measured breathing trace, selecting a sagittal slice and downsampling to the resolution of the MR cines. A motion model was built using the first half of the simulated cine data. Model accuracy and error in predicting the remaining frames of the cine were evaluated. Results: Mean difference between model predicted and deformably registered lung tissue positions for the 28 second preview MR cine acquired before treatment was 0.81 +/− 0.30 mm. The model was used to predict two minutes of the subsequent treatment cine with a mean accuracy of 1.59 +/− 0.63 mm. Conclusion: Inplane motion models were built using MR cine images and evaluated for accuracy and ability to predict future respiratory motion from breathing surrogate measurements. Examination of long term predictive ability is ongoing. The technique was applied to simulated CT cines for further validation, and the authors are currently investigating use of in-plane models to update pre-existing volumetric motion models used for generation of breathing-gated CT planning images.« less
A boosted optimal linear learner for retinal vessel segmentation
NASA Astrophysics Data System (ADS)
Poletti, E.; Grisan, E.
2014-03-01
Ocular fundus images provide important information about retinal degeneration, which may be related to acute pathologies or to early signs of systemic diseases. An automatic and quantitative assessment of vessel morphological features, such as diameters and tortuosity, can improve clinical diagnosis and evaluation of retinopathy. At variance with available methods, we propose a data-driven approach, in which the system learns a set of optimal discriminative convolution kernels (linear learner). The set is progressively built based on an ADA-boost sample weighting scheme, providing seamless integration between linear learner estimation and classification. In order to capture the vessel appearance changes at different scales, the kernels are estimated on a pyramidal decomposition of the training samples. The set is employed as a rotating bank of matched filters, whose response is used by the boosted linear classifier to provide a classification of each image pixel into the two classes of interest (vessel/background). We tested the approach fundus images available from the DRIVE dataset. We show that the segmentation performance yields an accuracy of 0.94.
Froula, D H; Boni, R; Bedzyk, M; Craxton, R S; Ehrne, F; Ivancic, S; Jungquist, R; Shoup, M J; Theobald, W; Weiner, D; Kugland, N L; Rushford, M C
2012-10-01
A 10-ps, 263-nm (4ω) laser is being built to probe plasmas produced on the OMEGA EP [J. H. Kelly, L. J. Waxer, V. Bagnoud, I. A. Begishev, J. Bromage, B. E. Kruschwitz, T. E. Kessler, S. J. Loucks, D. N. Maywar, R. L. McCrory et al., J. Phys. IV France 133, 75-80 (2006)]. A suite of optical diagnostics (schlieren, interferometry, and grid image refractometry) has been designed to diagnose and characterize a wide variety of plasmas. Light scattered by the probe beam is collected by an f/4 catadioptric telescope and a transport system is designed to image with a near-diffraction-limited resolution (~1 - μm full width at half maximum) over a 5-mm field of view to a diagnostic table. The transport system provides a contrast greater than 1 : 10(4) with respect to all wavelengths outside of the 263 ± 2 nm measurement range.
Speckle Imaging at Gemini and the DCT
NASA Astrophysics Data System (ADS)
Horch, E. P.; Löbb, J.; Howell, S. B.; van Altena, W. F.; Henry, T. J.; van Belle, G. T.
2018-01-01
A program of speckle observations at Lowell Observatory's Discovery Channel Telescope (DCT) and the Gemini North and South Telescopes will be described. It has featured the Differential Speckle Survey Instrument (DSSI), built at Southern Connecticut State University in 2008. DSSI is a dual-port system that records speckle images in two colors simultaneously and produces diffraction limited images to V˜ 16.5 mag at Gemini and V˜ 14.5 mag at the DCT. Of the several science projects that are being pursued at these telescopes, three will be highlighted here. The first is high-resolution follow-up observations for Kepler and K2 exoplanet missions, the second is a study of metal-poor spectroscopic binaries in an attempt to resolve these systems and determine their visual orbits en route to making mass determinations, and the third is a systematic survey of nearby late-type dwarfs, where the multiplicity fraction will be directly measured and compared to that of G dwarfs. The current status of these projects is discussed and some representative results are given.
NASA Astrophysics Data System (ADS)
Froula, D. H.; Boni, R.; Bedzyk, M.; Craxton, R. S.; Ehrne, F.; Ivancic, S.; Jungquist, R.; Shoup, M. J.; Theobald, W.; Weiner, D.; Kugland, N. L.; Rushford, M. C.
2012-10-01
A 10-ps, 263-nm (4ω) laser is being built to probe plasmas produced on the OMEGA EP [J. H. Kelly, L. J. Waxer, V. Bagnoud, I. A. Begishev, J. Bromage, B. E. Kruschwitz, T. E. Kessler, S. J. Loucks, D. N. Maywar, R. L. McCrory et al., J. Phys. IV France 133, 75-80 (2006)], 10.1051/jp4:2006133015. A suite of optical diagnostics (schlieren, interferometry, and grid image refractometry) has been designed to diagnose and characterize a wide variety of plasmas. Light scattered by the probe beam is collected by an f/4 catadioptric telescope and a transport system is designed to image with a near-diffraction-limited resolution (˜1 - μm full width at half maximum) over a 5-mm field of view to a diagnostic table. The transport system provides a contrast greater than 1 : 104 with respect to all wavelengths outside of the 263 ± 2 nm measurement range.
NASA Technical Reports Server (NTRS)
2005-01-01
[figure removed for brevity, see original site] Context image for PIA03287 Windstreak This beautiful windstreak is located on the lava flows from Arsia Mons. Image information: VIS instrument. Latitude -17.0N, Longitude 229.2E. 17 meter/pixel resolution. Note: this THEMIS visual image has not been radiometrically nor geometrically calibrated for this preliminary release. An empirical correction has been performed to remove instrumental effects. A linear shift has been applied in the cross-track and down-track direction to approximate spacecraft and planetary motion. Fully calibrated and geometrically projected images will be released through the Planetary Data System in accordance with Project policies at a later time. NASA's Jet Propulsion Laboratory manages the 2001 Mars Odyssey mission for NASA's Office of Space Science, Washington, D.C. The Thermal Emission Imaging System (THEMIS) was developed by Arizona State University, Tempe, in collaboration with Raytheon Santa Barbara Remote Sensing. The THEMIS investigation is led by Dr. Philip Christensen at Arizona State University. Lockheed Martin Astronautics, Denver, is the prime contractor for the Odyssey project, and developed and built the orbiter. Mission operations are conducted jointly from Lockheed Martin and from JPL, a division of the California Institute of Technology in Pasadena.NASA Technical Reports Server (NTRS)
2006-01-01
[figure removed for brevity, see original site] Context image for PIA03676 Linear Clouds This interesting deposit is located on the floor of Becquerel Crater. Image information: VIS instrument. Latitude 21.3N, Longitude 352.2E. 18 meter/pixel resolution. Note: this THEMIS visual image has not been radiometrically nor geometrically calibrated for this preliminary release. An empirical correction has been performed to remove instrumental effects. A linear shift has been applied in the cross-track and down-track direction to approximate spacecraft and planetary motion. Fully calibrated and geometrically projected images will be released through the Planetary Data System in accordance with Project policies at a later time. NASA's Jet Propulsion Laboratory manages the 2001 Mars Odyssey mission for NASA's Office of Space Science, Washington, D.C. The Thermal Emission Imaging System (THEMIS) was developed by Arizona State University, Tempe, in collaboration with Raytheon Santa Barbara Remote Sensing. The THEMIS investigation is led by Dr. Philip Christensen at Arizona State University. Lockheed Martin Astronautics, Denver, is the prime contractor for the Odyssey project, and developed and built the orbiter. Mission operations are conducted jointly from Lockheed Martin and from JPL, a division of the California Institute of Technology in Pasadena.Imaging Beyond What Man Can See
NASA Technical Reports Server (NTRS)
May, George; Mitchell, Brian
2004-01-01
Three lightweight, portable hyperspectral sensor systems have been built that capture energy from 200 to 1700 nanometers (ultravio1et to shortwave infrared). The sensors incorporate a line scanning technique that requires no relative movement between the target and the sensor. This unique capability, combined with portability, opens up new uses of hyperspectral imaging for laboratory and field environments. Each system has a GUI-based software package that allows the user to communicate with the imaging device for setting spatial resolution, spectral bands and other parameters. NASA's Space Partnership Development has sponsored these innovative developments and their application to human problems on Earth and in space. Hyperspectral datasets have been captured and analyzed in numerous areas including precision agriculture, food safety, biomedical imaging, and forensics. Discussion on research results will include realtime detection of food contaminants, molds and toxin research on corn, identifying counterfeit documents, non-invasive wound monitoring and aircraft applications. Future research will include development of a thermal infrared hyperspectral sensor that will support natural resource applications on Earth and thermal analyses during long duration space flight. This paper incorporates a variety of disciplines and imaging technologies that have been linked together to allow the expansion of remote sensing across both traditional and non-traditional boundaries.
NASA Technical Reports Server (NTRS)
2006-01-01
[figure removed for brevity, see original site] Context image for PIA02160 Landslide This large landslide is located within Ganges Chasma. Image information: VIS instrument. Latitude -7.6N, Longitude 315.8E. 17 meter/pixel resolution. Note: this THEMIS visual image has not been radiometrically nor geometrically calibrated for this preliminary release. An empirical correction has been performed to remove instrumental effects. A linear shift has been applied in the cross-track and down-track direction to approximate spacecraft and planetary motion. Fully calibrated and geometrically projected images will be released through the Planetary Data System in accordance with Project policies at a later time. NASA's Jet Propulsion Laboratory manages the 2001 Mars Odyssey mission for NASA's Office of Space Science, Washington, D.C. The Thermal Emission Imaging System (THEMIS) was developed by Arizona State University, Tempe, in collaboration with Raytheon Santa Barbara Remote Sensing. The THEMIS investigation is led by Dr. Philip Christensen at Arizona State University. Lockheed Martin Astronautics, Denver, is the prime contractor for the Odyssey project, and developed and built the orbiter. Mission operations are conducted jointly from Lockheed Martin and from JPL, a division of the California Institute of Technology in Pasadena.NASA Technical Reports Server (NTRS)
2006-01-01
[figure removed for brevity, see original site] Context image for PIA06088 Crater Landslide This landslide occurs in an unnamed crater southeast of Millochau Crater. Image information: VIS instrument. Latitude -24.4N, Longitude 87.5E. 17 meter/pixel resolution. Note: this THEMIS visual image has not been radiometrically nor geometrically calibrated for this preliminary release. An empirical correction has been performed to remove instrumental effects. A linear shift has been applied in the cross-track and down-track direction to approximate spacecraft and planetary motion. Fully calibrated and geometrically projected images will be released through the Planetary Data System in accordance with Project policies at a later time. NASA's Jet Propulsion Laboratory manages the 2001 Mars Odyssey mission for NASA's Office of Space Science, Washington, D.C. The Thermal Emission Imaging System (THEMIS) was developed by Arizona State University, Tempe, in collaboration with Raytheon Santa Barbara Remote Sensing. The THEMIS investigation is led by Dr. Philip Christensen at Arizona State University. Lockheed Martin Astronautics, Denver, is the prime contractor for the Odyssey project, and developed and built the orbiter. Mission operations are conducted jointly from Lockheed Martin and from JPL, a division of the California Institute of Technology in Pasadena.NASA Technical Reports Server (NTRS)
2005-01-01
[figure removed for brevity, see original site] Context image for PIA03582 Landslide This landslide occurred in Coprates Chasma. Image information: VIS instrument. Latitude 12.6S, Longitude 296.9E. 17 meter/pixel resolution. Note: this THEMIS visual image has not been radiometrically nor geometrically calibrated for this preliminary release. An empirical correction has been performed to remove instrumental effects. A linear shift has been applied in the cross-track and down-track direction to approximate spacecraft and planetary motion. Fully calibrated and geometrically projected images will be released through the Planetary Data System in accordance with Project policies at a later time. NASA's Jet Propulsion Laboratory manages the 2001 Mars Odyssey mission for NASA's Office of Space Science, Washington, D.C. The Thermal Emission Imaging System (THEMIS) was developed by Arizona State University, Tempe, in collaboration with Raytheon Santa Barbara Remote Sensing. The THEMIS investigation is led by Dr. Philip Christensen at Arizona State University. Lockheed Martin Astronautics, Denver, is the prime contractor for the Odyssey project, and developed and built the orbiter. Mission operations are conducted jointly from Lockheed Martin and from JPL, a division of the California Institute of Technology in Pasadena.Research on the feature set construction method for spherical stereo vision
NASA Astrophysics Data System (ADS)
Zhu, Junchao; Wan, Li; Röning, Juha; Feng, Weijia
2015-01-01
Spherical stereo vision is a kind of stereo vision system built by fish-eye lenses, which discussing the stereo algorithms conform to the spherical model. Epipolar geometry is the theory which describes the relationship of the two imaging plane in cameras for the stereo vision system based on perspective projection model. However, the epipolar in uncorrected fish-eye image will not be a line but an arc which intersects at the poles. It is polar curve. In this paper, the theory of nonlinear epipolar geometry will be explored and the method of nonlinear epipolar rectification will be proposed to eliminate the vertical parallax between two fish-eye images. Maximally Stable Extremal Region (MSER) utilizes grayscale as independent variables, and uses the local extremum of the area variation as the testing results. It is demonstrated in literatures that MSER is only depending on the gray variations of images, and not relating with local structural characteristics and resolution of image. Here, MSER will be combined with the nonlinear epipolar rectification method proposed in this paper. The intersection of the rectified epipolar and the corresponding MSER region is determined as the feature set of spherical stereo vision. Experiments show that this study achieved the expected results.
NASA Technical Reports Server (NTRS)
2005-01-01
[figure removed for brevity, see original site] Context image for PIA03637 Galle Cr. Dunes These dunes are located on the floor of Galle Crater. Image information: VIS instrument. Latitude 51.5S, Longitude 329.0E. 17 meter/pixel resolution. Note: this THEMIS visual image has not been radiometrically nor geometrically calibrated for this preliminary release. An empirical correction has been performed to remove instrumental effects. A linear shift has been applied in the cross-track and down-track direction to approximate spacecraft and planetary motion. Fully calibrated and geometrically projected images will be released through the Planetary Data System in accordance with Project policies at a later time. NASA's Jet Propulsion Laboratory manages the 2001 Mars Odyssey mission for NASA's Office of Space Science, Washington, D.C. The Thermal Emission Imaging System (THEMIS) was developed by Arizona State University, Tempe, in collaboration with Raytheon Santa Barbara Remote Sensing. The THEMIS investigation is led by Dr. Philip Christensen at Arizona State University. Lockheed Martin Astronautics, Denver, is the prime contractor for the Odyssey project, and developed and built the orbiter. Mission operations are conducted jointly from Lockheed Martin and from JPL, a division of the California Institute of Technology in Pasadena.NASA Astrophysics Data System (ADS)
Qin, Yuanwei; Xiao, Xiangming; Dong, Jinwei; Chen, Bangqian; Liu, Fang; Zhang, Geli; Zhang, Yao; Wang, Jie; Wu, Xiaocui
2017-02-01
Built-up area supports human settlements and activities, and its spatial distribution and temporal dynamics have significant impacts on ecosystem services and global environment change. To date, most of urban remote sensing has generated the maps of impervious surfaces, and limited effort has been made to explicitly identify the area, location and density of built-up in the complex and fragmented landscapes based on the freely available datasets. In this study, we took the lower Yangtze River Delta (Landsat Path/Row: 118/038), China, where extensive urbanization and industrialization have occurred, as a case study site. We analyzed the structure and optical features of typical land cover types from (1) the HH and HV gamma-naught imagery from the Advanced Land Observation Satellite (ALOS) Phased Array type L-band Synthetic Aperture Radar (PALSAR), and (2) time series Landsat imagery. We proposed a pixel- and rule-based decision tree approach to identify and map built-up area at 30-m resolution from 2007 to 2010, using PALSAR HH gamma-naught and Landsat annual maximum Normalized Difference Vegetation Index (NDVImax). The accuracy assessment showed that the resultant annual maps of built-up had relatively high user (87-93%) and producer accuracies (91-95%) from 2007 to 2010. The built-up area was 2805 km2 in 2010, about 16% of the total land area of the study site. The annual maps of built-up in 2007-2010 show relatively small changes in the urban core regions, but large outward expansion along the peri-urban regions. The average annual increase of built-up areas was about 80 km2 per year from 2007 to 2010. Our annual maps of built-up in the lower Yangtze River Delta clearly complement the existing maps of impervious surfaces in the region. This study provides a promising new approach to identify and map built-up area, which is critical to investigate the interactions between human activities and ecosystem services in urban-rural systems.
Design of embedded endoscopic ultrasonic imaging system
NASA Astrophysics Data System (ADS)
Li, Ming; Zhou, Hao; Wen, Shijie; Chen, Xiodong; Yu, Daoyin
2008-12-01
Endoscopic ultrasonic imaging system is an important component in the endoscopic ultrasonography system (EUS). Through the ultrasonic probe, the characteristics of the fault histology features of digestive organs is detected by EUS, and then received by the reception circuit which making up of amplifying, gain compensation, filtering and A/D converter circuit, in the form of ultrasonic echo. Endoscopic ultrasonic imaging system is the back-end processing system of the EUS, with the function of receiving digital ultrasonic echo modulated by the digestive tract wall from the reception circuit, acquiring and showing the fault histology features in the form of image and characteristic data after digital signal processing, such as demodulation, etc. Traditional endoscopic ultrasonic imaging systems are mainly based on image acquisition and processing chips, which connecting to personal computer with USB2.0 circuit, with the faults of expensive, complicated structure, poor portability, and difficult to popularize. To against the shortcomings above, this paper presents the methods of digital signal acquisition and processing specially based on embedded technology with the core hardware structure of ARM and FPGA for substituting the traditional design with USB2.0 and personal computer. With built-in FIFO and dual-buffer, FPGA implement the ping-pong operation of data storage, simultaneously transferring the image data into ARM through the EBI bus by DMA function, which is controlled by ARM to carry out the purpose of high-speed transmission. The ARM system is being chosen to implement the responsibility of image display every time DMA transmission over and actualizing system control with the drivers and applications running on the embedded operating system Windows CE, which could provide a stable, safe and reliable running platform for the embedded device software. Profiting from the excellent graphical user interface (GUI) and good performance of Windows CE, we can not only clearly show 511×511 pixels ultrasonic echo images through application program, but also provide a simple and friendly operating interface with mouse and touch screen which is more convenient than the traditional endoscopic ultrasonic imaging system. Including core and peripheral circuits of FPGA and ARM, power network circuit and LCD display circuit, we designed the whole embedded system, achieving the desired purpose by implementing ultrasonic image display properly after the experimental verification, solving the problem of hugeness and complexity of the traditional endoscopic ultrasonic imaging system.
Lawrence and Kelly's hands on controls in the Destiny laboratory module
2005-08-05
S114-E-7493 (5 August 2005) --- This image features a close-up view the hands of astronauts Wendy B. Lawrence, STS-114 mission specialist, and James M. Kelly, pilot, at the Mobile Service System (MSS) and Canadarm2 controls in the Destiny laboratory of the International Space Station while Space Shuttle Discovery was docked to the Station. The two were re-stowing the Italian-built Raffaello Multi-Purpose Logistics Module (MPLM) in the cargo bay.
Long-life 3-axis satellite attitude sensing, phase 1
NASA Technical Reports Server (NTRS)
Arild, Tor
1987-01-01
The purpose was to investigate the feasibility of new, moderate-cost, high reliability navigation sensors for high-altitude satellites, using stellar sources to obviate the use of gyroscopic devices. The primary investigation focused on the need for developing a star tracker model to replace an old star tracker which is still needed for current probe and satellite programs. One innovative element of the proposed star tracker was the design, development, and testing of technology components related to a phase scrambler plate. The purpose of the phase scrambler plate is to convert the impulse response of the optical system from a point image to a uniformly bright, square, angularly large, in-focus image of the star source. A collimated star source was built and tested. A breadboard star tracker with an 8 x 8 degree field of view was designed and built. It was tested in normal quad-cell mode (without the phase scrambler plate) and with the phase scrambler plate. Although the phase scrambler plate was crudely made, the performance of the star tracker breadboard was greatly improved using the phase scrambler plate, instead of system defocus. If further developed, the phase scrambler plate may be added as a low-cost retroconversion to any objective lens to greatly improve quad-cell or CCD array tracking; applications include star trackers, laser metrology, laser machining optics, and surveying instrumentation.
Photoacoustic Imaging with a Commercial Ultrasound System and a Custom Probe
Wang, Xueding; Fowlkes, J. Brian; Cannata, Jonathan M.; Hu, Changhong; Carson, Paul L.
2010-01-01
Building photoacoustic imaging (PAI) systems by using stand-alone ultrasound (US) units makes it convenient to take advantage of the state-of-the-art ultrasonic technologies. However, the sometimes limited receiving sensitivity and the comparatively narrow bandwidth of commercial US probes may not be sufficient to acquire high quality photoacoustic images. In this work, a high-speed PAI system has been developed using a commercial US unit and a custom built 128-element piezoelectric-polymer array (PPA) probe using a P(VDF-TrFE) film and flexible circuit to define the elements. Since the US unit supports simultaneous signal acquisition from 64 parallel receive channels, PAI data for synthetic image formation from a 64 or 128 element array aperture can be acquired after a single or dual laser firing, respectively. Therefore, 2D B-scan imaging can be achieved with a maximum frame rate up to 10 Hz, limited only by the laser repetition rate. The uniquely properties of P(VDF-TrFE) facilitated a wide -6 dB receiving bandwidth of over 120 % for the array. A specially designed 128-channel preamplifier board made the connection between the array and the system cable which not only enabled element electrical impedance matching but also further elevated the signal-to-noise ratio (SNR) to further enhance the detection of weak photoacoustic signals. Through the experiments on phantoms and rabbit ears, the good performance of this PAI system was demonstrated. PMID:21276653
Sochi, Russia Winter Olympic Sites Coastal Cluster
2014-02-05
This image acquired by NASA Terra spacecraft shows the Sochi Olympic Park Coastal Cluster, the circular area on the shoreline in the bottom center of the image, which was built for Olympic indoor sports.
A smartphone-based chip-scale microscope using ambient illumination.
Lee, Seung Ah; Yang, Changhuei
2014-08-21
Portable chip-scale microscopy devices can potentially address various imaging needs in mobile healthcare and environmental monitoring. Here, we demonstrate the adaptation of a smartphone's camera to function as a compact lensless microscope. Unlike other chip-scale microscopy schemes, this method uses ambient illumination as its light source and does not require the incorporation of a dedicated light source. The method is based on the shadow imaging technique where the sample is placed on the surface of the image sensor, which captures direct shadow images under illumination. To improve the image resolution beyond the pixel size, we perform pixel super-resolution reconstruction with multiple images at different angles of illumination, which are captured while the user is manually tilting the device around any ambient light source, such as the sun or a lamp. The lensless imaging scheme allows for sub-micron resolution imaging over an ultra-wide field-of-view (FOV). Image acquisition and reconstruction are performed on the device using a custom-built Android application, constructing a stand-alone imaging device for field applications. We discuss the construction of the device using a commercial smartphone and demonstrate the imaging capabilities of our system.
A smartphone-based chip-scale microscope using ambient illumination
Lee, Seung Ah; Yang, Changhuei
2014-01-01
Portable chip-scale microscopy devices can potentially address various imaging needs in mobile healthcare and environmental monitoring. Here, we demonstrate the adaptation of a smartphone’s camera to function as a compact lensless microscope. Unlike other chip-scale microscopy schemes, this method uses ambient illumination as its light source and does not require the incorporation of a dedicated light source. The method is based on the shadow imaging technique where the sample is placed on the surface of the image sensor, which captures direct shadow images under illumination. To improve the imaging resolution beyond the pixel size, we perform pixel super-resolution reconstruction with multiple images at different angles of illumination, which are captured while the user is manually tilting the device around any ambient light source, such as the sun or a lamp. The lensless imaging scheme allows for sub-micron resolution imaging over an ultra-wide field-of-view (FOV). Image acquisition and reconstruction is performed on the device using a custom-built android application, constructing a stand-alone imaging device for field applications. We discuss the construction of the device using a commercial smartphone and demonstrate the imaging capabilities of our system. PMID:24964209
Body-wide anatomy recognition in PET/CT images
NASA Astrophysics Data System (ADS)
Wang, Huiqian; Udupa, Jayaram K.; Odhner, Dewey; Tong, Yubing; Zhao, Liming; Torigian, Drew A.
2015-03-01
With the rapid growth of positron emission tomography/computed tomography (PET/CT)-based medical applications, body-wide anatomy recognition on whole-body PET/CT images becomes crucial for quantifying body-wide disease burden. This, however, is a challenging problem and seldom studied due to unclear anatomy reference frame and low spatial resolution of PET images as well as low contrast and spatial resolution of the associated low-dose CT images. We previously developed an automatic anatomy recognition (AAR) system [15] whose applicability was demonstrated on diagnostic computed tomography (CT) and magnetic resonance (MR) images in different body regions on 35 objects. The aim of the present work is to investigate strategies for adapting the previous AAR system to low-dose CT and PET images toward automated body-wide disease quantification. Our adaptation of the previous AAR methodology to PET/CT images in this paper focuses on 16 objects in three body regions - thorax, abdomen, and pelvis - and consists of the following steps: collecting whole-body PET/CT images from existing patient image databases, delineating all objects in these images, modifying the previous hierarchical models built from diagnostic CT images to account for differences in appearance in low-dose CT and PET images, automatically locating objects in these images following object hierarchy, and evaluating performance. Our preliminary evaluations indicate that the performance of the AAR approach on low-dose CT images achieves object localization accuracy within about 2 voxels, which is comparable to the accuracies achieved on diagnostic contrast-enhanced CT images. Object recognition on low-dose CT images from PET/CT examinations without requiring diagnostic contrast-enhanced CT seems feasible.
T/R Multi-Chip MMIC Modules for 150 GHz
NASA Technical Reports Server (NTRS)
Samoska, Lorene A.; Pukala, David M.; Soria, Mary M.; Sadowy, Gregory A.
2009-01-01
Modules containing multiple monolithic microwave integrated-circuit (MMIC) chips have been built as prototypes of transmitting/receiving (T/R) modules for millimeter-wavelength radar systems, including phased-array radar systems to be used for diverse purposes that could include guidance and avoidance of hazards for landing spacecraft, imaging systems for detecting hidden weapons, and hazard-avoidance systems for automobiles. Whereas prior landing radar systems have operated at frequencies around 35 GHz, the integrated circuits in this module operate in a frequency band centered at about 150 GHz. The higher frequency (and, hence, shorter wavelength), is expected to make it possible to obtain finer spatial resolution while also using smaller antennas and thereby reducing the sizes and masses of the affected systems.
A combined scanning tunneling microscope-atomic layer deposition tool.
Mack, James F; Van Stockum, Philip B; Iwadate, Hitoshi; Prinz, Fritz B
2011-12-01
We have built a combined scanning tunneling microscope-atomic layer deposition (STM-ALD) tool that performs in situ imaging of deposition. It operates from room temperature up to 200 °C, and at pressures from 1 × 10(-6) Torr to 1 × 10(-2) Torr. The STM-ALD system has a complete passive vibration isolation system that counteracts both seismic and acoustic excitations. The instrument can be used as an observation tool to monitor the initial growth phases of ALD in situ, as well as a nanofabrication tool by applying an electric field with the tip to laterally pattern deposition. In this paper, we describe the design of the tool and demonstrate its capability for atomic resolution STM imaging, atomic layer deposition, and the combination of the two techniques for in situ characterization of deposition.
A novel method for real-time edge-enhancement and its application to pattern recognition
NASA Astrophysics Data System (ADS)
Ge, Huayong; Bai, Enjian; Fan, Hong
2010-11-01
The coupling gain coefficient g is redefined and deduced based on coupling theory, the variant of coupling gain coefficient g for different ΓL and r is analyzed. A new optical system is proposed for image edge-enhancement. It recycles the back signal to amplify the edge signal, which has the advantages of high throughput efficiency and brightness. The optical system is designed and built, and the edge-enhanced image of hand bone is captured electronically by CCD camera. The principle of optical correlation is demonstrated, 3-D correlation distribution of letter H with and without edge-enhancement is simulated, the discrimination capability Iac and the full-width at half maximum intensity (FWHM) are compared for two kinds of correlators. The analysis shows that edge-enhancement preprocessing can improve the performance of correlator effectively.
Esenaliev, Rinat O.
2017-01-01
Abstract. Optoacoustic (photoacoustic) diagnostic modality is a technique that combines high optical contrast and ultrasound spatial resolution. We proposed using the optoacoustic technique for a number of applications, including cancer detection, monitoring of thermotherapy (hyperthermia, coagulation, and freezing), monitoring of cerebral blood oxygenation in patients with traumatic brain injury, neonatal patients, fetuses during late-stage labor, central venous oxygenation monitoring, and total hemoglobin concentration monitoring as well as hematoma detection and characterization. We developed and built optical parametric oscillator-based systems and multiwavelength, fiber-coupled highly compact, laser diode-based systems for optoacoustic imaging, monitoring, and sensing. To provide sufficient output pulse energy, a specially designed fiber-optic system was built and incorporated in ultrasensitive, wideband optoacoustic probes. We performed preclinical and clinical tests of the systems and the optoacoustic probes in backward mode for most of the applications and in forward mode for the breast cancer and cerebral applications. The high pulse energy and repetition rate allowed for rapid data acquisition with high signal-to-noise ratio from cerebral blood vessels, such as the superior sagittal sinus, central veins, and peripheral veins and arteries, as well as from intracranial hematomas. The optoacoustic systems were capable of automatic, real-time, continuous measurements of blood oxygenation in these blood vessels. PMID:28444150
NASA Astrophysics Data System (ADS)
Esenaliev, Rinat O.
2017-09-01
Optoacoustic (photoacoustic) diagnostic modality is a technique that combines high optical contrast and ultrasound spatial resolution. We proposed using the optoacoustic technique for a number of applications, including cancer detection, monitoring of thermotherapy (hyperthermia, coagulation, and freezing), monitoring of cerebral blood oxygenation in patients with traumatic brain injury, neonatal patients, fetuses during late-stage labor, central venous oxygenation monitoring, and total hemoglobin concentration monitoring as well as hematoma detection and characterization. We developed and built optical parametric oscillator-based systems and multiwavelength, fiber-coupled highly compact, laser diode-based systems for optoacoustic imaging, monitoring, and sensing. To provide sufficient output pulse energy, a specially designed fiber-optic system was built and incorporated in ultrasensitive, wideband optoacoustic probes. We performed preclinical and clinical tests of the systems and the optoacoustic probes in backward mode for most of the applications and in forward mode for the breast cancer and cerebral applications. The high pulse energy and repetition rate allowed for rapid data acquisition with high signal-to-noise ratio from cerebral blood vessels, such as the superior sagittal sinus, central veins, and peripheral veins and arteries, as well as from intracranial hematomas. The optoacoustic systems were capable of automatic, real-time, continuous measurements of blood oxygenation in these blood vessels.
NASA Astrophysics Data System (ADS)
Deshpande, Ruchi; DeMarco, John; Liu, Brent J.
2015-03-01
We have developed a comprehensive DICOM RT specific database of retrospective treatment planning data for radiation therapy of head and neck cancer. Further, we have designed and built an imaging informatics module that utilizes this database to perform data mining. The end-goal of this data mining system is to provide radiation therapy decision support for incoming head and neck cancer patients, by identifying best practices from previous patients who had the most similar tumor geometries. Since the performance of such systems often depends on the size and quality of the retrospective database, we have also placed an emphasis on developing infrastructure and strategies to encourage data sharing and participation from multiple institutions. The infrastructure and decision support algorithm have both been tested and evaluated with 51 sets of retrospective treatment planning data of head and neck cancer patients. We will present the overall design and architecture of our system, an overview of our decision support mechanism as well as the results of our evaluation.
An adiabatic demagnetization refrigerator for SIRTF
NASA Technical Reports Server (NTRS)
Timbie, P. T.; Bernstein, G. M.; Richards, P. L.
1989-01-01
An adiabatic demagnetization refrigerator (ADR) has been proposed to cool bolometric infrared detectors on the multiband imaging photometer of the Space Infrared Telescope Facility (SIRTF). One such refrigerator has been built which uses a ferric ammonium alum salt pill suspended by nylon threads in a 3-T solenoid. The resonant modes of this suspension are above 100 Hz. The heat leak to the salt pill is less than 0.5 microW. The system has a hold time at 0.1K of more than 12 h. The cold stage temperature is regulated with a feedback loop that controls the magnetic field. A second, similar refrigerator is being built at a SIRTF prototype to fly on a ballon-borne telescope. It will use a ferromagnetic shield. The possibility of using a high-Tc solenoid-actuated heat switch is also discussed.