Sample records for facilitates deep-tissue imaging

  1. Deep Tissue Fluorescent Imaging in Scattering Specimens Using Confocal Microscopy

    PubMed Central

    Clendenon, Sherry G.; Young, Pamela A.; Ferkowicz, Michael; Phillips, Carrie; Dunn, Kenneth W.

    2015-01-01

    In scattering specimens, multiphoton excitation and nondescanned detection improve imaging depth by a factor of 2 or more over confocal microscopy; however, imaging depth is still limited by scattering. We applied the concept of clearing to deep tissue imaging of highly scattering specimens. Clearing is a remarkably effective approach to improving image quality at depth using either confocal or multiphoton microscopy. Tissue clearing appears to eliminate the need for multiphoton excitation for deep tissue imaging. PMID:21729357

  2. Energy-Looping Nanoparticles: Harnessing Excited-State Absorption for Deep-Tissue Imaging.

    PubMed

    Levy, Elizabeth S; Tajon, Cheryl A; Bischof, Thomas S; Iafrati, Jillian; Fernandez-Bravo, Angel; Garfield, David J; Chamanzar, Maysamreza; Maharbiz, Michel M; Sohal, Vikaas S; Schuck, P James; Cohen, Bruce E; Chan, Emory M

    2016-09-27

    Near infrared (NIR) microscopy enables noninvasive imaging in tissue, particularly in the NIR-II spectral range (1000-1400 nm) where attenuation due to tissue scattering and absorption is minimized. Lanthanide-doped upconverting nanocrystals are promising deep-tissue imaging probes due to their photostable emission in the visible and NIR, but these materials are not efficiently excited at NIR-II wavelengths due to the dearth of lanthanide ground-state absorption transitions in this window. Here, we develop a class of lanthanide-doped imaging probes that harness an energy-looping mechanism that facilitates excitation at NIR-II wavelengths, such as 1064 nm, that are resonant with excited-state absorption transitions but not ground-state absorption. Using computational methods and combinatorial screening, we have identified Tm(3+)-doped NaYF4 nanoparticles as efficient looping systems that emit at 800 nm under continuous-wave excitation at 1064 nm. Using this benign excitation with standard confocal microscopy, energy-looping nanoparticles (ELNPs) are imaged in cultured mammalian cells and through brain tissue without autofluorescence. The 1 mm imaging depths and 2 μm feature sizes are comparable to those demonstrated by state-of-the-art multiphoton techniques, illustrating that ELNPs are a promising class of NIR probes for high-fidelity visualization in cells and tissue.

  3. Shaping field for deep tissue microscopy

    NASA Astrophysics Data System (ADS)

    Colon, J.; Lim, H.

    2015-05-01

    Information capacity of a lossless image-forming system is a conserved property determined by two imaging parameters - the resolution and the field of view (FOV). Adaptive optics improves the former by manipulating the phase, or wavefront, in the pupil plane. Here we describe a homologous approach, namely adaptive field microscopy, which aims to enhance the FOV by controlling the phase, or defocus, in the focal plane. In deep tissue imaging, the useful FOV can be severely limited if the region of interest is buried in a thick sample and not perpendicular to the optic axis. One must acquire many z-scans and reconstruct by post-processing, which exposes tissue to excessive radiation and is also time consuming. We demonstrate the effective FOV can be substantially enhanced by dynamic control of the image plane. Specifically, the tilt of the image plane is continuously adjusted in situ to match the oblique orientation of the sample plane within tissue. The utility of adaptive field microscopy is tested for imaging tissue with non-planar morphology. Ocular tissue of small animals was imaged by two-photon excited fluorescence. Our results show that adaptive field microscopy can utilize the full FOV. The freedom to adjust the image plane to account for the geometrical variations of sample could be extremely useful for 3D biological imaging. Furthermore, it could facilitate rapid surveillance of cellular features within deep tissue while avoiding photo damages, making it suitable for in vivo imaging.

  4. Automatic tissue image segmentation based on image processing and deep learning

    NASA Astrophysics Data System (ADS)

    Kong, Zhenglun; Luo, Junyi; Xu, Shengpu; Li, Ting

    2018-02-01

    Image segmentation plays an important role in multimodality imaging, especially in fusion structural images offered by CT, MRI with functional images collected by optical technologies or other novel imaging technologies. Plus, image segmentation also provides detailed structure description for quantitative visualization of treating light distribution in the human body when incorporated with 3D light transport simulation method. Here we used image enhancement, operators, and morphometry methods to extract the accurate contours of different tissues such as skull, cerebrospinal fluid (CSF), grey matter (GM) and white matter (WM) on 5 fMRI head image datasets. Then we utilized convolutional neural network to realize automatic segmentation of images in a deep learning way. We also introduced parallel computing. Such approaches greatly reduced the processing time compared to manual and semi-automatic segmentation and is of great importance in improving speed and accuracy as more and more samples being learned. Our results can be used as a criteria when diagnosing diseases such as cerebral atrophy, which is caused by pathological changes in gray matter or white matter. We demonstrated the great potential of such image processing and deep leaning combined automatic tissue image segmentation in personalized medicine, especially in monitoring, and treatments.

  5. Photoacoustic design parameter optimization for deep tissue imaging by numerical simulation

    NASA Astrophysics Data System (ADS)

    Wang, Zhaohui; Ha, Seunghan; Kim, Kang

    2012-02-01

    A new design of light illumination scheme for deep tissue photoacoustic (PA) imaging, a light catcher, is proposed and evaluated by in silico simulation. Finite element (FE)-based numerical simulation model was developed for photoacoustic (PA) imaging in soft tissues. In this in silico simulation using a commercially available FE simulation package (COMSOL MultiphysicsTM, COMSOL Inc., USA), a short-pulsed laser point source (pulse length of 5 ns) was placed in water on the tissue surface. Overall, four sets of simulation models were integrated together to describe the physical principles of PA imaging. Light energy transmission through background tissues from the laser source to the target tissue or contrast agent was described by diffusion equation. The absorption of light energy and its conversion to heat by target tissue or contrast agent was modeled using bio-heat equation. The heat then causes the stress and strain change, and the resulting displacement of the target surface produces acoustic pressure. The created wide-band acoustic pressure will propagate through background tissues to the ultrasound detector, which is governed by acoustic wave equation. Both optical and acoustical parameters in soft tissues such as scattering, absorption, and attenuation are incorporated in tissue models. PA imaging performance with different design parameters of the laser source and energy delivery scheme was investigated. The laser light illumination into the deep tissues can be significantly improved by up to 134.8% increase of fluence rate by introducing a designed compact light catcher with highly reflecting inner surface surrounding the light source. The optimized parameters through this simulation will guide the design of PA system for deep tissue imaging, and help to form the base protocols of experimental evaluations in vitro and in vivo.

  6. Volumetric imaging of fast biological dynamics in deep tissue via wavefront engineering

    NASA Astrophysics Data System (ADS)

    Kong, Lingjie; Tang, Jianyong; Cui, Meng

    2016-03-01

    To reveal fast biological dynamics in deep tissue, we combine two wavefront engineering methods that were developed in our laboratory, namely optical phase-locked ultrasound lens (OPLUL) based volumetric imaging and iterative multiphoton adaptive compensation technique (IMPACT). OPLUL is used to generate oscillating defocusing wavefront for fast axial scanning, and IMPACT is used to compensate the wavefront distortions for deep tissue imaging. We show its promising applications in neuroscience and immunology.

  7. Quantum dots versus organic fluorophores in fluorescent deep-tissue imaging--merits and demerits.

    PubMed

    Bakalova, Rumiana; Zhelev, Zhivko; Gadjeva, Veselina

    2008-12-01

    The use of fluorescence in deep-tissue imaging is rapidly expanding in last several years. The progress in fluorescent molecular probes and fluorescent imaging techniques gives an opportunity to detect single cells and even molecular targets in live organisms. The highly sensitive and high-speed fluorescent molecular sensors and detection devices allow the application of fluorescence in functional imaging. With the development of novel bright fluorophores based on nanotechnologies and 3D fluorescence scanners with high spatial and temporal resolution, the fluorescent imaging has a potential to become an alternative of the other non-invasive imaging techniques as magnetic resonance imaging, positron-emission tomography, X-ray, computing tomography. The fluorescent imaging has also a potential to give a real map of human anatomy and physiology. The current review outlines the advantages of fluorescent nanoparticles over conventional organic dyes in deep-tissue imaging in vivo and defines the major requirements to the "perfect fluorophore". The analysis proceeds from the basic principles of fluorescence and major characteristics of fluorophores, light-tissue interactions, and major limitations of fluorescent deep-tissue imaging. The article is addressed to a broad readership - from specialists in this field to university students.

  8. Confocal multispot microscope for fast and deep imaging in semicleared tissues

    NASA Astrophysics Data System (ADS)

    Adam, Marie-Pierre; Müllenbroich, Marie Caroline; Di Giovanna, Antonino Paolo; Alfieri, Domenico; Silvestri, Ludovico; Sacconi, Leonardo; Pavone, Francesco Saverio

    2018-02-01

    Although perfectly transparent specimens are imaged faster with light-sheet microscopy, less transparent samples are often imaged with two-photon microscopy leveraging its robustness to scattering; however, at the price of increased acquisition times. Clearing methods that are capable of rendering strongly scattering samples such as brain tissue perfectly transparent specimens are often complex, costly, and time intensive, even though for many applications a slightly lower level of tissue transparency is sufficient and easily achieved with simpler and faster methods. Here, we present a microscope type that has been geared toward the imaging of semicleared tissue by combining multispot two-photon excitation with rolling shutter wide-field detection to image deep and fast inside semicleared mouse brain. We present a theoretical and experimental evaluation of the point spread function and contrast as a function of shutter size. Finally, we demonstrate microscope performance in fixed brain slices by imaging dendritic spines up to 400-μm deep.

  9. Deep lamellar keratoplasty on air with lyophilised tissue.

    PubMed Central

    Chau, G K; Dilly, S A; Sheard, C E; Rostron, C K

    1992-01-01

    Deep lamellar keratoplasty on air involves injecting air into the corneal stroma to expand it to several times its normal thickness. This method is designed to facilitate dissection of the deep stroma and reduce the risk of perforation of Descemet's membrane when carrying out deep lamellar keratoplasty. We have modified the technique by using prelathed freeze-dried donor tissue and report our results in a series of patients with corneal stromal scarring owing to a variety of corneal problems, namely, keratoconus, pterygium, and herpes zoster ophthalmicus. All patients achieved best corrected postoperative visual acuity of 6/12 or better without problems associated with graft failure or rejection. Histopathological examination of corneal tissue following air injection showed surgical emphysema within the cornea and separation of deep stromal fibres from the underlying Descemet's membrane. Images PMID:1477037

  10. Multiplexed 3D FRET imaging in deep tissue of live embryos

    PubMed Central

    Zhao, Ming; Wan, Xiaoyang; Li, Yu; Zhou, Weibin; Peng, Leilei

    2015-01-01

    Current deep tissue microscopy techniques are mostly restricted to intensity mapping of fluorophores, which significantly limit their applications in investigating biochemical processes in vivo. We present a deep tissue multiplexed functional imaging method that probes multiple Förster resonant energy transfer (FRET) sensors in live embryos with high spatial resolution. The method simultaneously images fluorescence lifetimes in 3D with multiple excitation lasers. Through quantitative analysis of triple-channel intensity and lifetime images, we demonstrated that Ca2+ and cAMP levels of live embryos expressing dual FRET sensors can be monitored simultaneously at microscopic resolution. The method is compatible with a broad range of FRET sensors currently available for probing various cellular biochemical functions. It opens the door to imaging complex cellular circuitries in whole live organisms. PMID:26387920

  11. Deep-tissue focal fluorescence imaging with digitally time-reversed ultrasound-encoded light

    PubMed Central

    Wang, Ying Min; Judkewitz, Benjamin; DiMarzio, Charles A.; Yang, Changhuei

    2012-01-01

    Fluorescence imaging is one of the most important research tools in biomedical sciences. However, scattering of light severely impedes imaging of thick biological samples beyond the ballistic regime. Here we directly show focusing and high-resolution fluorescence imaging deep inside biological tissues by digitally time-reversing ultrasound-tagged light with high optical gain (~5×105). We confirm the presence of a time-reversed optical focus along with a diffuse background—a corollary of partial phase conjugation—and develop an approach for dynamic background cancellation. To illustrate the potential of our method, we image complex fluorescent objects and tumour microtissues at an unprecedented depth of 2.5 mm in biological tissues at a lateral resolution of 36 μm×52 μm and an axial resolution of 657 μm. Our results set the stage for a range of deep-tissue imaging applications in biomedical research and medical diagnostics. PMID:22735456

  12. Resolution enhancement in deep-tissue nanoparticle imaging based on plasmonic saturated excitation microscopy

    NASA Astrophysics Data System (ADS)

    Deka, Gitanjal; Nishida, Kentaro; Mochizuki, Kentaro; Ding, Hou-Xian; Fujita, Katsumasa; Chu, Shi-Wei

    2018-03-01

    Recently, many resolution enhancing techniques are demonstrated, but most of them are severely limited for deep tissue applications. For example, wide-field based localization techniques lack the ability of optical sectioning, and structured light based techniques are susceptible to beam distortion due to scattering/aberration. Saturated excitation (SAX) microscopy, which relies on temporal modulation that is less affected when penetrating into tissues, should be the best candidate for deep-tissue resolution enhancement. Nevertheless, although fluorescence saturation has been successfully adopted in SAX, it is limited by photobleaching, and its practical resolution enhancement is less than two-fold. Recently, we demonstrated plasmonic SAX which provides bleaching-free imaging with three-fold resolution enhancement. Here we show that the three-fold resolution enhancement is sustained throughout the whole working distance of an objective, i.e., 200 μm, which is the deepest super-resolution record to our knowledge, and is expected to extend into deeper tissues. In addition, SAX offers the advantage of background-free imaging by rejecting unwanted scattering background from biological tissues. This study provides an inspirational direction toward deep-tissue super-resolution imaging and has the potential in tumor monitoring and beyond.

  13. Deep-tissue reporter-gene imaging with fluorescence and optoacoustic tomography: a performance overview.

    PubMed

    Deliolanis, Nikolaos C; Ale, Angelique; Morscher, Stefan; Burton, Neal C; Schaefer, Karin; Radrich, Karin; Razansky, Daniel; Ntziachristos, Vasilis

    2014-10-01

    A primary enabling feature of near-infrared fluorescent proteins (FPs) and fluorescent probes is the ability to visualize deeper in tissues than in the visible. The purpose of this work is to find which is the optimal visualization method that can exploit the advantages of this novel class of FPs in full-scale pre-clinical molecular imaging studies. Nude mice were stereotactically implanted with near-infrared FP expressing glioma cells to from brain tumors. The feasibility and performance metrics of FPs were compared between planar epi-illumination and trans-illumination fluorescence imaging, as well as to hybrid Fluorescence Molecular Tomography (FMT) system combined with X-ray CT and Multispectral Optoacoustic (or Photoacoustic) Tomography (MSOT). It is shown that deep-seated glioma brain tumors are possible to visualize both with fluorescence and optoacoustic imaging. Fluorescence imaging is straightforward and has good sensitivity; however, it lacks resolution. FMT-XCT can provide an improved rough resolution of ∼1 mm in deep tissue, while MSOT achieves 0.1 mm resolution in deep tissue and has comparable sensitivity. We show imaging capacity that can shift the visualization paradigm in biological discovery. The results are relevant not only to reporter gene imaging, but stand as cross-platform comparison for all methods imaging near infrared fluorescent contrast agents.

  14. Applications of two-photon fluorescence microscopy in deep-tissue imaging

    NASA Astrophysics Data System (ADS)

    Dong, Chen-Yuan; Yu, Betty; Hsu, Lily L.; Kaplan, Peter D.; Blankschstein, D.; Langer, Robert; So, Peter T. C.

    2000-07-01

    Based on the non-linear excitation of fluorescence molecules, two-photon fluorescence microscopy has become a significant new tool for biological imaging. The point-like excitation characteristic of this technique enhances image quality by the virtual elimination of off-focal fluorescence. Furthermore, sample photodamage is greatly reduced because fluorescence excitation is limited to the focal region. For deep tissue imaging, two-photon microscopy has the additional benefit in the greatly improved imaging depth penetration. Since the near- infrared laser sources used in two-photon microscopy scatter less than their UV/glue-green counterparts, in-depth imaging of highly scattering specimen can be greatly improved. In this work, we will present data characterizing both the imaging characteristics (point-spread-functions) and tissue samples (skin) images using this novel technology. In particular, we will demonstrate how blind deconvolution can be used further improve two-photon image quality and how this technique can be used to study mechanisms of chemically-enhanced, transdermal drug delivery.

  15. In vivo deep tissue fluorescence imaging of the murine small intestine and colon

    NASA Astrophysics Data System (ADS)

    Crosignani, Viera; Dvornikov, Alexander; Aguilar, Jose S.; Stringari, Chiara; Edwards, Roberts; Mantulin, Williams; Gratton, Enrico

    2012-03-01

    Recently we described a novel technical approach with enhanced fluorescence detection capabilities in two-photon microscopy that achieves deep tissue imaging, while maintaining micron resolution. This technique was applied to in vivo imaging of murine small intestine and colon. Individuals with Inflammatory Bowel Disease (IBD), commonly presenting as Crohn's disease or Ulcerative Colitis, are at increased risk for developing colorectal cancer. We have developed a Giα2 gene knock out mouse IBD model that develops colitis and colon cancer. The challenge is to study the disease in the whole animal, while maintaining high resolution imaging at millimeter depth. In the Giα2-/- mice, we have been successful in imaging Lgr5-GFP positive stem cell reporters that are found in crypts of niche structures, as well as deeper structures, in the small intestine and colon at depths greater than 1mm. In parallel with these in vivo deep tissue imaging experiments, we have also pursued autofluorescence FLIM imaging of the colon and small intestine-at more shallow depths (roughly 160μm)- on commercial two photon microscopes with excellent structural correlation (in overlapping tissue regions) between the different technologies.

  16. In Vivo Deep Tissue Fluorescence and Magnetic Imaging Employing Hybrid Nanostructures.

    PubMed

    Ortgies, Dirk H; de la Cueva, Leonor; Del Rosal, Blanca; Sanz-Rodríguez, Francisco; Fernández, Nuria; Iglesias-de la Cruz, M Carmen; Salas, Gorka; Cabrera, David; Teran, Francisco J; Jaque, Daniel; Martín Rodríguez, Emma

    2016-01-20

    Breakthroughs in nanotechnology have made it possible to integrate different nanoparticles in one single hybrid nanostructure (HNS), constituting multifunctional nanosized sensors, carriers, and probes with great potential in the life sciences. In addition, such nanostructures could also offer therapeutic capabilities to achieve a wider variety of multifunctionalities. In this work, the encapsulation of both magnetic and infrared emitting nanoparticles into a polymeric matrix leads to a magnetic-fluorescent HNS with multimodal magnetic-fluorescent imaging abilities. The magnetic-fluorescent HNS are capable of simultaneous magnetic resonance imaging and deep tissue infrared fluorescence imaging, overcoming the tissue penetration limits of classical visible-light based optical imaging as reported here in living mice. Additionally, their applicability for magnetic heating in potential hyperthermia treatments is assessed.

  17. Automated classification of multiphoton microscopy images of ovarian tissue using deep learning.

    PubMed

    Huttunen, Mikko J; Hassan, Abdurahman; McCloskey, Curtis W; Fasih, Sijyl; Upham, Jeremy; Vanderhyden, Barbara C; Boyd, Robert W; Murugkar, Sangeeta

    2018-06-01

    Histopathological image analysis of stained tissue slides is routinely used in tumor detection and classification. However, diagnosis requires a highly trained pathologist and can thus be time-consuming, labor-intensive, and potentially risk bias. Here, we demonstrate a potential complementary approach for diagnosis. We show that multiphoton microscopy images from unstained, reproductive tissues can be robustly classified using deep learning techniques. We fine-train four pretrained convolutional neural networks using over 200 murine tissue images based on combined second-harmonic generation and two-photon excitation fluorescence contrast, to classify the tissues either as healthy or associated with high-grade serous carcinoma with over 95% sensitivity and 97% specificity. Our approach shows promise for applications involving automated disease diagnosis. It could also be readily applied to other tissues, diseases, and related classification problems. (2018) COPYRIGHT Society of Photo-Optical Instrumentation Engineers (SPIE).

  18. Super-nonlinear fluorescence microscopy for high-contrast deep tissue imaging

    NASA Astrophysics Data System (ADS)

    Wei, Lu; Zhu, Xinxin; Chen, Zhixing; Min, Wei

    2014-02-01

    Two-photon excited fluorescence microscopy (TPFM) offers the highest penetration depth with subcellular resolution in light microscopy, due to its unique advantage of nonlinear excitation. However, a fundamental imaging-depth limit, accompanied by a vanishing signal-to-background contrast, still exists for TPFM when imaging deep into scattering samples. Formally, the focusing depth, at which the in-focus signal and the out-of-focus background are equal to each other, is defined as the fundamental imaging-depth limit. To go beyond this imaging-depth limit of TPFM, we report a new class of super-nonlinear fluorescence microscopy for high-contrast deep tissue imaging, including multiphoton activation and imaging (MPAI) harnessing novel photo-activatable fluorophores, stimulated emission reduced fluorescence (SERF) microscopy by adding a weak laser beam for stimulated emission, and two-photon induced focal saturation imaging with preferential depletion of ground-state fluorophores at focus. The resulting image contrasts all exhibit a higher-order (third- or fourth- order) nonlinear signal dependence on laser intensity than that in the standard TPFM. Both the physical principles and the imaging demonstrations will be provided for each super-nonlinear microscopy. In all these techniques, the created super-nonlinearity significantly enhances the imaging contrast and concurrently extends the imaging depth-limit of TPFM. Conceptually different from conventional multiphoton processes mediated by virtual states, our strategy constitutes a new class of fluorescence microscopy where high-order nonlinearity is mediated by real population transfer.

  19. Molecular imaging needles: dual-modality optical coherence tomography and fluorescence imaging of labeled antibodies deep in tissue

    PubMed Central

    Scolaro, Loretta; Lorenser, Dirk; Madore, Wendy-Julie; Kirk, Rodney W.; Kramer, Anne S.; Yeoh, George C.; Godbout, Nicolas; Sampson, David D.; Boudoux, Caroline; McLaughlin, Robert A.

    2015-01-01

    Molecular imaging using optical techniques provides insight into disease at the cellular level. In this paper, we report on a novel dual-modality probe capable of performing molecular imaging by combining simultaneous three-dimensional optical coherence tomography (OCT) and two-dimensional fluorescence imaging in a hypodermic needle. The probe, referred to as a molecular imaging (MI) needle, may be inserted tens of millimeters into tissue. The MI needle utilizes double-clad fiber to carry both imaging modalities, and is interfaced to a 1310-nm OCT system and a fluorescence imaging subsystem using an asymmetrical double-clad fiber coupler customized to achieve high fluorescence collection efficiency. We present, to the best of our knowledge, the first dual-modality OCT and fluorescence needle probe with sufficient sensitivity to image fluorescently labeled antibodies. Such probes enable high-resolution molecular imaging deep within tissue. PMID:26137379

  20. Fast deep-tissue multispectral optoacoustic tomography (MSOT) for preclinical imaging of cancer and cardiovascular disease

    NASA Astrophysics Data System (ADS)

    Taruttis, Adrian; Razansky, Daniel; Ntziachristos, Vasilis

    2012-02-01

    Optoacoustic imaging has enabled the visualization of optical contrast at high resolutions in deep tissue. Our Multispectral optoacoustic tomography (MSOT) imaging results reveal internal tissue heterogeneity, where the underlying distribution of specific endogenous and exogenous sources of absorption can be resolved in detail. Technical advances in cardiac imaging allow motion-resolved multispectral measurements of the heart, opening the way for studies of cardiovascular disease. We further demonstrate the fast characterization of the pharmacokinetic profiles of lightabsorbing agents. Overall, our MSOT findings indicate new possibilities in high resolution imaging of functional and molecular parameters.

  1. Study of Tissue Phantoms, Tissues, and Contrast Agent with the Biophotoacoustic Radar and Comparison to Ultrasound Imaging for Deep Subsurface Imaging

    NASA Astrophysics Data System (ADS)

    Alwi, R.; Telenkov, S.; Mandelis, A.; Gu, F.

    2012-11-01

    In this study, the imaging capability of our wide-spectrum frequency-domain photoacoustic (FD-PA) imaging alias "photoacoustic radar" methodology for imaging of soft tissues is explored. A practical application of the mathematical correlation processing method with relatively long (1 ms) frequency-modulated optical excitation is demonstrated for reconstruction of the spatial location of the PA sources. Image comparison with ultrasound (US) modality was investigated to see the complementarity between the two techniques. The obtained results with a phased array probe on tissue phantoms and their comparison to US images demonstrated that the FD-PA technique has strong potential for deep subsurface imaging with excellent contrast and high signal-to-noise ratio. FD-PA images of blood vessels in a human wrist and an in vivo subcutaneous tumor in a rat model are presented. As in other imaging modalities, the employment of contrast agents is desirable to improve the capability of medical diagnostics. Therefore, this study also evaluated and characterized the use of Food and Drug Administration (FDA)-approved superparamagnetic iron oxide nanoparticles (SPION) as PA contrast agents.

  2. A deep learning approach to estimate chemically-treated collagenous tissue nonlinear anisotropic stress-strain responses from microscopy images.

    PubMed

    Liang, Liang; Liu, Minliang; Sun, Wei

    2017-11-01

    Biological collagenous tissues comprised of networks of collagen fibers are suitable for a broad spectrum of medical applications owing to their attractive mechanical properties. In this study, we developed a noninvasive approach to estimate collagenous tissue elastic properties directly from microscopy images using Machine Learning (ML) techniques. Glutaraldehyde-treated bovine pericardium (GLBP) tissue, widely used in the fabrication of bioprosthetic heart valves and vascular patches, was chosen to develop a representative application. A Deep Learning model was designed and trained to process second harmonic generation (SHG) images of collagen networks in GLBP tissue samples, and directly predict the tissue elastic mechanical properties. The trained model is capable of identifying the overall tissue stiffness with a classification accuracy of 84%, and predicting the nonlinear anisotropic stress-strain curves with average regression errors of 0.021 and 0.031. Thus, this study demonstrates the feasibility and great potential of using the Deep Learning approach for fast and noninvasive assessment of collagenous tissue elastic properties from microstructural images. In this study, we developed, to our best knowledge, the first Deep Learning-based approach to estimate the elastic properties of collagenous tissues directly from noninvasive second harmonic generation images. The success of this study holds promise for the use of Machine Learning techniques to noninvasively and efficiently estimate the mechanical properties of many structure-based biological materials, and it also enables many potential applications such as serving as a quality control tool to select tissue for the manufacturing of medical devices (e.g. bioprosthetic heart valves). Copyright © 2017 Acta Materialia Inc. Published by Elsevier Ltd. All rights reserved.

  3. High-Resolution Ultrasound-Switchable Fluorescence Imaging in Centimeter-Deep Tissue Phantoms with High Signal-To-Noise Ratio and High Sensitivity via Novel Contrast Agents.

    PubMed

    Cheng, Bingbing; Bandi, Venugopal; Wei, Ming-Yuan; Pei, Yanbo; D'Souza, Francis; Nguyen, Kytai T; Hong, Yi; Yuan, Baohong

    2016-01-01

    For many years, investigators have sought after high-resolution fluorescence imaging in centimeter-deep tissue because many interesting in vivo phenomena-such as the presence of immune system cells, tumor angiogenesis, and metastasis-may be located deep in tissue. Previously, we developed a new imaging technique to achieve high spatial resolution in sub-centimeter deep tissue phantoms named continuous-wave ultrasound-switchable fluorescence (CW-USF). The principle is to use a focused ultrasound wave to externally and locally switch on and off the fluorophore emission from a small volume (close to ultrasound focal volume). By making improvements in three aspects of this technique: excellent near-infrared USF contrast agents, a sensitive frequency-domain USF imaging system, and an effective signal processing algorithm, for the first time this study has achieved high spatial resolution (~ 900 μm) in 3-centimeter-deep tissue phantoms with high signal-to-noise ratio (SNR) and high sensitivity (3.4 picomoles of fluorophore in a volume of 68 nanoliters can be detected). We have achieved these results in both tissue-mimic phantoms and porcine muscle tissues. We have also demonstrated multi-color USF to image and distinguish two fluorophores with different wavelengths, which might be very useful for simultaneously imaging of multiple targets and observing their interactions in the future. This work has opened the door for future studies of high-resolution centimeter-deep tissue fluorescence imaging.

  4. Deep Learning MR Imaging-based Attenuation Correction for PET/MR Imaging.

    PubMed

    Liu, Fang; Jang, Hyungseok; Kijowski, Richard; Bradshaw, Tyler; McMillan, Alan B

    2018-02-01

    Purpose To develop and evaluate the feasibility of deep learning approaches for magnetic resonance (MR) imaging-based attenuation correction (AC) (termed deep MRAC) in brain positron emission tomography (PET)/MR imaging. Materials and Methods A PET/MR imaging AC pipeline was built by using a deep learning approach to generate pseudo computed tomographic (CT) scans from MR images. A deep convolutional auto-encoder network was trained to identify air, bone, and soft tissue in volumetric head MR images coregistered to CT data for training. A set of 30 retrospective three-dimensional T1-weighted head images was used to train the model, which was then evaluated in 10 patients by comparing the generated pseudo CT scan to an acquired CT scan. A prospective study was carried out for utilizing simultaneous PET/MR imaging for five subjects by using the proposed approach. Analysis of covariance and paired-sample t tests were used for statistical analysis to compare PET reconstruction error with deep MRAC and two existing MR imaging-based AC approaches with CT-based AC. Results Deep MRAC provides an accurate pseudo CT scan with a mean Dice coefficient of 0.971 ± 0.005 for air, 0.936 ± 0.011 for soft tissue, and 0.803 ± 0.021 for bone. Furthermore, deep MRAC provides good PET results, with average errors of less than 1% in most brain regions. Significantly lower PET reconstruction errors were realized with deep MRAC (-0.7% ± 1.1) compared with Dixon-based soft-tissue and air segmentation (-5.8% ± 3.1) and anatomic CT-based template registration (-4.8% ± 2.2). Conclusion The authors developed an automated approach that allows generation of discrete-valued pseudo CT scans (soft tissue, bone, and air) from a single high-spatial-resolution diagnostic-quality three-dimensional MR image and evaluated it in brain PET/MR imaging. This deep learning approach for MR imaging-based AC provided reduced PET reconstruction error relative to a CT-based standard within the brain compared

  5. Deep-tissue two-photon imaging in brain and peripheral nerve with a compact high-pulse energy ytterbium fiber laser

    NASA Astrophysics Data System (ADS)

    Fontaine, Arjun K.; Kirchner, Matthew S.; Caldwell, John H.; Weir, Richard F.; Gibson, Emily A.

    2018-02-01

    Two-photon microscopy is a powerful tool of current scientific research, allowing optical visualization of structures below the surface of tissues. This is of particular value in neuroscience, where optically accessing regions within the brain is critical for the continued advancement in understanding of neural circuits. However, two-photon imaging at significant depths have typically used Ti:Sapphire based amplifiers that are prohibitively expensive and bulky. In this study, we demonstrate deep tissue two-photon imaging using a compact, inexpensive, turnkey operated Ytterbium fiber laser (Y-Fi, KM Labs). The laser is based on all-normal dispersion (ANDi) that provides short pulse durations and high pulse energies. Depth measurements obtained in ex vivo mouse cortex exceed those obtainable with standard two-photon microscopes using Ti:Sapphire lasers. In addition to demonstrating the capability of deep-tissue imaging in the brain, we investigated imaging depth in highly-scattering white matter with measurements in sciatic nerve showing limited optical penetration of heavily myelinated nerve tissue relative to grey matter.

  6. High-Resolution Ultrasound-Switchable Fluorescence Imaging in Centimeter-Deep Tissue Phantoms with High Signal-To-Noise Ratio and High Sensitivity via Novel Contrast Agents

    PubMed Central

    Cheng, Bingbing; Bandi, Venugopal; Wei, Ming-Yuan; Pei, Yanbo; D’Souza, Francis; Nguyen, Kytai T.; Hong, Yi; Yuan, Baohong

    2016-01-01

    For many years, investigators have sought after high-resolution fluorescence imaging in centimeter-deep tissue because many interesting in vivo phenomena—such as the presence of immune system cells, tumor angiogenesis, and metastasis—may be located deep in tissue. Previously, we developed a new imaging technique to achieve high spatial resolution in sub-centimeter deep tissue phantoms named continuous-wave ultrasound-switchable fluorescence (CW-USF). The principle is to use a focused ultrasound wave to externally and locally switch on and off the fluorophore emission from a small volume (close to ultrasound focal volume). By making improvements in three aspects of this technique: excellent near-infrared USF contrast agents, a sensitive frequency-domain USF imaging system, and an effective signal processing algorithm, for the first time this study has achieved high spatial resolution (~ 900 μm) in 3-centimeter-deep tissue phantoms with high signal-to-noise ratio (SNR) and high sensitivity (3.4 picomoles of fluorophore in a volume of 68 nanoliters can be detected). We have achieved these results in both tissue-mimic phantoms and porcine muscle tissues. We have also demonstrated multi-color USF to image and distinguish two fluorophores with different wavelengths, which might be very useful for simultaneously imaging of multiple targets and observing their interactions in the future. This work has opened the door for future studies of high-resolution centimeter-deep tissue fluorescence imaging. PMID:27829050

  7. Automated assessment of breast tissue density in non-contrast 3D CT images without image segmentation based on a deep CNN

    NASA Astrophysics Data System (ADS)

    Zhou, Xiangrong; Kano, Takuya; Koyasu, Hiromi; Li, Shuo; Zhou, Xinxin; Hara, Takeshi; Matsuo, Masayuki; Fujita, Hiroshi

    2017-03-01

    This paper describes a novel approach for the automatic assessment of breast density in non-contrast three-dimensional computed tomography (3D CT) images. The proposed approach trains and uses a deep convolutional neural network (CNN) from scratch to classify breast tissue density directly from CT images without segmenting the anatomical structures, which creates a bottleneck in conventional approaches. Our scheme determines breast density in a 3D breast region by decomposing the 3D region into several radial 2D-sections from the nipple, and measuring the distribution of breast tissue densities on each 2D section from different orientations. The whole scheme is designed as a compact network without the need for post-processing and provides high robustness and computational efficiency in clinical settings. We applied this scheme to a dataset of 463 non-contrast CT scans obtained from 30- to 45-year-old-women in Japan. The density of breast tissue in each CT scan was assigned to one of four categories (glandular tissue within the breast <25%, 25%-50%, 50%-75%, and >75%) by a radiologist as ground truth. We used 405 CT scans for training a deep CNN and the remaining 58 CT scans for testing the performance. The experimental results demonstrated that the findings of the proposed approach and those of the radiologist were the same in 72% of the CT scans among the training samples and 76% among the testing samples. These results demonstrate the potential use of deep CNN for assessing breast tissue density in non-contrast 3D CT images.

  8. Deep learning based tissue analysis predicts outcome in colorectal cancer.

    PubMed

    Bychkov, Dmitrii; Linder, Nina; Turkki, Riku; Nordling, Stig; Kovanen, Panu E; Verrill, Clare; Walliander, Margarita; Lundin, Mikael; Haglund, Caj; Lundin, Johan

    2018-02-21

    Image-based machine learning and deep learning in particular has recently shown expert-level accuracy in medical image classification. In this study, we combine convolutional and recurrent architectures to train a deep network to predict colorectal cancer outcome based on images of tumour tissue samples. The novelty of our approach is that we directly predict patient outcome, without any intermediate tissue classification. We evaluate a set of digitized haematoxylin-eosin-stained tumour tissue microarray (TMA) samples from 420 colorectal cancer patients with clinicopathological and outcome data available. The results show that deep learning-based outcome prediction with only small tissue areas as input outperforms (hazard ratio 2.3; CI 95% 1.79-3.03; AUC 0.69) visual histological assessment performed by human experts on both TMA spot (HR 1.67; CI 95% 1.28-2.19; AUC 0.58) and whole-slide level (HR 1.65; CI 95% 1.30-2.15; AUC 0.57) in the stratification into low- and high-risk patients. Our results suggest that state-of-the-art deep learning techniques can extract more prognostic information from the tissue morphology of colorectal cancer than an experienced human observer.

  9. Transmission in near-infrared optical windows for deep brain imaging.

    PubMed

    Shi, Lingyan; Sordillo, Laura A; Rodríguez-Contreras, Adrián; Alfano, Robert

    2016-01-01

    Near-infrared (NIR) radiation has been employed using one- and two-photon excitation of fluorescence imaging at wavelengths 650-950 nm (optical window I) for deep brain imaging; however, longer wavelengths in NIR have been overlooked due to a lack of suitable NIR-low band gap semiconductor imaging detectors and/or femtosecond laser sources. This research introduces three new optical windows in NIR and demonstrates their potential for deep brain tissue imaging. The transmittances are measured in rat brain tissue in the second (II, 1,100-1,350 nm), third (III, 1,600-1,870 nm), and fourth (IV, centered at 2,200 nm) NIR optical tissue windows. The relationship between transmission and tissue thickness is measured and compared with the theory. Due to a reduction in scattering and minimal absorption, window III is shown to be the best for deep brain imaging, and windows II and IV show similar but better potential for deep imaging than window I. © 2015 WILEY-VCH Verlag GmbH & Co. KGaA, Weinheim.

  10. Versatile Polymer Nanoparticles as Two-Photon-Triggered Photosensitizers for Simultaneous Cellular, Deep-Tissue Imaging, and Photodynamic Therapy.

    PubMed

    Guo, Liang; Ge, Jiechao; Liu, Qian; Jia, Qingyan; Zhang, Hongyan; Liu, Weimin; Niu, Guangle; Liu, Sha; Gong, Jianru; Hackbarth, Steffen; Wang, Pengfei

    2017-06-01

    Clinical applications of current photodynamic therapy (PDT) photosensitizers (PSs) are often limited by their absorption in the UV-vis range that possesses limited tissue penetration ability, leading to ineffective therapeutic response for deep-seated tumors. Alternatively, two-photon excited PS (TPE-PS) using NIR light triggered is one the most promising candidates for PDT improvement. Herein, multimodal polymer nanoparticles (PNPs) from polythiophene derivative as two-photon fluorescence imaging as well as two-photon-excited PDT agent are developed. The prepared PNPs exhibit excellent water dispersibility, high photostability and pH stability, strong fluorescence brightness, and low dark toxicity. More importantly, the PNPs also possess other outstanding features including: (1) the high 1 O 2 quantum yield; (2) the strong two-photon-induced fluorescence and efficient 1 O 2 generation; (3) the specific accumulation in lysosomes of HeLa cells; and (4) the imaging detection depth up to 2100 µm in the mock tissue under two-photon. The multifunctional PNPs are promising candidates as TPE-PDT agent for simultaneous cellular, deep-tissue imaging, and highly efficient in vivo PDT of cancer. © 2017 WILEY-VCH Verlag GmbH & Co. KGaA, Weinheim.

  11. Breaking the acoustic diffraction limit via nonlinear effect and thermal confinement for potential deep-tissue high-resolution imaging

    PubMed Central

    Yuan, Baohong; Pei, Yanbo; Kandukuri, Jayanth

    2013-01-01

    Our recently developed ultrasound-switchable fluorescence (USF) imaging technique showed that it was feasible to conduct high-resolution fluorescence imaging in a centimeter-deep turbid medium. Because the spatial resolution of this technique highly depends on the ultrasound-induced temperature focal size (UTFS), minimization of UTFS becomes important for further improving the spatial resolution USF technique. In this study, we found that UTFS can be significantly reduced below the diffraction-limited acoustic intensity focal size via nonlinear acoustic effects and thermal confinement by appropriately controlling ultrasound power and exposure time, which can be potentially used for deep-tissue high-resolution imaging. PMID:23479498

  12. Deep learning for tissue microarray image-based outcome prediction in patients with colorectal cancer

    NASA Astrophysics Data System (ADS)

    Bychkov, Dmitrii; Turkki, Riku; Haglund, Caj; Linder, Nina; Lundin, Johan

    2016-03-01

    Recent advances in computer vision enable increasingly accurate automated pattern classification. In the current study we evaluate whether a convolutional neural network (CNN) can be trained to predict disease outcome in patients with colorectal cancer based on images of tumor tissue microarray samples. We compare the prognostic accuracy of CNN features extracted from the whole, unsegmented tissue microarray spot image, with that of CNN features extracted from the epithelial and non-epithelial compartments, respectively. The prognostic accuracy of visually assessed histologic grade is used as a reference. The image data set consists of digitized hematoxylin-eosin (H and E) stained tissue microarray samples obtained from 180 patients with colorectal cancer. The patient samples represent a variety of histological grades, have data available on a series of clinicopathological variables including long-term outcome and ground truth annotations performed by experts. The CNN features extracted from images of the epithelial tissue compartment significantly predicted outcome (hazard ratio (HR) 2.08; CI95% 1.04-4.16; area under the curve (AUC) 0.66) in a test set of 60 patients, as compared to the CNN features extracted from unsegmented images (HR 1.67; CI95% 0.84-3.31, AUC 0.57) and visually assessed histologic grade (HR 1.96; CI95% 0.99-3.88, AUC 0.61). As a conclusion, a deep-learning classifier can be trained to predict outcome of colorectal cancer based on images of H and E stained tissue microarray samples and the CNN features extracted from the epithelial compartment only resulted in a prognostic discrimination comparable to that of visually determined histologic grade.

  13. Deep Learning in Medical Image Analysis.

    PubMed

    Shen, Dinggang; Wu, Guorong; Suk, Heung-Il

    2017-06-21

    This review covers computer-assisted analysis of images in the field of medical imaging. Recent advances in machine learning, especially with regard to deep learning, are helping to identify, classify, and quantify patterns in medical images. At the core of these advances is the ability to exploit hierarchical feature representations learned solely from data, instead of features designed by hand according to domain-specific knowledge. Deep learning is rapidly becoming the state of the art, leading to enhanced performance in various medical applications. We introduce the fundamentals of deep learning methods and review their successes in image registration, detection of anatomical and cellular structures, tissue segmentation, computer-aided disease diagnosis and prognosis, and so on. We conclude by discussing research issues and suggesting future directions for further improvement.

  14. Deep two-photon microscopic imaging through brain tissue using the second singlet state from fluorescent agent chlorophyll α in spinach leaf

    NASA Astrophysics Data System (ADS)

    Shi, Lingyan; Rodríguez-Contreras, Adrián; Budansky, Yury; Pu, Yang; An Nguyen, Thien; Alfano, Robert R.

    2014-06-01

    Two-photon (2P) excitation of the second singlet (S) state was studied to achieve deep optical microscopic imaging in brain tissue when both the excitation (800 nm) and emission (685 nm) wavelengths lie in the "tissue optical window" (650 to 950 nm). S2 state technique was used to investigate chlorophyll α (Chl α) fluorescence inside a spinach leaf under a thick layer of freshly sliced rat brain tissue in combination with 2P microscopic imaging. Strong emission at the peak wavelength of 685 nm under the 2P S state of Chl α enabled the imaging depth up to 450 μm through rat brain tissue.

  15. Deep two-photon microscopic imaging through brain tissue using the second singlet state from fluorescent agent chlorophyll α in spinach leaf.

    PubMed

    Shi, Lingyan; Rodríguez-Contreras, Adrián; Budansky, Yury; Pu, Yang; Nguyen, Thien An; Alfano, Robert R

    2014-06-01

    Two-photon (2P) excitation of the second singlet (S₂) state was studied to achieve deep optical microscopic imaging in brain tissue when both the excitation (800 nm) and emission (685 nm) wavelengths lie in the "tissue optical window" (650 to 950 nm). S₂ state technique was used to investigate chlorophyll α (Chl α) fluorescence inside a spinach leaf under a thick layer of freshly sliced rat brain tissue in combination with 2P microscopic imaging. Strong emission at the peak wavelength of 685 nm under the 2P S₂ state of Chl α enabled the imaging depth up to 450 μm through rat brain tissue.

  16. Deep Learning in Medical Image Analysis

    PubMed Central

    Shen, Dinggang; Wu, Guorong; Suk, Heung-Il

    2016-01-01

    The computer-assisted analysis for better interpreting images have been longstanding issues in the medical imaging field. On the image-understanding front, recent advances in machine learning, especially, in the way of deep learning, have made a big leap to help identify, classify, and quantify patterns in medical images. Specifically, exploiting hierarchical feature representations learned solely from data, instead of handcrafted features mostly designed based on domain-specific knowledge, lies at the core of the advances. In that way, deep learning is rapidly proving to be the state-of-the-art foundation, achieving enhanced performances in various medical applications. In this article, we introduce the fundamentals of deep learning methods; review their successes to image registration, anatomical/cell structures detection, tissue segmentation, computer-aided disease diagnosis or prognosis, and so on. We conclude by raising research issues and suggesting future directions for further improvements. PMID:28301734

  17. Realistic tissue visualization using photoacoustic image

    NASA Astrophysics Data System (ADS)

    Cho, Seonghee; Managuli, Ravi; Jeon, Seungwan; Kim, Jeesu; Kim, Chulhong

    2018-02-01

    Visualization methods are very important in biomedical imaging. As a technology that understands life, biomedical imaging has the unique advantage of providing the most intuitive information in the image. This advantage of biomedical imaging can be greatly improved by choosing a special visualization method. This is more complicated in volumetric data. Volume data has the advantage of containing 3D spatial information. Unfortunately, the data itself cannot directly represent the potential value. Because images are always displayed in 2D space, visualization is the key and creates the real value of volume data. However, image processing of 3D data requires complicated algorithms for visualization and high computational burden. Therefore, specialized algorithms and computing optimization are important issues in volume data. Photoacoustic-imaging is a unique imaging modality that can visualize the optical properties of deep tissue. Because the color of the organism is mainly determined by its light absorbing component, photoacoustic data can provide color information of tissue, which is closer to real tissue color. In this research, we developed realistic tissue visualization using acoustic-resolution photoacoustic volume data. To achieve realistic visualization, we designed specialized color transfer function, which depends on the depth of the tissue from the skin. We used direct ray casting method and processed color during computing shader parameter. In the rendering results, we succeeded in obtaining similar texture results from photoacoustic data. The surface reflected rays were visualized in white, and the reflected color from the deep tissue was visualized red like skin tissue. We also implemented the CUDA algorithm in an OpenGL environment for real-time interactive imaging.

  18. Spatial Statistics for Segmenting Histological Structures in H&E Stained Tissue Images.

    PubMed

    Nguyen, Luong; Tosun, Akif Burak; Fine, Jeffrey L; Lee, Adrian V; Taylor, D Lansing; Chennubhotla, S Chakra

    2017-07-01

    Segmenting a broad class of histological structures in transmitted light and/or fluorescence-based images is a prerequisite for determining the pathological basis of cancer, elucidating spatial interactions between histological structures in tumor microenvironments (e.g., tumor infiltrating lymphocytes), facilitating precision medicine studies with deep molecular profiling, and providing an exploratory tool for pathologists. This paper focuses on segmenting histological structures in hematoxylin- and eosin-stained images of breast tissues, e.g., invasive carcinoma, carcinoma in situ, atypical and normal ducts, adipose tissue, and lymphocytes. We propose two graph-theoretic segmentation methods based on local spatial color and nuclei neighborhood statistics. For benchmarking, we curated a data set of 232 high-power field breast tissue images together with expertly annotated ground truth. To accurately model the preference for histological structures (ducts, vessels, tumor nets, adipose, etc.) over the remaining connective tissue and non-tissue areas in ground truth annotations, we propose a new region-based score for evaluating segmentation algorithms. We demonstrate the improvement of our proposed methods over the state-of-the-art algorithms in both region- and boundary-based performance measures.

  19. Self-recognition in corals facilitates deep-sea habitat engineering

    USGS Publications Warehouse

    Hennige, Sebastian J; Morrison, Cheryl L.; Form, Armin U.; Buscher, Janina; Kamenos, Nicholas A.; Roberts, J. Murray

    2014-01-01

    The ability of coral reefs to engineer complex three-dimensional habitats is central to their success and the rich biodiversity they support. In tropical reefs, encrusting coralline algae bind together substrates and dead coral framework to make continuous reef structures, but beyond the photic zone, the cold-water coral Lophelia pertusa also forms large biogenic reefs, facilitated by skeletal fusion. Skeletal fusion in tropical corals can occur in closely related or juvenile individuals as a result of non-aggressive skeletal overgrowth or allogeneic tissue fusion, but contact reactions in many species result in mortality if there is no ‘self-recognition’ on a broad species level. This study reveals areas of ‘flawless’ skeletal fusion in Lophelia pertusa, potentially facilitated by allogeneic tissue fusion, are identified as having small aragonitic crystals or low levels of crystal organisation, and strong molecular bonding. Regardless of the mechanism, the recognition of ‘self’ between adjacent L. pertusa colonies leads to no observable mortality, facilitates ecosystem engineering and reduces aggression-related energetic expenditure in an environment where energy conservation is crucial. The potential for self-recognition at a species level, and subsequent skeletal fusion in framework-forming cold-water corals is an important first step in understanding their significance as ecological engineers in deep-seas worldwide.

  20. Predictive validity of granulation tissue color measured by digital image analysis for deep pressure ulcer healing: a multicenter prospective cohort study.

    PubMed

    Iizaka, Shinji; Kaitani, Toshiko; Sugama, Junko; Nakagami, Gojiro; Naito, Ayumi; Koyanagi, Hiroe; Konya, Chizuko; Sanada, Hiromi

    2013-01-01

    This multicenter prospective cohort study examined the predictive validity of granulation tissue color evaluated by digital image analysis for deep pressure ulcer healing. Ninety-one patients with deep pressure ulcers were followed for 3 weeks. From a wound photograph taken at baseline, an image representing the granulation red index (GRI) was processed in which a redder color represented higher values. We calculated the average GRI over granulation tissue and the proportion of pixels exceeding the threshold intensity of 80 for the granulation tissue surface (%GRI80) and wound surface (%wound red index 80). In the receiver operating characteristics curve analysis, most GRI parameters had adequate discriminative values for both improvement of the DESIGN-R total score and wound closure. Ulcers were categorized by the obtained cutoff points of the average GRI (≤80, >80), %GRI80 (≤55, >55-80, >80%), and %wound red index 80 (≤25, >25-50, >50%). In the linear mixed model, higher classes for all GRI parameters showed significantly greater relative improvement in overall wound severity during the 3 weeks after adjustment for patient characteristics and wound locations. Assessment of granulation tissue color by digital image analysis will be useful as an objective monitoring tool for granulation tissue quality or surrogate outcomes of pressure ulcer healing. © 2012 by the Wound Healing Society.

  1. A simple model for deep tissue attenuation correction and large organ analysis of Cerenkov luminescence imaging

    NASA Astrophysics Data System (ADS)

    Habte, Frezghi; Natarajan, Arutselvan; Paik, David S.; Gambhir, Sanjiv S.

    2014-03-01

    Cerenkov luminescence imaging (CLI) is an emerging cost effective modality that uses conventional small animal optical imaging systems and clinically available radionuclide probes for light emission. CLI has shown good correlation with PET for organs of high uptake such as kidney, spleen, thymus and subcutaneous tumors in mouse models. However, CLI has limitations for deep tissue quantitative imaging since the blue-weighted spectral characteristics of Cerenkov radiation attenuates highly by mammalian tissue. Large organs such as the liver have also shown higher signal due to the contribution of emission of light from a greater thickness of tissue. In this study, we developed a simple model that estimates the effective tissue attenuation coefficient in order to correct the CLI signal intensity with a priori estimated depth and thickness of specific organs. We used several thin slices of ham to build a phantom with realistic attenuation. We placed radionuclide sources inside the phantom at different tissue depths and imaged it using an IVIS Spectrum (Perkin-Elmer, Waltham, MA, USA) and Inveon microPET (Preclinical Solutions Siemens, Knoxville, TN). We also performed CLI and PET of mouse models and applied the proposed attenuation model to correct CLI measurements. Using calibration factors obtained from phantom study that converts the corrected CLI measurements to %ID/g, we obtained an average difference of less that 10% for spleen and less than 35% for liver compared to conventional PET measurements. Hence, the proposed model has a capability of correcting the CLI signal to provide comparable measurements with PET data.

  2. Adult stem cell lineage tracing and deep tissue imaging

    PubMed Central

    Fink, Juergen; Andersson-Rolf, Amanda; Koo, Bon-Kyoung

    2015-01-01

    Lineage tracing is a widely used method for understanding cellular dynamics in multicellular organisms during processes such as development, adult tissue maintenance, injury repair and tumorigenesis. Advances in tracing or tracking methods, from light microscopy-based live cell tracking to fluorescent label-tracing with two-photon microscopy, together with emerging tissue clearing strategies and intravital imaging approaches have enabled scientists to decipher adult stem and progenitor cell properties in various tissues and in a wide variety of biological processes. Although technical advances have enabled time-controlled genetic labeling and simultaneous live imaging, a number of obstacles still need to be overcome. In this review, we aim to provide an in-depth description of the traditional use of lineage tracing as well as current strategies and upcoming new methods of labeling and imaging. [BMB Reports 2015; 48(12): 655-667] PMID:26634741

  3. An optical clearing technique for plant tissues allowing deep imaging and compatible with fluorescence microscopy.

    PubMed

    Warner, Cherish A; Biedrzycki, Meredith L; Jacobs, Samuel S; Wisser, Randall J; Caplan, Jeffrey L; Sherrier, D Janine

    2014-12-01

    We report on a nondestructive clearing technique that enhances transmission of light through specimens from diverse plant species, opening unique opportunities for microscope-enabled plant research. After clearing, plant organs and thick tissue sections are amenable to deep imaging. The clearing method is compatible with immunocytochemistry techniques and can be used in concert with common fluorescent probes, including widely adopted protein tags such as GFP, which has fluorescence that is preserved during the clearing process. © 2014 American Society of Plant Biologists. All Rights Reserved.

  4. Adaptive focus for deep tissue using diffuse backscatter

    NASA Astrophysics Data System (ADS)

    Kress, Jeremy; Pourrezaei, Kambiz

    2014-02-01

    A system integrating high density diffuse optical imaging with adaptive optics using MEMS for deep tissue interaction is presented. In this system, a laser source is scanned over a high density fiber bundle using Digital Micromirror Device (DMD) and channeled to a tissue phantom. Backscatter is then collected from the tissue phantom by a high density fiber array of different fiber type and channeled to CMOS sensor for image acquisition. Intensity focus is directly verified using a second CMOS sensor which measures intensity transmitted though the tissue phantom. A set of training patterns are displayed on the DMD and backscatter is numerically fit to the transmission intensity. After the training patterns are displayed, adaptive focus is performed using only the backscatter and fitting functions. Additionally, tissue reconstruction and prediction of interference focusing by photoacoustic and optical tomographic methods is discussed. Finally, potential NIR applications such as in-vivo adaptive neural photostimulation and cancer targeting are discussed.

  5. 3D printed optical phantoms and deep tissue imaging for in vivo applications including oral surgery

    NASA Astrophysics Data System (ADS)

    Bentz, Brian Z.; Costas, Alfonso; Gaind, Vaibhav; Garcia, Jose M.; Webb, Kevin J.

    2017-03-01

    Progress in developing optical imaging for biomedical applications requires customizable and often complex objects known as "phantoms" for testing, evaluation, and calibration. This work demonstrates that 3D printing is an ideal method for fabricating such objects, allowing intricate inhomogeneities to be placed at exact locations in complex or anatomically realistic geometries, a process that is difficult or impossible using molds. We show printed mouse phantoms we have fabricated for developing deep tissue fluorescence imaging methods, and measurements of both their optical and mechanical properties. Additionally, we present a printed phantom of the human mouth that we use to develop an artery localization method to assist in oral surgery.

  6. A metadata-aware application for remote scoring and exchange of tissue microarray images

    PubMed Central

    2013-01-01

    Background The use of tissue microarrays (TMA) and advances in digital scanning microscopy has enabled the collection of thousands of tissue images. There is a need for software tools to annotate, query and share this data amongst researchers in different physical locations. Results We have developed an open source web-based application for remote scoring of TMA images, which exploits the value of Microsoft Silverlight Deep Zoom to provide a intuitive interface for zooming and panning around digital images. We use and extend existing XML-based standards to ensure that the data collected can be archived and that our system is interoperable with other standards-compliant systems. Conclusion The application has been used for multi-centre scoring of TMA slides composed of tissues from several Phase III breast cancer trials and ten different studies participating in the International Breast Cancer Association Consortium (BCAC). The system has enabled researchers to simultaneously score large collections of TMA and export the standardised data to integrate with pathological and clinical outcome data, thereby facilitating biomarker discovery. PMID:23635078

  7. Deep tissue massage: What are we talking about?

    PubMed

    Koren, Yogev; Kalichman, Leonid

    2018-04-01

    Massage is a common treatment in complementary and integrative medicine. Deep tissue massage, a form of therapeutic massage, has become more and more popular in recent years. Hence, the use of massage generally and deep tissue massage specifically, should be evaluated as any other modality of therapy to establish its efficacy and safety. To determine the definitions used for deep tissue massage in the scientific literature and to review the current scientific evidence for its efficacy and safety. Narrative review. There is no commonly accepted definition of deep tissue massage in the literature. The definition most frequently used is the intention of the therapist. We suggest separating the definitions of deep massage and deep tissue massage as follows: deep massage should be used to describe the intention of the therapist to treat deep tissue by using any form of massage and deep tissue massage should be used to describe a specific and independent method of massage therapy, utilizing the specific set of principles and techniques as defined by Riggs: "The understanding of the layers of the body, and the ability to work with tissue in these layers to relax, lengthen, and release holding patterns in the most effective and energy efficient way possible within the client's parameters of comfort". Heterogeneity of techniques and protocols used in published studies have made it difficult to draw any clear conclusions. Favorable outcomes may result from deep tissue massage in pain populations and patients with decreased range of motion. In addition, several rare serious adverse events were found related to deep tissue massage, probably as a result of the forceful application of massage therapy. Future research of deep tissue massage should be based on a common definition, classification system and the use of common comparators as controls. Copyright © 2017 Elsevier Ltd. All rights reserved.

  8. Technical Note: Deep learning based MRAC using rapid ultra-short echo time imaging.

    PubMed

    Jang, Hyungseok; Liu, Fang; Zhao, Gengyan; Bradshaw, Tyler; McMillan, Alan B

    2018-05-15

    In this study, we explore the feasibility of a novel framework for MR-based attenuation correction for PET/MR imaging based on deep learning via convolutional neural networks, which enables fully automated and robust estimation of a pseudo CT image based on ultrashort echo time (UTE), fat, and water images obtained by a rapid MR acquisition. MR images for MRAC are acquired using dual echo ramped hybrid encoding (dRHE), where both UTE and out-of-phase echo images are obtained within a short single acquisition (35 sec). Tissue labeling of air, soft tissue, and bone in the UTE image is accomplished via a deep learning network that was pre-trained with T1-weighted MR images. UTE images are used as input to the network, which was trained using labels derived from co-registered CT images. The tissue labels estimated by deep learning are refined by a conditional random field based correction. The soft tissue labels are further separated into fat and water components using the two-point Dixon method. The estimated bone, air, fat, and water images are then assigned appropriate Hounsfield units, resulting in a pseudo CT image for PET attenuation correction. To evaluate the proposed MRAC method, PET/MR imaging of the head was performed on 8 human subjects, where Dice similarity coefficients of the estimated tissue labels and relative PET errors were evaluated through comparison to a registered CT image. Dice coefficients for air (within the head), soft tissue, and bone labels were 0.76±0.03, 0.96±0.006, and 0.88±0.01. In PET quantification, the proposed MRAC method produced relative PET errors less than 1% within most brain regions. The proposed MRAC method utilizing deep learning with transfer learning and an efficient dRHE acquisition enables reliable PET quantification with accurate and rapid pseudo CT generation. This article is protected by copyright. All rights reserved. This article is protected by copyright. All rights reserved.

  9. Deep convolutional neural network and 3D deformable approach for tissue segmentation in musculoskeletal magnetic resonance imaging.

    PubMed

    Liu, Fang; Zhou, Zhaoye; Jang, Hyungseok; Samsonov, Alexey; Zhao, Gengyan; Kijowski, Richard

    2018-04-01

    To describe and evaluate a new fully automated musculoskeletal tissue segmentation method using deep convolutional neural network (CNN) and three-dimensional (3D) simplex deformable modeling to improve the accuracy and efficiency of cartilage and bone segmentation within the knee joint. A fully automated segmentation pipeline was built by combining a semantic segmentation CNN and 3D simplex deformable modeling. A CNN technique called SegNet was applied as the core of the segmentation method to perform high resolution pixel-wise multi-class tissue classification. The 3D simplex deformable modeling refined the output from SegNet to preserve the overall shape and maintain a desirable smooth surface for musculoskeletal structure. The fully automated segmentation method was tested using a publicly available knee image data set to compare with currently used state-of-the-art segmentation methods. The fully automated method was also evaluated on two different data sets, which include morphological and quantitative MR images with different tissue contrasts. The proposed fully automated segmentation method provided good segmentation performance with segmentation accuracy superior to most of state-of-the-art methods in the publicly available knee image data set. The method also demonstrated versatile segmentation performance on both morphological and quantitative musculoskeletal MR images with different tissue contrasts and spatial resolutions. The study demonstrates that the combined CNN and 3D deformable modeling approach is useful for performing rapid and accurate cartilage and bone segmentation within the knee joint. The CNN has promising potential applications in musculoskeletal imaging. Magn Reson Med 79:2379-2391, 2018. © 2017 International Society for Magnetic Resonance in Medicine. © 2017 International Society for Magnetic Resonance in Medicine.

  10. Automatical and accurate segmentation of cerebral tissues in fMRI dataset with combination of image processing and deep learning

    NASA Astrophysics Data System (ADS)

    Kong, Zhenglun; Luo, Junyi; Xu, Shengpu; Li, Ting

    2018-02-01

    Image segmentation plays an important role in medical science. One application is multimodality imaging, especially the fusion of structural imaging with functional imaging, which includes CT, MRI and new types of imaging technology such as optical imaging to obtain functional images. The fusion process require precisely extracted structural information, in order to register the image to it. Here we used image enhancement, morphometry methods to extract the accurate contours of different tissues such as skull, cerebrospinal fluid (CSF), grey matter (GM) and white matter (WM) on 5 fMRI head image datasets. Then we utilized convolutional neural network to realize automatic segmentation of images in deep learning way. Such approach greatly reduced the processing time compared to manual and semi-automatic segmentation and is of great importance in improving speed and accuracy as more and more samples being learned. The contours of the borders of different tissues on all images were accurately extracted and 3D visualized. This can be used in low-level light therapy and optical simulation software such as MCVM. We obtained a precise three-dimensional distribution of brain, which offered doctors and researchers quantitative volume data and detailed morphological characterization for personal precise medicine of Cerebral atrophy/expansion. We hope this technique can bring convenience to visualization medical and personalized medicine.

  11. Hyperspectral imaging in SWIR: from stain-free microscopy to deep tissue imaging (Conference Presentation)

    NASA Astrophysics Data System (ADS)

    Berezin, Mikhail Y.

    2016-03-01

    Recent advances in relatively unexplored short wave infrared (SWIR) range from 800-1600 nm detectors make wide-field imaging in this spectral range attractive to biology. The distinct advantages of SWIR region over the visible and near infrared (NIR) in tissue analysis are two-fold: (i) high abundance endogenous chromophores (i.e. water and lipids) enable tissue component differentiation based on wavelength-dependent absorption properties and (ii) the weak scattering of tissue permits better resolution of imaging in thick specimens. When combined with high spectral resolution, SWIR imaging produces a spectroscopic image, where every pixel corresponds to the entire high-resolution spectrum. This hyperspectral (HS) approach provides rich information about the relative abundance of individual chromophores and their interactions that contribute to the intensity and location of the optical signal. The presentation discusses the challenges in the SWIR-HS instrument design and data analysis and demonstrates some of the promising applications of this technology in life science and medicine.

  12. Hyperspectral Image Enhancement and Mixture Deep-Learning Classification of Corneal Epithelium Injuries

    PubMed Central

    Md Noor, Siti Salwa; Michael, Kaleena; Marshall, Stephen; Ren, Jinchang

    2017-01-01

    In our preliminary study, the reflectance signatures obtained from hyperspectral imaging (HSI) of normal and abnormal corneal epithelium tissues of porcine show similar morphology with subtle differences. Here we present image enhancement algorithms that can be used to improve the interpretability of data into clinically relevant information to facilitate diagnostics. A total of 25 corneal epithelium images without the application of eye staining were used. Three image feature extraction approaches were applied for image classification: (i) image feature classification from histogram using a support vector machine with a Gaussian radial basis function (SVM-GRBF); (ii) physical image feature classification using deep-learning Convolutional Neural Networks (CNNs) only; and (iii) the combined classification of CNNs and SVM-Linear. The performance results indicate that our chosen image features from the histogram and length-scale parameter were able to classify with up to 100% accuracy; particularly, at CNNs and CNNs-SVM, by employing 80% of the data sample for training and 20% for testing. Thus, in the assessment of corneal epithelium injuries, HSI has high potential as a method that could surpass current technologies regarding speed, objectivity, and reliability. PMID:29144388

  13. Hyperspectral Image Enhancement and Mixture Deep-Learning Classification of Corneal Epithelium Injuries.

    PubMed

    Noor, Siti Salwa Md; Michael, Kaleena; Marshall, Stephen; Ren, Jinchang

    2017-11-16

    In our preliminary study, the reflectance signatures obtained from hyperspectral imaging (HSI) of normal and abnormal corneal epithelium tissues of porcine show similar morphology with subtle differences. Here we present image enhancement algorithms that can be used to improve the interpretability of data into clinically relevant information to facilitate diagnostics. A total of 25 corneal epithelium images without the application of eye staining were used. Three image feature extraction approaches were applied for image classification: (i) image feature classification from histogram using a support vector machine with a Gaussian radial basis function (SVM-GRBF); (ii) physical image feature classification using deep-learning Convolutional Neural Networks (CNNs) only; and (iii) the combined classification of CNNs and SVM-Linear. The performance results indicate that our chosen image features from the histogram and length-scale parameter were able to classify with up to 100% accuracy; particularly, at CNNs and CNNs-SVM, by employing 80% of the data sample for training and 20% for testing. Thus, in the assessment of corneal epithelium injuries, HSI has high potential as a method that could surpass current technologies regarding speed, objectivity, and reliability.

  14. Clutter elimination for deep clinical optoacoustic imaging using localised vibration tagging (LOVIT)☆

    PubMed Central

    Jaeger, Michael; Bamber, Jeffrey C.; Frenz, Martin

    2013-01-01

    This paper investigates a novel method which allows clutter elimination in deep optoacoustic imaging. Clutter significantly limits imaging depth in clinical optoacoustic imaging, when irradiation optics and ultrasound detector are integrated in a handheld probe for flexible imaging of the human body. Strong optoacoustic transients generated at the irradiation site obscure weak signals from deep inside the tissue, either directly by propagating towards the probe, or via acoustic scattering. In this study we demonstrate that signals of interest can be distinguished from clutter by tagging them at the place of origin with localised tissue vibration induced by the acoustic radiation force in a focused ultrasonic beam. We show phantom results where this technique allowed almost full clutter elimination and thus strongly improved contrast for deep imaging. Localised vibration tagging by means of acoustic radiation force is especially promising for integration into ultrasound systems that already have implemented radiation force elastography. PMID:25302147

  15. Three-Photon Luminescence of Gold Nanorods and Its Applications for High Contrast Tissue and Deep In Vivo Brain Imaging

    PubMed Central

    Wang, Shaowei; Xi, Wang; Cai, Fuhong; Zhao, Xinyuan; Xu, Zhengping; Qian, Jun; He, Sailing

    2015-01-01

    Gold nanoparticles can be used as contrast agents for bio-imaging applications. Here we studied multi-photon luminescence (MPL) of gold nanorods (GNRs), under the excitation of femtosecond (fs) lasers. GNRs functionalized with polyethylene glycol (PEG) molecules have high chemical and optical stability, and can be used as multi-photon luminescent nanoprobes for deep in vivo imaging of live animals. We have found that the depth of in vivo imaging is dependent upon the transmission and focal capability of the excitation light interacting with the GNRs. Our study focused on the comparison of MPL from GNRs with two different aspect ratios, as well as their ex vivo and in vivo imaging effects under 760 nm and 1000 nm excitation, respectively. Both of these wavelengths were located at an optically transparent window of biological tissue (700-1000 nm). PEGylated GNRs, which were intravenously injected into mice via the tail vein and accumulated in major organs and tumor tissue, showed high image contrast due to distinct three-photon luminescence (3PL) signals upon irradiation of a 1000 nm fs laser. Concerning in vivo mouse brain imaging, the 3PL imaging depth of GNRs under 1000 nm fs excitation could reach 600 μm, which was approximately 170 μm deeper than the two-photon luminescence (2PL) imaging depth of GNRs with a fs excitation of 760 nm. PMID:25553113

  16. Three-photon tissue imaging using moxifloxacin.

    PubMed

    Lee, Seunghun; Lee, Jun Ho; Wang, Taejun; Jang, Won Hyuk; Yoon, Yeoreum; Kim, Bumju; Jun, Yong Woong; Kim, Myoung Joon; Kim, Ki Hean

    2018-06-20

    Moxifloxacin is an antibiotic used in clinics and has recently been used as a clinically compatible cell-labeling agent for two-photon (2P) imaging. Although 2P imaging with moxifloxacin labeling visualized cells inside tissues using enhanced fluorescence, the imaging depth was quite limited because of the relatively short excitation wavelength (<800 nm) used. In this study, the feasibility of three-photon (3P) excitation of moxifloxacin using a longer excitation wavelength and moxifloxacin-based 3P imaging were tested to increase the imaging depth. Moxifloxacin fluorescence via 3P excitation was detected at a >1000 nm excitation wavelength. After obtaining the excitation and emission spectra of moxifloxacin, moxifloxacin-based 3P imaging was applied to ex vivo mouse bladder and ex vivo mouse small intestine tissues and compared with moxifloxacin-based 2P imaging by switching the excitation wavelength of a Ti:sapphire oscillator between near 1030 and 780 nm. Both moxifloxacin-based 2P and 3P imaging visualized cellular structures in the tissues via moxifloxacin labeling, but the image contrast was better with 3P imaging than with 2P imaging at the same imaging depths. The imaging speed and imaging depth of moxifloxacin-based 3P imaging using a Ti:sapphire oscillator were limited by insufficient excitation power. Therefore, we constructed a new system for moxifloxacin-based 3P imaging using a high-energy Yb fiber laser at 1030 nm and used it for in vivo deep tissue imaging of a mouse small intestine. Moxifloxacin-based 3P imaging could be useful for clinical applications with enhanced imaging depth.

  17. A Non-Invasive Deep Tissue PH Monitor.

    DTIC Science & Technology

    1995-08-11

    disturbances in acid-base regulation may have serious effects on metabolic activity, circulation, and the central nervous system. Currently, acid-base...to tissue ischemia than is arterial pH. Consequently, a non-invasive deep tissue pH monitor has enormous value as a mechanism for rapid and effective ...achieved, and improve our understanding of what physical effects are important to successful non-invasive deep tissue pH monitoring. This last statement

  18. Deep Convolutional Neural Networks for Multi-Modality Isointense Infant Brain Image Segmentation

    PubMed Central

    Zhang, Wenlu; Li, Rongjian; Deng, Houtao; Wang, Li; Lin, Weili; Ji, Shuiwang; Shen, Dinggang

    2015-01-01

    The segmentation of infant brain tissue images into white matter (WM), gray matter (GM), and cerebrospinal fluid (CSF) plays an important role in studying early brain development in health and disease. In the isointense stage (approximately 6–8 months of age), WM and GM exhibit similar levels of intensity in both T1 and T2 MR images, making the tissue segmentation very challenging. Only a small number of existing methods have been designed for tissue segmentation in this isointense stage; however, they only used a single T1 or T2 images, or the combination of T1 and T2 images. In this paper, we propose to use deep convolutional neural networks (CNNs) for segmenting isointense stage brain tissues using multi-modality MR images. CNNs are a type of deep models in which trainable filters and local neighborhood pooling operations are applied alternatingly on the raw input images, resulting in a hierarchy of increasingly complex features. Specifically, we used multimodality information from T1, T2, and fractional anisotropy (FA) images as inputs and then generated the segmentation maps as outputs. The multiple intermediate layers applied convolution, pooling, normalization, and other operations to capture the highly nonlinear mappings between inputs and outputs. We compared the performance of our approach with that of the commonly used segmentation methods on a set of manually segmented isointense stage brain images. Results showed that our proposed model significantly outperformed prior methods on infant brain tissue segmentation. In addition, our results indicated that integration of multi-modality images led to significant performance improvement. PMID:25562829

  19. Bioluminescence resonance energy transfer (BRET) imaging of protein–protein interactions within deep tissues of living subjects

    PubMed Central

    Dragulescu-Andrasi, Anca; Chan, Carmel T.; Massoud, Tarik F.; Gambhir, Sanjiv S.

    2011-01-01

    Identifying protein–protein interactions (PPIs) is essential for understanding various disease mechanisms and developing new therapeutic approaches. Current methods for assaying cellular intermolecular interactions are mainly used for cells in culture and have limited use for the noninvasive assessment of small animal disease models. Here, we describe red light-emitting reporter systems based on bioluminescence resonance energy transfer (BRET) that allow for assaying PPIs both in cell culture and deep tissues of small animals. These BRET systems consist of the recently developed Renilla reniformis luciferase (RLuc) variants RLuc8 and RLuc8.6, used as BRET donors, combined with two red fluorescent proteins, TagRFP and TurboFP635, as BRET acceptors. In addition to the native coelenterazine luciferase substrate, we used the synthetic derivative coelenterazine-v, which further red-shifts the emission maxima of Renilla luciferases by 35 nm. We show the use of these BRET systems for ratiometric imaging of both cells in culture and deep-tissue small animal tumor models and validate their applicability for studying PPIs in mice in the context of rapamycin-induced FK506 binding protein 12 (FKBP12)-FKBP12 rapamycin binding domain (FRB) association. These red light-emitting BRET systems have great potential for investigating PPIs in the context of drug screening and target validation applications. PMID:21730157

  20. Deep learning based classification for head and neck cancer detection with hyperspectral imaging in an animal model

    NASA Astrophysics Data System (ADS)

    Ma, Ling; Lu, Guolan; Wang, Dongsheng; Wang, Xu; Chen, Zhuo Georgia; Muller, Susan; Chen, Amy; Fei, Baowei

    2017-03-01

    Hyperspectral imaging (HSI) is an emerging imaging modality that can provide a noninvasive tool for cancer detection and image-guided surgery. HSI acquires high-resolution images at hundreds of spectral bands, providing big data to differentiating different types of tissue. We proposed a deep learning based method for the detection of head and neck cancer with hyperspectral images. Since the deep learning algorithm can learn the feature hierarchically, the learned features are more discriminative and concise than the handcrafted features. In this study, we adopt convolutional neural networks (CNN) to learn the deep feature of pixels for classifying each pixel into tumor or normal tissue. We evaluated our proposed classification method on the dataset containing hyperspectral images from 12 tumor-bearing mice. Experimental results show that our method achieved an average accuracy of 91.36%. The preliminary study demonstrated that our deep learning method can be applied to hyperspectral images for detecting head and neck tumors in animal models.

  1. Cellular imaging of deep organ using two-photon Bessel light-sheet nonlinear structured illumination microscopy

    PubMed Central

    Zhao, Ming; Zhang, Han; Li, Yu; Ashok, Amit; Liang, Rongguang; Zhou, Weibin; Peng, Leilei

    2014-01-01

    In vivo fluorescent cellular imaging of deep internal organs is highly challenging, because the excitation needs to penetrate through strong scattering tissue and the emission signal is degraded significantly by photon diffusion induced by tissue-scattering. We report that by combining two-photon Bessel light-sheet microscopy with nonlinear structured illumination microscopy (SIM), live samples up to 600 microns wide can be imaged by light-sheet microscopy with 500 microns penetration depth, and diffused background in deep tissue light-sheet imaging can be reduced to obtain clear images at cellular resolution in depth beyond 200 microns. We demonstrate in vivo two-color imaging of pronephric glomeruli and vasculature of zebrafish kidney, whose cellular structures located at the center of the fish body are revealed in high clarity by two-color two-photon Bessel light-sheet SIM. PMID:24876996

  2. A bright cyan-excitable orange fluorescent protein facilitates dual-emission microscopy and enhances bioluminescence imaging in vivo

    PubMed Central

    Chu, Jun; Oh, Young-Hee; Sens, Alex; Ataie, Niloufar; Dana, Hod; Macklin, John J.; Laviv, Tal; Welf, Erik S.; Dean, Kevin M.; Zhang, Feijie; Kim, Benjamin B.; Tang, Clement Tran; Hu, Michelle; Baird, Michelle A.; Davidson, Michael W.; Kay, Mark A.; Fiolka, Reto; Yasuda, Ryohei; Kim, Douglas S.; Ng, Ho-Leung; Lin, Michael Z.

    2016-01-01

    Orange-red fluorescent proteins (FPs) are widely used in biomedical research for multiplexed epifluorescence microscopy with GFP-based probes, but their different excitation requirements make multiplexing with new advanced microscopy methods difficult. Separately, orange-red FPs are useful for deep-tissue imaging in mammals due to the relative tissue transmissibility of orange-red light, but their dependence on illumination limits their sensitivity as reporters in deep tissues. Here we describe CyOFP1, a bright engineered orange-red FP that is excitable by cyan light. We show that CyOFP1 enables single-excitation multiplexed imaging with GFP-based probes in single-photon and two-photon microscopy, including time-lapse imaging in light-sheet systems. CyOFP1 also serves as an efficient acceptor for resonance energy transfer from the highly catalytic blue-emitting luciferase NanoLuc. An optimized fusion of CyOFP1 and NanoLuc, called Antares, functions as a highly sensitive bioluminescent reporter in vivo, producing substantially brighter signals from deep tissues than firefly luciferase and other bioluminescent proteins. PMID:27240196

  3. Spectroscopic Imaging of Deep Tissue through Photoacoustic Detection of Molecular Vibration

    PubMed Central

    Wang, Pu; Rajian, Justin R.; Cheng, Ji-Xin

    2013-01-01

    The quantized vibration of chemical bonds provides a way of imaging target molecules in a complex tissue environment. Photoacoustic detection of harmonic vibrational transitions provides an approach to visualize tissue content beyond the ballistic photon regime. This method involves pulsed laser excitation of overtone transitions in target molecules inside a tissue. Fast relaxation of the vibrational energy into heat results in a local temperature rise on the order of mK and a subsequent generation of acoustic waves detectable with an ultrasonic transducer. In this perspective, we review recent advances that demonstrate the advantages of vibration-based photoacoustic imaging and illustrate its potential in diagnosing cardiovascular plaques. An outlook into future development of vibrational photoacoustic endoscopy and tomography is provided. PMID:24073304

  4. Bond-selective imaging of deep tissue through the optical window between 1600 and 1850 nm.

    PubMed

    Wang, Pu; Wang, Han-Wei; Sturek, Michael; Cheng, Ji-Xin

    2012-01-01

    We report the employment of an optical window between 1600 nm and 1850 nm for bond-selective deep tissue imaging through harmonic vibrational excitation and acoustic detection of resultant pressure waves. In this window where a local minimum of water absorption resides, we found a 5 times enhancement of photoacoustic signal by first overtone excitation of the methylene group CH(2) at 1730 nm, compared to the second overtone excitation at 1210 nm. The enhancement allows 3D mapping of intramuscular fat with improved contrast and of lipid deposition inside an atherosclerotic artery wall in the presence of blood. Moreover, lipid and protein are differentiated based on the first overtone absorption profiles of CH(2) and methyl group CH(3) in this window. Copyright © 2012 WILEY-VCH Verlag GmbH & Co. KGaA, Weinheim.

  5. Photoacoustic diagnosis of burns in rats: two-dimensional photo-acoustic imaging of burned tissue

    NASA Astrophysics Data System (ADS)

    Yamazaki, Mutsuo; Sato, Shunichi; Saito, Daizo; Okada, Yoshiaki; Kurita, Akira; Kikuchi, Makoto; Ashida, Hiroshi; Obara, Minoru

    2003-06-01

    We previously reported that for rat burn models, deep dermal burns and deep burns can be well differentiated by measuring the propagation time of the photoacoustic signals originated from the blood in the healthy skin tissue under the damaged tissue layer. However, the diagnosis was based on point measurement in the wound, and therefore site-dependent information on the injuries was not obtained; such information is very important for diagnosis of extended burns. In the present study, we scanned a photoacoustic detector on the wound and constructed two-dimensional (2-D) images of the blood-originated photoacoustic signals for superficial dermal burns (SDB), deep dermal burns (DDB), deep burns (DB), and healthy skins (control) in rats. For each burn model, site-dependent variation of the signal was observed; the variation probably reflects the distribution of blood vessels in the skin tissue. In spite of the variation, clear differentiation was obtained between SDB, DDB, and DB from the 2D images. The images were constructed as a function of post burn time. Temporal signal variation will be also presented.

  6. Three-Dimensional Imaging of the Intracellular Fate of Plasmid DNA and Transgene Expression: ZsGreen1 and Tissue Clearing Method CUBIC Are an Optimal Combination for Multicolor Deep Imaging in Murine Tissues.

    PubMed

    Fumoto, Shintaro; Nishimura, Koyo; Nishida, Koyo; Kawakami, Shigeru

    2016-01-01

    Evaluation methods for determining the distribution of transgene expression in the body and the in vivo fate of viral and non-viral vectors are necessary for successful development of in vivo gene delivery systems. Here, we evaluated the spatial distribution of transgene expression using tissue clearing methods. After hydrodynamic injection of plasmid DNA into mice, whole tissues were subjected to tissue clearing. Tissue clearing followed by confocal laser scanning microscopy enabled evaluation of the three-dimensional distribution of transgene expression without preparation of tissue sections. Among the tested clearing methods (ClearT2, SeeDB, and CUBIC), CUBIC was the most suitable method for determining the spatial distribution of transgene expression in not only the liver but also other tissues such as the kidney and lung. In terms of the type of fluorescent protein, the observable depth for green fluorescent protein ZsGreen1 was slightly greater than that for red fluorescent protein tdTomato. We observed a depth of ~1.5 mm for the liver and 500 μm for other tissues without preparation of tissue sections. Furthermore, we succeeded in multicolor deep imaging of the intracellular fate of plasmid DNA in the murine liver. Thus, tissue clearing would be a powerful approach for determining the spatial distribution of plasmid DNA and transgene expression in various murine tissues.

  7. Photoacoustic physio-chemical analysis and its implementation in deep tissue with a catheter setup

    NASA Astrophysics Data System (ADS)

    Xu, Guan; Meng, Zhou-xian; Lin, Jian-die D.; Cheng, Qian; Wang, Xueding

    2015-03-01

    Photoacoustic (PA) measurements encode the information associated with both physical microstructures and chemical contents in biological tissues. A two-dimensional physio-chemical spectrogram (PCS) can be formulated by combining the power spectra of PA signals acquired at a series of optical wavelengths. The analysis of PCS, or namely PA physio-chemical analysis (PAPCA), enables the quantification of the relative concentrations and the spatial distributions of a variety of chemical components in the tissue. This study validated the feasibility of PAPCA in characterizing liver conditions during the progression of non-alcoholic fatty liver disease. A catheter based setup facilitating measurement in deep tissues was also tested.

  8. Fiber-based tunable repetition rate source for deep tissue two-photon fluorescence microscopy.

    PubMed

    Charan, Kriti; Li, Bo; Wang, Mengran; Lin, Charles P; Xu, Chris

    2018-05-01

    Deep tissue multiphoton imaging requires high peak power to enhance signal and low average power to prevent thermal damage. Both goals can be advantageously achieved through laser repetition rate tuning instead of simply adjusting the average power. We show that the ideal repetition rate for deep two-photon imaging in the mouse brain is between 1 and 10 MHz, and we present a fiber-based source with an arbitrarily tunable repetition rate within this range. The performance of the new source is compared to a mode-locked Ti:Sapphire (Ti:S) laser for in vivo imaging of mouse brain vasculature. At 2.5 MHz, the fiber source requires 5.1 times less average power to obtain the same signal as a standard Ti:S laser operating at 80 MHz.

  9. Fiber-based tunable repetition rate source for deep tissue two-photon fluorescence microscopy

    PubMed Central

    Charan, Kriti; Li, Bo; Wang, Mengran; Lin, Charles P.; Xu, Chris

    2018-01-01

    Deep tissue multiphoton imaging requires high peak power to enhance signal and low average power to prevent thermal damage. Both goals can be advantageously achieved through laser repetition rate tuning instead of simply adjusting the average power. We show that the ideal repetition rate for deep two-photon imaging in the mouse brain is between 1 and 10 MHz, and we present a fiber-based source with an arbitrarily tunable repetition rate within this range. The performance of the new source is compared to a mode-locked Ti:Sapphire (Ti:S) laser for in vivo imaging of mouse brain vasculature. At 2.5 MHz, the fiber source requires 5.1 times less average power to obtain the same signal as a standard Ti:S laser operating at 80 MHz. PMID:29760989

  10. Imaging deep skeletal muscle structure using a high-sensitivity ultrathin side-viewing optical coherence tomography needle probe

    PubMed Central

    Yang, Xiaojie; Lorenser, Dirk; McLaughlin, Robert A.; Kirk, Rodney W.; Edmond, Matthew; Simpson, M. Cather; Grounds, Miranda D.; Sampson, David D.

    2013-01-01

    We have developed an extremely miniaturized optical coherence tomography (OCT) needle probe (outer diameter 310 µm) with high sensitivity (108 dB) to enable minimally invasive imaging of cellular structure deep within skeletal muscle. Three-dimensional volumetric images were acquired from ex vivo mouse tissue, examining both healthy and pathological dystrophic muscle. Individual myofibers were visualized as striations in the images. Degradation of cellular structure in necrotic regions was seen as a loss of these striations. Tendon and connective tissue were also visualized. The observed structures were validated against co-registered hematoxylin and eosin (H&E) histology sections. These images of internal cellular structure of skeletal muscle acquired with an OCT needle probe demonstrate the potential of this technique to visualize structure at the microscopic level deep in biological tissue in situ. PMID:24466482

  11. Characterisation of signal enhancements achieved when utilizing a photon diode in deep Raman spectroscopy of tissue

    PubMed Central

    Vardaki, Martha Z.; Matousek, Pavel; Stone, Nicholas

    2016-01-01

    We characterise the performance of a beam enhancing element (‘photon diode’) for use in deep Raman spectroscopy (DRS) of biological tissues. The optical component enhances the number of laser photons coupled into a tissue sample by returning escaping photons back into it at the illumination zone. The method is compatible with transmission Raman spectroscopy, a deep Raman spectroscopy concept, and its implementation leads to considerable enhancement of detected Raman photon rates. In the past, the enhancement concept was demonstrated with a variety of samples (pharmaceutical tablets, tissue, etc) but it was not systematically characterized with biological tissues. In this study, we investigate the enhancing properties of the photon diode in the transmission Raman geometry as a function of: a) the depth and b) the optical properties of tissue samples. Liquid tissue phantoms were employed to facilitate systematic variation of optical properties. These were chosen to mimic optical properties of human tissues, including breast and prostate. The obtained results evidence that a photon diode can enhance Raman signals of tissues by a maximum of × 2.4, although it can also decrease the signals created towards the back of samples that exhibit high scattering or absorption properties. PMID:27375932

  12. Fiber-optic fluorescence imaging

    PubMed Central

    Flusberg, Benjamin A; Cocker, Eric D; Piyawattanametha, Wibool; Jung, Juergen C; Cheung, Eunice L M; Schnitzer, Mark J

    2010-01-01

    Optical fibers guide light between separate locations and enable new types of fluorescence imaging. Fiber-optic fluorescence imaging systems include portable handheld microscopes, flexible endoscopes well suited for imaging within hollow tissue cavities and microendoscopes that allow minimally invasive high-resolution imaging deep within tissue. A challenge in the creation of such devices is the design and integration of miniaturized optical and mechanical components. Until recently, fiber-based fluorescence imaging was mainly limited to epifluorescence and scanning confocal modalities. Two new classes of photonic crystal fiber facilitate ultrashort pulse delivery for fiber-optic two-photon fluorescence imaging. An upcoming generation of fluorescence imaging devices will be based on microfabricated device components. PMID:16299479

  13. Relevance of deep learning to facilitate the diagnosis of HER2 status in breast cancer

    NASA Astrophysics Data System (ADS)

    Vandenberghe, Michel E.; Scott, Marietta L. J.; Scorer, Paul W.; Söderberg, Magnus; Balcerzak, Denis; Barker, Craig

    2017-04-01

    Tissue biomarker scoring by pathologists is central to defining the appropriate therapy for patients with cancer. Yet, inter-pathologist variability in the interpretation of ambiguous cases can affect diagnostic accuracy. Modern artificial intelligence methods such as deep learning have the potential to supplement pathologist expertise to ensure constant diagnostic accuracy. We developed a computational approach based on deep learning that automatically scores HER2, a biomarker that defines patient eligibility for anti-HER2 targeted therapies in breast cancer. In a cohort of 71 breast tumour resection samples, automated scoring showed a concordance of 83% with a pathologist. The twelve discordant cases were then independently reviewed, leading to a modification of diagnosis from initial pathologist assessment for eight cases. Diagnostic discordance was found to be largely caused by perceptual differences in assessing HER2 expression due to high HER2 staining heterogeneity. This study provides evidence that deep learning aided diagnosis can facilitate clinical decision making in breast cancer by identifying cases at high risk of misdiagnosis.

  14. In vivo tomographic imaging of deep seated cancer using fluorescence lifetime contrast

    PubMed Central

    Rice, William L.; Shcherbakova, Daria M; Verkusha, Vladislav V.; Kumar, Anand T.N.

    2015-01-01

    Preclinical cancer research would benefit from non-invasive imaging methods that allow tracking and visualization of early stage metastasis in vivo. While fluorescent proteins revolutionized intravital microscopy, two major challenges which still remain are tissue autofluorescence and hemoglobin absorption, which act to limit intravital optical techniques to large or subcutaneous tumors. Here we employ time-domain technology for the effective separation of tissue autofluorescence from extrinsic fluorophores, based on their distinct fluorescence lifetimes. Additionally, we employ cancer cells labelled with near infra-red fluorescent proteins (iRFP) to allow deep-tissue imaging. Our results demonstrate that time-domain imaging allows the detection of metastasis in deep-seated organs of living mice with a more than 20-fold increase in sensitivity compared to conventional continuous wave techniques. Furthermore, the distinct fluorescence lifetimes of each iRFP enables lifetime multiplexing of three different tumors, each expressing unique iRFP labels in the same animal. Fluorescence tomographic reconstructions reveal 3D distributions of iRFP720-expressing cancer cells in lungs and brain of live mice, allowing ready longitudinal monitoring of cancer cell fate with greater sensitivity than otherwise currently possible. PMID:25670171

  15. Overview of deep learning in medical imaging.

    PubMed

    Suzuki, Kenji

    2017-09-01

    The use of machine learning (ML) has been increasing rapidly in the medical imaging field, including computer-aided diagnosis (CAD), radiomics, and medical image analysis. Recently, an ML area called deep learning emerged in the computer vision field and became very popular in many fields. It started from an event in late 2012, when a deep-learning approach based on a convolutional neural network (CNN) won an overwhelming victory in the best-known worldwide computer vision competition, ImageNet Classification. Since then, researchers in virtually all fields, including medical imaging, have started actively participating in the explosively growing field of deep learning. In this paper, the area of deep learning in medical imaging is overviewed, including (1) what was changed in machine learning before and after the introduction of deep learning, (2) what is the source of the power of deep learning, (3) two major deep-learning models: a massive-training artificial neural network (MTANN) and a convolutional neural network (CNN), (4) similarities and differences between the two models, and (5) their applications to medical imaging. This review shows that ML with feature input (or feature-based ML) was dominant before the introduction of deep learning, and that the major and essential difference between ML before and after deep learning is the learning of image data directly without object segmentation or feature extraction; thus, it is the source of the power of deep learning, although the depth of the model is an important attribute. The class of ML with image input (or image-based ML) including deep learning has a long history, but recently gained popularity due to the use of the new terminology, deep learning. There are two major models in this class of ML in medical imaging, MTANN and CNN, which have similarities as well as several differences. In our experience, MTANNs were substantially more efficient in their development, had a higher performance, and required a

  16. Ultrafast ultrasound localization microscopy for deep super-resolution vascular imaging

    NASA Astrophysics Data System (ADS)

    Errico, Claudia; Pierre, Juliette; Pezet, Sophie; Desailly, Yann; Lenkei, Zsolt; Couture, Olivier; Tanter, Mickael

    2015-11-01

    Non-invasive imaging deep into organs at microscopic scales remains an open quest in biomedical imaging. Although optical microscopy is still limited to surface imaging owing to optical wave diffusion and fast decorrelation in tissue, revolutionary approaches such as fluorescence photo-activated localization microscopy led to a striking increase in resolution by more than an order of magnitude in the last decade. In contrast with optics, ultrasonic waves propagate deep into organs without losing their coherence and are much less affected by in vivo decorrelation processes. However, their resolution is impeded by the fundamental limits of diffraction, which impose a long-standing trade-off between resolution and penetration. This limits clinical and preclinical ultrasound imaging to a sub-millimetre scale. Here we demonstrate in vivo that ultrasound imaging at ultrafast frame rates (more than 500 frames per second) provides an analogue to optical localization microscopy by capturing the transient signal decorrelation of contrast agents—inert gas microbubbles. Ultrafast ultrasound localization microscopy allowed both non-invasive sub-wavelength structural imaging and haemodynamic quantification of rodent cerebral microvessels (less than ten micrometres in diameter) more than ten millimetres below the tissue surface, leading to transcranial whole-brain imaging within short acquisition times (tens of seconds). After intravenous injection, single echoes from individual microbubbles were detected through ultrafast imaging. Their localization, not limited by diffraction, was accumulated over 75,000 images, yielding 1,000,000 events per coronal plane and statistically independent pixels of ten micrometres in size. Precise temporal tracking of microbubble positions allowed us to extract accurately in-plane velocities of the blood flow with a large dynamic range (from one millimetre per second to several centimetres per second). These results pave the way for deep non

  17. Ultrafast ultrasound localization microscopy for deep super-resolution vascular imaging.

    PubMed

    Errico, Claudia; Pierre, Juliette; Pezet, Sophie; Desailly, Yann; Lenkei, Zsolt; Couture, Olivier; Tanter, Mickael

    2015-11-26

    Non-invasive imaging deep into organs at microscopic scales remains an open quest in biomedical imaging. Although optical microscopy is still limited to surface imaging owing to optical wave diffusion and fast decorrelation in tissue, revolutionary approaches such as fluorescence photo-activated localization microscopy led to a striking increase in resolution by more than an order of magnitude in the last decade. In contrast with optics, ultrasonic waves propagate deep into organs without losing their coherence and are much less affected by in vivo decorrelation processes. However, their resolution is impeded by the fundamental limits of diffraction, which impose a long-standing trade-off between resolution and penetration. This limits clinical and preclinical ultrasound imaging to a sub-millimetre scale. Here we demonstrate in vivo that ultrasound imaging at ultrafast frame rates (more than 500 frames per second) provides an analogue to optical localization microscopy by capturing the transient signal decorrelation of contrast agents--inert gas microbubbles. Ultrafast ultrasound localization microscopy allowed both non-invasive sub-wavelength structural imaging and haemodynamic quantification of rodent cerebral microvessels (less than ten micrometres in diameter) more than ten millimetres below the tissue surface, leading to transcranial whole-brain imaging within short acquisition times (tens of seconds). After intravenous injection, single echoes from individual microbubbles were detected through ultrafast imaging. Their localization, not limited by diffraction, was accumulated over 75,000 images, yielding 1,000,000 events per coronal plane and statistically independent pixels of ten micrometres in size. Precise temporal tracking of microbubble positions allowed us to extract accurately in-plane velocities of the blood flow with a large dynamic range (from one millimetre per second to several centimetres per second). These results pave the way for deep non

  18. Dual targeting luminescent gold nanoclusters for tumor imaging and deep tissue therapy.

    PubMed

    Chen, Dan; Li, Bowen; Cai, Songhua; Wang, Peng; Peng, Shuwen; Sheng, Yuanzhi; He, Yuanyuan; Gu, Yueqing; Chen, Haiyan

    2016-09-01

    Dual targeting towards both extracellular and intracellular receptors specific to tumor is a significant approach for cancer diagnosis and therapy. In the present study, a novel nano-platform (AuNC-cRGD-Apt) with dual targeting function was initially established by conjugating gold nanocluster (AuNC) with cyclic RGD (cRGD) that is specific to αvβ3integrins over-expressed on the surface of tumor tissues and aptamer AS1411 (Apt) that is of high affinity to nucleolin over-expressed in the cytoplasm and nucleus of tumor cells. Then, AuNC-cRGD-Apt was further functionalized with near infrared (NIR) fluorescence dye (MPA), giving a NIR fluorescent dual-targeting probe AuNC-MPA-cRGD-Apt. AuNC-MPA-cRGD-Apt displays low cytotoxicity and favorable tumor-targeting capability at both in vitro and in vivo level, suggesting its clinical potential for tumor imaging. Additionally, Doxorubicin (DOX), a widely used clinical chemotherapeutic drug that kill cancer cells by intercalating DNA in cellular nucleus, was immobilized onto AuNC-cRGD-Apt forming a pro-drug, AuNC-DOX-cRGD-Apt. The enhanced tumor affinity, deep tumor penetration and improved anti-tumor activity of this pro-drug were demonstrated in different tumor cell lines, tumor spheroid and tumor-bearing mouse models. Results in this study suggest not only the prospect of non-toxic AuNC modified with two targeting ligands for tumor targeted imaging, but also confirm the promising future of dual targeting AuNC as a core for the design of prodrug in the field of cancer therapy. Copyright © 2016 Elsevier Ltd. All rights reserved.

  19. Prediction of Tissue Outcome and Assessment of Treatment Effect in Acute Ischemic Stroke Using Deep Learning.

    PubMed

    Nielsen, Anne; Hansen, Mikkel Bo; Tietze, Anna; Mouridsen, Kim

    2018-06-01

    Treatment options for patients with acute ischemic stroke depend on the volume of salvageable tissue. This volume assessment is currently based on fixed thresholds and single imagine modalities, limiting accuracy. We wish to develop and validate a predictive model capable of automatically identifying and combining acute imaging features to accurately predict final lesion volume. Using acute magnetic resonance imaging, we developed and trained a deep convolutional neural network (CNN deep ) to predict final imaging outcome. A total of 222 patients were included, of which 187 were treated with rtPA (recombinant tissue-type plasminogen activator). The performance of CNN deep was compared with a shallow CNN based on the perfusion-weighted imaging biomarker Tmax (CNN Tmax ), a shallow CNN based on a combination of 9 different biomarkers (CNN shallow ), a generalized linear model, and thresholding of the diffusion-weighted imaging biomarker apparent diffusion coefficient (ADC) at 600×10 -6 mm 2 /s (ADC thres ). To assess whether CNN deep is capable of differentiating outcomes of ±intravenous rtPA, patients not receiving intravenous rtPA were included to train CNN deep, -rtpa to access a treatment effect. The networks' performances were evaluated using visual inspection, area under the receiver operating characteristic curve (AUC), and contrast. CNN deep yields significantly better performance in predicting final outcome (AUC=0.88±0.12) than generalized linear model (AUC=0.78±0.12; P =0.005), CNN Tmax (AUC=0.72±0.14; P <0.003), and ADC thres (AUC=0.66±0.13; P <0.0001) and a substantially better performance than CNN shallow (AUC=0.85±0.11; P =0.063). Measured by contrast, CNN deep improves the predictions significantly, showing superiority to all other methods ( P ≤0.003). CNN deep also seems to be able to differentiate outcomes based on treatment strategy with the volume of final infarct being significantly different ( P =0.048). The considerable prediction

  20. Seeing through Musculoskeletal Tissues: Improving In Situ Imaging of Bone and the Lacunar Canalicular System through Optical Clearing

    PubMed Central

    Berke, Ian M.; Miola, Joseph P.; David, Michael A.; Smith, Melanie K.; Price, Christopher

    2016-01-01

    In situ, cells of the musculoskeletal system reside within complex and often interconnected 3-D environments. Key to better understanding how 3-D tissue and cellular environments regulate musculoskeletal physiology, homeostasis, and health is the use of robust methodologies for directly visualizing cell-cell and cell-matrix architecture in situ. However, the use of standard optical imaging techniques is often of limited utility in deep imaging of intact musculoskeletal tissues due to the highly scattering nature of biological tissues. Drawing inspiration from recent developments in the deep-tissue imaging field, we describe the application of immersion based optical clearing techniques, which utilize the principle of refractive index (RI) matching between the clearing/mounting media and tissue under observation, to improve the deep, in situ imaging of musculoskeletal tissues. To date, few optical clearing techniques have been applied specifically to musculoskeletal tissues, and a systematic comparison of the clearing ability of optical clearing agents in musculoskeletal tissues has yet to be fully demonstrated. In this study we tested the ability of eight different aqueous and non-aqueous clearing agents, with RIs ranging from 1.45 to 1.56, to optically clear murine knee joints and cortical bone. We demonstrated and quantified the ability of these optical clearing agents to clear musculoskeletal tissues and improve both macro- and micro-scale imaging of musculoskeletal tissue across several imaging modalities (stereomicroscopy, spectroscopy, and one-, and two-photon confocal microscopy) and investigational techniques (dynamic bone labeling and en bloc tissue staining). Based upon these findings we believe that optical clearing, in combination with advanced imaging techniques, has the potential to complement classical musculoskeletal analysis techniques; opening the door for improved in situ investigation and quantification of musculoskeletal tissues. PMID:26930293

  1. Seeing through Musculoskeletal Tissues: Improving In Situ Imaging of Bone and the Lacunar Canalicular System through Optical Clearing.

    PubMed

    Berke, Ian M; Miola, Joseph P; David, Michael A; Smith, Melanie K; Price, Christopher

    2016-01-01

    In situ, cells of the musculoskeletal system reside within complex and often interconnected 3-D environments. Key to better understanding how 3-D tissue and cellular environments regulate musculoskeletal physiology, homeostasis, and health is the use of robust methodologies for directly visualizing cell-cell and cell-matrix architecture in situ. However, the use of standard optical imaging techniques is often of limited utility in deep imaging of intact musculoskeletal tissues due to the highly scattering nature of biological tissues. Drawing inspiration from recent developments in the deep-tissue imaging field, we describe the application of immersion based optical clearing techniques, which utilize the principle of refractive index (RI) matching between the clearing/mounting media and tissue under observation, to improve the deep, in situ imaging of musculoskeletal tissues. To date, few optical clearing techniques have been applied specifically to musculoskeletal tissues, and a systematic comparison of the clearing ability of optical clearing agents in musculoskeletal tissues has yet to be fully demonstrated. In this study we tested the ability of eight different aqueous and non-aqueous clearing agents, with RIs ranging from 1.45 to 1.56, to optically clear murine knee joints and cortical bone. We demonstrated and quantified the ability of these optical clearing agents to clear musculoskeletal tissues and improve both macro- and micro-scale imaging of musculoskeletal tissue across several imaging modalities (stereomicroscopy, spectroscopy, and one-, and two-photon confocal microscopy) and investigational techniques (dynamic bone labeling and en bloc tissue staining). Based upon these findings we believe that optical clearing, in combination with advanced imaging techniques, has the potential to complement classical musculoskeletal analysis techniques; opening the door for improved in situ investigation and quantification of musculoskeletal tissues.

  2. Blind CT image quality assessment via deep learning strategy: initial study

    NASA Astrophysics Data System (ADS)

    Li, Sui; He, Ji; Wang, Yongbo; Liao, Yuting; Zeng, Dong; Bian, Zhaoying; Ma, Jianhua

    2018-03-01

    Computed Tomography (CT) is one of the most important medical imaging modality. CT images can be used to assist in the detection and diagnosis of lesions and to facilitate follow-up treatment. However, CT images are vulnerable to noise. Actually, there are two major source intrinsically causing the CT data noise, i.e., the X-ray photo statistics and the electronic noise background. Therefore, it is necessary to doing image quality assessment (IQA) in CT imaging before diagnosis and treatment. Most of existing CT images IQA methods are based on human observer study. However, these methods are impractical in clinical for their complex and time-consuming. In this paper, we presented a blind CT image quality assessment via deep learning strategy. A database of 1500 CT images is constructed, containing 300 high-quality images and 1200 corresponding noisy images. Specifically, the high-quality images were used to simulate the corresponding noisy images at four different doses. Then, the images are scored by the experienced radiologists by the following attributes: image noise, artifacts, edge and structure, overall image quality, and tumor size and boundary estimation with five-point scale. We trained a network for learning the non-liner map from CT images to subjective evaluation scores. Then, we load the pre-trained model to yield predicted score from the test image. To demonstrate the performance of the deep learning network in IQA, correlation coefficients: Pearson Linear Correlation Coefficient (PLCC) and Spearman Rank Order Correlation Coefficient (SROCC) are utilized. And the experimental result demonstrate that the presented deep learning based IQA strategy can be used in the CT image quality assessment.

  3. High Definition Confocal Imaging Modalities for the Characterization of Tissue-Engineered Substitutes.

    PubMed

    Mayrand, Dominique; Fradette, Julie

    2018-01-01

    Optimal imaging methods are necessary in order to perform a detailed characterization of thick tissue samples from either native or engineered tissues. Tissue-engineered substitutes are featuring increasing complexity including multiple cell types and capillary-like networks. Therefore, technical approaches allowing the visualization of the inner structural organization and cellular composition of tissues are needed. This chapter describes an optical clearing technique which facilitates the detailed characterization of whole-mount samples from skin and adipose tissues (ex vivo tissues and in vitro tissue-engineered substitutes) when combined with spectral confocal microscopy and quantitative analysis on image renderings.

  4. Segmenting Brain Tissues from Chinese Visible Human Dataset by Deep-Learned Features with Stacked Autoencoder

    PubMed Central

    Zhao, Guangjun; Wang, Xuchu; Niu, Yanmin; Tan, Liwen; Zhang, Shao-Xiang

    2016-01-01

    Cryosection brain images in Chinese Visible Human (CVH) dataset contain rich anatomical structure information of tissues because of its high resolution (e.g., 0.167 mm per pixel). Fast and accurate segmentation of these images into white matter, gray matter, and cerebrospinal fluid plays a critical role in analyzing and measuring the anatomical structures of human brain. However, most existing automated segmentation methods are designed for computed tomography or magnetic resonance imaging data, and they may not be applicable for cryosection images due to the imaging difference. In this paper, we propose a supervised learning-based CVH brain tissues segmentation method that uses stacked autoencoder (SAE) to automatically learn the deep feature representations. Specifically, our model includes two successive parts where two three-layer SAEs take image patches as input to learn the complex anatomical feature representation, and then these features are sent to Softmax classifier for inferring the labels. Experimental results validated the effectiveness of our method and showed that it outperformed four other classical brain tissue detection strategies. Furthermore, we reconstructed three-dimensional surfaces of these tissues, which show their potential in exploring the high-resolution anatomical structures of human brain. PMID:27057543

  5. Segmenting Brain Tissues from Chinese Visible Human Dataset by Deep-Learned Features with Stacked Autoencoder.

    PubMed

    Zhao, Guangjun; Wang, Xuchu; Niu, Yanmin; Tan, Liwen; Zhang, Shao-Xiang

    2016-01-01

    Cryosection brain images in Chinese Visible Human (CVH) dataset contain rich anatomical structure information of tissues because of its high resolution (e.g., 0.167 mm per pixel). Fast and accurate segmentation of these images into white matter, gray matter, and cerebrospinal fluid plays a critical role in analyzing and measuring the anatomical structures of human brain. However, most existing automated segmentation methods are designed for computed tomography or magnetic resonance imaging data, and they may not be applicable for cryosection images due to the imaging difference. In this paper, we propose a supervised learning-based CVH brain tissues segmentation method that uses stacked autoencoder (SAE) to automatically learn the deep feature representations. Specifically, our model includes two successive parts where two three-layer SAEs take image patches as input to learn the complex anatomical feature representation, and then these features are sent to Softmax classifier for inferring the labels. Experimental results validated the effectiveness of our method and showed that it outperformed four other classical brain tissue detection strategies. Furthermore, we reconstructed three-dimensional surfaces of these tissues, which show their potential in exploring the high-resolution anatomical structures of human brain.

  6. Deep learning and three-compartment breast imaging in breast cancer diagnosis

    NASA Astrophysics Data System (ADS)

    Drukker, Karen; Huynh, Benjamin Q.; Giger, Maryellen L.; Malkov, Serghei; Avila, Jesus I.; Fan, Bo; Joe, Bonnie; Kerlikowske, Karla; Drukteinis, Jennifer S.; Kazemi, Leila; Pereira, Malesa M.; Shepherd, John

    2017-03-01

    We investigated whether deep learning has potential to aid in the diagnosis of breast cancer when applied to mammograms and biologic tissue composition images derived from three-compartment (3CB) imaging. The dataset contained diagnostic mammograms and 3CB images (water, lipid, and protein content) of biopsy-sampled BIRADS 4 and 5 lesions in 195 patients. In 58 patients, the lesion manifested as a mass (13 malignant vs. 45 benign), in 87 as microcalcifications (19 vs. 68), and in 56 as (focal) asymmetry or architectural distortion (11 vs. 45). Six patients had both a mass and calcifications. For each mammogram and corresponding 3CB images, a 128x128 region of interest containing the lesion was selected by an expert radiologist and used directly as input to a deep learning method pretrained on a very large independent set of non-medical images. We used a nested leave-one-out-by-case (patient) model selection and classification protocol. The area under the ROC curve (AUC) for the task of distinguishing between benign and malignant lesions was used as performance metric. For the cases with mammographic masses, the AUC increased from 0.83 (mammograms alone) to 0.89 (mammograms+3CB, p=.162). For the microcalcification and asymmetry/architectural distortion cases the AUC increased from 0.84 to 0.91 (p=.116) and from 0.61 to 0.87 (p=.006), respectively. Our results indicate great potential for the application of deep learning methods in the diagnosis of breast cancer and additional knowledge of the biologic tissue composition appeared to improve performance, especially for lesions mammographically manifesting as asymmetries or architectural distortions.

  7. Sex differences and hormonal modulation of deep tissue pain

    PubMed Central

    Traub, Richard J.; Ji, Yaping

    2013-01-01

    Women disproportionately suffer from many deep tissue pain conditions. Experimental studies show that women have lower pain thresholds, higher pain ratings and less tolerance to a range of painful stimuli. Most clinical and epidemiological reports suggest female gonadal hormones modulate pain for some, but not all, conditions. Similarly, animal studies support greater nociceptive sensitivity in females in many deep tissue pain models. Gonadal hormones modulate responses in primary afferents, dorsal horn neurons and supraspinal sites, but the direction of modulation is variable. This review will examine sex differences in deep tissue pain in humans and animals focusing on the role of gonadal hormones (mainly estradiol) as an underlying component of the modulation of pain sensitivity. PMID:23872333

  8. Combining deep learning and coherent anti-Stokes Raman scattering imaging for automated differential diagnosis of lung cancer

    NASA Astrophysics Data System (ADS)

    Weng, Sheng; Xu, Xiaoyun; Li, Jiasong; Wong, Stephen T. C.

    2017-10-01

    Lung cancer is the most prevalent type of cancer and the leading cause of cancer-related deaths worldwide. Coherent anti-Stokes Raman scattering (CARS) is capable of providing cellular-level images and resolving pathologically related features on human lung tissues. However, conventional means of analyzing CARS images requires extensive image processing, feature engineering, and human intervention. This study demonstrates the feasibility of applying a deep learning algorithm to automatically differentiate normal and cancerous lung tissue images acquired by CARS. We leverage the features learned by pretrained deep neural networks and retrain the model using CARS images as the input. We achieve 89.2% accuracy in classifying normal, small-cell carcinoma, adenocarcinoma, and squamous cell carcinoma lung images. This computational method is a step toward on-the-spot diagnosis of lung cancer and can be further strengthened by the efforts aimed at miniaturizing the CARS technique for fiber-based microendoscopic imaging.

  9. Deep Imaging Survey

    NASA Image and Video Library

    2003-07-25

    This is the first Deep Imaging Survey image taken by NASA Galaxy Evolution Explorer. On June 22 and 23, 2003, the spacecraft obtained this near ultraviolet image of the Groth region by adding multiple orbits for a total exposure time of 14,000 seconds. Tens of thousands of objects can be identified in this picture. http://photojournal.jpl.nasa.gov/catalog/PIA04627

  10. An activity-based near-infrared glucuronide trapping probe for imaging β-glucuronidase expression in deep tissues.

    PubMed

    Cheng, Ta-Chun; Roffler, Steve R; Tzou, Shey-Cherng; Chuang, Kuo-Hsiang; Su, Yu-Cheng; Chuang, Chih-Hung; Kao, Chien-Han; Chen, Chien-Shu; Harn, I-Hong; Liu, Kuan-Yi; Cheng, Tian-Lu; Leu, Yu-Ling

    2012-02-15

    β-glucuronidase is an attractive reporter and prodrug-converting enzyme. The development of near-IR (NIR) probes for imaging of β-glucuronidase activity would be ideal to allow estimation of reporter expression and for personalized glucuronide prodrug cancer therapy in preclinical studies. However, NIR glucuronide probes are not yet available. In this work, we developed two fluorescent probes for detection of β-glucuronidase activity, one for the NIR range (containing IR-820 dye) and the other for the visible range [containing fluorescein isothiocyanate (FITC)], by utilizing a difluoromethylphenol-glucuronide moiety (TrapG) to trap the fluorochromes in the vicinity of the active enzyme. β-glucuronidase-mediated hydrolysis of the glucuronyl bond of TrapG generates a highly reactive alkylating group that facilitates the attachment of the fluorochrome to nucleophilic moieties located near β-glucuronidase-expressing sites. FITC-TrapG was selectively trapped on purified β-glucuronidase or β-glucuronidase-expressing CT26 cells (CT26/mβG) but not on bovine serum albumin or non-β-glucuronidase-expressing CT26 cells used as controls. β-glucuronidase-activated FITC-TrapG did not interfere with β-glucuronidase activity and could label bystander proteins near β-glucuronidase. Both FITC-TrapG and NIR-TrapG specifically imaged subcutaneous CT26/mβG tumors, but only NIR-TrapG could image CT26/mβG tumors transplanted deep in the liver. Thus NIR-TrapG may provide a valuable tool for visualizing β-glucuronidase activity in vivo.

  11. Deep Neural Networks Based Recognition of Plant Diseases by Leaf Image Classification.

    PubMed

    Sladojevic, Srdjan; Arsenovic, Marko; Anderla, Andras; Culibrk, Dubravko; Stefanovic, Darko

    2016-01-01

    The latest generation of convolutional neural networks (CNNs) has achieved impressive results in the field of image classification. This paper is concerned with a new approach to the development of plant disease recognition model, based on leaf image classification, by the use of deep convolutional networks. Novel way of training and the methodology used facilitate a quick and easy system implementation in practice. The developed model is able to recognize 13 different types of plant diseases out of healthy leaves, with the ability to distinguish plant leaves from their surroundings. According to our knowledge, this method for plant disease recognition has been proposed for the first time. All essential steps required for implementing this disease recognition model are fully described throughout the paper, starting from gathering images in order to create a database, assessed by agricultural experts. Caffe, a deep learning framework developed by Berkley Vision and Learning Centre, was used to perform the deep CNN training. The experimental results on the developed model achieved precision between 91% and 98%, for separate class tests, on average 96.3%.

  12. Deep Neural Networks Based Recognition of Plant Diseases by Leaf Image Classification

    PubMed Central

    Sladojevic, Srdjan; Arsenovic, Marko; Culibrk, Dubravko; Stefanovic, Darko

    2016-01-01

    The latest generation of convolutional neural networks (CNNs) has achieved impressive results in the field of image classification. This paper is concerned with a new approach to the development of plant disease recognition model, based on leaf image classification, by the use of deep convolutional networks. Novel way of training and the methodology used facilitate a quick and easy system implementation in practice. The developed model is able to recognize 13 different types of plant diseases out of healthy leaves, with the ability to distinguish plant leaves from their surroundings. According to our knowledge, this method for plant disease recognition has been proposed for the first time. All essential steps required for implementing this disease recognition model are fully described throughout the paper, starting from gathering images in order to create a database, assessed by agricultural experts. Caffe, a deep learning framework developed by Berkley Vision and Learning Centre, was used to perform the deep CNN training. The experimental results on the developed model achieved precision between 91% and 98%, for separate class tests, on average 96.3%. PMID:27418923

  13. Deep and Structured Robust Information Theoretic Learning for Image Analysis.

    PubMed

    Deng, Yue; Bao, Feng; Deng, Xuesong; Wang, Ruiping; Kong, Youyong; Dai, Qionghai

    2016-07-07

    This paper presents a robust information theoretic (RIT) model to reduce the uncertainties, i.e. missing and noisy labels, in general discriminative data representation tasks. The fundamental pursuit of our model is to simultaneously learn a transformation function and a discriminative classifier that maximize the mutual information of data and their labels in the latent space. In this general paradigm, we respectively discuss three types of the RIT implementations with linear subspace embedding, deep transformation and structured sparse learning. In practice, the RIT and deep RIT are exploited to solve the image categorization task whose performances will be verified on various benchmark datasets. The structured sparse RIT is further applied to a medical image analysis task for brain MRI segmentation that allows group-level feature selections on the brain tissues.

  14. Relevance of deep learning to facilitate the diagnosis of HER2 status in breast cancer

    PubMed Central

    Vandenberghe, Michel E.; Scott, Marietta L. J.; Scorer, Paul W.; Söderberg, Magnus; Balcerzak, Denis; Barker, Craig

    2017-01-01

    Tissue biomarker scoring by pathologists is central to defining the appropriate therapy for patients with cancer. Yet, inter-pathologist variability in the interpretation of ambiguous cases can affect diagnostic accuracy. Modern artificial intelligence methods such as deep learning have the potential to supplement pathologist expertise to ensure constant diagnostic accuracy. We developed a computational approach based on deep learning that automatically scores HER2, a biomarker that defines patient eligibility for anti-HER2 targeted therapies in breast cancer. In a cohort of 71 breast tumour resection samples, automated scoring showed a concordance of 83% with a pathologist. The twelve discordant cases were then independently reviewed, leading to a modification of diagnosis from initial pathologist assessment for eight cases. Diagnostic discordance was found to be largely caused by perceptual differences in assessing HER2 expression due to high HER2 staining heterogeneity. This study provides evidence that deep learning aided diagnosis can facilitate clinical decision making in breast cancer by identifying cases at high risk of misdiagnosis. PMID:28378829

  15. RGD-conjugated two-photon absorbing near-IR emitting fluorescent probes for tumor vascular imaging (Conference Presentation)

    NASA Astrophysics Data System (ADS)

    Belfield, Kevin D.; Yue, Xiling; Morales, Alma R.; Githaiga, Grace W.; Woodward, Adam W.; Tang, Simon; Sawada, Junko; Komatsu, Masanobu; Liu, Xuan

    2016-03-01

    Observation of the activation and inhibition of angiogenesis processes is important in the progression of cancer. Application of targeting peptides, such as a small peptide that contains adjacent L-arginine (R), glycine (G) and L-aspartic acid (D) residues can afford high selectivity and deep penetration in vessel imaging. To facilitate deep tissue vasculature imaging, probes that can be excited via two-photon absorption (2PA) in the near-infrared (NIR) and subsequently emit in the NIR are essential. In this study, the enhancement of tissue image quality with RGD conjugates was investigated with new NIR-emitting pyranyl fluorophore derivatives in two-photon fluorescence microscopy. Linear and nonlinear photophysical properties of the new probes were comprehensively characterized; significantly the probes exhibited good 2PA over a broad spectral range from 700-1100 nm. Cell and tissue images were then acquired and examined, revealing deep penetration and high contrast with the new pyranyl RGD-conjugates up to 350 μm in tumor tissue.

  16. DeepInfer: open-source deep learning deployment toolkit for image-guided therapy

    NASA Astrophysics Data System (ADS)

    Mehrtash, Alireza; Pesteie, Mehran; Hetherington, Jorden; Behringer, Peter A.; Kapur, Tina; Wells, William M.; Rohling, Robert; Fedorov, Andriy; Abolmaesumi, Purang

    2017-03-01

    Deep learning models have outperformed some of the previous state-of-the-art approaches in medical image analysis. Instead of using hand-engineered features, deep models attempt to automatically extract hierarchical representations at multiple levels of abstraction from the data. Therefore, deep models are usually considered to be more flexible and robust solutions for image analysis problems compared to conventional computer vision models. They have demonstrated significant improvements in computer-aided diagnosis and automatic medical image analysis applied to such tasks as image segmentation, classification and registration. However, deploying deep learning models often has a steep learning curve and requires detailed knowledge of various software packages. Thus, many deep models have not been integrated into the clinical research work ows causing a gap between the state-of-the-art machine learning in medical applications and evaluation in clinical research procedures. In this paper, we propose "DeepInfer" - an open-source toolkit for developing and deploying deep learning models within the 3D Slicer medical image analysis platform. Utilizing a repository of task-specific models, DeepInfer allows clinical researchers and biomedical engineers to deploy a trained model selected from the public registry, and apply it to new data without the need for software development or configuration. As two practical use cases, we demonstrate the application of DeepInfer in prostate segmentation for targeted MRI-guided biopsy and identification of the target plane in 3D ultrasound for spinal injections.

  17. DeepInfer: Open-Source Deep Learning Deployment Toolkit for Image-Guided Therapy.

    PubMed

    Mehrtash, Alireza; Pesteie, Mehran; Hetherington, Jorden; Behringer, Peter A; Kapur, Tina; Wells, William M; Rohling, Robert; Fedorov, Andriy; Abolmaesumi, Purang

    2017-02-11

    Deep learning models have outperformed some of the previous state-of-the-art approaches in medical image analysis. Instead of using hand-engineered features, deep models attempt to automatically extract hierarchical representations at multiple levels of abstraction from the data. Therefore, deep models are usually considered to be more flexible and robust solutions for image analysis problems compared to conventional computer vision models. They have demonstrated significant improvements in computer-aided diagnosis and automatic medical image analysis applied to such tasks as image segmentation, classification and registration. However, deploying deep learning models often has a steep learning curve and requires detailed knowledge of various software packages. Thus, many deep models have not been integrated into the clinical research workflows causing a gap between the state-of-the-art machine learning in medical applications and evaluation in clinical research procedures. In this paper, we propose "DeepInfer" - an open-source toolkit for developing and deploying deep learning models within the 3D Slicer medical image analysis platform. Utilizing a repository of task-specific models, DeepInfer allows clinical researchers and biomedical engineers to deploy a trained model selected from the public registry, and apply it to new data without the need for software development or configuration. As two practical use cases, we demonstrate the application of DeepInfer in prostate segmentation for targeted MRI-guided biopsy and identification of the target plane in 3D ultrasound for spinal injections.

  18. DeepInfer: Open-Source Deep Learning Deployment Toolkit for Image-Guided Therapy

    PubMed Central

    Mehrtash, Alireza; Pesteie, Mehran; Hetherington, Jorden; Behringer, Peter A.; Kapur, Tina; Wells, William M.; Rohling, Robert; Fedorov, Andriy; Abolmaesumi, Purang

    2017-01-01

    Deep learning models have outperformed some of the previous state-of-the-art approaches in medical image analysis. Instead of using hand-engineered features, deep models attempt to automatically extract hierarchical representations at multiple levels of abstraction from the data. Therefore, deep models are usually considered to be more flexible and robust solutions for image analysis problems compared to conventional computer vision models. They have demonstrated significant improvements in computer-aided diagnosis and automatic medical image analysis applied to such tasks as image segmentation, classification and registration. However, deploying deep learning models often has a steep learning curve and requires detailed knowledge of various software packages. Thus, many deep models have not been integrated into the clinical research workflows causing a gap between the state-of-the-art machine learning in medical applications and evaluation in clinical research procedures. In this paper, we propose “DeepInfer” – an open-source toolkit for developing and deploying deep learning models within the 3D Slicer medical image analysis platform. Utilizing a repository of task-specific models, DeepInfer allows clinical researchers and biomedical engineers to deploy a trained model selected from the public registry, and apply it to new data without the need for software development or configuration. As two practical use cases, we demonstrate the application of DeepInfer in prostate segmentation for targeted MRI-guided biopsy and identification of the target plane in 3D ultrasound for spinal injections. PMID:28615794

  19. Combining deep learning and coherent anti-Stokes Raman scattering imaging for automated differential diagnosis of lung cancer.

    PubMed

    Weng, Sheng; Xu, Xiaoyun; Li, Jiasong; Wong, Stephen T C

    2017-10-01

    Lung cancer is the most prevalent type of cancer and the leading cause of cancer-related deaths worldwide. Coherent anti-Stokes Raman scattering (CARS) is capable of providing cellular-level images and resolving pathologically related features on human lung tissues. However, conventional means of analyzing CARS images requires extensive image processing, feature engineering, and human intervention. This study demonstrates the feasibility of applying a deep learning algorithm to automatically differentiate normal and cancerous lung tissue images acquired by CARS. We leverage the features learned by pretrained deep neural networks and retrain the model using CARS images as the input. We achieve 89.2% accuracy in classifying normal, small-cell carcinoma, adenocarcinoma, and squamous cell carcinoma lung images. This computational method is a step toward on-the-spot diagnosis of lung cancer and can be further strengthened by the efforts aimed at miniaturizing the CARS technique for fiber-based microendoscopic imaging. (2017) COPYRIGHT Society of Photo-Optical Instrumentation Engineers (SPIE).

  20. Assessment of deep tissue hyperalgesia in the groin - a method comparison of electrical vs. pressure stimulation.

    PubMed

    Aasvang, E K; Werner, M U; Kehlet, H

    2014-09-01

    Deep pain complaints are more frequent than cutaneous in post-surgical patients, and a prevalent finding in quantitative sensory testing studies. However, the preferred assessment method - pressure algometry - is indirect and tissue unspecific, hindering advances in treatment and preventive strategies. Thus, there is a need for development of methods with direct stimulation of suspected hyperalgesic tissues to identify the peripheral origin of nociceptive input. We compared the reliability of an ultrasound-guided needle stimulation protocol of electrical detection and pain thresholds to pressure algometry, by performing identical test-retest sequences 10 days apart, in deep tissues in the groin region. Electrical stimulation was performed by five up-and-down staircase series of single impulses of 0.04 ms duration, starting from 0 mA in increments of 0.2 mA until a threshold was reached and descending until sensation was lost. Method reliability was assessed by Bland-Altman plots, descriptive statistics, coefficients of variance and intraclass correlation coefficients. The electrical stimulation method was comparable to pressure algometry regarding 10 days test-retest repeatability, but with superior same-day reliability for electrical stimulation (P < 0.05). Between-subject variance rather than within-subject variance was the main source for test variation. There were no systematic differences in electrical thresholds across tissues and locations (P > 0.05). The presented tissue-specific direct deep tissue electrical stimulation technique has equal or superior reliability compared with the indirect tissue-unspecific stimulation by pressure algometry. This method may facilitate advances in mechanism based preventive and treatment strategies in acute and chronic post-surgical pain states. © 2014 The Acta Anaesthesiologica Scandinavica Foundation. Published by John Wiley & Sons Ltd.

  1. Recent progress in tissue optical clearing for spectroscopic application

    NASA Astrophysics Data System (ADS)

    Sdobnov, A. Yu.; Darvin, M. E.; Genina, E. A.; Bashkatov, A. N.; Lademann, J.; Tuchin, V. V.

    2018-05-01

    This paper aims to review recent progress in optical clearing of the skin and over naturally turbid biological tissues and blood using this technique in vivo and in vitro with multiphoton microscopy, confocal Raman microscopy, confocal microscopy, NIR spectroscopy, optical coherence tomography, and laser speckle contrast imaging. Basic principles of the technique, its safety, advantages and limitations are discussed. The application of optical clearing agent on a tissue allows for controlling the optical properties of tissue. Optical clearing-induced reduction of tissue scattering significantly facilitates the observation of deep-located tissue regions, at the same time improving the resolution and image contrast for a variety of optical imaging methods suitable for clinical applications, such as diagnostics and laser treatment of skin diseases, mucosal tumor imaging, laser disruption of pathological abnormalities, etc. Structural images of different skin layers obtained ex vivo for porcine ear skin samples at application of Omnipaque™ and glycerol solutions during 60 min. Red color corresponds to TPEAF signal channel. Green color corresponds to SHG signal channel.

  2. Groth Deep Image

    NASA Image and Video Library

    2003-07-25

    This ultraviolet color blowup of the Groth Deep Image was taken by NASA Galaxy Evolution Explorer on June 22 and June 23, 2003. Many hundreds of galaxies are detected in this portion of the image. NASA astronomers believe the faint red galaxies are 6 billion light years away. http://photojournal.jpl.nasa.gov/catalog/PIA04625

  3. In vivo mapping of current density distribution in brain tissues during deep brain stimulation (DBS)

    NASA Astrophysics Data System (ADS)

    Sajib, Saurav Z. K.; Oh, Tong In; Kim, Hyung Joong; Kwon, Oh In; Woo, Eung Je

    2017-01-01

    New methods for in vivo mapping of brain responses during deep brain stimulation (DBS) are indispensable to secure clinical applications. Assessment of current density distribution, induced by internally injected currents, may provide an alternative method for understanding the therapeutic effects of electrical stimulation. The current flow and pathway are affected by internal conductivity, and can be imaged using magnetic resonance-based conductivity imaging methods. Magnetic resonance electrical impedance tomography (MREIT) is an imaging method that can enable highly resolved mapping of electromagnetic tissue properties such as current density and conductivity of living tissues. In the current study, we experimentally imaged current density distribution of in vivo canine brains by applying MREIT to electrical stimulation. The current density maps of three canine brains were calculated from the measured magnetic flux density data. The absolute current density values of brain tissues, including gray matter, white matter, and cerebrospinal fluid were compared to assess the active regions during DBS. The resulting current density in different tissue types may provide useful information about current pathways and volume activation for adjusting surgical planning and understanding the therapeutic effects of DBS.

  4. Automatic segmentation of the prostate on CT images using deep learning and multi-atlas fusion

    NASA Astrophysics Data System (ADS)

    Ma, Ling; Guo, Rongrong; Zhang, Guoyi; Tade, Funmilayo; Schuster, David M.; Nieh, Peter; Master, Viraj; Fei, Baowei

    2017-02-01

    Automatic segmentation of the prostate on CT images has many applications in prostate cancer diagnosis and therapy. However, prostate CT image segmentation is challenging because of the low contrast of soft tissue on CT images. In this paper, we propose an automatic segmentation method by combining a deep learning method and multi-atlas refinement. First, instead of segmenting the whole image, we extract the region of interesting (ROI) to delete irrelevant regions. Then, we use the convolutional neural networks (CNN) to learn the deep features for distinguishing the prostate pixels from the non-prostate pixels in order to obtain the preliminary segmentation results. CNN can automatically learn the deep features adapting to the data, which are different from some handcrafted features. Finally, we select some similar atlases to refine the initial segmentation results. The proposed method has been evaluated on a dataset of 92 prostate CT images. Experimental results show that our method achieved a Dice similarity coefficient of 86.80% as compared to the manual segmentation. The deep learning based method can provide a useful tool for automatic segmentation of the prostate on CT images and thus can have a variety of clinical applications.

  5. A Deep Learning Approach to Digitally Stain Optical Coherence Tomography Images of the Optic Nerve Head.

    PubMed

    Devalla, Sripad Krishna; Chin, Khai Sing; Mari, Jean-Martial; Tun, Tin A; Strouthidis, Nicholas G; Aung, Tin; Thiéry, Alexandre H; Girard, Michaël J A

    2018-01-01

    To develop a deep learning approach to digitally stain optical coherence tomography (OCT) images of the optic nerve head (ONH). A horizontal B-scan was acquired through the center of the ONH using OCT (Spectralis) for one eye of each of 100 subjects (40 healthy and 60 glaucoma). All images were enhanced using adaptive compensation. A custom deep learning network was then designed and trained with the compensated images to digitally stain (i.e., highlight) six tissue layers of the ONH. The accuracy of our algorithm was assessed (against manual segmentations) using the dice coefficient, sensitivity, specificity, intersection over union (IU), and accuracy. We studied the effect of compensation, number of training images, and performance comparison between glaucoma and healthy subjects. For images it had not yet assessed, our algorithm was able to digitally stain the retinal nerve fiber layer + prelamina, the RPE, all other retinal layers, the choroid, and the peripapillary sclera and lamina cribrosa. For all tissues, the dice coefficient, sensitivity, specificity, IU, and accuracy (mean) were 0.84 ± 0.03, 0.92 ± 0.03, 0.99 ± 0.00, 0.89 ± 0.03, and 0.94 ± 0.02, respectively. Our algorithm performed significantly better when compensated images were used for training (P < 0.001). Besides offering a good reliability, digital staining also performed well on OCT images of both glaucoma and healthy individuals. Our deep learning algorithm can simultaneously stain the neural and connective tissues of the ONH, offering a framework to automatically measure multiple key structural parameters of the ONH that may be critical to improve glaucoma management.

  6. Biodynamic imaging for phenotypic profiling of three-dimensional tissue culture

    NASA Astrophysics Data System (ADS)

    Sun, Hao; Merrill, Daniel; An, Ran; Turek, John; Matei, Daniela; Nolte, David D.

    2017-01-01

    Three-dimensional (3-D) tissue culture represents a more biologically relevant environment for testing new drugs compared to conventional two-dimensional cancer cell culture models. Biodynamic imaging is a high-content 3-D optical imaging technology based on low-coherence interferometry and digital holography that uses dynamic speckle as high-content image contrast to probe deep inside 3-D tissue. Speckle contrast is shown to be a scaling function of the acquisition time relative to the persistence time of intracellular transport and hence provides a measure of cellular activity. Cellular responses of 3-D multicellular spheroids to paclitaxel are compared among three different growth techniques: rotating bioreactor (BR), hanging-drop (HD), and nonadherent (U-bottom, UB) plate spheroids, compared with ex vivo living tissues. HD spheroids have the most homogeneous tissue, whereas BR spheroids display large sample-to-sample variability as well as spatial heterogeneity. The responses of BR-grown tumor spheroids to paclitaxel are more similar to those of ex vivo biopsies than the responses of spheroids grown using HD or plate methods. The rate of mitosis inhibition by application of taxol is measured through tissue dynamics spectroscopic imaging, demonstrating the ability to monitor antimitotic chemotherapy. These results illustrate the potential use of low-coherence digital holography for 3-D pharmaceutical screening applications.

  7. Optical toolkits for in vivo deep tissue laser scanning microscopy: a primer

    NASA Astrophysics Data System (ADS)

    Lee, Woei Ming; McMenamin, Thomas; Li, Yongxiao

    2018-06-01

    Life at the microscale is animated and multifaceted. The impact of dynamic in vivo microscopy in small animals has opened up opportunities to peer into a multitude of biological processes at the cellular scale in their native microenvironments. Laser scanning microscopy (LSM) coupled with targeted fluorescent proteins has become an indispensable tool to enable dynamic imaging in vivo at high temporal and spatial resolutions. In the last few decades, the technique has been translated from imaging cells in thin samples to mapping cells in the thick biological tissue of living organisms. Here, we sought to provide a concise overview of the design considerations of a LSM that enables cellular and subcellular imaging in deep tissue. Individual components under review include: long working distance microscope objectives, laser scanning technologies, adaptive optics devices, beam shaping technologies and photon detectors, with an emphasis on more recent advances. The review will conclude with the latest innovations in automated optical microscopy, which would impact tracking and quantification of heterogeneous populations of cells in vivo.

  8. Deep, noninvasive imaging and surgical guidance of submillimeter tumors using targeted M13-stabilized single-walled carbon nanotubes

    PubMed Central

    Ghosh, Debadyuti; Bagley, Alexander F.; Na, Young Jeong; Birrer, Michael J.; Bhatia, Sangeeta N.; Belcher, Angela M.

    2014-01-01

    Highly sensitive detection of small, deep tumors for early diagnosis and surgical interventions remains a challenge for conventional imaging modalities. Second-window near-infrared light (NIR2, 950–1,400 nm) is promising for in vivo fluorescence imaging due to deep tissue penetration and low tissue autofluorescence. With their intrinsic fluorescence in the NIR2 regime and lack of photobleaching, single-walled carbon nanotubes (SWNTs) are potentially attractive contrast agents to detect tumors. Here, targeted M13 virus-stabilized SWNTs are used to visualize deep, disseminated tumors in vivo. This targeted nanoprobe, which uses M13 to stably display both tumor-targeting peptides and an SWNT imaging probe, demonstrates excellent tumor-to-background uptake and exhibits higher signal-to-noise performance compared with visible and near-infrared (NIR1) dyes for delineating tumor nodules. Detection and excision of tumors by a gynecological surgeon improved with SWNT image guidance and led to the identification of submillimeter tumors. Collectively, these findings demonstrate the promise of targeted SWNT nanoprobes for noninvasive disease monitoring and guided surgery. PMID:25214538

  9. Deep, noninvasive imaging and surgical guidance of submillimeter tumors using targeted M13-stabilized single-walled carbon nanotubes.

    PubMed

    Ghosh, Debadyuti; Bagley, Alexander F; Na, Young Jeong; Birrer, Michael J; Bhatia, Sangeeta N; Belcher, Angela M

    2014-09-23

    Highly sensitive detection of small, deep tumors for early diagnosis and surgical interventions remains a challenge for conventional imaging modalities. Second-window near-infrared light (NIR2, 950-1,400 nm) is promising for in vivo fluorescence imaging due to deep tissue penetration and low tissue autofluorescence. With their intrinsic fluorescence in the NIR2 regime and lack of photobleaching, single-walled carbon nanotubes (SWNTs) are potentially attractive contrast agents to detect tumors. Here, targeted M13 virus-stabilized SWNTs are used to visualize deep, disseminated tumors in vivo. This targeted nanoprobe, which uses M13 to stably display both tumor-targeting peptides and an SWNT imaging probe, demonstrates excellent tumor-to-background uptake and exhibits higher signal-to-noise performance compared with visible and near-infrared (NIR1) dyes for delineating tumor nodules. Detection and excision of tumors by a gynecological surgeon improved with SWNT image guidance and led to the identification of submillimeter tumors. Collectively, these findings demonstrate the promise of targeted SWNT nanoprobes for noninvasive disease monitoring and guided surgery.

  10. Time-lapse imaging of disease progression in deep brain areas using fluorescence microendoscopy

    PubMed Central

    Barretto, Robert P. J.; Ko, Tony H.; Jung, Juergen C.; Wang, Tammy J.; Capps, George; Waters, Allison C.; Ziv, Yaniv; Attardo, Alessio; Recht, Lawrence; Schnitzer, Mark J.

    2013-01-01

    The combination of intravital microscopy and animal models of disease has propelled studies of disease mechanisms and treatments. However, many disorders afflict tissues inaccessible to light microscopy in live subjects. Here we introduce cellular-level time-lapse imaging deep within the live mammalian brain by one- and two-photon fluorescence microendoscopy over multiple weeks. Bilateral imaging sites allowed longitudinal comparisons within individual subjects, including of normal and diseased tissues. Using this approach we tracked CA1 hippocampal pyramidal neuron dendrites in adult mice, revealing these dendrites' extreme stability (>8,000 day mean lifetime) and rare examples of their structural alterations. To illustrate disease studies, we tracked deep lying gliomas by observing tumor growth, visualizing three-dimensional vasculature structure, and determining microcirculatory speeds. Average erythrocyte speeds in gliomas declined markedly as the disease advanced, notwithstanding significant increases in capillary diameters. Time-lapse microendoscopy will be applicable to studies of numerous disorders, including neurovascular, neurological, cancerous, and trauma-induced conditions. PMID:21240263

  11. Computational ghost imaging using deep learning

    NASA Astrophysics Data System (ADS)

    Shimobaba, Tomoyoshi; Endo, Yutaka; Nishitsuji, Takashi; Takahashi, Takayuki; Nagahama, Yuki; Hasegawa, Satoki; Sano, Marie; Hirayama, Ryuji; Kakue, Takashi; Shiraki, Atsushi; Ito, Tomoyoshi

    2018-04-01

    Computational ghost imaging (CGI) is a single-pixel imaging technique that exploits the correlation between known random patterns and the measured intensity of light transmitted (or reflected) by an object. Although CGI can obtain two- or three-dimensional images with a single or a few bucket detectors, the quality of the reconstructed images is reduced by noise due to the reconstruction of images from random patterns. In this study, we improve the quality of CGI images using deep learning. A deep neural network is used to automatically learn the features of noise-contaminated CGI images. After training, the network is able to predict low-noise images from new noise-contaminated CGI images.

  12. Identifying mitosis deep in tissue using dynamic light scattering fluctuation spectroscopy

    NASA Astrophysics Data System (ADS)

    An, Ran; Jeong, Kwan; Turek, John; Nolte, David

    2012-03-01

    In the cell cycle, mitosis is the most dramatic phase, especially in Telophase and Cytokinesis. For single cells and cell monolayer, there are precise microscopic studies of mitosis, while for 3-D tissue such as tumor spheroids the light signal is obscured by the high background of diffusely scattered light. Therefore, the mitosis phase cannot be detected deep inside 3-D tissue using conventional microscopic techniques. In this work, we detect mitosis in living tissue using Tissue Dynamic Imaging (TDI). We trace depth-gated dynamic speckles from a tumor spheroid (up to 1mm in diameter) using coherence-gated digital holography imaging. Frequency-versus-time spectrograms depend on specific types of perturbation such as cell shape change, membrane undulation and cell organelles movements. By using these spectral responses as functional finger prints, we can identify mitosis events from different voxels at a specified depth inside tumor spheroids. By performing B-scans of the tumor spheroid, we generate 3-D mitosis maps (or movies) for the entire tumor spheroids. We show that for healthy tumor spheroids, the mitosis events only happen within the proliferating shell. We also compare results when anti-cancer drugs are applied to arrest, release and synchronize mitosis. This shows the application of TDI for drug screening. The technique can identify and monitor complex motilities inside 3-D tissue with a strong potential for drug diagnosis and developmental biology studies.

  13. A simultaneous multimodal imaging system for tissue functional parameters

    NASA Astrophysics Data System (ADS)

    Ren, Wenqi; Zhang, Zhiwu; Wu, Qiang; Zhang, Shiwu; Xu, Ronald

    2014-02-01

    Simultaneous and quantitative assessment of skin functional characteristics in different modalities will facilitate diagnosis and therapy in many clinical applications such as wound healing. However, many existing clinical practices and multimodal imaging systems are subjective, qualitative, sequential for multimodal data collection, and need co-registration between different modalities. To overcome these limitations, we developed a multimodal imaging system for quantitative, non-invasive, and simultaneous imaging of cutaneous tissue oxygenation and blood perfusion parameters. The imaging system integrated multispectral and laser speckle imaging technologies into one experimental setup. A Labview interface was developed for equipment control, synchronization, and image acquisition. Advanced algorithms based on a wide gap second derivative reflectometry and laser speckle contrast analysis (LASCA) were developed for accurate reconstruction of tissue oxygenation and blood perfusion respectively. Quantitative calibration experiments and a new style of skinsimulating phantom were designed to verify the accuracy and reliability of the imaging system. The experimental results were compared with a Moor tissue oxygenation and perfusion monitor. For In vivo testing, a post-occlusion reactive hyperemia (PORH) procedure in human subject and an ongoing wound healing monitoring experiment using dorsal skinfold chamber models were conducted to validate the usability of our system for dynamic detection of oxygenation and perfusion parameters. In this study, we have not only setup an advanced multimodal imaging system for cutaneous tissue oxygenation and perfusion parameters but also elucidated its potential for wound healing assessment in clinical practice.

  14. Deep Learning Nuclei Detection in Digitized Histology Images by Superpixels.

    PubMed

    Sornapudi, Sudhir; Stanley, Ronald Joe; Stoecker, William V; Almubarak, Haidar; Long, Rodney; Antani, Sameer; Thoma, George; Zuna, Rosemary; Frazier, Shelliane R

    2018-01-01

    Advances in image analysis and computational techniques have facilitated automatic detection of critical features in histopathology images. Detection of nuclei is critical for squamous epithelium cervical intraepithelial neoplasia (CIN) classification into normal, CIN1, CIN2, and CIN3 grades. In this study, a deep learning (DL)-based nuclei segmentation approach is investigated based on gathering localized information through the generation of superpixels using a simple linear iterative clustering algorithm and training with a convolutional neural network. The proposed approach was evaluated on a dataset of 133 digitized histology images and achieved an overall nuclei detection (object-based) accuracy of 95.97%, with demonstrated improvement over imaging-based and clustering-based benchmark techniques. The proposed DL-based nuclei segmentation Method with superpixel analysis has shown improved segmentation results in comparison to state-of-the-art methods.

  15. Deep Learning Nuclei Detection in Digitized Histology Images by Superpixels

    PubMed Central

    Sornapudi, Sudhir; Stanley, Ronald Joe; Stoecker, William V.; Almubarak, Haidar; Long, Rodney; Antani, Sameer; Thoma, George; Zuna, Rosemary; Frazier, Shelliane R.

    2018-01-01

    Background: Advances in image analysis and computational techniques have facilitated automatic detection of critical features in histopathology images. Detection of nuclei is critical for squamous epithelium cervical intraepithelial neoplasia (CIN) classification into normal, CIN1, CIN2, and CIN3 grades. Methods: In this study, a deep learning (DL)-based nuclei segmentation approach is investigated based on gathering localized information through the generation of superpixels using a simple linear iterative clustering algorithm and training with a convolutional neural network. Results: The proposed approach was evaluated on a dataset of 133 digitized histology images and achieved an overall nuclei detection (object-based) accuracy of 95.97%, with demonstrated improvement over imaging-based and clustering-based benchmark techniques. Conclusions: The proposed DL-based nuclei segmentation Method with superpixel analysis has shown improved segmentation results in comparison to state-of-the-art methods. PMID:29619277

  16. The effect of adjunctive noncontact low frequency ultrasound on deep tissue pressure injury.

    PubMed

    Honaker, Jeremy S; Forston, Michael R; Davis, Emily A; Weisner, Michelle M; Morgan, Jennifer A; Sacca, Emily

    2016-11-01

    The optimal treatment for deep tissue pressure injuries has not been determined. Deep tissue pressure injuries represent a more ominous early stage pressure injury that may evolve into full thickness ulceration despite implementing the standard of care for pressure injury. A longitudinal prospective historical case control study design was used to determine the effectiveness of noncontact low frequency ultrasound plus standard of care (treatment group) in comparison to standard of care (control group) in reducing deep tissue pressure injury severity, total surface area, and final pressure injury stage. The Honaker Suspected Deep Tissue Injury Severity Scale (range 3-18[more severe]) was used to determine deep tissue pressure injury severity at enrollment (Time 1) and discharge (Time 2). A total of 60 subjects (Treatment = 30; Control= 30) were enrolled in the study. In comparison to the control group mean deep tissue pressure injury total surface area change at Time 2 (0.3 cm 2 ), the treatment group had a greater decrease (8.8 cm 2 ) that was significant (t = 2.41, p = 0.014, r 2  = 0.10). In regards to the Honaker Suspected Deep Tissue Injury Severity Scale scores, the treatment group had a significantly lower score (7.6) in comparison to the control group (11.9) at time 2, with a mean difference of 4.6 (t = 6.146, p = 0.0001, r 2  = 0.39). When considering the final pressure ulcer stage at Time 2, the control group were mostly composed of unstageable pressure ulcer (57%) and deep tissue pressure injury severity (27%). In contrast, the treatment group final pressure ulcer stages were less severe and were mostly composed of stage 2 pressure injury (50%) and deep tissue pressure injury severity (23%) were the most common at time 2. The results of this study have shown that deep tissue pressure injury severity treated with noncontact low frequency ultrasound within 5 days of onset and in conjunction with standard of care may improve

  17. Hello World Deep Learning in Medical Imaging.

    PubMed

    Lakhani, Paras; Gray, Daniel L; Pett, Carl R; Nagy, Paul; Shih, George

    2018-05-03

    There is recent popularity in applying machine learning to medical imaging, notably deep learning, which has achieved state-of-the-art performance in image analysis and processing. The rapid adoption of deep learning may be attributed to the availability of machine learning frameworks and libraries to simplify their use. In this tutorial, we provide a high-level overview of how to build a deep neural network for medical image classification, and provide code that can help those new to the field begin their informatics projects.

  18. Biodynamic imaging for phenotypic profiling of three-dimensional tissue culture

    PubMed Central

    Sun, Hao; Merrill, Daniel; An, Ran; Turek, John; Matei, Daniela; Nolte, David D.

    2017-01-01

    Abstract. Three-dimensional (3-D) tissue culture represents a more biologically relevant environment for testing new drugs compared to conventional two-dimensional cancer cell culture models. Biodynamic imaging is a high-content 3-D optical imaging technology based on low-coherence interferometry and digital holography that uses dynamic speckle as high-content image contrast to probe deep inside 3-D tissue. Speckle contrast is shown to be a scaling function of the acquisition time relative to the persistence time of intracellular transport and hence provides a measure of cellular activity. Cellular responses of 3-D multicellular spheroids to paclitaxel are compared among three different growth techniques: rotating bioreactor (BR), hanging-drop (HD), and nonadherent (U-bottom, UB) plate spheroids, compared with ex vivo living tissues. HD spheroids have the most homogeneous tissue, whereas BR spheroids display large sample-to-sample variability as well as spatial heterogeneity. The responses of BR-grown tumor spheroids to paclitaxel are more similar to those of ex vivo biopsies than the responses of spheroids grown using HD or plate methods. The rate of mitosis inhibition by application of taxol is measured through tissue dynamics spectroscopic imaging, demonstrating the ability to monitor antimitotic chemotherapy. These results illustrate the potential use of low-coherence digital holography for 3-D pharmaceutical screening applications. PMID:28301634

  19. Autofluorescence-based diagnostic UV imaging of tissues and cells

    NASA Astrophysics Data System (ADS)

    Renkoski, Timothy E.

    Cancer is the second leading cause of death in the United States, and its early diagnosis is critical to improving treatment options and patient outcomes. In autofluorescence (AF) imaging, light of controlled wavelengths is projected onto tissue, absorbed by specific molecules, and re-emitted at longer wavelengths. Images of re-emitted light are used together with spectral information to infer tissue functional information and diagnosis. This dissertation describes AF imaging studies of three different organs using data collected from fresh human surgical specimens. In the ovary study, illumination was at 365 nm, and images were captured at 8 emission wavelengths. Measurements from a multispectral imaging system and fiber optic probe were used to map tissue diagnosis at every image pixel. For the colon and pancreas studies, instrumentation was developed extending AF imaging capability to sub-300 nm excitation. Images excited in the deep UV revealed tryptophan and protein content which are believed to change with disease state. Several excitation wavelength bands from 280 nm to 440 nm were investigated. Microscopic AF images collected in the pancreas study included both cultured and primary cells. Several findings are reported. A method of transforming fiber optic probe spectra for direct comparison with imager spectra was devised. Normalization of AF data by green reflectance data was found useful in correcting hemoglobin absorption. Ratio images, both AF and reflectance, were formulated to highlight growths in the colon. Novel tryptophan AF images were found less useful for colon diagnostics than the new ratio techniques. Microscopic tryptophan AF images produce useful visualization of cellular protein content, but their diagnostic value requires further study.

  20. In vivo three-photon imaging of deep cerebellum

    NASA Astrophysics Data System (ADS)

    Wang, Mengran; Wang, Tianyu; Wu, Chunyan; Li, Bo; Ouzounov, Dimitre G.; Sinefeld, David; Guru, Akash; Nam, Hyung-Song; Capecchi, Mario R.; Warden, Melissa R.; Xu, Chris

    2018-02-01

    We demonstrate three-photon microscopy (3PM) of mouse cerebellum at 1 mm depth by imaging both blood vessels and neurons. We compared 3PM and 2PM in the mouse cerebellum for imaging green (using excitation sources at 1300 nm and 920 nm, respectively) and red fluorescence (using excitation sources at 1680 nm and 1064 nm, respectively). 3PM enabled deeper imaging than 2PM because the use of longer excitation wavelength reduces the scattering in biological tissue and the higher order nonlinear excitation provides better 3D localization. To illustrate these two advantages quantitatively, we measured the signal decay as well as the signal-to-background ratio (SBR) as a function of depth. We performed 2-photon imaging from the brain surface all the way down to the area where the SBR reaches 1, while at the same depth, 3PM still has SBR above 30. The segmented decay curve shows that the mouse cerebellum has different effective attenuation lengths at different depths, indicating heterogeneous tissue property for this brain region. We compared the third harmonic generation (THG) signal, which is used to visualize myelinated fibers, with the decay curve. We found that the regions with shorter effective attenuation lengths correspond to the regions with more fibers. Our results indicate that the widespread, non-uniformly distributed myelinated fibers adds heterogeneity to mouse cerebellum, which poses additional challenges in deep imaging of this brain region.

  1. Searching for prostate cancer by fully automated magnetic resonance imaging classification: deep learning versus non-deep learning.

    PubMed

    Wang, Xinggang; Yang, Wei; Weinreb, Jeffrey; Han, Juan; Li, Qiubai; Kong, Xiangchuang; Yan, Yongluan; Ke, Zan; Luo, Bo; Liu, Tao; Wang, Liang

    2017-11-13

    Prostate cancer (PCa) is a major cause of death since ancient time documented in Egyptian Ptolemaic mummy imaging. PCa detection is critical to personalized medicine and varies considerably under an MRI scan. 172 patients with 2,602 morphologic images (axial 2D T2-weighted imaging) of the prostate were obtained. A deep learning with deep convolutional neural network (DCNN) and a non-deep learning with SIFT image feature and bag-of-word (BoW), a representative method for image recognition and analysis, were used to distinguish pathologically confirmed PCa patients from prostate benign conditions (BCs) patients with prostatitis or prostate benign hyperplasia (BPH). In fully automated detection of PCa patients, deep learning had a statistically higher area under the receiver operating characteristics curve (AUC) than non-deep learning (P = 0.0007 < 0.001). The AUCs were 0.84 (95% CI 0.78-0.89) for deep learning method and 0.70 (95% CI 0.63-0.77) for non-deep learning method, respectively. Our results suggest that deep learning with DCNN is superior to non-deep learning with SIFT image feature and BoW model for fully automated PCa patients differentiation from prostate BCs patients. Our deep learning method is extensible to image modalities such as MR imaging, CT and PET of other organs.

  2. Deep Tissue Photoacoustic Imaging Using a Miniaturized 2-D Capacitive Micromachined Ultrasonic Transducer Array

    PubMed Central

    Kothapalli, Sri-Rajasekhar; Ma, Te-Jen; Vaithilingam, Srikant; Oralkan, Ömer

    2014-01-01

    In this paper, we demonstrate 3-D photoacoustic imaging (PAI) of light absorbing objects embedded as deep as 5 cm inside strong optically scattering phantoms using a miniaturized (4 mm × 4 mm × 500 µm), 2-D capacitive micromachined ultrasonic transducer (CMUT) array of 16 × 16 elements with a center frequency of 5.5 MHz. Two-dimensional tomographic images and 3-D volumetric images of the objects placed at different depths are presented. In addition, we studied the sensitivity of CMUT-based PAI to the concentration of indocyanine green dye at 5 cm depth inside the phantom. Under optimized experimental conditions, the objects at 5 cm depth can be imaged with SNR of about 35 dB and a spatial resolution of approximately 500 µm. Results demonstrate that CMUTs with integrated front-end amplifier circuits are an attractive choice for achieving relatively high depth sensitivity for PAI. PMID:22249594

  3. Wishart Deep Stacking Network for Fast POLSAR Image Classification.

    PubMed

    Jiao, Licheng; Liu, Fang

    2016-05-11

    Inspired by the popular deep learning architecture - Deep Stacking Network (DSN), a specific deep model for polarimetric synthetic aperture radar (POLSAR) image classification is proposed in this paper, which is named as Wishart Deep Stacking Network (W-DSN). First of all, a fast implementation of Wishart distance is achieved by a special linear transformation, which speeds up the classification of POLSAR image and makes it possible to use this polarimetric information in the following Neural Network (NN). Then a single-hidden-layer neural network based on the fast Wishart distance is defined for POLSAR image classification, which is named as Wishart Network (WN) and improves the classification accuracy. Finally, a multi-layer neural network is formed by stacking WNs, which is in fact the proposed deep learning architecture W-DSN for POLSAR image classification and improves the classification accuracy further. In addition, the structure of WN can be expanded in a straightforward way by adding hidden units if necessary, as well as the structure of the W-DSN. As a preliminary exploration on formulating specific deep learning architecture for POLSAR image classification, the proposed methods may establish a simple but clever connection between POLSAR image interpretation and deep learning. The experiment results tested on real POLSAR image show that the fast implementation of Wishart distance is very efficient (a POLSAR image with 768000 pixels can be classified in 0.53s), and both the single-hidden-layer architecture WN and the deep learning architecture W-DSN for POLSAR image classification perform well and work efficiently.

  4. Reconstruction of initial pressure from limited view photoacoustic images using deep learning

    NASA Astrophysics Data System (ADS)

    Waibel, Dominik; Gröhl, Janek; Isensee, Fabian; Kirchner, Thomas; Maier-Hein, Klaus; Maier-Hein, Lena

    2018-02-01

    Quantification of tissue properties with photoacoustic (PA) imaging typically requires a highly accurate representation of the initial pressure distribution in tissue. Almost all PA scanners reconstruct the PA image only from a partial scan of the emitted sound waves. Especially handheld devices, which have become increasingly popular due to their versatility and ease of use, only provide limited view data because of their geometry. Owing to such limitations in hardware as well as to the acoustic attenuation in tissue, state-of-the-art reconstruction methods deliver only approximations of the initial pressure distribution. To overcome the limited view problem, we present a machine learning-based approach to the reconstruction of initial pressure from limited view PA data. Our method involves a fully convolutional deep neural network based on a U-Net-like architecture with pixel-wise regression loss on the acquired PA images. It is trained and validated on in silico data generated with Monte Carlo simulations. In an initial study we found an increase in accuracy over the state-of-the-art when reconstructing simulated linear-array scans of blood vessels.

  5. Single myelin fiber imaging in living rodents without labeling by deep optical coherence microscopy

    NASA Astrophysics Data System (ADS)

    Ben Arous, Juliette; Binding, Jonas; Léger, Jean-François; Casado, Mariano; Topilko, Piotr; Gigan, Sylvain; Claude Boccara, A.; Bourdieu, Laurent

    2011-11-01

    Myelin sheath disruption is responsible for multiple neuropathies in the central and peripheral nervous system. Myelin imaging has thus become an important diagnosis tool. However, in vivo imaging has been limited to either low-resolution techniques unable to resolve individual fibers or to low-penetration imaging of single fibers, which cannot provide quantitative information about large volumes of tissue, as required for diagnostic purposes. Here, we perform myelin imaging without labeling and at micron-scale resolution with >300-μm penetration depth on living rodents. This was achieved with a prototype [termed deep optical coherence microscopy (deep-OCM)] of a high-numerical aperture infrared full-field optical coherence microscope, which includes aberration correction for the compensation of refractive index mismatch and high-frame-rate interferometric measurements. We were able to measure the density of individual myelinated fibers in the rat cortex over a large volume of gray matter. In the peripheral nervous system, deep-OCM allows, after minor surgery, in situ imaging of single myelinated fibers over a large fraction of the sciatic nerve. This allows quantitative comparison of normal and Krox20 mutant mice, in which myelination in the peripheral nervous system is impaired. This opens promising perspectives for myelin chronic imaging in demyelinating diseases and for minimally invasive medical diagnosis.

  6. Single myelin fiber imaging in living rodents without labeling by deep optical coherence microscopy.

    PubMed

    Ben Arous, Juliette; Binding, Jonas; Léger, Jean-François; Casado, Mariano; Topilko, Piotr; Gigan, Sylvain; Boccara, A Claude; Bourdieu, Laurent

    2011-11-01

    Myelin sheath disruption is responsible for multiple neuropathies in the central and peripheral nervous system. Myelin imaging has thus become an important diagnosis tool. However, in vivo imaging has been limited to either low-resolution techniques unable to resolve individual fibers or to low-penetration imaging of single fibers, which cannot provide quantitative information about large volumes of tissue, as required for diagnostic purposes. Here, we perform myelin imaging without labeling and at micron-scale resolution with >300-μm penetration depth on living rodents. This was achieved with a prototype [termed deep optical coherence microscopy (deep-OCM)] of a high-numerical aperture infrared full-field optical coherence microscope, which includes aberration correction for the compensation of refractive index mismatch and high-frame-rate interferometric measurements. We were able to measure the density of individual myelinated fibers in the rat cortex over a large volume of gray matter. In the peripheral nervous system, deep-OCM allows, after minor surgery, in situ imaging of single myelinated fibers over a large fraction of the sciatic nerve. This allows quantitative comparison of normal and Krox20 mutant mice, in which myelination in the peripheral nervous system is impaired. This opens promising perspectives for myelin chronic imaging in demyelinating diseases and for minimally invasive medical diagnosis.

  7. Thermalnet: a Deep Convolutional Network for Synthetic Thermal Image Generation

    NASA Astrophysics Data System (ADS)

    Kniaz, V. V.; Gorbatsevich, V. S.; Mizginov, V. A.

    2017-05-01

    Deep convolutional neural networks have dramatically changed the landscape of the modern computer vision. Nowadays methods based on deep neural networks show the best performance among image recognition and object detection algorithms. While polishing of network architectures received a lot of scholar attention, from the practical point of view the preparation of a large image dataset for a successful training of a neural network became one of major challenges. This challenge is particularly profound for image recognition in wavelengths lying outside the visible spectrum. For example no infrared or radar image datasets large enough for successful training of a deep neural network are available to date in public domain. Recent advances of deep neural networks prove that they are also capable to do arbitrary image transformations such as super-resolution image generation, grayscale image colorisation and imitation of style of a given artist. Thus a natural question arise: how could be deep neural networks used for augmentation of existing large image datasets? This paper is focused on the development of the Thermalnet deep convolutional neural network for augmentation of existing large visible image datasets with synthetic thermal images. The Thermalnet network architecture is inspired by colorisation deep neural networks.

  8. Automated Grading of Gliomas using Deep Learning in Digital Pathology Images: A modular approach with ensemble of convolutional neural networks.

    PubMed

    Ertosun, Mehmet Günhan; Rubin, Daniel L

    2015-01-01

    Brain glioma is the most common primary malignant brain tumors in adults with different pathologic subtypes: Lower Grade Glioma (LGG) Grade II, Lower Grade Glioma (LGG) Grade III, and Glioblastoma Multiforme (GBM) Grade IV. The survival and treatment options are highly dependent of this glioma grade. We propose a deep learning-based, modular classification pipeline for automated grading of gliomas using digital pathology images. Whole tissue digitized images of pathology slides obtained from The Cancer Genome Atlas (TCGA) were used to train our deep learning modules. Our modular pipeline provides diagnostic quality statistics, such as precision, sensitivity and specificity, of the individual deep learning modules, and (1) facilitates training given the limited data in this domain, (2) enables exploration of different deep learning structures for each module, (3) leads to developing less complex modules that are simpler to analyze, and (4) provides flexibility, permitting use of single modules within the framework or use of other modeling or machine learning applications, such as probabilistic graphical models or support vector machines. Our modular approach helps us meet the requirements of minimum accuracy levels that are demanded by the context of different decision points within a multi-class classification scheme. Convolutional Neural Networks are trained for each module for each sub-task with more than 90% classification accuracies on validation data set, and achieved classification accuracy of 96% for the task of GBM vs LGG classification, 71% for further identifying the grade of LGG into Grade II or Grade III on independent data set coming from new patients from the multi-institutional repository.

  9. Automated Grading of Gliomas using Deep Learning in Digital Pathology Images: A modular approach with ensemble of convolutional neural networks

    PubMed Central

    Ertosun, Mehmet Günhan; Rubin, Daniel L.

    2015-01-01

    Brain glioma is the most common primary malignant brain tumors in adults with different pathologic subtypes: Lower Grade Glioma (LGG) Grade II, Lower Grade Glioma (LGG) Grade III, and Glioblastoma Multiforme (GBM) Grade IV. The survival and treatment options are highly dependent of this glioma grade. We propose a deep learning-based, modular classification pipeline for automated grading of gliomas using digital pathology images. Whole tissue digitized images of pathology slides obtained from The Cancer Genome Atlas (TCGA) were used to train our deep learning modules. Our modular pipeline provides diagnostic quality statistics, such as precision, sensitivity and specificity, of the individual deep learning modules, and (1) facilitates training given the limited data in this domain, (2) enables exploration of different deep learning structures for each module, (3) leads to developing less complex modules that are simpler to analyze, and (4) provides flexibility, permitting use of single modules within the framework or use of other modeling or machine learning applications, such as probabilistic graphical models or support vector machines. Our modular approach helps us meet the requirements of minimum accuracy levels that are demanded by the context of different decision points within a multi-class classification scheme. Convolutional Neural Networks are trained for each module for each sub-task with more than 90% classification accuracies on validation data set, and achieved classification accuracy of 96% for the task of GBM vs LGG classification, 71% for further identifying the grade of LGG into Grade II or Grade III on independent data set coming from new patients from the multi-institutional repository. PMID:26958289

  10. End-to-end deep neural network for optical inversion in quantitative photoacoustic imaging.

    PubMed

    Cai, Chuangjian; Deng, Kexin; Ma, Cheng; Luo, Jianwen

    2018-06-15

    An end-to-end deep neural network, ResU-net, is developed for quantitative photoacoustic imaging. A residual learning framework is used to facilitate optimization and to gain better accuracy from considerably increased network depth. The contracting and expanding paths enable ResU-net to extract comprehensive context information from multispectral initial pressure images and, subsequently, to infer a quantitative image of chromophore concentration or oxygen saturation (sO 2 ). According to our numerical experiments, the estimations of sO 2 and indocyanine green concentration are accurate and robust against variations in both optical property and object geometry. An extremely short reconstruction time of 22 ms is achieved.

  11. Deep Undercooling of Tissue Water and Winter Hardiness Limitations in Timberline Flora 1

    PubMed Central

    Becwar, Michael R.; Rajashekar, Channa; Bristow, Katherine J. Hansen; Burke, Michael J.

    1981-01-01

    Deep undercooled tissue water, which froze near −40 C, was found in winter collected stem and leaf tissue of the dominant timberline tree species of the Colorado Rocky Mountains, Engelmann spruce (Picea engelmannii (Parry) Engelm.) and subalpine fir (Abies lasiocarpa (Hook.) Nutt.), and in numerous other woody species in and below the subalpine vegetation zone. Previous work on numerous woody plants indicates that deep undercooling in xylem makes probable a −40 C winter hardiness limit in stem tissue. Visual injury determinations and electrolyte loss measurements on stem tissue revealed injury near −40 C associated with the freezing of the deep undercooled stem tissue water. These results suggest that the winter hardiness limit of this woody flora is near −40 C. The relevance of deep undercooling in relation to timberline, the upper elevational limit of the subalpine forest, is discussed. PMID:16661852

  12. Do Facilitated Online Dual Credit Classes Result in Deep Learning?

    ERIC Educational Resources Information Center

    Stark Education Partnership, 2015

    2015-01-01

    This study, with funding from the Jennings Foundation, sought to answer the following broad research question: Do facilitated online dual credit courses result in deep learning? The answer to this question is key to addressing barriers many students face in bridging from high school to college. This report includes a descriptive case study that…

  13. Wireless power transfer to deep-tissue microimplants

    PubMed Central

    Yeh, Alexander J.; Neofytou, Evgenios; Kim, Sanghoek; Tanabe, Yuji; Patlolla, Bhagat; Beygui, Ramin E.; Poon, Ada S. Y.

    2014-01-01

    The ability to implant electronic systems in the human body has led to many medical advances. Progress in semiconductor technology paved the way for devices at the scale of a millimeter or less (“microimplants”), but the miniaturization of the power source remains challenging. Although wireless powering has been demonstrated, energy transfer beyond superficial depths in tissue has so far been limited by large coils (at least a centimeter in diameter) unsuitable for a microimplant. Here, we show that this limitation can be overcome by a method, termed midfield powering, to create a high-energy density region deep in tissue inside of which the power-harvesting structure can be made extremely small. Unlike conventional near-field (inductively coupled) coils, for which coupling is limited by exponential field decay, a patterned metal plate is used to induce spatially confined and adaptive energy transport through propagating modes in tissue. We use this method to power a microimplant (2 mm, 70 mg) capable of closed-chest wireless control of the heart that is orders of magnitude smaller than conventional pacemakers. With exposure levels below human safety thresholds, milliwatt levels of power can be transferred to a deep-tissue (>5 cm) microimplant for both complex electronic function and physiological stimulation. The approach developed here should enable new generations of implantable systems that can be integrated into the body at minimal cost and risk. PMID:24843161

  14. Wireless power transfer to deep-tissue microimplants.

    PubMed

    Ho, John S; Yeh, Alexander J; Neofytou, Evgenios; Kim, Sanghoek; Tanabe, Yuji; Patlolla, Bhagat; Beygui, Ramin E; Poon, Ada S Y

    2014-06-03

    The ability to implant electronic systems in the human body has led to many medical advances. Progress in semiconductor technology paved the way for devices at the scale of a millimeter or less ("microimplants"), but the miniaturization of the power source remains challenging. Although wireless powering has been demonstrated, energy transfer beyond superficial depths in tissue has so far been limited by large coils (at least a centimeter in diameter) unsuitable for a microimplant. Here, we show that this limitation can be overcome by a method, termed midfield powering, to create a high-energy density region deep in tissue inside of which the power-harvesting structure can be made extremely small. Unlike conventional near-field (inductively coupled) coils, for which coupling is limited by exponential field decay, a patterned metal plate is used to induce spatially confined and adaptive energy transport through propagating modes in tissue. We use this method to power a microimplant (2 mm, 70 mg) capable of closed-chest wireless control of the heart that is orders of magnitude smaller than conventional pacemakers. With exposure levels below human safety thresholds, milliwatt levels of power can be transferred to a deep-tissue (>5 cm) microimplant for both complex electronic function and physiological stimulation. The approach developed here should enable new generations of implantable systems that can be integrated into the body at minimal cost and risk.

  15. Quantitative characterization of viscoelastic behavior in tissue-mimicking phantoms and ex vivo animal tissues.

    PubMed

    Maccabi, Ashkan; Shin, Andrew; Namiri, Nikan K; Bajwa, Neha; St John, Maie; Taylor, Zachary D; Grundfest, Warren; Saddik, George N

    2018-01-01

    Viscoelasticity of soft tissue is often related to pathology, and therefore, has become an important diagnostic indicator in the clinical assessment of suspect tissue. Surgeons, particularly within head and neck subsites, typically use palpation techniques for intra-operative tumor detection. This detection method, however, is highly subjective and often fails to detect small or deep abnormalities. Vibroacoustography (VA) and similar methods have previously been used to distinguish tissue with high-contrast, but a firm understanding of the main contrast mechanism has yet to be verified. The contributions of tissue mechanical properties in VA images have been difficult to verify given the limited literature on viscoelastic properties of various normal and diseased tissue. This paper aims to investigate viscoelasticity theory and present a detailed description of viscoelastic experimental results obtained in tissue-mimicking phantoms (TMPs) and ex vivo tissues to verify the main contrast mechanism in VA and similar imaging modalities. A spherical-tip micro-indentation technique was employed with the Hertzian model to acquire absolute, quantitative, point measurements of the elastic modulus (E), long term shear modulus (η), and time constant (τ) in homogeneous TMPs and ex vivo tissue in rat liver and porcine liver and gallbladder. Viscoelastic differences observed between porcine liver and gallbladder tissue suggest that imaging modalities which utilize the mechanical properties of tissue as a primary contrast mechanism can potentially be used to quantitatively differentiate between proximate organs in a clinical setting. These results may facilitate more accurate tissue modeling and add information not currently available to the field of systems characterization and biomedical research.

  16. Quantitative characterization of viscoelastic behavior in tissue-mimicking phantoms and ex vivo animal tissues

    PubMed Central

    Shin, Andrew; Namiri, Nikan K.; Bajwa, Neha; St. John, Maie; Taylor, Zachary D.; Grundfest, Warren; Saddik, George N.

    2018-01-01

    Viscoelasticity of soft tissue is often related to pathology, and therefore, has become an important diagnostic indicator in the clinical assessment of suspect tissue. Surgeons, particularly within head and neck subsites, typically use palpation techniques for intra-operative tumor detection. This detection method, however, is highly subjective and often fails to detect small or deep abnormalities. Vibroacoustography (VA) and similar methods have previously been used to distinguish tissue with high-contrast, but a firm understanding of the main contrast mechanism has yet to be verified. The contributions of tissue mechanical properties in VA images have been difficult to verify given the limited literature on viscoelastic properties of various normal and diseased tissue. This paper aims to investigate viscoelasticity theory and present a detailed description of viscoelastic experimental results obtained in tissue-mimicking phantoms (TMPs) and ex vivo tissues to verify the main contrast mechanism in VA and similar imaging modalities. A spherical-tip micro-indentation technique was employed with the Hertzian model to acquire absolute, quantitative, point measurements of the elastic modulus (E), long term shear modulus (η), and time constant (τ) in homogeneous TMPs and ex vivo tissue in rat liver and porcine liver and gallbladder. Viscoelastic differences observed between porcine liver and gallbladder tissue suggest that imaging modalities which utilize the mechanical properties of tissue as a primary contrast mechanism can potentially be used to quantitatively differentiate between proximate organs in a clinical setting. These results may facilitate more accurate tissue modeling and add information not currently available to the field of systems characterization and biomedical research. PMID:29373598

  17. Deep convolutional neural networks for classifying head and neck cancer using hyperspectral imaging

    NASA Astrophysics Data System (ADS)

    Halicek, Martin; Lu, Guolan; Little, James V.; Wang, Xu; Patel, Mihir; Griffith, Christopher C.; El-Deiry, Mark W.; Chen, Amy Y.; Fei, Baowei

    2017-06-01

    Surgical cancer resection requires an accurate and timely diagnosis of the cancer margins in order to achieve successful patient remission. Hyperspectral imaging (HSI) has emerged as a useful, noncontact technique for acquiring spectral and optical properties of tissue. A convolutional neural network (CNN) classifier is developed to classify excised, squamous-cell carcinoma, thyroid cancer, and normal head and neck tissue samples using HSI. The CNN classification was validated by the manual annotation of a pathologist specialized in head and neck cancer. The preliminary results of 50 patients indicate the potential of HSI and deep learning for automatic tissue-labeling of surgical specimens of head and neck patients.

  18. Quicksilver: Fast predictive image registration - A deep learning approach.

    PubMed

    Yang, Xiao; Kwitt, Roland; Styner, Martin; Niethammer, Marc

    2017-09-01

    This paper introduces Quicksilver, a fast deformable image registration method. Quicksilver registration for image-pairs works by patch-wise prediction of a deformation model based directly on image appearance. A deep encoder-decoder network is used as the prediction model. While the prediction strategy is general, we focus on predictions for the Large Deformation Diffeomorphic Metric Mapping (LDDMM) model. Specifically, we predict the momentum-parameterization of LDDMM, which facilitates a patch-wise prediction strategy while maintaining the theoretical properties of LDDMM, such as guaranteed diffeomorphic mappings for sufficiently strong regularization. We also provide a probabilistic version of our prediction network which can be sampled during the testing time to calculate uncertainties in the predicted deformations. Finally, we introduce a new correction network which greatly increases the prediction accuracy of an already existing prediction network. We show experimental results for uni-modal atlas-to-image as well as uni-/multi-modal image-to-image registrations. These experiments demonstrate that our method accurately predicts registrations obtained by numerical optimization, is very fast, achieves state-of-the-art registration results on four standard validation datasets, and can jointly learn an image similarity measure. Quicksilver is freely available as an open-source software. Copyright © 2017 Elsevier Inc. All rights reserved.

  19. Groth Deep Locations Image

    NASA Image and Video Library

    2003-07-25

    NASA's Galaxy Evolution Explorer photographed this ultraviolet color blowup of the Groth Deep Image on June 22 and June 23, 2003. Hundreds of galaxies are detected in this portion of the image, and the faint red galaxies are believed to be 6 billion light years away. The white boxes show the location of these distant galaxies, of which more than a 100 can be detected in this image. NASA astronomers expect to detect 10,000 such galaxies after extrapolating to the full image at a deeper exposure level. http://photojournal.jpl.nasa.gov/catalog/PIA04626

  20. Common-path biodynamic imaging for dynamic fluctuation spectroscopy of 3D living tissue

    NASA Astrophysics Data System (ADS)

    Li, Zhe; Turek, John; Nolte, David D.

    2017-03-01

    Biodynamic imaging is a novel 3D optical imaging technology based on short-coherence digital holography that measures intracellular motions of cells inside their natural microenvironments. Here both common-path and Mach-Zehnder designs are presented. Biological tissues such as tumor spheroids and ex vivo biopsies are used as targets, and backscattered light is collected as signal. Drugs are applied to samples, and their effects are evaluated by identifying biomarkers that capture intracellular dynamics from the reconstructed holograms. Through digital holography and coherence gating, information from different depths of the samples can be extracted, enabling the deep-tissue measurement of the responses to drugs.

  1. Frequency-locked pulse sequencer for high-frame-rate monochromatic tissue motion imaging.

    PubMed

    Azar, Reza Zahiri; Baghani, Ali; Salcudean, Septimiu E; Rohling, Robert

    2011-04-01

    To overcome the inherent low frame rate of conventional ultrasound, we have previously presented a system that can be implemented on conventional ultrasound scanners for high-frame-rate imaging of monochromatic tissue motion. The system employs a sector subdivision technique in the sequencer to increase the acquisition rate. To eliminate the delays introduced during data acquisition, a motion phase correction algorithm has also been introduced to create in-phase displacement images. Previous experimental results from tissue- mimicking phantoms showed that the system can achieve effective frame rates of up to a few kilohertz on conventional ultrasound systems. In this short communication, we present a new pulse sequencing strategy that facilitates high-frame-rate imaging of monochromatic motion such that the acquired echo signals are inherently in-phase. The sequencer uses the knowledge of the excitation frequency to synchronize the acquisition of the entire imaging plane to that of an external exciter. This sequencing approach eliminates any need for synchronization or phase correction and has applications in tissue elastography, which we demonstrate with tissue-mimicking phantoms. © 2011 IEEE

  2. Deep learning classifier with optical coherence tomography images for early dental caries detection

    NASA Astrophysics Data System (ADS)

    Karimian, Nima; Salehi, Hassan S.; Mahdian, Mina; Alnajjar, Hisham; Tadinada, Aditya

    2018-02-01

    Dental caries is a microbial disease that results in localized dissolution of the mineral content of dental tissue. Despite considerable decline in the incidence of dental caries, it remains a major health problem in many societies. Early detection of incipient lesions at initial stages of demineralization can result in the implementation of non-surgical preventive approaches to reverse the demineralization process. In this paper, we present a novel approach combining deep convolutional neural networks (CNN) and optical coherence tomography (OCT) imaging modality for classification of human oral tissues to detect early dental caries. OCT images of oral tissues with various densities were input to a CNN classifier to determine variations in tissue densities resembling the demineralization process. The CNN automatically learns a hierarchy of increasingly complex features and a related classifier directly from training data sets. The initial CNN layer parameters were randomly selected. The training set is split into minibatches, with 10 OCT images per batch. Given a batch of training patches, the CNN employs two convolutional and pooling layers to extract features and then classify each patch based on the probabilities from the SoftMax classification layer (output-layer). Afterward, the CNN calculates the error between the classification result and the reference label, and then utilizes the backpropagation process to fine-tune all the layer parameters to minimize this error using batch gradient descent algorithm. We validated our proposed technique on ex-vivo OCT images of human oral tissues (enamel, cortical-bone, trabecular-bone, muscular-tissue, and fatty-tissue), which attested to effectiveness of our proposed method.

  3. Transurethral illumination probe design for deep photoacoustic imaging of prostate

    NASA Astrophysics Data System (ADS)

    Ai, Min; Salcudean, Tim; Rohling, Robert; Abolmaesumi, Purang; Tang, Shuo

    2018-02-01

    Photoacoustic (PA) imaging with internal light illumination through optical fiber could enable imaging of internal organs at deep penetration. We have developed a transurethral probe with a multimode fiber inserted in a rigid cystoscope sheath for illuminating the prostate. At the distal end, the fiber tip is processed to diffuse light circumferentially over 2 cm length. A parabolic cylinder mirror then reflects the light to form a rectangular-shaped parallel beam which has at least 1 cm2 at the probe surface. The relatively large rectangular beam size can reduce the laser fluence rate on the urethral wall and thus reduce the potential of tissue damage. A 3 cm optical penetration in chicken tissue is achieved at a fluence rate around 7 mJ/cm2 . For further validation, a prostate phantom was built with similar optical properties of the human prostate. A 1.5 cm penetration depth is achieved in the prostate mimicking phantom at 10 mJ/cm2 fluence rate. PA imaging of prostate can potentially be carried out in the future by combining a transrectal ultrasound transducer and the transurethral illumination.

  4. Stable microwave radiometry system for long term monitoring of deep tissue temperature

    NASA Astrophysics Data System (ADS)

    Stauffer, Paul R.; Rodriques, Dario B.; Salahi, Sara; Topsakal, Erdem; Oliveira, Tiago R.; Prakash, Aniruddh; D'Isidoro, Fabio; Reudink, Douglas; Snow, Brent W.; Maccarini, Paolo F.

    2013-02-01

    Background: There are numerous clinical applications for non-invasive monitoring of deep tissue temperature. We present the design and experimental performance of a miniature radiometric thermometry system for measuring volume average temperature of tissue regions located up to 5cm deep in the body. Methods: We constructed a miniature sensor consisting of EMI-shielded log spiral microstrip antenna with high gain onaxis and integrated high-sensitivity 1.35GHz total power radiometer with 500 MHz bandwidth. We tested performance of the radiometry system in both simulated and experimental multilayer phantom models of several intended clinical measurement sites: i) brown adipose tissue (BAT) depots within 2cm of the skin surface, ii) 3-5cm deep kidney, and iii) human brain underlying intact scalp and skull. The physical models included layers of circulating tissue-mimicking liquids controlled at different temperatures to characterize our ability to quantify small changes in target temperature at depth under normothermic surface tissues. Results: We report SAR patterns that characterize the sense region of a 2.6cm diameter receive antenna, and radiometric power measurements as a function of deep tissue temperature that quantify radiometer sensitivity. The data demonstrate: i) our ability to accurately track temperature rise in realistic tissue targets such as urine refluxed from prewarmed bladder into kidney, and 10°C drop in brain temperature underlying normothermic scalp and skull, and ii) long term accuracy and stability of +0.4°C over 4.5 hours as needed for monitoring core body temperature over extended surgery or monitoring effects of brown fat metabolism over an extended sleep/wake cycle. Conclusions: A non-invasive sensor consisting of 2.6cm diameter receive antenna and integral 1.35GHz total power radiometer has demonstrated sufficient sensitivity to track clinically significant changes in temperature of deep tissue targets underlying normothermic surface

  5. Stable Microwave Radiometry System for Long Term Monitoring of Deep Tissue Temperature.

    PubMed

    Stauffer, Paul R; Rodriques, Dario B; Salahi, Sara; Topsakal, Erdem; Oliveira, Tiago R; Prakash, Aniruddh; D'Isidoro, Fabio; Reudink, Douglas; Snow, Brent W; Maccarini, Paolo F

    2013-02-26

    There are numerous clinical applications for non-invasive monitoring of deep tissue temperature. We present the design and experimental performance of a miniature radiometric thermometry system for measuring volume average temperature of tissue regions located up to 5cm deep in the body. We constructed a miniature sensor consisting of EMI-shielded log spiral microstrip antenna with high gain on-axis and integrated high-sensitivity 1.35GHz total power radiometer with 500 MHz bandwidth. We tested performance of the radiometry system in both simulated and experimental multilayer phantom models of several intended clinical measurement sites: i) brown adipose tissue (BAT) depots within 2cm of the skin surface, ii) 3-5cm deep kidney, and iii) human brain underlying intact scalp and skull. The physical models included layers of circulating tissue-mimicking liquids controlled at different temperatures to characterize our ability to quantify small changes in target temperature at depth under normothermic surface tissues. We report SAR patterns that characterize the sense region of a 2.6cm diameter receive antenna, and radiometric power measurements as a function of deep tissue temperature that quantify radiometer sensitivity. The data demonstrate: i) our ability to accurately track temperature rise in realistic tissue targets such as urine refluxed from prewarmed bladder into kidney, and 10°C drop in brain temperature underlying normothermic scalp and skull, and ii) long term accuracy and stability of ∓0.4°C over 4.5 hours as needed for monitoring core body temperature over extended surgery or monitoring effects of brown fat metabolism over an extended sleep/wake cycle. A non-invasive sensor consisting of 2.6cm diameter receive antenna and integral 1.35GHz total power radiometer has demonstrated sufficient sensitivity to track clinically significant changes in temperature of deep tissue targets underlying normothermic surface tissues for clinical applications like the

  6. Factors affecting the microbiological condition of the deep tissues of mechanically tenderized beef.

    PubMed

    Gill, C O; McGinnis, J C

    2005-04-01

    Whole or halved top butt prime beef cuts were treated in two types of mechanical tenderizing machines that both pierced the meat with thin blades but that used blades of different forms. Aerobes on meat surfaces and in the deep tissues of cuts after treatments were counted. When cuts were treated at a laboratory using a Lumar machine, the contamination of deep tissues increased significantly (P < 0.01) with increasing numbers of aerobic bacteria on meat surfaces and decreased significantly (P < 0.001) with increasing distance from the incised surface. However, contamination did not increase significantly (P > 0.1) with repeated incising of the meat. When halved cuts were incised one or eight times using a commercially cleaned Ross machine at a retail store, the numbers of aerobes recovered from deep tissues were similar with both treatments. When halved cuts were treated in one or other machine, deep tissue contamination was greater with the Lumar machine than with the Ross machine. Contamination of deep tissues as a result of tenderizing by piercing with thin blades can be minimized if the blades are designed to limit the number of bacteria carried into the meat and the microbiological condition of incised surface is well controlled.

  7. Deep tissue volume imaging of birefringence through fibre-optic needle probes for the delineation of breast tumour

    NASA Astrophysics Data System (ADS)

    Villiger, Martin; Lorenser, Dirk; McLaughlin, Robert A.; Quirk, Bryden C.; Kirk, Rodney W.; Bouma, Brett E.; Sampson, David D.

    2016-07-01

    Identifying tumour margins during breast-conserving surgeries is a persistent challenge. We have previously developed miniature needle probes that could enable intraoperative volume imaging with optical coherence tomography. In many situations, however, scattering contrast alone is insufficient to clearly identify and delineate malignant regions. Additional polarization-sensitive measurements provide the means to assess birefringence, which is elevated in oriented collagen fibres and may offer an intrinsic biomarker to differentiate tumour from benign tissue. Here, we performed polarization-sensitive optical coherence tomography through miniature imaging needles and developed an algorithm to efficiently reconstruct images of the depth-resolved tissue birefringence free of artefacts. First ex vivo imaging of breast tumour samples revealed excellent contrast between lowly birefringent malignant regions, and stromal tissue, which is rich in oriented collagen and exhibits higher birefringence, as confirmed with co-located histology. The ability to clearly differentiate between tumour and uninvolved stroma based on intrinsic contrast could prove decisive for the intraoperative assessment of tumour margins.

  8. Deep brain two-photon NIR fluorescence imaging for study of Alzheimer's disease

    NASA Astrophysics Data System (ADS)

    Chen, Congping; Liang, Zhuoyi; Zhou, Biao; Ip, Nancy Y.; Qu, Jianan Y.

    2018-02-01

    Amyloid depositions in the brain represent the characteristic hallmarks of Alzheimer's disease (AD) pathology. The abnormal accumulation of extracellular amyloid-beta (Aβ) and resulting toxic amyloid plaques are considered to be responsible for the clinical deficits including cognitive decline and memory loss. In vivo two-photon fluorescence imaging of amyloid plaques in live AD mouse model through a chronic imaging window (thinned skull or craniotomy) provides a mean to greatly facilitate the study of the pathological mechanism of AD owing to its high spatial resolution and long-term continuous monitoring. However, the imaging depth for amyloid plaques is largely limited to upper cortical layers due to the short-wavelength fluorescence emission of commonly used amyloid probes. In this work, we reported that CRANAD-3, a near-infrared (NIR) probe for amyloid species with excitation wavelength at 900 nm and emission wavelength around 650 nm, has great advantages over conventionally used probes and is well suited for twophoton deep imaging of amyloid plaques in AD mouse brain. Compared with a commonly used MeO-X04 probe, the imaging depth of CRANAD-3 is largely extended for open skull cranial window. Furthermore, by using two-photon excited fluorescence spectroscopic imaging, we characterized the intrinsic fluorescence of the "aging pigment" lipofuscin in vivo, which has distinct spectra from CRANAD-3 labeled plaques. This study reveals the unique potential of NIR probes for in vivo, high-resolution and deep imaging of brain amyloid in Alzheimer's disease.

  9. Zero-Echo-Time and Dixon Deep Pseudo-CT (ZeDD CT): Direct Generation of Pseudo-CT Images for Pelvic PET/MRI Attenuation Correction Using Deep Convolutional Neural Networks with Multiparametric MRI.

    PubMed

    Leynes, Andrew P; Yang, Jaewon; Wiesinger, Florian; Kaushik, Sandeep S; Shanbhag, Dattesh D; Seo, Youngho; Hope, Thomas A; Larson, Peder E Z

    2018-05-01

    Accurate quantification of uptake on PET images depends on accurate attenuation correction in reconstruction. Current MR-based attenuation correction methods for body PET use a fat and water map derived from a 2-echo Dixon MRI sequence in which bone is neglected. Ultrashort-echo-time or zero-echo-time (ZTE) pulse sequences can capture bone information. We propose the use of patient-specific multiparametric MRI consisting of Dixon MRI and proton-density-weighted ZTE MRI to directly synthesize pseudo-CT images with a deep learning model: we call this method ZTE and Dixon deep pseudo-CT (ZeDD CT). Methods: Twenty-six patients were scanned using an integrated 3-T time-of-flight PET/MRI system. Helical CT images of the patients were acquired separately. A deep convolutional neural network was trained to transform ZTE and Dixon MR images into pseudo-CT images. Ten patients were used for model training, and 16 patients were used for evaluation. Bone and soft-tissue lesions were identified, and the SUV max was measured. The root-mean-squared error (RMSE) was used to compare the MR-based attenuation correction with the ground-truth CT attenuation correction. Results: In total, 30 bone lesions and 60 soft-tissue lesions were evaluated. The RMSE in PET quantification was reduced by a factor of 4 for bone lesions (10.24% for Dixon PET and 2.68% for ZeDD PET) and by a factor of 1.5 for soft-tissue lesions (6.24% for Dixon PET and 4.07% for ZeDD PET). Conclusion: ZeDD CT produces natural-looking and quantitatively accurate pseudo-CT images and reduces error in pelvic PET/MRI attenuation correction compared with standard methods. © 2018 by the Society of Nuclear Medicine and Molecular Imaging.

  10. Imaging Strategies for Tissue Engineering Applications

    PubMed Central

    Nam, Seung Yun; Ricles, Laura M.; Suggs, Laura J.

    2015-01-01

    Tissue engineering has evolved with multifaceted research being conducted using advanced technologies, and it is progressing toward clinical applications. As tissue engineering technology significantly advances, it proceeds toward increasing sophistication, including nanoscale strategies for material construction and synergetic methods for combining with cells, growth factors, or other macromolecules. Therefore, to assess advanced tissue-engineered constructs, tissue engineers need versatile imaging methods capable of monitoring not only morphological but also functional and molecular information. However, there is no single imaging modality that is suitable for all tissue-engineered constructs. Each imaging method has its own range of applications and provides information based on the specific properties of the imaging technique. Therefore, according to the requirements of the tissue engineering studies, the most appropriate tool should be selected among a variety of imaging modalities. The goal of this review article is to describe available biomedical imaging methods to assess tissue engineering applications and to provide tissue engineers with criteria and insights for determining the best imaging strategies. Commonly used biomedical imaging modalities, including X-ray and computed tomography, positron emission tomography and single photon emission computed tomography, magnetic resonance imaging, ultrasound imaging, optical imaging, and emerging techniques and multimodal imaging, will be discussed, focusing on the latest trends of their applications in recent tissue engineering studies. PMID:25012069

  11. Co-encapsulation of magnetic nanoparticles and doxorubicin into biodegradable microcarriers for deep tissue targeting by vascular MRI navigation.

    PubMed

    Pouponneau, Pierre; Leroux, Jean-Christophe; Soulez, Gilles; Gaboury, Louis; Martel, Sylvain

    2011-05-01

    Magnetic tumor targeting with external magnets is a promising method to increase the delivery of cytotoxic agents to tumor cells while reducing side effects. However, this approach suffers from intrinsic limitations, such as the inability to target areas within deep tissues, due mainly to a strong decrease of the magnetic field magnitude away from the magnets. Magnetic resonance navigation (MRN) involving the endovascular steering of therapeutic magnetic microcarriers (TMMC) represents a clinically viable alternative to reach deep tissues. MRN is achieved with an upgraded magnetic resonance imaging (MRI) scanner. In this proof-of-concept preclinical study, the preparation and steering of TMMC which were designed by taking into consideration the constraints of MRN and liver chemoembolization are reported. TMMC were biodegradable microparticles loaded with iron-cobalt nanoparticles and doxorubicin (DOX). These particles displayed high saturation magnetization (Ms = 72 emu g(-1)), MRI tracking compatibility (strong contrast on T2∗-weighted images), appropriate size for the blood vessel embolization (∼50 μm), and sustained release of DOX (over several days). The TMMC were successfully steered in vitro and in vivo in the rabbit model. In vivo targeting of the right or left liver lobes was achieved by MRN through the hepatic artery located 4 cm beneath the skin. Parameters such as flow velocity, TMMC release site in the artery, magnetic gradient and TMMC properties, affected the steering efficiency. These data illustrate the potential of MRN to improve drug targeting in deep tissues. Copyright © 2011 Elsevier Ltd. All rights reserved.

  12. Deep convolutional neural networks for automatic classification of gastric carcinoma using whole slide images in digital histopathology.

    PubMed

    Sharma, Harshita; Zerbe, Norman; Klempert, Iris; Hellwich, Olaf; Hufnagl, Peter

    2017-11-01

    Deep learning using convolutional neural networks is an actively emerging field in histological image analysis. This study explores deep learning methods for computer-aided classification in H&E stained histopathological whole slide images of gastric carcinoma. An introductory convolutional neural network architecture is proposed for two computerized applications, namely, cancer classification based on immunohistochemical response and necrosis detection based on the existence of tumor necrosis in the tissue. Classification performance of the developed deep learning approach is quantitatively compared with traditional image analysis methods in digital histopathology requiring prior computation of handcrafted features, such as statistical measures using gray level co-occurrence matrix, Gabor filter-bank responses, LBP histograms, gray histograms, HSV histograms and RGB histograms, followed by random forest machine learning. Additionally, the widely known AlexNet deep convolutional framework is comparatively analyzed for the corresponding classification problems. The proposed convolutional neural network architecture reports favorable results, with an overall classification accuracy of 0.6990 for cancer classification and 0.8144 for necrosis detection. Copyright © 2017 Elsevier Ltd. All rights reserved.

  13. Stokes-polarimetry imaging of tissue

    NASA Astrophysics Data System (ADS)

    Wu, Paul J.

    A novel Stokes-polarimetry imaging system and technique was developed to quantify fully the polarization properties of light remitted from tissue. The uniqueness of the system and technique is established in the incident polarization. Here, the diffuse illumination is varied and controlled with the intention to improve the visibility of tissue structures. Since light retains some polarization even after multiple-scattering events, the polarization of remitted light depends upon the interactions within the material. Differentiation between tissue structures is accomplished by two-dimensional mapping of the imaged area using metrics such as the degree of linear polarization, degree of circular polarization, ellipticity, and Stokes parameters. While Stokes-polarimetry imaging can be applied to a variety of tissues and conditions, this thesis focuses on tissue types associated with the disease endometriosis. The current standard in diagnosing endometriosis is visual laparoscopy with tissue biopsy. The documented correlation between laparoscopy inspection and histological confirmation of suspected lesions was at best 67%. Endometrial lesions vary greatly in their appearance and depth of infiltration. Although laparoscopy permits tissue to be assessed by color and texture, to advance beyond the state-of-the-art, a new imaging modality involving polarized light was investigated; in particular, Stokes-polarimetry imaging was used to determine the polarization signature of light that interacted with tissue. Basic science studies were conducted on rat tails embedded within turbid gelatin. The purpose of these experiments was to determine how identification of sub-surface structures could be improved. Experimental results indicate image contrast among various structures such as tendon, soft tissue and intervertebral discs. Stokes-polarimetry imaging experiments were performed on various tissues associated with endometriosis to obtain a baseline characterization for each

  14. Dual-modality endoscopic probe for tissue surface shape reconstruction and hyperspectral imaging enabled by deep neural networks.

    PubMed

    Lin, Jianyu; Clancy, Neil T; Qi, Ji; Hu, Yang; Tatla, Taran; Stoyanov, Danail; Maier-Hein, Lena; Elson, Daniel S

    2018-06-15

    Surgical guidance and decision making could be improved with accurate and real-time measurement of intra-operative data including shape and spectral information of the tissue surface. In this work, a dual-modality endoscopic system has been proposed to enable tissue surface shape reconstruction and hyperspectral imaging (HSI). This system centers around a probe comprised of an incoherent fiber bundle, whose fiber arrangement is different at the two ends, and miniature imaging optics. For 3D reconstruction with structured light (SL), a light pattern formed of randomly distributed spots with different colors is projected onto the tissue surface, creating artificial texture. Pattern decoding with a Convolutional Neural Network (CNN) model and a customized feature descriptor enables real-time 3D surface reconstruction at approximately 12 frames per second (FPS). In HSI mode, spatially sparse hyperspectral signals from the tissue surface can be captured with a slit hyperspectral imager in a single snapshot. A CNN based super-resolution model, namely "super-spectral-resolution" network (SSRNet), has also been developed to estimate pixel-level dense hypercubes from the endoscope cameras standard RGB images and the sparse hyperspectral signals, at approximately 2 FPS. The probe, with a 2.1 mm diameter, enables the system to be used with endoscope working channels. Furthermore, since data acquisition in both modes can be accomplished in one snapshot, operation of this system in clinical applications is minimally affected by tissue surface movement and deformation. The whole apparatus has been validated on phantoms and tissue (ex vivo and in vivo), while initial measurements on patients during laryngeal surgery show its potential in real-world clinical applications. Copyright © 2018 Elsevier B.V. All rights reserved.

  15. Sensory processing of deep tissue nociception in the rat spinal cord and thalamic ventrobasal complex.

    PubMed

    Sikandar, Shafaq; West, Steven J; McMahon, Stephen B; Bennett, David L; Dickenson, Anthony H

    2017-07-01

    Sensory processing of deep somatic tissue constitutes an important component of the nociceptive system, yet associated central processing pathways remain poorly understood. Here, we provide a novel electrophysiological characterization and immunohistochemical analysis of neural activation in the lateral spinal nucleus (LSN). These neurons show evoked activity to deep, but not cutaneous, stimulation. The evoked responses of neurons in the LSN can be sensitized to somatosensory stimulation following intramuscular hypertonic saline, an acute model of muscle pain, suggesting this is an important spinal relay site for the processing of deep tissue nociceptive inputs. Neurons of the thalamic ventrobasal complex (VBC) mediate both cutaneous and deep tissue sensory processing, but in contrast to the lateral spinal nucleus our electrophysiological studies do not suggest the existence of a subgroup of cells that selectively process deep tissue inputs. The sensitization of polymodal and thermospecific VBC neurons to mechanical somatosensory stimulation following acute muscle stimulation with hypertonic saline suggests differential roles of thalamic subpopulations in mediating cutaneous and deep tissue nociception in pathological states. Overall, our studies at both the spinal (lateral spinal nucleus) and supraspinal (thalamic ventrobasal complex) levels suggest a convergence of cutaneous and deep somatosensory inputs onto spinothalamic pathways, which are unmasked by activation of muscle nociceptive afferents to produce consequent phenotypic alterations in spinal and thalamic neural coding of somatosensory stimulation. A better understanding of the sensory pathways involved in deep tissue nociception, as well as the degree of labeled line and convergent pathways for cutaneous and deep somatosensory inputs, is fundamental to developing targeted analgesic therapies for deep pain syndromes. © 2017 University College London. Physiological Reports published by Wiley Periodicals

  16. High resolution and deep tissue imaging using a near infrared acoustic resolution photoacoustic microscopy

    NASA Astrophysics Data System (ADS)

    Moothanchery, Mohesh; Sharma, Arunima; Periyasamy, Vijitha; Pramanik, Manojit

    2018-02-01

    It is always a great challenge for pure optical techniques to maintain good resolution and imaging depth at the same time. Photoacoustic imaging is an emerging technique which can overcome the limitation by pulsed light illumination and acoustic detection. Here, we report a Near Infrared Acoustic-Resolution Photoacoustic Microscopy (NIR-AR-PAM) systm with 30 MHz transducer and 1064 nm illumination which can achieve a lateral resolution of around 88 μm and imaging depth of 9.2 mm. Compared to visible light NIR beam can penetrate deeper in biological tissue due to weaker optical attenuation. In this work, we also demonstrated the in vivo imaging capabilty of NIRARPAM by near infrared detection of SLN with black ink as exogenous photoacoustic contrast agent in a rodent model.

  17. Optical coherence microscopy for deep tissue imaging of the cerebral cortex with intrinsic contrast

    PubMed Central

    Srinivasan, Vivek J.; Radhakrishnan, Harsha; Jiang, James Y.; Barry, Scott; Cable, Alex E.

    2012-01-01

    In vivo optical microscopic imaging techniques have recently emerged as important tools for the study of neurobiological development and pathophysiology. In particular, two-photon microscopy has proved to be a robust and highly flexible method for in vivo imaging in highly scattering tissue. However, two-photon imaging typically requires extrinsic dyes or contrast agents, and imaging depths are limited to a few hundred microns. Here we demonstrate Optical Coherence Microscopy (OCM) for in vivo imaging of neuronal cell bodies and cortical myelination up to depths of ~1.3 mm in the rat neocortex. Imaging does not require the administration of exogenous dyes or contrast agents, and is achieved through intrinsic scattering contrast and image processing alone. Furthermore, using OCM we demonstrate in vivo, quantitative measurements of optical properties (index of refraction and attenuation coefficient) in the cortex, and correlate these properties with laminar cellular architecture determined from the images. Lastly, we show that OCM enables direct visualization of cellular changes during cell depolarization and may therefore provide novel optical markers of cell viability. PMID:22330462

  18. Using deep learning in image hyper spectral segmentation, classification, and detection

    NASA Astrophysics Data System (ADS)

    Zhao, Xiuying; Su, Zhenyu

    2018-02-01

    Recent years have shown that deep learning neural networks are a valuable tool in the field of computer vision. Deep learning method can be used in applications like remote sensing such as Land cover Classification, Detection of Vehicle in Satellite Images, Hyper spectral Image classification. This paper addresses the use of the deep learning artificial neural network in Satellite image segmentation. Image segmentation plays an important role in image processing. The hue of the remote sensing image often has a large hue difference, which will result in the poor display of the images in the VR environment. Image segmentation is a pre processing technique applied to the original images and splits the image into many parts which have different hue to unify the color. Several computational models based on supervised, unsupervised, parametric, probabilistic region based image segmentation techniques have been proposed. Recently, one of the machine learning technique known as, deep learning with convolution neural network has been widely used for development of efficient and automatic image segmentation models. In this paper, we focus on study of deep neural convolution network and its variants for automatic image segmentation rather than traditional image segmentation strategies.

  19. Magnetic Resonance Imaging of Human Tissue-Engineered Adipose Substitutes

    PubMed Central

    Proulx, Maryse; Aubin, Kim; Lagueux, Jean; Audet, Pierre; Auger, Michèle

    2015-01-01

    Adipose tissue (AT) substitutes are being developed to answer the strong demand in reconstructive surgery. To facilitate the validation of their functional performance in vivo, and to avoid resorting to excessive number of animals, it is crucial at this stage to develop biomedical imaging methodologies, enabling the follow-up of reconstructed AT substitutes. Until now, biomedical imaging of AT substitutes has scarcely been reported in the literature. Therefore, the optimal parameters enabling good resolution, appropriate contrast, and graft delineation, as well as blood perfusion validation, must be studied and reported. In this study, human adipose substitutes produced from adipose-derived stem/stromal cells using the self-assembly approach of tissue engineering were implanted into athymic mice. The fate of the reconstructed AT substitutes implanted in vivo was successfully followed by magnetic resonance imaging (MRI), which is the imaging modality of choice for visualizing soft ATs. T1-weighted images allowed clear delineation of the grafts, followed by volume integration. The magnetic resonance (MR) signal of reconstructed AT was studied in vitro by proton nuclear magnetic resonance (1H-NMR). This confirmed the presence of a strong triglyceride peak of short longitudinal proton relaxation time (T1) values (200±53 ms) in reconstructed AT substitutes (total T1=813±76 ms), which establishes a clear signal difference between adjacent muscle, connective tissue, and native fat (total T1 ∼300 ms). Graft volume retention was followed up to 6 weeks after implantation, revealing a gradual resorption rate averaging at 44% of initial substitute's volume. In addition, vascular perfusion measured by dynamic contrast-enhanced-MRI confirmed the graft's vascularization postimplantation (14 and 21 days after grafting). Histological analysis of the grafted tissues revealed the persistence of numerous adipocytes without evidence of cysts or tissue necrosis. This study

  20. Swept-source optical coherence tomography powered by a 1.3-μm vertical cavity surface emitting laser enables 2.3-mm-deep brain imaging in mice in vivo

    NASA Astrophysics Data System (ADS)

    Choi, Woo June; Wang, Ruikang K.

    2015-10-01

    We report noninvasive, in vivo optical imaging deep within a mouse brain by swept-source optical coherence tomography (SS-OCT), enabled by a 1.3-μm vertical cavity surface emitting laser (VCSEL). VCSEL SS-OCT offers a constant signal sensitivity of 105 dB throughout an entire depth of 4.25 mm in air, ensuring an extended usable imaging depth range of more than 2 mm in turbid biological tissue. Using this approach, we show deep brain imaging in mice with an open-skull cranial window preparation, revealing intact mouse brain anatomy from the superficial cerebral cortex to the deep hippocampus. VCSEL SS-OCT would be applicable to small animal studies for the investigation of deep tissue compartments in living brains where diseases such as dementia and tumor can take their toll.

  1. NiftyNet: a deep-learning platform for medical imaging.

    PubMed

    Gibson, Eli; Li, Wenqi; Sudre, Carole; Fidon, Lucas; Shakir, Dzhoshkun I; Wang, Guotai; Eaton-Rosen, Zach; Gray, Robert; Doel, Tom; Hu, Yipeng; Whyntie, Tom; Nachev, Parashkev; Modat, Marc; Barratt, Dean C; Ourselin, Sébastien; Cardoso, M Jorge; Vercauteren, Tom

    2018-05-01

    Medical image analysis and computer-assisted intervention problems are increasingly being addressed with deep-learning-based solutions. Established deep-learning platforms are flexible but do not provide specific functionality for medical image analysis and adapting them for this domain of application requires substantial implementation effort. Consequently, there has been substantial duplication of effort and incompatible infrastructure developed across many research groups. This work presents the open-source NiftyNet platform for deep learning in medical imaging. The ambition of NiftyNet is to accelerate and simplify the development of these solutions, and to provide a common mechanism for disseminating research outputs for the community to use, adapt and build upon. The NiftyNet infrastructure provides a modular deep-learning pipeline for a range of medical imaging applications including segmentation, regression, image generation and representation learning applications. Components of the NiftyNet pipeline including data loading, data augmentation, network architectures, loss functions and evaluation metrics are tailored to, and take advantage of, the idiosyncracies of medical image analysis and computer-assisted intervention. NiftyNet is built on the TensorFlow framework and supports features such as TensorBoard visualization of 2D and 3D images and computational graphs by default. We present three illustrative medical image analysis applications built using NiftyNet infrastructure: (1) segmentation of multiple abdominal organs from computed tomography; (2) image regression to predict computed tomography attenuation maps from brain magnetic resonance images; and (3) generation of simulated ultrasound images for specified anatomical poses. The NiftyNet infrastructure enables researchers to rapidly develop and distribute deep learning solutions for segmentation, regression, image generation and representation learning applications, or extend the platform to new

  2. Nonlinear Deep Kernel Learning for Image Annotation.

    PubMed

    Jiu, Mingyuan; Sahbi, Hichem

    2017-02-08

    Multiple kernel learning (MKL) is a widely used technique for kernel design. Its principle consists in learning, for a given support vector classifier, the most suitable convex (or sparse) linear combination of standard elementary kernels. However, these combinations are shallow and often powerless to capture the actual similarity between highly semantic data, especially for challenging classification tasks such as image annotation. In this paper, we redefine multiple kernels using deep multi-layer networks. In this new contribution, a deep multiple kernel is recursively defined as a multi-layered combination of nonlinear activation functions, each one involves a combination of several elementary or intermediate kernels, and results into a positive semi-definite deep kernel. We propose four different frameworks in order to learn the weights of these networks: supervised, unsupervised, kernel-based semisupervised and Laplacian-based semi-supervised. When plugged into support vector machines (SVMs), the resulting deep kernel networks show clear gain, compared to several shallow kernels for the task of image annotation. Extensive experiments and analysis on the challenging ImageCLEF photo annotation benchmark, the COREL5k database and the Banana dataset validate the effectiveness of the proposed method.

  3. Identifying Medical Diagnoses and Treatable Diseases by Image-Based Deep Learning.

    PubMed

    Kermany, Daniel S; Goldbaum, Michael; Cai, Wenjia; Valentim, Carolina C S; Liang, Huiying; Baxter, Sally L; McKeown, Alex; Yang, Ge; Wu, Xiaokang; Yan, Fangbing; Dong, Justin; Prasadha, Made K; Pei, Jacqueline; Ting, Magdalene Y L; Zhu, Jie; Li, Christina; Hewett, Sierra; Dong, Jason; Ziyar, Ian; Shi, Alexander; Zhang, Runze; Zheng, Lianghong; Hou, Rui; Shi, William; Fu, Xin; Duan, Yaou; Huu, Viet A N; Wen, Cindy; Zhang, Edward D; Zhang, Charlotte L; Li, Oulan; Wang, Xiaobo; Singer, Michael A; Sun, Xiaodong; Xu, Jie; Tafreshi, Ali; Lewis, M Anthony; Xia, Huimin; Zhang, Kang

    2018-02-22

    The implementation of clinical-decision support algorithms for medical imaging faces challenges with reliability and interpretability. Here, we establish a diagnostic tool based on a deep-learning framework for the screening of patients with common treatable blinding retinal diseases. Our framework utilizes transfer learning, which trains a neural network with a fraction of the data of conventional approaches. Applying this approach to a dataset of optical coherence tomography images, we demonstrate performance comparable to that of human experts in classifying age-related macular degeneration and diabetic macular edema. We also provide a more transparent and interpretable diagnosis by highlighting the regions recognized by the neural network. We further demonstrate the general applicability of our AI system for diagnosis of pediatric pneumonia using chest X-ray images. This tool may ultimately aid in expediting the diagnosis and referral of these treatable conditions, thereby facilitating earlier treatment, resulting in improved clinical outcomes. VIDEO ABSTRACT. Copyright © 2018 Elsevier Inc. All rights reserved.

  4. Microscopic medical image classification framework via deep learning and shearlet transform.

    PubMed

    Rezaeilouyeh, Hadi; Mollahosseini, Ali; Mahoor, Mohammad H

    2016-10-01

    Cancer is the second leading cause of death in US after cardiovascular disease. Image-based computer-aided diagnosis can assist physicians to efficiently diagnose cancers in early stages. Existing computer-aided algorithms use hand-crafted features such as wavelet coefficients, co-occurrence matrix features, and recently, histogram of shearlet coefficients for classification of cancerous tissues and cells in images. These hand-crafted features often lack generalizability since every cancerous tissue and cell has a specific texture, structure, and shape. An alternative approach is to use convolutional neural networks (CNNs) to learn the most appropriate feature abstractions directly from the data and handle the limitations of hand-crafted features. A framework for breast cancer detection and prostate Gleason grading using CNN trained on images along with the magnitude and phase of shearlet coefficients is presented. Particularly, we apply shearlet transform on images and extract the magnitude and phase of shearlet coefficients. Then we feed shearlet features along with the original images to our CNN consisting of multiple layers of convolution, max pooling, and fully connected layers. Our experiments show that using the magnitude and phase of shearlet coefficients as extra information to the network can improve the accuracy of detection and generalize better compared to the state-of-the-art methods that rely on hand-crafted features. This study expands the application of deep neural networks into the field of medical image analysis, which is a difficult domain considering the limited medical data available for such analysis.

  5. Enhanced detection of myeloperoxidase activity in deep tissues through luminescent excitation of near-infrared nanoparticles.

    PubMed

    Zhang, Ning; Francis, Kevin P; Prakash, Arun; Ansaldi, Daniel

    2013-04-01

    A previous study reported the use of luminol for the detection of myeloperoxidase (MPO) activity using optical imaging in infiltrating neutrophils under inflammatory disease conditions. The detection is based on a photon-emitting reaction between luminol and an MPO metabolite. Because of tissue absorption and scattering, however, luminol-emitted blue light can be efficiently detected from superficial inflammatory foci only. In this study we report a chemiluminescence resonance energy transfer (CRET) methodology in which luminol-generated blue light excites nanoparticles to emit light in the near-infrared spectral range, resulting in remarkable improvement of MPO detectability in vivo. CRET caused a 37-fold increase in luminescence emission over luminol alone in detecting MPO activity in lung tissues after lipopolysaccharide challenge. We demonstrated a dependence of the chemiluminescent signal on MPO activity using MPO-deficient mice. In addition, co-administration of 4-aminobenzoic acid hydrazide (4-ABAH), an irreversible inhibitor of MPO, significantly attenuated luminescent emission from inflamed lungs. Inhibition of nitric oxide synthase with a nonspecific inhibitor, L-NAME, had no effect on luminol-mediated chemiluminescence production. Pretreatment of mice with MLN120B, a selective inhibitor of IKK-2, resulted in suppression of neutrophil infiltration to the lung tissues and reduction of MPO activity. We also demonstrated that CRET can effectively detect MPO activity at deep tissue tumor foci due to tumor development-associated neutrophil infiltration. We developed a sensitive MPO detection methodology that provides a means for visualizing and quantifying oxidative stress in deep tissue. This method is amenable to rapid evaluation of anti-inflammatory agents in animal models.

  6. Photoacoustic and ultrasound imaging of cancellous bone tissue.

    PubMed

    Yang, Lifeng; Lashkari, Bahman; Tan, Joel W Y; Mandelis, Andreas

    2015-07-01

    We used ultrasound (US) and photoacoustic (PA) imaging modalities to characterize cattle trabecular bones. The PA signals were generated with an 805-nm continuous wave laser used for optimally deep optical penetration depth. The detector for both modalities was a 2.25-MHz US transducer with a lateral resolution of ~1 mm at its focal point. Using a lateral pixel size much larger than the size of the trabeculae, raster scanning generated PA images related to the averaged values of the optical and thermoelastic properties, as well as density measurements in the focal volume. US backscatter yielded images related to mechanical properties and density in the focal volume. The depth of interest was selected by time-gating the signals for both modalities. The raster scanned PA and US images were compared with microcomputed tomography (μCT) images averaged over the same volume to generate similar spatial resolution as US and PA. The comparison revealed correlations between PA and US modalities with the mineral volume fraction of the bone tissue. Various features and properties of these modalities such as detectable depth, resolution, and sensitivity are discussed.

  7. Biodynamic Doppler imaging of subcellular motion inside 3D living tissue culture and biopsies (Conference Presentation)

    NASA Astrophysics Data System (ADS)

    Nolte, David D.

    2016-03-01

    Biodynamic imaging is an emerging 3D optical imaging technology that probes up to 1 mm deep inside three-dimensional living tissue using short-coherence dynamic light scattering to measure the intracellular motions of cells inside their natural microenvironments. Biodynamic imaging is label-free and non-invasive. The information content of biodynamic imaging is captured through tissue dynamics spectroscopy that displays the changes in the Doppler signatures from intracellular constituents in response to applied compounds. The affected dynamic intracellular mechanisms include organelle transport, membrane undulations, cytoskeletal restructuring, strain at cellular adhesions, cytokinesis, mitosis, exo- and endo-cytosis among others. The development of 3D high-content assays such as biodynamic profiling can become a critical new tool for assessing efficacy of drugs and the suitability of specific types of tissue growth for drug discovery and development. The use of biodynamic profiling to predict clinical outcome of living biopsies to cancer therapeutics can be developed into a phenotypic companion diagnostic, as well as a new tool for therapy selection in personalized medicine. This invited talk will present an overview of the optical, physical and physiological processes involved in biodynamic imaging. Several different biodynamic imaging modalities include motility contrast imaging (MCI), tissue-dynamics spectroscopy (TDS) and tissue-dynamics imaging (TDI). A wide range of potential applications will be described that include process monitoring for 3D tissue culture, drug discovery and development, cancer therapy selection, embryo assessment for in-vitro fertilization and artificial reproductive technologies, among others.

  8. 3D Printer Generated Tissue iMolds for Cleared Tissue Using Single- and Multi-Photon Microscopy for Deep Tissue Evaluation.

    PubMed

    Miller, Sean J; Rothstein, Jeffrey D

    2017-01-01

    Pathological analyses and methodology has recently undergone a dramatic revolution. With the creation of tissue clearing methods such as CLARITY and CUBIC, groups can now achieve complete transparency in tissue samples in nano-porous hydrogels. Cleared tissue is then imagined in a semi-aqueous medium that matches the refractive index of the objective being used. However, one major challenge is the ability to control tissue movement during imaging and to relocate precise locations post sequential clearing and re-staining. Using 3D printers, we designed tissue molds that fit precisely around the specimen being imaged. First, images are taken of the specimen, followed by importing and design of a structural mold, then printed with affordable plastics by a 3D printer. With our novel design, we have innovated tissue molds called innovative molds (iMolds) that can be generated in any laboratory and are customized for any organ, tissue, or bone matter being imaged. Furthermore, the inexpensive and reusable tissue molds are made compatible for any microscope such as single and multi-photon confocal with varying stage dimensions. Excitingly, iMolds can also be generated to hold multiple organs in one mold, making reconstruction and imaging much easier. Taken together, with iMolds it is now possible to image cleared tissue in clearing medium while limiting movement and being able to relocate precise anatomical and cellular locations on sequential imaging events in any basic laboratory. This system provides great potential for screening widespread effects of therapeutics and disease across entire organ systems.

  9. Flexible needle with integrated optical coherence tomography probe for imaging during transbronchial tissue aspiration

    NASA Astrophysics Data System (ADS)

    Li, Jiawen; Quirk, Bryden C.; Noble, Peter B.; Kirk, Rodney W.; Sampson, David D.; McLaughlin, Robert A.

    2017-10-01

    Transbronchial needle aspiration (TBNA) of small lesions or lymph nodes in the lung may result in nondiagnostic tissue samples. We demonstrate the integration of an optical coherence tomography (OCT) probe into a 19-gauge flexible needle for lung tissue aspiration. This probe allows simultaneous visualization and aspiration of the tissue. By eliminating the need for insertion and withdrawal of a separate imaging probe, this integrated design minimizes the risk of dislodging the needle from the lesion prior to aspiration and may facilitate more accurate placement of the needle. Results from in situ imaging in a sheep lung show clear distinction between solid tissue and two typical constituents of nondiagnostic samples (adipose and lung parenchyma). Clinical translation of this OCT-guided aspiration needle holds promise for improving the diagnostic yield of TBNA.

  10. Image-guided tissue engineering

    PubMed Central

    Ballyns, Jeffrey J; Bonassar, Lawrence J

    2009-01-01

    Replication of anatomic shape is a significant challenge in developing implants for regenerative medicine. This has lead to significant interest in using medical imaging techniques such as magnetic resonance imaging and computed tomography to design tissue engineered constructs. Implementation of medical imaging and computer aided design in combination with technologies for rapid prototyping of living implants enables the generation of highly reproducible constructs with spatial resolution up to 25 μm. In this paper, we review the medical imaging modalities available and a paradigm for choosing a particular imaging technique. We also present fabrication techniques and methodologies for producing cellular engineered constructs. Finally, we comment on future challenges involved with image guided tissue engineering and efforts to generate engineered constructs ready for implantation. PMID:19583811

  11. Improved QD-BRET conjugates for detection and imaging

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Xing Yun; So, Min-kyung; Koh, Ai Leen

    2008-08-01

    Self-illuminating quantum dots, also known as QD-BRET conjugates, are a new class of quantum dot bioconjugates which do not need external light for excitation. Instead, light emission relies on the bioluminescence resonance energy transfer from the attached Renilla luciferase enzyme, which emits light upon the oxidation of its substrate. QD-BRET combines the advantages of the QDs (such as superior brightness and photostability, tunable emission, multiplexing) as well as the high sensitivity of bioluminescence imaging, thus holding the promise for improved deep tissue in vivo imaging. Although studies have demonstrated the superior sensitivity and deep tissue imaging potential, the stability ofmore » the QD-BRET conjugates in biological environment needs to be improved for long-term imaging studies such as in vivo cell tracking. In this study, we seek to improve the stability of QD-BRET probes through polymeric encapsulation with a polyacrylamide gel. Results show that encapsulation caused some activity loss, but significantly improved both the in vitro serum stability and in vivo stability when subcutaneously injected into the animal. Stable QD-BRET probes should further facilitate their applications for both in vitro testing as well as in vivo cell tracking studies.« less

  12. Tissue microstructure estimation using a deep network inspired by a dictionary-based framework.

    PubMed

    Ye, Chuyang

    2017-12-01

    Diffusion magnetic resonance imaging (dMRI) captures the anisotropic pattern of water displacement in the neuronal tissue and allows noninvasive investigation of the complex tissue microstructure. A number of biophysical models have been proposed to relate the tissue organization with the observed diffusion signals, so that the tissue microstructure can be inferred. The Neurite Orientation Dispersion and Density Imaging (NODDI) model has been a popular choice and has been widely used for many neuroscientific studies. It models the diffusion signal with three compartments that are characterized by distinct diffusion properties, and the parameters in the model describe tissue microstructure. In NODDI, these parameters are estimated in a maximum likelihood framework, where the nonlinear model fitting is computationally intensive. Therefore, efforts have been made to develop efficient and accurate algorithms for NODDI microstructure estimation, which is still an open problem. In this work, we propose a deep network based approach that performs end-to-end estimation of NODDI microstructure, which is named Microstructure Estimation using a Deep Network (MEDN). MEDN comprises two cascaded stages and is motivated by the AMICO algorithm, where the NODDI microstructure estimation is formulated in a dictionary-based framework. The first stage computes the coefficients of the dictionary. It resembles the solution to a sparse reconstruction problem, where the iterative process in conventional estimation approaches is unfolded and truncated, and the weights are learned instead of predetermined by the dictionary. In the second stage, microstructure properties are computed from the output of the first stage, which resembles the weighted sum of normalized dictionary coefficients in AMICO, and the weights are also learned. Because spatial consistency of diffusion signals can be used to reduce the effect of noise, we also propose MEDN+, which is an extended version of MEDN. MEDN

  13. Deep Learning in Nuclear Medicine and Molecular Imaging: Current Perspectives and Future Directions.

    PubMed

    Choi, Hongyoon

    2018-04-01

    Recent advances in deep learning have impacted various scientific and industrial fields. Due to the rapid application of deep learning in biomedical data, molecular imaging has also started to adopt this technique. In this regard, it is expected that deep learning will potentially affect the roles of molecular imaging experts as well as clinical decision making. This review firstly offers a basic overview of deep learning particularly for image data analysis to give knowledge to nuclear medicine physicians and researchers. Because of the unique characteristics and distinctive aims of various types of molecular imaging, deep learning applications can be different from other fields. In this context, the review deals with current perspectives of deep learning in molecular imaging particularly in terms of development of biomarkers. Finally, future challenges of deep learning application for molecular imaging and future roles of experts in molecular imaging will be discussed.

  14. Distributed deep learning networks among institutions for medical imaging.

    PubMed

    Chang, Ken; Balachandar, Niranjan; Lam, Carson; Yi, Darvin; Brown, James; Beers, Andrew; Rosen, Bruce; Rubin, Daniel L; Kalpathy-Cramer, Jayashree

    2018-03-29

    Deep learning has become a promising approach for automated support for clinical diagnosis. When medical data samples are limited, collaboration among multiple institutions is necessary to achieve high algorithm performance. However, sharing patient data often has limitations due to technical, legal, or ethical concerns. In this study, we propose methods of distributing deep learning models as an attractive alternative to sharing patient data. We simulate the distribution of deep learning models across 4 institutions using various training heuristics and compare the results with a deep learning model trained on centrally hosted patient data. The training heuristics investigated include ensembling single institution models, single weight transfer, and cyclical weight transfer. We evaluated these approaches for image classification in 3 independent image collections (retinal fundus photos, mammography, and ImageNet). We find that cyclical weight transfer resulted in a performance that was comparable to that of centrally hosted patient data. We also found that there is an improvement in the performance of cyclical weight transfer heuristic with a high frequency of weight transfer. We show that distributing deep learning models is an effective alternative to sharing patient data. This finding has implications for any collaborative deep learning study.

  15. Nonlinear spectral imaging of biological tissues

    NASA Astrophysics Data System (ADS)

    Palero, J. A.

    2007-07-01

    The work presented in this thesis demonstrates live high resolution 3D imaging of tissue in its native state and environment. The nonlinear interaction between focussed femtosecond light pulses and the biological tissue results in the emission of natural autofluorescence and second-harmonic signal. Because biological intrinsic emission is generally very weak and extends from the ultraviolet to the visible spectral range, a broad-spectral range and high sensitivity 3D spectral imaging system is developed. Imaging the spectral characteristics of the biological intrinsic emission reveals the structure and biochemistry of the cells and extra-cellular components. By using different methods in visualizing the spectral images, discrimination between different tissue structures is achieved without the use of any stain or fluorescent label. For instance, RGB real color spectral images of the intrinsic emission of mouse skin tissues show blue cells, green hair follicles, and purple collagen fibers. The color signature of each tissue component is directly related to its characteristic emission spectrum. The results of this study show that skin tissue nonlinear intrinsic emission is mainly due to the autofluorescence of reduced nicotinamide adenine dinucleotide (phosphate), flavins, keratin, melanin, phospholipids, elastin and collagen and nonlinear Raman scattering and second-harmonic generation in Type I collagen. In vivo time-lapse spectral imaging is implemented to study metabolic changes in epidermal cells in tissues. Optical scattering in tissues, a key factor in determining the maximum achievable imaging depth, is also investigated in this work.

  16. A deep-learning classifier identifies patients with clinical heart failure using whole-slide images of H&E tissue

    PubMed Central

    Peyster, Eliot G.; Frank, Renee; Margulies, Kenneth B.; Feldman, Michael D.

    2018-01-01

    Over 26 million people worldwide suffer from heart failure annually. When the cause of heart failure cannot be identified, endomyocardial biopsy (EMB) represents the gold-standard for the evaluation of disease. However, manual EMB interpretation has high inter-rater variability. Deep convolutional neural networks (CNNs) have been successfully applied to detect cancer, diabetic retinopathy, and dermatologic lesions from images. In this study, we develop a CNN classifier to detect clinical heart failure from H&E stained whole-slide images from a total of 209 patients, 104 patients were used for training and the remaining 105 patients for independent testing. The CNN was able to identify patients with heart failure or severe pathology with a 99% sensitivity and 94% specificity on the test set, outperforming conventional feature-engineering approaches. Importantly, the CNN outperformed two expert pathologists by nearly 20%. Our results suggest that deep learning analytics of EMB can be used to predict cardiac outcome. PMID:29614076

  17. A deep-learning classifier identifies patients with clinical heart failure using whole-slide images of H&E tissue.

    PubMed

    Nirschl, Jeffrey J; Janowczyk, Andrew; Peyster, Eliot G; Frank, Renee; Margulies, Kenneth B; Feldman, Michael D; Madabhushi, Anant

    2018-01-01

    Over 26 million people worldwide suffer from heart failure annually. When the cause of heart failure cannot be identified, endomyocardial biopsy (EMB) represents the gold-standard for the evaluation of disease. However, manual EMB interpretation has high inter-rater variability. Deep convolutional neural networks (CNNs) have been successfully applied to detect cancer, diabetic retinopathy, and dermatologic lesions from images. In this study, we develop a CNN classifier to detect clinical heart failure from H&E stained whole-slide images from a total of 209 patients, 104 patients were used for training and the remaining 105 patients for independent testing. The CNN was able to identify patients with heart failure or severe pathology with a 99% sensitivity and 94% specificity on the test set, outperforming conventional feature-engineering approaches. Importantly, the CNN outperformed two expert pathologists by nearly 20%. Our results suggest that deep learning analytics of EMB can be used to predict cardiac outcome.

  18. 3D printing facilitated scaffold-free tissue unit fabrication.

    PubMed

    Tan, Yu; Richards, Dylan J; Trusk, Thomas C; Visconti, Richard P; Yost, Michael J; Kindy, Mark S; Drake, Christopher J; Argraves, William Scott; Markwald, Roger R; Mei, Ying

    2014-06-01

    Tissue spheroids hold great potential in tissue engineering as building blocks to assemble into functional tissues. To date, agarose molds have been extensively used to facilitate fusion process of tissue spheroids. As a molding material, agarose typically requires low temperature plates for gelation and/or heated dispenser units. Here, we proposed and developed an alginate-based, direct 3D mold-printing technology: 3D printing microdroplets of alginate solution into biocompatible, bio-inert alginate hydrogel molds for the fabrication of scaffold-free tissue engineering constructs. Specifically, we developed a 3D printing technology to deposit microdroplets of alginate solution on calcium containing substrates in a layer-by-layer fashion to prepare ring-shaped 3D hydrogel molds. Tissue spheroids composed of 50% endothelial cells and 50% smooth muscle cells were robotically placed into the 3D printed alginate molds using a 3D printer, and were found to rapidly fuse into toroid-shaped tissue units. Histological and immunofluorescence analysis indicated that the cells secreted collagen type I playing a critical role in promoting cell-cell adhesion, tissue formation and maturation.

  19. A survey on deep learning in medical image analysis.

    PubMed

    Litjens, Geert; Kooi, Thijs; Bejnordi, Babak Ehteshami; Setio, Arnaud Arindra Adiyoso; Ciompi, Francesco; Ghafoorian, Mohsen; van der Laak, Jeroen A W M; van Ginneken, Bram; Sánchez, Clara I

    2017-12-01

    Deep learning algorithms, in particular convolutional networks, have rapidly become a methodology of choice for analyzing medical images. This paper reviews the major deep learning concepts pertinent to medical image analysis and summarizes over 300 contributions to the field, most of which appeared in the last year. We survey the use of deep learning for image classification, object detection, segmentation, registration, and other tasks. Concise overviews are provided of studies per application area: neuro, retinal, pulmonary, digital pathology, breast, cardiac, abdominal, musculoskeletal. We end with a summary of the current state-of-the-art, a critical discussion of open challenges and directions for future research. Copyright © 2017 Elsevier B.V. All rights reserved.

  20. Distribution, composition and functions of gelatinous tissues in deep-sea fishes.

    PubMed

    Gerringer, Mackenzie E; Drazen, Jeffrey C; Linley, Thomas D; Summers, Adam P; Jamieson, Alan J; Yancey, Paul H

    2017-12-01

    Many deep-sea fishes have a gelatinous layer, or subdermal extracellular matrix, below the skin or around the spine. We document the distribution of gelatinous tissues across fish families (approx. 200 species in ten orders), then review and investigate their composition and function. Gelatinous tissues from nine species were analysed for water content (96.53 ± 1.78% s.d.), ionic composition, osmolality, protein (0.39 ± 0.23%), lipid (0.69 ± 0.56%) and carbohydrate (0.61 ± 0.28%). Results suggest that gelatinous tissues are mostly extracellular fluid, which may allow animals to grow inexpensively. Further, almost all gelatinous tissues floated in cold seawater, thus their lower density than seawater may contribute to buoyancy in some species. We also propose a new hypothesis: gelatinous tissues, which are inexpensive to grow, may sometimes be a method to increase swimming efficiency by fairing the transition from trunk to tail. Such a layer is particularly prominent in hadal snailfishes (Liparidae); therefore, a robotic snailfish model was designed and constructed to analyse the influence of gelatinous tissues on locomotory performance. The model swam faster with a watery layer, representing gelatinous tissue, around the tail than without. Results suggest that the tissues may, in addition to providing buoyancy and low-cost growth, aid deep-sea fish locomotion.

  1. Distribution, composition and functions of gelatinous tissues in deep-sea fishes

    PubMed Central

    Drazen, Jeffrey C.; Linley, Thomas D.; Summers, Adam P.; Jamieson, Alan J.; Yancey, Paul H.

    2017-01-01

    Many deep-sea fishes have a gelatinous layer, or subdermal extracellular matrix, below the skin or around the spine. We document the distribution of gelatinous tissues across fish families (approx. 200 species in ten orders), then review and investigate their composition and function. Gelatinous tissues from nine species were analysed for water content (96.53 ± 1.78% s.d.), ionic composition, osmolality, protein (0.39 ± 0.23%), lipid (0.69 ± 0.56%) and carbohydrate (0.61 ± 0.28%). Results suggest that gelatinous tissues are mostly extracellular fluid, which may allow animals to grow inexpensively. Further, almost all gelatinous tissues floated in cold seawater, thus their lower density than seawater may contribute to buoyancy in some species. We also propose a new hypothesis: gelatinous tissues, which are inexpensive to grow, may sometimes be a method to increase swimming efficiency by fairing the transition from trunk to tail. Such a layer is particularly prominent in hadal snailfishes (Liparidae); therefore, a robotic snailfish model was designed and constructed to analyse the influence of gelatinous tissues on locomotory performance. The model swam faster with a watery layer, representing gelatinous tissue, around the tail than without. Results suggest that the tissues may, in addition to providing buoyancy and low-cost growth, aid deep-sea fish locomotion. PMID:29308245

  2. Iterative deep convolutional encoder-decoder network for medical image segmentation.

    PubMed

    Jung Uk Kim; Hak Gu Kim; Yong Man Ro

    2017-07-01

    In this paper, we propose a novel medical image segmentation using iterative deep learning framework. We have combined an iterative learning approach and an encoder-decoder network to improve segmentation results, which enables to precisely localize the regions of interest (ROIs) including complex shapes or detailed textures of medical images in an iterative manner. The proposed iterative deep convolutional encoder-decoder network consists of two main paths: convolutional encoder path and convolutional decoder path with iterative learning. Experimental results show that the proposed iterative deep learning framework is able to yield excellent medical image segmentation performances for various medical images. The effectiveness of the proposed method has been proved by comparing with other state-of-the-art medical image segmentation methods.

  3. Radiometric Thermometry for Wearable Deep Tissue Monitoring

    NASA Astrophysics Data System (ADS)

    Momenroodaki, Parisa

    Microwave thermometry is an attractive non-invasive method for measuring internal body temperature. This approach has the potential of enabling a wearable device that can continuously monitor core body temperature. There are a number of health-related applications in both diagnostics and therapy, which can benefit from the knowledge of core body temperature. However,there are a limited number of device solutions, which are usually not wearable or cannot continuously monitor internal body temperature non-invasively. In this thesis, a possible path toward implementing such a thermometer is presented. The device operates in the "quiet" frequency band of 1.4 GHz which is chosen as a compromise between sensing depth and radio frequency interference (RFI). A major challenge in microwave thermometry is detecting small temperature variations of deep tissue layers from surface (skin) measurements. The type and thickness of tissue materials significantly affect the design of the probe, which has the function of receiving black-body radiation from tissues beneath it and coupling the power to a sensitive radiometric receiver. High dielectric constant contrast between skin, fat (/bone), and muscle layers suggests structures with dominant tangential component of the electric field, such as a patch or slot. Adding a layer of low-loss,low-dielectric constant superstrate can further reduce the contribution of superficial tissue layers in the received thermal noise. Several probe types are designed using a full-wave electromagnetic simulator, with a goal of maximizing the power reception from deep tissue layers. The designs are validated with a second software tool and various measurements. A stable, narrow-band, and highly sensitive radiometer is developed, enabling the device to operate in a non-shielded RF environment.To use the microwave thermometer in a RF congested environment, not only narrow-band probe and radiometers are used but an additional probe is introduced for

  4. Use of kurtosis for locating deep blood vessels in raw speckle imaging using a homogeneity representation.

    PubMed

    Peregrina-Barreto, Hayde; Perez-Corona, Elizabeth; Rangel-Magdaleno, Jose; Ramos-Garcia, Ruben; Chiu, Roger; Ramirez-San-Juan, Julio C

    2017-06-01

    Visualization of deep blood vessels in speckle images is an important task as it is used to analyze the dynamics of the blood flow and the health status of biological tissue. Laser speckle imaging is a wide-field optical technique to measure relative blood flow speed based on the local speckle contrast analysis. However, it has been reported that this technique is limited to certain deep blood vessels (about ? = 300 ?? ? m ) because of the high scattering of the sample; beyond this depth, the quality of the vessel’s image decreases. The use of a representation based on homogeneity values, computed from the co-occurrence matrix, is proposed as it provides an improved vessel definition and its corresponding diameter. Moreover, a methodology is proposed for automatic blood vessel location based on the kurtosis analysis. Results were obtained from the different skin phantoms, showing that it is possible to identify the vessel region for different morphologies, even up to 900 ?? ? m in depth.

  5. Characterization of water molecular state in in-vivo thick tissues using diffuse optical spectroscopic imaging

    NASA Astrophysics Data System (ADS)

    Chung, So Hyun

    Structural changes in water molecules are related to physiological, anatomical and pathological properties of tissues. Near infrared (NIR) optical absorption methods are sensitive to water; however, detailed characterization of water in thick tissues is difficult to achieve because subtle spectral shifts can be obscured by multiple light scattering. In the NIR, a water absorption peak is observed around 975 nm. The precise NIR peak's shape and position are highly sensitive to water molecular disposition. A bound water index (BWI) was developed that quantifies the spectral shift and shape changes observed in tissue water absorption spectra measured by broadband diffuse optical spectroscopic imaging (DOSI). DOSI quantitatively measures light absorption and scattering spectra in cm-deep tissues and therefore reveals bound water spectral shifts. BWI as a water state index was validated by comparing broadband DOSI to MRI and a conductivity cell using bound water phantoms. Non-invasive BWI measurements of malignant and normal tissues in 18 subjects showed a significantly higher fraction of free water in malignant tissues (p<0.0001) compared to normal tissues. BWI showed potential as a prognostic index based on high correlations with tumor grade and size. An algorithm for absolute temperature measurements in deep tissues was developed based on resolving opposing effects of water vibrational frequency shifts due to macromolecular binding. DOSI measures absolute temperature with a difference of 1.1+/-0.91°C from a thermistor. Deep tissue temperature measured in forearms during cold-stress was consistent with previously reported invasively-measured deep tissue temperature. Finally, the BWI was compared to Apparent Diffusion Coefficient (ADC) of diffusion weighted MRI in 9 breast cancer patients. The BWI and ADC correlated (R=0.8, p=<0.01) and both parameters decreased with increasing bulk water content in cancer tissues. Although BWI and ADC are positively correlated in

  6. Deep kernel learning method for SAR image target recognition

    NASA Astrophysics Data System (ADS)

    Chen, Xiuyuan; Peng, Xiyuan; Duan, Ran; Li, Junbao

    2017-10-01

    With the development of deep learning, research on image target recognition has made great progress in recent years. Remote sensing detection urgently requires target recognition for military, geographic, and other scientific research. This paper aims to solve the synthetic aperture radar image target recognition problem by combining deep and kernel learning. The model, which has a multilayer multiple kernel structure, is optimized layer by layer with the parameters of Support Vector Machine and a gradient descent algorithm. This new deep kernel learning method improves accuracy and achieves competitive recognition results compared with other learning methods.

  7. Image quality assessment using deep convolutional networks

    NASA Astrophysics Data System (ADS)

    Li, Yezhou; Ye, Xiang; Li, Yong

    2017-12-01

    This paper proposes a method of accurately assessing image quality without a reference image by using a deep convolutional neural network. Existing training based methods usually utilize a compact set of linear filters for learning features of images captured by different sensors to assess their quality. These methods may not be able to learn the semantic features that are intimately related with the features used in human subject assessment. Observing this drawback, this work proposes training a deep convolutional neural network (CNN) with labelled images for image quality assessment. The ReLU in the CNN allows non-linear transformations for extracting high-level image features, providing a more reliable assessment of image quality than linear filters. To enable the neural network to take images of any arbitrary size as input, the spatial pyramid pooling (SPP) is introduced connecting the top convolutional layer and the fully-connected layer. In addition, the SPP makes the CNN robust to object deformations to a certain extent. The proposed method taking an image as input carries out an end-to-end learning process, and outputs the quality of the image. It is tested on public datasets. Experimental results show that it outperforms existing methods by a large margin and can accurately assess the image quality on images taken by different sensors of varying sizes.

  8. Application of deep learning to the classification of images from colposcopy.

    PubMed

    Sato, Masakazu; Horie, Koji; Hara, Aki; Miyamoto, Yuichiro; Kurihara, Kazuko; Tomio, Kensuke; Yokota, Harushige

    2018-03-01

    The objective of the present study was to investigate whether deep learning could be applied successfully to the classification of images from colposcopy. For this purpose, a total of 158 patients who underwent conization were enrolled, and medical records and data from the gynecological oncology database were retrospectively reviewed. Deep learning was performed with the Keras neural network and TensorFlow libraries. Using preoperative images from colposcopy as the input data and deep learning technology, the patients were classified into three groups [severe dysplasia, carcinoma in situ (CIS) and invasive cancer (IC)]. A total of 485 images were obtained for the analysis, of which 142 images were of severe dysplasia (2.9 images/patient), 257 were of CIS (3.3 images/patient), and 86 were of IC (4.1 images/patient). Of these, 233 images were captured with a green filter, and the remaining 252 were captured without a green filter. Following the application of L2 regularization, L1 regularization, dropout and data augmentation, the accuracy of the validation dataset was ~50%. Although the present study is preliminary, the results indicated that deep learning may be applied to classify colposcopy images.

  9. Application of deep learning to the classification of images from colposcopy

    PubMed Central

    Sato, Masakazu; Horie, Koji; Hara, Aki; Miyamoto, Yuichiro; Kurihara, Kazuko; Tomio, Kensuke; Yokota, Harushige

    2018-01-01

    The objective of the present study was to investigate whether deep learning could be applied successfully to the classification of images from colposcopy. For this purpose, a total of 158 patients who underwent conization were enrolled, and medical records and data from the gynecological oncology database were retrospectively reviewed. Deep learning was performed with the Keras neural network and TensorFlow libraries. Using preoperative images from colposcopy as the input data and deep learning technology, the patients were classified into three groups [severe dysplasia, carcinoma in situ (CIS) and invasive cancer (IC)]. A total of 485 images were obtained for the analysis, of which 142 images were of severe dysplasia (2.9 images/patient), 257 were of CIS (3.3 images/patient), and 86 were of IC (4.1 images/patient). Of these, 233 images were captured with a green filter, and the remaining 252 were captured without a green filter. Following the application of L2 regularization, L1 regularization, dropout and data augmentation, the accuracy of the validation dataset was ~50%. Although the present study is preliminary, the results indicated that deep learning may be applied to classify colposcopy images. PMID:29456725

  10. Photoacoustic image reconstruction via deep learning

    NASA Astrophysics Data System (ADS)

    Antholzer, Stephan; Haltmeier, Markus; Nuster, Robert; Schwab, Johannes

    2018-02-01

    Applying standard algorithms to sparse data problems in photoacoustic tomography (PAT) yields low-quality images containing severe under-sampling artifacts. To some extent, these artifacts can be reduced by iterative image reconstruction algorithms which allow to include prior knowledge such as smoothness, total variation (TV) or sparsity constraints. These algorithms tend to be time consuming as the forward and adjoint problems have to be solved repeatedly. Further, iterative algorithms have additional drawbacks. For example, the reconstruction quality strongly depends on a-priori model assumptions about the objects to be recovered, which are often not strictly satisfied in practical applications. To overcome these issues, in this paper, we develop direct and efficient reconstruction algorithms based on deep learning. As opposed to iterative algorithms, we apply a convolutional neural network, whose parameters are trained before the reconstruction process based on a set of training data. For actual image reconstruction, a single evaluation of the trained network yields the desired result. Our presented numerical results (using two different network architectures) demonstrate that the proposed deep learning approach reconstructs images with a quality comparable to state of the art iterative reconstruction methods.

  11. Jet-imagesdeep learning edition

    DOE PAGES

    de Oliveira, Luke; Kagan, Michael; Mackey, Lester; ...

    2016-07-13

    Building on the notion of a particle physics detector as a camera and the collimated streams of high energy particles, or jets, it measures as an image, we investigate the potential of machine learning techniques based on deep learning architectures to identify highly boosted W bosons. Modern deep learning algorithms trained on jet images can out-perform standard physically-motivated feature driven approaches to jet tagging. We develop techniques for visualizing how these features are learned by the network and what additional information is used to improve performance. Finally, this interplay between physically-motivated feature driven tools and supervised learning algorithms is generalmore » and can be used to significantly increase the sensitivity to discover new particles and new forces, and gain a deeper understanding of the physics within jets.« less

  12. Jet-imagesdeep learning edition

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    de Oliveira, Luke; Kagan, Michael; Mackey, Lester

    Building on the notion of a particle physics detector as a camera and the collimated streams of high energy particles, or jets, it measures as an image, we investigate the potential of machine learning techniques based on deep learning architectures to identify highly boosted W bosons. Modern deep learning algorithms trained on jet images can out-perform standard physically-motivated feature driven approaches to jet tagging. We develop techniques for visualizing how these features are learned by the network and what additional information is used to improve performance. Finally, this interplay between physically-motivated feature driven tools and supervised learning algorithms is generalmore » and can be used to significantly increase the sensitivity to discover new particles and new forces, and gain a deeper understanding of the physics within jets.« less

  13. Imaging of Alkaline Phosphatase Activity in Bone Tissue

    PubMed Central

    Gade, Terence P.; Motley, Matthew W.; Beattie, Bradley J.; Bhakta, Roshni; Boskey, Adele L.; Koutcher, Jason A.; Mayer-Kuckuk, Philipp

    2011-01-01

    The purpose of this study was to develop a paradigm for quantitative molecular imaging of bone cell activity. We hypothesized the feasibility of non-invasive imaging of the osteoblast enzyme alkaline phosphatase (ALP) using a small imaging molecule in combination with 19Flourine magnetic resonance spectroscopic imaging (19FMRSI). 6, 8-difluoro-4-methylumbelliferyl phosphate (DiFMUP), a fluorinated ALP substrate that is activatable to a fluorescent hydrolysis product was utilized as a prototype small imaging molecule. The molecular structure of DiFMUP includes two Fluorine atoms adjacent to a phosphate group allowing it and its hydrolysis product to be distinguished using 19Fluorine magnetic resonance spectroscopy (19FMRS) and 19FMRSI. ALP-mediated hydrolysis of DiFMUP was tested on osteoblastic cells and bone tissue, using serial measurements of fluorescence activity. Extracellular activation of DiFMUP on ALP-positive mouse bone precursor cells was observed. Concurringly, DiFMUP was also activated on bone derived from rat tibia. Marked inhibition of the cell and tissue activation of DiFMUP was detected after the addition of the ALP inhibitor levamisole. 19FMRS and 19FMRSI were applied for the non-invasive measurement of DiFMUP hydrolysis. 19FMRS revealed a two-peak spectrum representing DiFMUP with an associated chemical shift for the hydrolysis product. Activation of DiFMUP by ALP yielded a characteristic pharmacokinetic profile, which was quantifiable using non-localized 19FMRS and enabled the development of a pharmacokinetic model of ALP activity. Application of 19FMRSI facilitated anatomically accurate, non-invasive imaging of ALP concentration and activity in rat bone. Thus, 19FMRSI represents a promising approach for the quantitative imaging of bone cell activity during bone formation with potential for both preclinical and clinical applications. PMID:21799916

  14. Three Dimensional Imaging of Paraffin Embedded Human Lung Tissue Samples by Micro-Computed Tomography

    PubMed Central

    Scott, Anna E.; Vasilescu, Dragos M.; Seal, Katherine A. D.; Keyes, Samuel D.; Mavrogordato, Mark N.; Hogg, James C.; Sinclair, Ian; Warner, Jane A.; Hackett, Tillie-Louise; Lackie, Peter M.

    2015-01-01

    Background Understanding the three-dimensional (3-D) micro-architecture of lung tissue can provide insights into the pathology of lung disease. Micro computed tomography (µCT) has previously been used to elucidate lung 3D histology and morphometry in fixed samples that have been stained with contrast agents or air inflated and dried. However, non-destructive microstructural 3D imaging of formalin-fixed paraffin embedded (FFPE) tissues would facilitate retrospective analysis of extensive tissue archives of lung FFPE lung samples with linked clinical data. Methods FFPE human lung tissue samples (n = 4) were scanned using a Nikon metrology µCT scanner. Semi-automatic techniques were used to segment the 3D structure of airways and blood vessels. Airspace size (mean linear intercept, Lm) was measured on µCT images and on matched histological sections from the same FFPE samples imaged by light microscopy to validate µCT imaging. Results The µCT imaging protocol provided contrast between tissue and paraffin in FFPE samples (15mm x 7mm). Resolution (voxel size 6.7 µm) in the reconstructed images was sufficient for semi-automatic image segmentation of airways and blood vessels as well as quantitative airspace analysis. The scans were also used to scout for regions of interest, enabling time-efficient preparation of conventional histological sections. The Lm measurements from µCT images were not significantly different to those from matched histological sections. Conclusion We demonstrated how non-destructive imaging of routinely prepared FFPE samples by laboratory µCT can be used to visualize and assess the 3D morphology of the lung including by morphometric analysis. PMID:26030902

  15. Deep learning of symmetrical discrepancies for computer-aided detection of mammographic masses

    NASA Astrophysics Data System (ADS)

    Kooi, Thijs; Karssemeijer, Nico

    2017-03-01

    When humans identify objects in images, context is an important cue; a cheetah is more likely to be a domestic cat when a television set is recognised in the background. Similar principles apply to the analysis of medical images. The detection of diseases that manifest unilaterally in symmetrical organs or organ pairs can in part be facilitated by a search for symmetrical discrepancies in or between the organs in question. During a mammographic exam, images are recorded of each breast and absence of a certain structure around the same location in the contralateral image will render the area under scrutiny more suspicious and conversely, the presence of similar tissue less so. In this paper, we present a fusion scheme for a deep Convolutional Neural Network (CNN) architecture with the goal to optimally capture such asymmetries. The method is applied to the domain of mammography CAD, but can be relevant to other medical image analysis tasks where symmetry is important such as lung, prostate or brain images.

  16. Ontology-based, Tissue MicroArray oriented, image centered tissue bank

    PubMed Central

    Viti, Federica; Merelli, Ivan; Caprera, Andrea; Lazzari, Barbara; Stella, Alessandra; Milanesi, Luciano

    2008-01-01

    Background Tissue MicroArray technique is becoming increasingly important in pathology for the validation of experimental data from transcriptomic analysis. This approach produces many images which need to be properly managed, if possible with an infrastructure able to support tissue sharing between institutes. Moreover, the available frameworks oriented to Tissue MicroArray provide good storage for clinical patient, sample treatment and block construction information, but their utility is limited by the lack of data integration with biomolecular information. Results In this work we propose a Tissue MicroArray web oriented system to support researchers in managing bio-samples and, through the use of ontologies, enables tissue sharing aimed at the design of Tissue MicroArray experiments and results evaluation. Indeed, our system provides ontological description both for pre-analysis tissue images and for post-process analysis image results, which is crucial for information exchange. Moreover, working on well-defined terms it is then possible to query web resources for literature articles to integrate both pathology and bioinformatics data. Conclusions Using this system, users associate an ontology-based description to each image uploaded into the database and also integrate results with the ontological description of biosequences identified in every tissue. Moreover, it is possible to integrate the ontological description provided by the user with a full compliant gene ontology definition, enabling statistical studies about correlation between the analyzed pathology and the most commonly related biological processes. PMID:18460177

  17. Deep Learning in Medical Imaging: General Overview

    PubMed Central

    Lee, June-Goo; Jun, Sanghoon; Cho, Young-Won; Lee, Hyunna; Kim, Guk Bae

    2017-01-01

    The artificial neural network (ANN)–a machine learning technique inspired by the human neuronal synapse system–was introduced in the 1950s. However, the ANN was previously limited in its ability to solve actual problems, due to the vanishing gradient and overfitting problems with training of deep architecture, lack of computing power, and primarily the absence of sufficient data to train the computer system. Interest in this concept has lately resurfaced, due to the availability of big data, enhanced computing power with the current graphics processing units, and novel algorithms to train the deep neural network. Recent studies on this technology suggest its potentially to perform better than humans in some visual and auditory recognition tasks, which may portend its applications in medicine and healthcare, especially in medical imaging, in the foreseeable future. This review article offers perspectives on the history, development, and applications of deep learning technology, particularly regarding its applications in medical imaging. PMID:28670152

  18. Deep Learning in Medical Imaging: General Overview.

    PubMed

    Lee, June-Goo; Jun, Sanghoon; Cho, Young-Won; Lee, Hyunna; Kim, Guk Bae; Seo, Joon Beom; Kim, Namkug

    2017-01-01

    The artificial neural network (ANN)-a machine learning technique inspired by the human neuronal synapse system-was introduced in the 1950s. However, the ANN was previously limited in its ability to solve actual problems, due to the vanishing gradient and overfitting problems with training of deep architecture, lack of computing power, and primarily the absence of sufficient data to train the computer system. Interest in this concept has lately resurfaced, due to the availability of big data, enhanced computing power with the current graphics processing units, and novel algorithms to train the deep neural network. Recent studies on this technology suggest its potentially to perform better than humans in some visual and auditory recognition tasks, which may portend its applications in medicine and healthcare, especially in medical imaging, in the foreseeable future. This review article offers perspectives on the history, development, and applications of deep learning technology, particularly regarding its applications in medical imaging.

  19. Photoacoustic tomography of foreign bodies in soft biological tissue.

    PubMed

    Cai, Xin; Kim, Chulhong; Pramanik, Manojit; Wang, Lihong V

    2011-04-01

    In detecting small foreign bodies in soft biological tissue, ultrasound imaging suffers from poor sensitivity (52.6%) and specificity (47.2%). Hence, alternative imaging methods are needed. Photoacoustic (PA) imaging takes advantage of strong optical absorption contrast and high ultrasonic resolution. A PA imaging system is employed to detect foreign bodies in biological tissues. To achieve deep penetration, we use near-infrared light ranging from 750 to 800 nm and a 5-MHz spherically focused ultrasonic transducer. PA images were obtained from various targets including glass, wood, cloth, plastic, and metal embedded more than 1 cm deep in chicken tissue. The locations and sizes of the targets from the PA images agreed well with those of the actual samples. Spectroscopic PA imaging was also performed on the objects. These results suggest that PA imaging can potentially be a useful intraoperative imaging tool to identify foreign bodies.

  20. Classification of breast MRI lesions using small-size training sets: comparison of deep learning approaches

    NASA Astrophysics Data System (ADS)

    Amit, Guy; Ben-Ari, Rami; Hadad, Omer; Monovich, Einat; Granot, Noa; Hashoul, Sharbell

    2017-03-01

    Diagnostic interpretation of breast MRI studies requires meticulous work and a high level of expertise. Computerized algorithms can assist radiologists by automatically characterizing the detected lesions. Deep learning approaches have shown promising results in natural image classification, but their applicability to medical imaging is limited by the shortage of large annotated training sets. In this work, we address automatic classification of breast MRI lesions using two different deep learning approaches. We propose a novel image representation for dynamic contrast enhanced (DCE) breast MRI lesions, which combines the morphological and kinetics information in a single multi-channel image. We compare two classification approaches for discriminating between benign and malignant lesions: training a designated convolutional neural network and using a pre-trained deep network to extract features for a shallow classifier. The domain-specific trained network provided higher classification accuracy, compared to the pre-trained model, with an area under the ROC curve of 0.91 versus 0.81, and an accuracy of 0.83 versus 0.71. Similar accuracy was achieved in classifying benign lesions, malignant lesions, and normal tissue images. The trained network was able to improve accuracy by using the multi-channel image representation, and was more robust to reductions in the size of the training set. A small-size convolutional neural network can learn to accurately classify findings in medical images using only a few hundred images from a few dozen patients. With sufficient data augmentation, such a network can be trained to outperform a pre-trained out-of-domain classifier. Developing domain-specific deep-learning models for medical imaging can facilitate technological advancements in computer-aided diagnosis.

  1. Scratching the surface: the processing of pain from deep tissues.

    PubMed

    Sikandar, Shafaq; Aasvang, Eske Kvanner; Dickenson, Anthony H

    2016-04-01

    Although most pain research focuses on skin, muscles, joints and viscerae are major sources of pain. We discuss the mechanisms of deep pains arising from somatic and visceral structures and how this can lead to widespread manifestations and chronification. We include how both altered peripheral and central sensory neurotransmission lead to deep pain states and comment on key areas such as top-down modulation where little is known. It is vital that the clinical characterization of deep pain in patients is improved to allow for back translation to preclinical models so that the missing links can be ascertained. The contribution of deeper somatic and visceral tissues to various chronic pain syndromes is common but there is much we need to know.

  2. In vivo multiphoton imaging of a diverse array of fluorophores to investigate deep neurovascular structure

    PubMed Central

    Miller, David R.; Hassan, Ahmed M.; Jarrett, Jeremy W.; Medina, Flor A.; Perillo, Evan P.; Hagan, Kristen; Shams Kazmi, S. M.; Clark, Taylor A.; Sullender, Colin T.; Jones, Theresa A.; Zemelman, Boris V.; Dunn, Andrew K.

    2017-01-01

    We perform high-resolution, non-invasive, in vivo deep-tissue imaging of the mouse neocortex using multiphoton microscopy with a high repetition rate optical parametric amplifier laser source tunable between λ=1,100 and 1,400 nm. By combining the high repetition rate (511 kHz) and high pulse energy (400 nJ) of our amplifier laser system, we demonstrate imaging of vasculature labeled with Texas Red and Indocyanine Green, and neurons expressing tdTomato and yellow fluorescent protein. We measure the blood flow speed of a single capillary at a depth of 1.2 mm, and image vasculature to a depth of 1.53 mm with fine axial steps (5 μm) and reasonable acquisition times. The high image quality enabled analysis of vascular morphology at depths to 1.45 mm. PMID:28717582

  3. Speckle contrast optical tomography: A new method for deep tissue three-dimensional tomography of blood flow

    PubMed Central

    Varma, Hari M.; Valdes, Claudia P.; Kristoffersen, Anna K.; Culver, Joseph P.; Durduran, Turgut

    2014-01-01

    A novel tomographic method based on the laser speckle contrast, speckle contrast optical tomography (SCOT) is introduced that allows us to reconstruct three dimensional distribution of blood flow in deep tissues. This method is analogous to the diffuse optical tomography (DOT) but for deep tissue blood flow. We develop a reconstruction algorithm based on first Born approximation to generate three dimensional distribution of flow using the experimental data obtained from tissue simulating phantoms. PMID:24761306

  4. Tissues segmentation based on multi spectral medical images

    NASA Astrophysics Data System (ADS)

    Li, Ya; Wang, Ying

    2017-11-01

    Each band image contains the most obvious tissue feature according to the optical characteristics of different tissues in different specific bands for multispectral medical images. In this paper, the tissues were segmented by their spectral information at each multispectral medical images. Four Local Binary Patter descriptors were constructed to extract blood vessels based on the gray difference between the blood vessels and their neighbors. The segmented tissue in each band image was merged to a clear image.

  5. Near-infrared spectroscopic tissue imaging for medical applications

    DOEpatents

    Demos,; Stavros, Staggs [Livermore, CA; Michael, C [Tracy, CA

    2006-03-21

    Near infrared imaging using elastic light scattering and tissue autofluorescence are explored for medical applications. The approach involves imaging using cross-polarized elastic light scattering and tissue autofluorescence in the Near Infra-Red (NIR) coupled with image processing and inter-image operations to differentiate human tissue components.

  6. Near-infrared spectroscopic tissue imaging for medical applications

    DOEpatents

    Demos, Stavros [Livermore, CA; Staggs, Michael C [Tracy, CA

    2006-12-12

    Near infrared imaging using elastic light scattering and tissue autofluorescence are explored for medical applications. The approach involves imaging using cross-polarized elastic light scattering and tissue autofluorescence in the Near Infra-Red (NIR) coupled with image processing and inter-image operations to differentiate human tissue components.

  7. Penetration depth of photons in biological tissues from hyperspectral imaging in shortwave infrared in transmission and reflection geometries.

    PubMed

    Zhang, Hairong; Salo, Daniel; Kim, David M; Komarov, Sergey; Tai, Yuan-Chuan; Berezin, Mikhail Y

    2016-12-01

    Measurement of photon penetration in biological tissues is a central theme in optical imaging. A great number of endogenous tissue factors such as absorption, scattering, and anisotropy affect the path of photons in tissue, making it difficult to predict the penetration depth at different wavelengths. Traditional studies evaluating photon penetration at different wavelengths are focused on tissue spectroscopy that does not take into account the heterogeneity within the sample. This is especially critical in shortwave infrared where the individual vibration-based absorption properties of the tissue molecules are affected by nearby tissue components. We have explored the depth penetration in biological tissues from 900 to 1650 nm using Monte–Carlo simulation and a hyperspectral imaging system with Michelson spatial contrast as a metric of light penetration. Chromatic aberration-free hyperspectral images in transmission and reflection geometries were collected with a spectral resolution of 5.27 nm and a total acquisition time of 3 min. Relatively short recording time minimized artifacts from sample drying. Results from both transmission and reflection geometries consistently revealed that the highest spatial contrast in the wavelength range for deep tissue lies within 1300 to 1375 nm; however, in heavily pigmented tissue such as the liver, the range 1550 to 1600 nm is also prominent.

  8. Deep Learning and Its Applications in Biomedicine.

    PubMed

    Cao, Chensi; Liu, Feng; Tan, Hai; Song, Deshou; Shu, Wenjie; Li, Weizhong; Zhou, Yiming; Bo, Xiaochen; Xie, Zhi

    2018-02-01

    Advances in biological and medical technologies have been providing us explosive volumes of biological and physiological data, such as medical images, electroencephalography, genomic and protein sequences. Learning from these data facilitates the understanding of human health and disease. Developed from artificial neural networks, deep learning-based algorithms show great promise in extracting features and learning patterns from complex data. The aim of this paper is to provide an overview of deep learning techniques and some of the state-of-the-art applications in the biomedical field. We first introduce the development of artificial neural network and deep learning. We then describe two main components of deep learning, i.e., deep learning architectures and model optimization. Subsequently, some examples are demonstrated for deep learning applications, including medical image classification, genomic sequence analysis, as well as protein structure classification and prediction. Finally, we offer our perspectives for the future directions in the field of deep learning. Copyright © 2018. Production and hosting by Elsevier B.V.

  9. Coherence-Gated Sensorless Adaptive Optics Multiphoton Retinal Imaging.

    PubMed

    Cua, Michelle; Wahl, Daniel J; Zhao, Yuan; Lee, Sujin; Bonora, Stefano; Zawadzki, Robert J; Jian, Yifan; Sarunic, Marinko V

    2016-09-07

    Multiphoton microscopy enables imaging deep into scattering tissues. The efficient generation of non-linear optical effects is related to both the pulse duration (typically on the order of femtoseconds) and the size of the focused spot. Aberrations introduced by refractive index inhomogeneity in the sample distort the wavefront and enlarge the focal spot, which reduces the multiphoton signal. Traditional approaches to adaptive optics wavefront correction are not effective in thick or multi-layered scattering media. In this report, we present sensorless adaptive optics (SAO) using low-coherence interferometric detection of the excitation light for depth-resolved aberration correction of two-photon excited fluorescence (TPEF) in biological tissue. We demonstrate coherence-gated SAO TPEF using a transmissive multi-actuator adaptive lens for in vivo imaging in a mouse retina. This configuration has significant potential for reducing the laser power required for adaptive optics multiphoton imaging, and for facilitating integration with existing systems.

  10. Noncontact diffuse correlation spectroscopy for noninvasive deep tissue blood flow measurement

    NASA Astrophysics Data System (ADS)

    Lin, Yu; He, Lian; Shang, Yu; Yu, Guoqiang

    2012-01-01

    A noncontact diffuse correlation spectroscopy (DCS) probe has been developed using two separated optical paths for the source and detector. This unique design avoids the interference between the source and detector and allows large source-detector separations for deep tissue blood flow measurements. The noncontact probe has been calibrated against a contact probe in a tissue-like phantom solution and human muscle tissues; flow changes concurrently measured by the two probes are highly correlated in both phantom (R2=0.89, p<10-5) and real-tissue (R2=0.77, p<10-5, n=9) tests. The noncontact DCS holds promise for measuring blood flow in vulnerable (e.g., pressure ulcer) and soft (e.g., breast) tissues without distorting tissue hemodynamic properties.

  11. Nondestructive cryomicro-CT imaging enables structural and molecular analysis of human lung tissue.

    PubMed

    Vasilescu, Dragoş M; Phillion, André B; Tanabe, Naoya; Kinose, Daisuke; Paige, David F; Kantrowitz, Jacob J; Liu, Gang; Liu, Hanqiao; Fishbane, Nick; Verleden, Stijn E; Vanaudenaerde, Bart M; Lenburg, Marc; Stevenson, Christopher S; Spira, Avrum; Cooper, Joel D; Hackett, Tillie-Louise; Hogg, James C

    2017-01-01

    Micro-computed tomography (CT) enables three-dimensional (3D) imaging of complex soft tissue structures, but current protocols used to achieve this goal preclude cellular and molecular phenotyping of the tissue. Here we describe a radiolucent cryostage that permits micro-CT imaging of unfixed frozen human lung samples at an isotropic voxel size of (11 µm) 3 under conditions where the sample is maintained frozen at -30°C during imaging. The cryostage was tested for thermal stability to maintain samples frozen up to 8 h. This report describes the methods used to choose the materials required for cryostage construction and demonstrates that whole genome mRNA integrity and expression are not compromised by exposure to micro-CT radiation and that the tissue can be used for immunohistochemistry. The new cryostage provides a novel method enabling integration of 3D tissue structure with cellular and molecular analysis to facilitate the identification of molecular determinants of disease. The described micro-CT cryostage provides a novel way to study the three-dimensional lung structure preserved without the effects of fixatives while enabling subsequent studies of the cellular matrix composition and gene expression. This approach will, for the first time, enable researchers to study structural changes of lung tissues that occur with disease and correlate them with changes in gene or protein signatures. Copyright © 2017 the American Physiological Society.

  12. Study of CT image texture using deep learning techniques

    NASA Astrophysics Data System (ADS)

    Dutta, Sandeep; Fan, Jiahua; Chevalier, David

    2018-03-01

    For CT imaging, reduction of radiation dose while improving or maintaining image quality (IQ) is currently a very active research and development topic. Iterative Reconstruction (IR) approaches have been suggested to be able to offer better IQ to dose ratio compared to the conventional Filtered Back Projection (FBP) reconstruction. However, it has been widely reported that often CT image texture from IR is different compared to that from FBP. Researchers have proposed different figure of metrics to quantitate the texture from different reconstruction methods. But there is still a lack of practical and robust method in the field for texture description. This work applied deep learning method for CT image texture study. Multiple dose scans of a 20cm diameter cylindrical water phantom was performed on Revolution CT scanner (GE Healthcare, Waukesha) and the images were reconstructed with FBP and four different IR reconstruction settings. The training images generated were randomly allotted (80:20) to a training and validation set. An independent test set of 256-512 images/class were collected with the same scan and reconstruction settings. Multiple deep learning (DL) networks with Convolution, RELU activation, max-pooling, fully-connected, global average pooling and softmax activation layers were investigated. Impact of different image patch size for training was investigated. Original pixel data as well as normalized image data were evaluated. DL models were reliably able to classify CT image texture with accuracy up to 99%. Results show that the deep learning techniques suggest that CT IR techniques may help lower the radiation dose compared to FBP.

  13. Deep Hashing for Scalable Image Search.

    PubMed

    Lu, Jiwen; Liong, Venice Erin; Zhou, Jie

    2017-05-01

    In this paper, we propose a new deep hashing (DH) approach to learn compact binary codes for scalable image search. Unlike most existing binary codes learning methods, which usually seek a single linear projection to map each sample into a binary feature vector, we develop a deep neural network to seek multiple hierarchical non-linear transformations to learn these binary codes, so that the non-linear relationship of samples can be well exploited. Our model is learned under three constraints at the top layer of the developed deep network: 1) the loss between the compact real-valued code and the learned binary vector is minimized, 2) the binary codes distribute evenly on each bit, and 3) different bits are as independent as possible. To further improve the discriminative power of the learned binary codes, we extend DH into supervised DH (SDH) and multi-label SDH by including a discriminative term into the objective function of DH, which simultaneously maximizes the inter-class variations and minimizes the intra-class variations of the learned binary codes with the single-label and multi-label settings, respectively. Extensive experimental results on eight widely used image search data sets show that our proposed methods achieve very competitive results with the state-of-the-arts.

  14. Single-shot T2 mapping using overlapping-echo detachment planar imaging and a deep convolutional neural network.

    PubMed

    Cai, Congbo; Wang, Chao; Zeng, Yiqing; Cai, Shuhui; Liang, Dong; Wu, Yawen; Chen, Zhong; Ding, Xinghao; Zhong, Jianhui

    2018-04-24

    An end-to-end deep convolutional neural network (CNN) based on deep residual network (ResNet) was proposed to efficiently reconstruct reliable T 2 mapping from single-shot overlapping-echo detachment (OLED) planar imaging. The training dataset was obtained from simulations that were carried out on SPROM (Simulation with PRoduct Operator Matrix) software developed by our group. The relationship between the original OLED image containing two echo signals and the corresponding T 2 mapping was learned by ResNet training. After the ResNet was trained, it was applied to reconstruct the T 2 mapping from simulation and in vivo human brain data. Although the ResNet was trained entirely on simulated data, the trained network was generalized well to real human brain data. The results from simulation and in vivo human brain experiments show that the proposed method significantly outperforms the echo-detachment-based method. Reliable T 2 mapping with higher accuracy is achieved within 30 ms after the network has been trained, while the echo-detachment-based OLED reconstruction method took approximately 2 min. The proposed method will facilitate real-time dynamic and quantitative MR imaging via OLED sequence, and deep convolutional neural network has the potential to reconstruct maps from complex MRI sequences efficiently. © 2018 International Society for Magnetic Resonance in Medicine.

  15. Using deep learning for content-based medical image retrieval

    NASA Astrophysics Data System (ADS)

    Sun, Qinpei; Yang, Yuanyuan; Sun, Jianyong; Yang, Zhiming; Zhang, Jianguo

    2017-03-01

    Content-Based medical image retrieval (CBMIR) is been highly active research area from past few years. The retrieval performance of a CBMIR system crucially depends on the feature representation, which have been extensively studied by researchers for decades. Although a variety of techniques have been proposed, it remains one of the most challenging problems in current CBMIR research, which is mainly due to the well-known "semantic gap" issue that exists between low-level image pixels captured by machines and high-level semantic concepts perceived by human[1]. Recent years have witnessed some important advances of new techniques in machine learning. One important breakthrough technique is known as "deep learning". Unlike conventional machine learning methods that are often using "shallow" architectures, deep learning mimics the human brain that is organized in a deep architecture and processes information through multiple stages of transformation and representation. This means that we do not need to spend enormous energy to extract features manually. In this presentation, we propose a novel framework which uses deep learning to retrieval the medical image to improve the accuracy and speed of a CBIR in integrated RIS/PACS.

  16. 3D Deep Learning Angiography (3D-DLA) from C-arm Conebeam CT.

    PubMed

    Montoya, J C; Li, Y; Strother, C; Chen, G-H

    2018-05-01

    Deep learning is a branch of artificial intelligence that has demonstrated unprecedented performance in many medical imaging applications. Our purpose was to develop a deep learning angiography method to generate 3D cerebral angiograms from a single contrast-enhanced C-arm conebeam CT acquisition in order to reduce image artifacts and radiation dose. A set of 105 3D rotational angiography examinations were randomly selected from an internal data base. All were acquired using a clinical system in conjunction with a standard injection protocol. More than 150 million labeled voxels from 35 subjects were used for training. A deep convolutional neural network was trained to classify each image voxel into 3 tissue types (vasculature, bone, and soft tissue). The trained deep learning angiography model was then applied for tissue classification into a validation cohort of 8 subjects and a final testing cohort of the remaining 62 subjects. The final vasculature tissue class was used to generate the 3D deep learning angiography images. To quantify the generalization error of the trained model, we calculated the accuracy, sensitivity, precision, and Dice similarity coefficients for vasculature classification in relevant anatomy. The 3D deep learning angiography and clinical 3D rotational angiography images were subjected to a qualitative assessment for the presence of intersweep motion artifacts. Vasculature classification accuracy and 95% CI in the testing dataset were 98.7% (98.3%-99.1%). No residual signal from osseous structures was observed for any 3D deep learning angiography testing cases except for small regions in the otic capsule and nasal cavity compared with 37% (23/62) of the 3D rotational angiographies. Deep learning angiography accurately recreated the vascular anatomy of the 3D rotational angiography reconstructions without a mask. Deep learning angiography reduced misregistration artifacts induced by intersweep motion, and it reduced radiation exposure

  17. Deep UV Native Fluorescence Imaging of Antarctic Cryptoendolithic Communities

    NASA Technical Reports Server (NTRS)

    Storrie-Lombardi, M. C.; Douglas, S.; Sun, H.; McDonald, G. D.; Bhartia, R.; Nealson, K. H.; Hug, W. F.

    2001-01-01

    An interdisciplinary team at the Jet Propulsion Laboratory Center for Life Detection has embarked on a project to provide in situ chemical and morphological characterization of Antarctic cryptoendolithic microbial communities. We present here in situ deep ultraviolet (UV) native fluorescence and environmental scanning electron microscopy images transiting 8.5 mm into a sandstone sample from the Antarctic Dry Valleys. The deep ultraviolet imaging system employs 224.3, 248.6, and 325 nm lasers to elicit differential fluorescence and resonance Raman responses from biomolecules and minerals. The 224.3 and 248.6 nm lasers elicit a fluorescence response from the aromatic amino and nucleic acids. Excitation at 325 nm may elicit activity from a variety of biomolecules, but is more likely to elicit mineral fluorescence. The resultant fluorescence images provide in situ chemical and morphological maps of microorganisms and the associated organic matrix. Visible broadband reflectance images provide orientation against the mineral background. Environmental scanning electron micrographs provided detailed morphological information. The technique has made possible the construction of detailed fluorescent maps extending from the surface of an Antarctic sandstone sample to a depth of 8.5 mm. The images detect no evidence of microbial life in the superficial 0.2 mm crustal layer. The black lichen component between 0.3 and 0.5 mm deep absorbs all wavelengths of both laser and broadband illumination. Filamentous deep ultraviolet native fluorescent activity dominates in the white layer between 0.6 mm and 5.0 mm from the surface. These filamentous forms are fungi that continue into the red (iron-rich) region of the sample extending from 5.0 to 8.5 mm. Using differential image subtraction techniques it is possible to identify fungal nuclei. The ultraviolet response is markedly attenuated in this region, apparently from the absorption of ultraviolet light by iron-rich particles coating

  18. Classification of radiolarian images with hand-crafted and deep features

    NASA Astrophysics Data System (ADS)

    Keçeli, Ali Seydi; Kaya, Aydın; Keçeli, Seda Uzunçimen

    2017-12-01

    Radiolarians are planktonic protozoa and are important biostratigraphic and paleoenvironmental indicators for paleogeographic reconstructions. Radiolarian paleontology still remains as a low cost and the one of the most convenient way to obtain dating of deep ocean sediments. Traditional methods for identifying radiolarians are time-consuming and cannot scale to the granularity or scope necessary for large-scale studies. Automated image classification will allow making these analyses promptly. In this study, a method for automatic radiolarian image classification is proposed on Scanning Electron Microscope (SEM) images of radiolarians to ease species identification of fossilized radiolarians. The proposed method uses both hand-crafted features like invariant moments, wavelet moments, Gabor features, basic morphological features and deep features obtained from a pre-trained Convolutional Neural Network (CNN). Feature selection is applied over deep features to reduce high dimensionality. Classification outcomes are analyzed to compare hand-crafted features, deep features, and their combinations. Results show that the deep features obtained from a pre-trained CNN are more discriminative comparing to hand-crafted ones. Additionally, feature selection utilizes to the computational cost of classification algorithms and have no negative effect on classification accuracy.

  19. Deep tissue single cell MSC ablation using a fiber laser source to evaluate therapeutic potential in osteogenesis imperfecta

    NASA Astrophysics Data System (ADS)

    Tehrani, Kayvan F.; Pendleton, Emily G.; Lin, Charles P.; Mortensen, Luke J.

    2016-04-01

    Osteogenesis imperfecta (OI) is a currently uncurable disease where a mutation in collagen type I yields brittle bones. One potential therapy is transplantation of mesenchymal stem cells (MSCs), but controlling and enhancing transplanted cell survival has proven challenging. Therefore, we use a 2- photon imaging system to study individual transplanted cells in the living bone marrow. We ablated cells deep in the bone marrow and observed minimal collateral damage to surrounding tissue. Future work will evaluate the local impact of transplanted MSCs on bone deposition in vivo.

  20. Time-reversed ultrasonically encoded (TRUE) focusing for deep-tissue optogenetic modulation

    NASA Astrophysics Data System (ADS)

    Brake, Joshua; Ruan, Haowen; Robinson, J. Elliott; Liu, Yan; Gradinaru, Viviana; Yang, Changhuei

    2018-02-01

    The problem of optical scattering was long thought to fundamentally limit the depth at which light could be focused through turbid media such as fog or biological tissue. However, recent work in the field of wavefront shaping has demonstrated that by properly shaping the input light field, light can be noninvasively focused to desired locations deep inside scattering media. This has led to the development of several new techniques which have the potential to enhance the capabilities of existing optical tools in biomedicine. Unfortunately, extending these methods to living tissue has a number of challenges related to the requirements for noninvasive guidestar operation, speed, and focusing fidelity. Of existing wavefront shaping methods, time-reversed ultrasonically encoded (TRUE) focusing is well suited for applications in living tissue since it uses ultrasound as a guidestar which enables noninvasive operation and provides compatibility with optical phase conjugation for high-speed operation. In this paper, we will discuss the results of our recent work to apply TRUE focusing for optogenetic modulation, which enables enhanced optogenetic stimulation deep in tissue with a 4-fold spatial resolution improvement in 800-micron thick acute brain slices compared to conventional focusing, and summarize future directions to further extend the impact of wavefront shaping technologies in biomedicine.

  1. A novel biomedical image indexing and retrieval system via deep preference learning.

    PubMed

    Pang, Shuchao; Orgun, Mehmet A; Yu, Zhezhou

    2018-05-01

    The traditional biomedical image retrieval methods as well as content-based image retrieval (CBIR) methods originally designed for non-biomedical images either only consider using pixel and low-level features to describe an image or use deep features to describe images but still leave a lot of room for improving both accuracy and efficiency. In this work, we propose a new approach, which exploits deep learning technology to extract the high-level and compact features from biomedical images. The deep feature extraction process leverages multiple hidden layers to capture substantial feature structures of high-resolution images and represent them at different levels of abstraction, leading to an improved performance for indexing and retrieval of biomedical images. We exploit the current popular and multi-layered deep neural networks, namely, stacked denoising autoencoders (SDAE) and convolutional neural networks (CNN) to represent the discriminative features of biomedical images by transferring the feature representations and parameters of pre-trained deep neural networks from another domain. Moreover, in order to index all the images for finding the similarly referenced images, we also introduce preference learning technology to train and learn a kind of a preference model for the query image, which can output the similarity ranking list of images from a biomedical image database. To the best of our knowledge, this paper introduces preference learning technology for the first time into biomedical image retrieval. We evaluate the performance of two powerful algorithms based on our proposed system and compare them with those of popular biomedical image indexing approaches and existing regular image retrieval methods with detailed experiments over several well-known public biomedical image databases. Based on different criteria for the evaluation of retrieval performance, experimental results demonstrate that our proposed algorithms outperform the state

  2. Deblurring adaptive optics retinal images using deep convolutional neural networks.

    PubMed

    Fei, Xiao; Zhao, Junlei; Zhao, Haoxin; Yun, Dai; Zhang, Yudong

    2017-12-01

    The adaptive optics (AO) can be used to compensate for ocular aberrations to achieve near diffraction limited high-resolution retinal images. However, many factors such as the limited aberration measurement and correction accuracy with AO, intraocular scatter, imaging noise and so on will degrade the quality of retinal images. Image post processing is an indispensable and economical method to make up for the limitation of AO retinal imaging procedure. In this paper, we proposed a deep learning method to restore the degraded retinal images for the first time. The method directly learned an end-to-end mapping between the blurred and restored retinal images. The mapping was represented as a deep convolutional neural network that was trained to output high-quality images directly from blurry inputs without any preprocessing. This network was validated on synthetically generated retinal images as well as real AO retinal images. The assessment of the restored retinal images demonstrated that the image quality had been significantly improved.

  3. Deblurring adaptive optics retinal images using deep convolutional neural networks

    PubMed Central

    Fei, Xiao; Zhao, Junlei; Zhao, Haoxin; Yun, Dai; Zhang, Yudong

    2017-01-01

    The adaptive optics (AO) can be used to compensate for ocular aberrations to achieve near diffraction limited high-resolution retinal images. However, many factors such as the limited aberration measurement and correction accuracy with AO, intraocular scatter, imaging noise and so on will degrade the quality of retinal images. Image post processing is an indispensable and economical method to make up for the limitation of AO retinal imaging procedure. In this paper, we proposed a deep learning method to restore the degraded retinal images for the first time. The method directly learned an end-to-end mapping between the blurred and restored retinal images. The mapping was represented as a deep convolutional neural network that was trained to output high-quality images directly from blurry inputs without any preprocessing. This network was validated on synthetically generated retinal images as well as real AO retinal images. The assessment of the restored retinal images demonstrated that the image quality had been significantly improved. PMID:29296496

  4. Imaging of non-osteochondral tissues in osteoarthritis.

    PubMed

    Guermazi, A; Roemer, F W; Crema, M D; Englund, M; Hayashi, D

    2014-10-01

    The aim of this review is to describe imaging techniques for evaluation of non-osteochondral structures such as the synovium, menisci in the knee, labrum in the hip, ligaments and muscles and to review the literature from recent clinical and epidemiological studies of OA. This is a non-systematic narrative review of published literature on imaging of non-osteochondral tissues in OA. PubMed and MEDLINE search for articles published up to 2014, using the keywords osteoarthritis, synovitis, meniscus, labrum, ligaments, plica, muscles, magnetic resonance imaging (MRI), ultrasound, computed tomography (CT), scintigraphy, and positron emission tomography (PET). Published literature showed imaging of non-osteochondral tissues in OA relies primarily on MRI and ultrasound. The use of semiquantitative and quantitative imaging biomarkers of non-osteochondral tissues in clinical and epidemiological OA studies is reported. We highlight studies that have compared both imaging methodologies directly, and those that have established a relationship between imaging biomarkers and clinical outcomes. We provide recommendations as to which imaging protocols should be used to assess disease-specific changes regarding synovium, meniscus in the knee, labrum in the hip, and ligaments, and highlight potential pitfalls in their usage. MRI and ultrasound are currently the most useful imaging modalities for evaluation of non-osteochondral tissues in OA. MRI evaluation of any tissue needs to be performed using appropriate MR pulse sequences. Ultrasound may be particularly useful for evaluation of small joints of the hand. Nuclear medicine and CT play a limited role in imaging of non-osteochondral tissues in OA. Copyright © 2014 Osteoarthritis Research Society International. Published by Elsevier Ltd. All rights reserved.

  5. Deep Learning for Lowtextured Image Matching

    NASA Astrophysics Data System (ADS)

    Kniaz, V. V.; Fedorenko, V. V.; Fomin, N. A.

    2018-05-01

    Low-textured objects pose challenges for an automatic 3D model reconstruction. Such objects are common in archeological applications of photogrammetry. Most of the common feature point descriptors fail to match local patches in featureless regions of an object. Hence, automatic documentation of the archeological process using Structure from Motion (SfM) methods is challenging. Nevertheless, such documentation is possible with the aid of a human operator. Deep learning-based descriptors have outperformed most of common feature point descriptors recently. This paper is focused on the development of a new Wide Image Zone Adaptive Robust feature Descriptor (WIZARD) based on the deep learning. We use a convolutional auto-encoder to compress discriminative features of a local path into a descriptor code. We build a codebook to perform point matching on multiple images. The matching is performed using the nearest neighbor search and a modified voting algorithm. We present a new "Multi-view Amphora" (Amphora) dataset for evaluation of point matching algorithms. The dataset includes images of an Ancient Greek vase found at Taman Peninsula in Southern Russia. The dataset provides color images, a ground truth 3D model, and a ground truth optical flow. We evaluated the WIZARD descriptor on the "Amphora" dataset to show that it outperforms the SIFT and SURF descriptors on the complex patch pairs.

  6. Effects of neuromuscular joint facilitation on bridging exercises with respect to deep muscle changes.

    PubMed

    Zhou, Bin; Huang, QiuChen; Zheng, Tao; Huo, Ming; Maruyama, Hitoshi

    2015-05-01

    [Purpose] This study examined the effects of neuromuscular joint facilitation on bridging exercises by assessing the cross-sectional area of the multifidus muscle and thickness of the musculus transversus abdominis. [Subjects] Twelve healthy men. [Methods] Four exercises were evaluated: (a) supine resting, (b) bridging resistance exercise involving posterior pelvic tilting, (c) bridging resistance exercise involving anterior pelvic tilting, and (d) bridging resistance exercise involving neuromuscular joint facilitation. The cross-sectional area of the multifidus muscle and thickness of the musculus transversus abdominis were measured during each exercise. [Results] The cross-sectional area of the multifidus muscle and thickness of the musculus transversus abdominis were significantly greater in the neuromuscular joint facilitation group than the others. [Conclusion] Neuromuscular joint facilitation intervention improves the function of deep muscles such as the multifidus muscle and musculus transversus abdominis. Therefore, it can be recommended for application in clinical treatments such as that for back pain.

  7. Incidence of deep vein thrombosis in patients undergoing breast reconstruction with autologous tissue transfer.

    PubMed

    Konoeda, Hisato; Yamaki, Takashi; Hamahata, Atsumori; Ochi, Masakazu; Osada, Atsuyoshi; Hasegawa, Yuki; Kirita, Miho; Sakurai, Hiroyuki

    2017-05-01

    Background Breast reconstruction is associated with multiple risk factors for venous thromboembolism. However, the incidence of deep vein thrombosis in patients undergoing breast reconstruction is uncertain. Objective The aim of this study was to prospectively evaluate the incidence of deep vein thrombosis in patients undergoing breast reconstruction using autologous tissue transfer and to identify potential risk factors for deep vein thrombosis. Methods Thirty-five patients undergoing breast reconstruction were enrolled. We measured patients' preoperative characteristics including age, body mass index (kg/m 2 ), and risk factors for deep vein thrombosis. The preoperative diameter of each venous segment in the deep veins was measured using duplex ultrasound. All patients received intermittent pneumatic pump and elastic compression stockings for postoperative thromboprophylaxis. Results Among the 35 patients evaluated, 11 (31.4%) were found to have deep vein thrombosis postoperatively, and one patient was found to have pulmonary embolism postoperatively. All instances of deep vein thrombosis developed in the calf and were asymptomatic. Ten of 11 patients underwent free flap transfer, and the remaining one patient received a latissimus dorsi pedicled flap. Deep vein thrombosis incidence did not significantly differ between patients with a free flap or pedicled flap (P = 0.13). Documented risk factors for deep vein thrombosis demonstrated no significant differences between patients with and without deep vein thrombosis. The diameter of the common femoral vein was significantly larger in patients who developed postoperative deep vein thrombosis than in those who did not ( P < 0.05). Conclusions The morbidity of deep vein thrombosis in patients who underwent breast reconstruction using autologous tissue transfer was relatively high. Since only the diameter of the common femoral vein was predictive of developing postoperative deep vein thrombosis, postoperative

  8. Geocoronal Imaging from the Deep Space Gateway

    NASA Astrophysics Data System (ADS)

    Waldrop, L.; Immel, T.; Clarke, J.; Fillingim, M.; Rider, K.; Qin, J.; Bhattacharyya, D.; Doe, R.

    2018-02-01

    UV imaging of geocoronal emission at high spatial and temporal resolution from deep space would provide crucial new constraints on global exospheric structure and dynamics, significantly advancing models of space weather and atmospheric escape.

  9. Computed tomography of deep fat masses in multiple symmetrical lipomatosis

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Enzi, G.; Biondetti, P.R.; Fiore, D.

    1982-07-01

    Deep fat masses were evaluated by computed tomography (CT) in 15 patients with multiple symmetrical lipomatosis. In 4 patients, peritracheal accumulations of fat were observed. In 3 of them, tracheal compression by lipomatous tissue was demonstrated: 2 were asymptomatic and the third severe respiratory insufficiency secondary to blockage of the air was by the vocal cords as the result of recurrent nerve palsy. In 6 patients, lipomatous tissue occupied the potential space between the spina scapulae and the trapezius, supraspinatus, and infraspinatus muscles. In 2, calcification of lipomatous masses was observed. There was no relationship between extension of subcutaneous fatmore » and accumulation at deep sites. CT facilitates early detection of peritracheal lipomatous tissue and is helpful in follow-up when deep fat accumulation is responsible for space-occupying lesions requiring surgery.« less

  10. Intelligent Detection of Structure from Remote Sensing Images Based on Deep Learning Method

    NASA Astrophysics Data System (ADS)

    Xin, L.

    2018-04-01

    Utilizing high-resolution remote sensing images for earth observation has become the common method of land use monitoring. It requires great human participation when dealing with traditional image interpretation, which is inefficient and difficult to guarantee the accuracy. At present, the artificial intelligent method such as deep learning has a large number of advantages in the aspect of image recognition. By means of a large amount of remote sensing image samples and deep neural network models, we can rapidly decipher the objects of interest such as buildings, etc. Whether in terms of efficiency or accuracy, deep learning method is more preponderant. This paper explains the research of deep learning method by a great mount of remote sensing image samples and verifies the feasibility of building extraction via experiments.

  11. Penetration depth of photons in biological tissues from hyperspectral imaging in shortwave infrared in transmission and reflection geometries

    PubMed Central

    Zhang, Hairong; Salo, Daniel; Kim, David M.; Komarov, Sergey; Tai, Yuan-Chuan; Berezin, Mikhail Y.

    2016-01-01

    Abstract. Measurement of photon penetration in biological tissues is a central theme in optical imaging. A great number of endogenous tissue factors such as absorption, scattering, and anisotropy affect the path of photons in tissue, making it difficult to predict the penetration depth at different wavelengths. Traditional studies evaluating photon penetration at different wavelengths are focused on tissue spectroscopy that does not take into account the heterogeneity within the sample. This is especially critical in shortwave infrared where the individual vibration-based absorption properties of the tissue molecules are affected by nearby tissue components. We have explored the depth penetration in biological tissues from 900 to 1650 nm using Monte–Carlo simulation and a hyperspectral imaging system with Michelson spatial contrast as a metric of light penetration. Chromatic aberration-free hyperspectral images in transmission and reflection geometries were collected with a spectral resolution of 5.27 nm and a total acquisition time of 3 min. Relatively short recording time minimized artifacts from sample drying. Results from both transmission and reflection geometries consistently revealed that the highest spatial contrast in the wavelength range for deep tissue lies within 1300 to 1375 nm; however, in heavily pigmented tissue such as the liver, the range 1550 to 1600 nm is also prominent. PMID:27930773

  12. Deep Marginalized Sparse Denoising Auto-Encoder for Image Denoising

    NASA Astrophysics Data System (ADS)

    Ma, Hongqiang; Ma, Shiping; Xu, Yuelei; Zhu, Mingming

    2018-01-01

    Stacked Sparse Denoising Auto-Encoder (SSDA) has been successfully applied to image denoising. As a deep network, the SSDA network with powerful data feature learning ability is superior to the traditional image denoising algorithms. However, the algorithm has high computational complexity and slow convergence rate in the training. To address this limitation, we present a method of image denoising based on Deep Marginalized Sparse Denoising Auto-Encoder (DMSDA). The loss function of Sparse Denoising Auto-Encoder is marginalized so that it satisfies both sparseness and marginality. The experimental results show that the proposed algorithm can not only outperform SSDA in the convergence speed and training time, but also has better denoising performance than the current excellent denoising algorithms, including both the subjective and objective evaluation of image denoising.

  13. Ship detection leveraging deep neural networks in WorldView-2 images

    NASA Astrophysics Data System (ADS)

    Yamamoto, T.; Kazama, Y.

    2017-10-01

    Interpretation of high-resolution satellite images has been so difficult that skilled interpreters must have checked the satellite images manually because of the following issues. One is the requirement of the high detection accuracy rate. The other is the variety of the target, taking ships for example, there are many kinds of ships, such as boat, cruise ship, cargo ship, aircraft carrier, and so on. Furthermore, there are similar appearance objects throughout the image; therefore, it is often difficult even for the skilled interpreters to distinguish what object the pixels really compose. In this paper, we explore the feasibility of object extraction leveraging deep learning with high-resolution satellite images, especially focusing on ship detection. We calculated the detection accuracy using the WorldView-2 images. First, we collected the training images labelled as "ship" and "not ship". After preparing the training data, we defined the deep neural network model to judge whether ships are existing or not, and trained them with about 50,000 training images for each label. Subsequently, we scanned the evaluation image with different resolution windows and extracted the "ship" images. Experimental result shows the effectiveness of the deep learning based object detection.

  14. Deep learning analyzes Helicobacter pylori infection by upper gastrointestinal endoscopy images.

    PubMed

    Itoh, Takumi; Kawahira, Hiroshi; Nakashima, Hirotaka; Yata, Noriko

    2018-02-01

    Helicobacter pylori (HP)-associated chronic gastritis can cause mucosal atrophy and intestinal metaplasia, both of which increase the risk of gastric cancer. The accurate diagnosis of HP infection during routine medical checks is important. We aimed to develop a convolutional neural network (CNN), which is a machine-learning algorithm similar to deep learning, capable of recognizing specific features of gastric endoscopy images. The goal behind developing such a system was to detect HP infection early, thus preventing gastric cancer.  For the development of the CNN, we used 179 upper gastrointestinal endoscopy images obtained from 139 patients (65 were HP-positive: ≥ 10 U/mL and 74 were HP-negative: < 3 U/mL on HP IgG antibody assessment). Of the 179 images, 149 were used as training images, and the remaining 30 (15 from HP-negative patients and 15 from HP-positive patients) were set aside to be used as test images. The 149 training images were subjected to data augmentation, which yielded 596 images. We used the CNN to create a learning tool that would recognize HP infection and assessed the decision accuracy of the CNN with the 30 test images by calculating the sensitivity, specificity, and area under the receiver operating characteristic (ROC) curve (AUC).  The sensitivity and specificity of the CNN for the detection of HP infection were 86.7 % and 86.7 %, respectively, and the AUC was 0.956.  CNN-aided diagnosis of HP infection seems feasible and is expected to facilitate and improve diagnosis during health check-ups.

  15. High-throughput sequencing and analysis of the gill tissue transcriptome from the deep-sea hydrothermal vent mussel Bathymodiolus azoricus

    PubMed Central

    2010-01-01

    have established the first tissue transcriptional analysis of a deep-sea hydrothermal vent animal and generated a searchable catalog of genes that provides a direct method of identifying and retrieving vast numbers of novel coding sequences which can be applied in gene expression profiling experiments from a non-conventional model organism. This provides the most comprehensive sequence resource for identifying novel genes currently available for a deep-sea vent organism, in particular, genes putatively involved in immune and inflammatory reactions in vent mussels. The characterization of the B. azoricus transcriptome will facilitate research into biological processes underlying physiological adaptations to hydrothermal vent environments and will provide a basis for expanding our understanding of genes putatively involved in adaptations processes during post-capture long term acclimatization experiments, at "sea-level" conditions, using B. azoricus as a model organism. PMID:20937131

  16. A fiber-optic-based imaging system for nondestructive assessment of cell-seeded tissue-engineered scaffolds.

    PubMed

    Hofmann, Matthias C; Whited, Bryce M; Criswell, Tracy; Rylander, Marissa Nichole; Rylander, Christopher G; Soker, Shay; Wang, Ge; Xu, Yong

    2012-09-01

    A major limitation in tissue engineering is the lack of nondestructive methods that assess the development of tissue scaffolds undergoing preconditioning in bioreactors. Due to significant optical scattering in most scaffolding materials, current microscope-based imaging methods cannot "see" through thick and optically opaque tissue constructs. To address this deficiency, we developed a fiber-optic-based imaging method that is capable of nondestructive imaging of fluorescently labeled cells through a thick and optically opaque scaffold, contained in a bioreactor. This imaging modality is based on the local excitation of fluorescent cells, the acquisition of fluorescence through the scaffold, and fluorescence mapping based on the position of the excitation light. To evaluate the capability and accuracy of the imaging system, human endothelial cells (ECs), stably expressing green fluorescent protein (GFP), were imaged through a fibrous scaffold. Without sacrificing the scaffolds, we nondestructively visualized the distribution of GFP-labeled cells through a ~500 μm thick scaffold with cell-level resolution and distinct localization. These results were similar to control images obtained using an optical microscope with direct line-of-sight access. Through a detailed quantitative analysis, we demonstrated that this method achieved a resolution on the order of 20-30 μm, with 10% or less deviation from standard optical microscopy. Furthermore, we demonstrated that the penetration depth of the imaging method exceeded that of confocal laser scanning microscopy by more than a factor of 2. Our imaging method also possesses a working distance (up to 8 cm) much longer than that of a standard confocal microscopy system, which can significantly facilitate bioreactor integration. This method will enable the nondestructive monitoring of ECs seeded on the lumen of a tissue-engineered vascular graft during preconditioning in vitro, as well as for other tissue-engineered constructs

  17. Improved Contrast-Enhanced Ultrasound Imaging With Multiplane-Wave Imaging.

    PubMed

    Gong, Ping; Song, Pengfei; Chen, Shigao

    2018-02-01

    Contrast-enhanced ultrasound (CEUS) imaging has great potential for use in new ultrasound clinical applications such as myocardial perfusion imaging and abdominal lesion characterization. In CEUS imaging, contrast agents (i.e., microbubbles) are used to improve contrast between blood and tissue because of their high nonlinearity under low ultrasound pressure. However, the quality of CEUS imaging sometimes suffers from a low signal-to-noise ratio (SNR) in deeper imaging regions when a low mechanical index (MI) is used to avoid microbubble disruption, especially for imaging at off-resonance transmit frequencies. In this paper, we propose a new strategy of combining CEUS sequences with the recently proposed multiplane-wave (MW) compounding method to improve the SNR of CEUS in deeper imaging regions without increasing MI or sacrificing frame rate. The MW-CEUS method emits multiple Hadamard-coded CEUS pulses in each transmission event (i.e., pulse-echo event). The received echo signals first undergo fundamental bandpass filtering (i.e., the filter is centered on the transmit frequency) to eliminate the microbubble's second-harmonic signals because they cannot be encoded by pulse inversion. The filtered signals are then Hadamard decoded and realigned in fast time to recover the signals as they would have been obtained using classic CEUS pulses, followed by designed recombination to cancel the linear tissue responses. The MW-CEUS method significantly improved contrast-to-tissue ratio and SNR of CEUS imaging by transmitting longer coded pulses. The image resolution was also preserved. The microbubble disruption ratio and motion artifacts in MW-CEUS were similar to those of classic CEUS imaging. In addition, the MW-CEUS sequence can be adapted to other transmission coding formats. These properties of MW-CEUS can potentially facilitate CEUS imaging for many clinical applications, especially assessing deep abdominal organs or the heart.

  18. Deep learning methods to guide CT image reconstruction and reduce metal artifacts

    NASA Astrophysics Data System (ADS)

    Gjesteby, Lars; Yang, Qingsong; Xi, Yan; Zhou, Ye; Zhang, Junping; Wang, Ge

    2017-03-01

    The rapidly-rising field of machine learning, including deep learning, has inspired applications across many disciplines. In medical imaging, deep learning has been primarily used for image processing and analysis. In this paper, we integrate a convolutional neural network (CNN) into the computed tomography (CT) image reconstruction process. Our first task is to monitor the quality of CT images during iterative reconstruction and decide when to stop the process according to an intelligent numerical observer instead of using a traditional stopping rule, such as a fixed error threshold or a maximum number of iterations. After training on ground truth images, the CNN was successful in guiding an iterative reconstruction process to yield high-quality images. Our second task is to improve a sinogram to correct for artifacts caused by metal objects. A large number of interpolation and normalization-based schemes were introduced for metal artifact reduction (MAR) over the past four decades. The NMAR algorithm is considered a state-of-the-art method, although residual errors often remain in the reconstructed images, especially in cases of multiple metal objects. Here we merge NMAR with deep learning in the projection domain to achieve additional correction in critical image regions. Our results indicate that deep learning can be a viable tool to address CT reconstruction challenges.

  19. Coherence-Gated Sensorless Adaptive Optics Multiphoton Retinal Imaging

    PubMed Central

    Cua, Michelle; Wahl, Daniel J.; Zhao, Yuan; Lee, Sujin; Bonora, Stefano; Zawadzki, Robert J.; Jian, Yifan; Sarunic, Marinko V.

    2016-01-01

    Multiphoton microscopy enables imaging deep into scattering tissues. The efficient generation of non-linear optical effects is related to both the pulse duration (typically on the order of femtoseconds) and the size of the focused spot. Aberrations introduced by refractive index inhomogeneity in the sample distort the wavefront and enlarge the focal spot, which reduces the multiphoton signal. Traditional approaches to adaptive optics wavefront correction are not effective in thick or multi-layered scattering media. In this report, we present sensorless adaptive optics (SAO) using low-coherence interferometric detection of the excitation light for depth-resolved aberration correction of two-photon excited fluorescence (TPEF) in biological tissue. We demonstrate coherence-gated SAO TPEF using a transmissive multi-actuator adaptive lens for in vivo imaging in a mouse retina. This configuration has significant potential for reducing the laser power required for adaptive optics multiphoton imaging, and for facilitating integration with existing systems. PMID:27599635

  20. Multimodal imaging of cutaneous wound tissue

    NASA Astrophysics Data System (ADS)

    Zhang, Shiwu; Gnyawali, Surya; Huang, Jiwei; Ren, Wenqi; Gordillo, Gayle; Sen, Chandan K.; Xu, Ronald

    2015-01-01

    Quantitative assessment of wound tissue ischemia, perfusion, and inflammation provides critical information for appropriate detection, staging, and treatment of chronic wounds. However, few methods are available for simultaneous assessment of these tissue parameters in a noninvasive and quantitative fashion. We integrated hyperspectral, laser speckle, and thermographic imaging modalities in a single-experimental setup for multimodal assessment of tissue oxygenation, perfusion, and inflammation characteristics. Algorithms were developed for appropriate coregistration between wound images acquired by different imaging modalities at different times. The multimodal wound imaging system was validated in an occlusion experiment, where oxygenation and perfusion maps of a healthy subject's upper extremity were continuously monitored during a postocclusive reactive hyperemia procedure and compared with standard measurements. The system was also tested in a clinical trial where a wound of three millimeters in diameter was introduced on a healthy subject's lower extremity and the healing process was continuously monitored. Our in vivo experiments demonstrated the clinical feasibility of multimodal cutaneous wound imaging.

  1. Deep Spitzer/IRAC Imaging of the Subaru Deep Field

    NASA Astrophysics Data System (ADS)

    Jiang, Linhua; Egami, Eiichi; Cohen, Seth; Fan, Xiaohui; Ly, Chun; Mechtley, Matthew; Windhorst, Rogier

    2013-10-01

    The last decade saw great progress in our understanding of the distant Universe as a number of objects at z > 6 were discovered. The Subaru Deep Field (SDF) project has played an important role on study of high-z galaxies. The SDF is unique: it covers a large area of 850 sq arcmin; it has extremely deep optical images in a series of broad and narrow bands; it has the largest sample of spectroscopically-confirmed galaxies known at z >= 6, including ~100 Lyman alpha emitters (LAEs) and ~50 Lyman break galaxies (LBGs). Here we propose to carry out deep IRAC imaging observations of the central 75% of the SDF. The proposed observations together with those from our previous Spitzer programs will reach a depth of ~10 hours, and enable the first complete census of physical properties and stellar populations of spectroscopically-confirmed galaxies at the end of cosmic reionization. IRAC data is the key to measure stellar masses and constrain stellar populations in high-z galaxies. From SED modeling with secure redshifts, we will characterize the physical properties of these galaxies, and trace their mass assembly and star formation history. In particular, it allows us, for the first time, to study stellar populations in a large sample of z >=6 LAEs. We will also address some critical questions, such as whether LAEs and LBGs represent physically different galaxy populations. All these will help us to understand the earliest galaxy formation and evolution, and better constrain the galaxy contribution to reionization. The IRAC data will also cover 10,000 emission-line selected galaxies at z < 1.5, 50,000 UV and mass selected LBGs at 1.5 < z < 3, and more than 5,000 LBGs at 3 < z < 6. It will have a legacy value for SDF-related programs.

  2. Detection of soft tissue densities from digital breast tomosynthesis: comparison of conventional and deep learning approaches

    NASA Astrophysics Data System (ADS)

    Fotin, Sergei V.; Yin, Yin; Haldankar, Hrishikesh; Hoffmeister, Jeffrey W.; Periaswamy, Senthil

    2016-03-01

    Computer-aided detection (CAD) has been used in screening mammography for many years and is likely to be utilized for digital breast tomosynthesis (DBT). Higher detection performance is desirable as it may have an impact on radiologist's decisions and clinical outcomes. Recently the algorithms based on deep convolutional architectures have been shown to achieve state of the art performance in object classification and detection. Similarly, we trained a deep convolutional neural network directly on patches sampled from two-dimensional mammography and reconstructed DBT volumes and compared its performance to a conventional CAD algorithm that is based on computation and classification of hand-engineered features. The detection performance was evaluated on the independent test set of 344 DBT reconstructions (GE SenoClaire 3D, iterative reconstruction algorithm) containing 328 suspicious and 115 malignant soft tissue densities including masses and architectural distortions. Detection sensitivity was measured on a region of interest (ROI) basis at the rate of five detection marks per volume. Moving from conventional to deep learning approach resulted in increase of ROI sensitivity from 0:832 +/- 0:040 to 0:893 +/- 0:033 for suspicious ROIs; and from 0:852 +/- 0:065 to 0:930 +/- 0:046 for malignant ROIs. These results indicate the high utility of deep feature learning in the analysis of DBT data and high potential of the method for broader medical image analysis tasks.

  3. Deep tissue imaging of microfracture and non-displaced fracture of bone using the second and third near-infrared therapeutic windows

    NASA Astrophysics Data System (ADS)

    Sordillo, Laura A.; Pu, Yang; Sordillo, P. P.; Budansky, Yury; Alfano, Robert R.

    2014-03-01

    Near-infrared (NIR) light in the wavelengths of 700 nm to 2,000 nm has three NIR optical, or therapeutic, windows, which allow for deeper depth penetration in scattering tissue media. Microfractures secondary to repetitive stress, particularly in the lower extremities, are an important problem for military recruits and athletes. They also frequently occur in the elderly, or in patients taking bisphosphonates or denosumab. Microfractures can be early predictors of a major bone fracture. Using the second and third NIR therapeutic windows, we investigated the results from images of chicken bone and human tibial bone with microfractures and non-displaced fractures with and without overlying tissues of various thicknesses. Images of bone with microfractures and non-displaced fractures with tissue show scattering photons in the third NIR window with wavelengths between 1,650 nm and 1,870 nm are diminished and absorption is increased slightly from and second NIR windows. Results from images of fractured bones show the attenuation length of light through tissue in the third optical window to be larger than in the second therapeutic window. Use of these windows may aid in the detection of bone microfractures, and thus reduce the incidence of major bone fracture in susceptible groups.

  4. Physics of tissue harmonic imaging by ultrasound

    NASA Astrophysics Data System (ADS)

    Jing, Yuan

    Tissue Harmonic Imaging (THI) is an imaging modality that is currently deployed on diagnostic ultrasound scanners. In THI the amplitude of the ultrasonic pulse that is used to probe the tissue is large enough that the pulse undergoes nonlinear distortion as it propagates into the tissue. One result of the distortion is that as the pulse propagates energy is shifted from the fundamental frequency of the source pulse into its higher harmonics. These harmonics will scatter off objects in the tissue and images formed from the scattered higher harmonics are considered to have superior quality to the images formed from the fundamental frequency. Processes that have been suggested as possibly responsible for the improved imaging in THI include: (1) reduced sensitivity to reverberation, (2) reduced sensitivity to aberration, and (3) reduction in side lobes. By using a combination of controlled experiments and numerical simulations, these three reasons have been investigated. A single element transducer and a clinical ultrasound scanner with a phased array transducer were used to image a commercial tissue-mimicking phantom with calibrated targets. The higher image quality achieved with THI was quantified in terms of spatial resolution and "clutter" signals. A three-dimensional model of the forward propagation of nonlinear sound beams in media with arbitrary spatial properties (a generalized KZK equation) was developed. A time-domain code for solving the KZK equation was validated with measurements of the acoustic field generated by the single element transducer and the phased array transducer. The code was used to investigate the impact of aberration using tissue-like media with three-dimensional variations in all acoustic properties. The three-dimensional maps of tissue properties were derived from the datasets available through the Visible Female project. The experiments and simulations demonstrated that second harmonic imaging (1) suffers less clutter associated with

  5. Activatable Fluorescence Probe via Self-Immolative Intramolecular Cyclization for Histone Deacetylase Imaging in Live Cells and Tissues.

    PubMed

    Liu, Xianjun; Xiang, Meihao; Tong, Zongxuan; Luo, Fengyan; Chen, Wen; Liu, Feng; Wang, Fenglin; Yu, Ru-Qin; Jiang, Jian-Hui

    2018-05-01

    Histone deacetylases (HDACs) play essential roles in transcription regulation and are valuable theranostic targets. However, there are no activatable fluorescent probes for imaging of HDAC activity in live cells. Here, we develop for the first time a novel activatable two-photon fluorescence probe that enables in situ imaging of HDAC activity in living cells and tissues. The probe is designed by conjugating an acetyl-lysine mimic substrate to a masked aldehyde-containing fluorophore via a cyanoester linker. Upon deacetylation by HDAC, the probe undergoes a rapid self-immolative intramolecular cyclization reaction, producing a cyanohydrin intermediate that is spontaneously rapidly decomposed into the highly fluorescent aldehyde-containing two-photon fluorophore. The probe is shown to exhibit high sensitivity, high specificity, and fast response for HDAC detection in vitro. Imaging studies reveal that the probe is able to directly visualize and monitor HDAC activity in living cells. Moreover, the probe is demonstrated to have the capability of two-photon imaging of HDAC activity in deep tissue slices up to 130 μm. This activatable fluorescent probe affords a useful tool for evaluating HDAC activity and screening HDAC-targeting drugs in both live cell and tissue assays.

  6. Geographical topic learning for social images with a deep neural network

    NASA Astrophysics Data System (ADS)

    Feng, Jiangfan; Xu, Xin

    2017-03-01

    The use of geographical tagging in social-media images is becoming a part of image metadata and a great interest for geographical information science. It is well recognized that geographical topic learning is crucial for geographical annotation. Existing methods usually exploit geographical characteristics using image preprocessing, pixel-based classification, and feature recognition. How to effectively exploit the high-level semantic feature and underlying correlation among different types of contents is a crucial task for geographical topic learning. Deep learning (DL) has recently demonstrated robust capabilities for image tagging and has been introduced into geoscience. It extracts high-level features computed from a whole image component, where the cluttered background may dominate spatial features in the deep representation. Therefore, a method of spatial-attentional DL for geographical topic learning is provided and we can regard it as a special case of DL combined with various deep networks and tuning tricks. Results demonstrated that the method is discriminative for different types of geographical topic learning. In addition, it outperforms other sequential processing models in a tagging task for a geographical image dataset.

  7. Phase contrast imaging of cochlear soft tissue.

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Smith, S.; Hwang, M.; Rau, C.

    A noninvasive technique to image soft tissue could expedite diagnosis and disease management in the auditory system. We propose inline phase contrast imaging with hard X-rays as a novel method that overcomes the limitations of conventional absorption radiography for imaging soft tissue. In this study, phase contrast imaging of mouse cochleae was performed at the Argonne National Laboratory Advanced Photon Source. The phase contrast tomographic reconstructions show soft tissue structures of the cochlea, including the inner pillar cells, the inner spiral sulcus, the tectorial membrane, the basilar membrane, and the Reissner's membrane. The results suggest that phase contrast X-ray imagingmore » and tomographic techniques hold promise to noninvasively image cochlear structures at an unprecedented cellular level.« less

  8. VoxResNet: Deep voxelwise residual networks for brain segmentation from 3D MR images.

    PubMed

    Chen, Hao; Dou, Qi; Yu, Lequan; Qin, Jing; Heng, Pheng-Ann

    2018-04-15

    Segmentation of key brain tissues from 3D medical images is of great significance for brain disease diagnosis, progression assessment and monitoring of neurologic conditions. While manual segmentation is time-consuming, laborious, and subjective, automated segmentation is quite challenging due to the complicated anatomical environment of brain and the large variations of brain tissues. We propose a novel voxelwise residual network (VoxResNet) with a set of effective training schemes to cope with this challenging problem. The main merit of residual learning is that it can alleviate the degradation problem when training a deep network so that the performance gains achieved by increasing the network depth can be fully leveraged. With this technique, our VoxResNet is built with 25 layers, and hence can generate more representative features to deal with the large variations of brain tissues than its rivals using hand-crafted features or shallower networks. In order to effectively train such a deep network with limited training data for brain segmentation, we seamlessly integrate multi-modality and multi-level contextual information into our network, so that the complementary information of different modalities can be harnessed and features of different scales can be exploited. Furthermore, an auto-context version of the VoxResNet is proposed by combining the low-level image appearance features, implicit shape information, and high-level context together for further improving the segmentation performance. Extensive experiments on the well-known benchmark (i.e., MRBrainS) of brain segmentation from 3D magnetic resonance (MR) images corroborated the efficacy of the proposed VoxResNet. Our method achieved the first place in the challenge out of 37 competitors including several state-of-the-art brain segmentation methods. Our method is inherently general and can be readily applied as a powerful tool to many brain-related studies, where accurate segmentation of brain

  9. Deep multi-scale convolutional neural network for hyperspectral image classification

    NASA Astrophysics Data System (ADS)

    Zhang, Feng-zhe; Yang, Xia

    2018-04-01

    In this paper, we proposed a multi-scale convolutional neural network for hyperspectral image classification task. Firstly, compared with conventional convolution, we utilize multi-scale convolutions, which possess larger respective fields, to extract spectral features of hyperspectral image. We design a deep neural network with a multi-scale convolution layer which contains 3 different convolution kernel sizes. Secondly, to avoid overfitting of deep neural network, dropout is utilized, which randomly sleeps neurons, contributing to improve the classification accuracy a bit. In addition, new skills like ReLU in deep learning is utilized in this paper. We conduct experiments on University of Pavia and Salinas datasets, and obtained better classification accuracy compared with other methods.

  10. Pitfalls of CT for deep neck abscess imaging assessment: a retrospective review of 162 cases.

    PubMed

    Chuang, S Y; Lin, H T; Wen, Y S; Hsu, F J

    2013-01-01

    To investigate the diagnostic value of contrast-enhanced computed tomography (CT) for the prediction of deep neck abscesses in different deep neck spaces and to evaluate the false-positive results. We retrospectively analysed the clinical charts, CT examinations, surgical findings, bacteriology, pathological examinations and complications of hospitalised patients with a diagnosis of deep neck abscess from 2004 to 2010. The positive predictive values (PPV) for the prediction of abscesses by CT scan in different deep neck spaces were calculated individually on the basis of surgical findings. A total of 162 patients were included in this study. All patients received both intravenous antibiotics and surgical drainage. The parapharyngeal space was the most commonly involved space. The overall PPV for the prediction of deep neck abscess with contrast-enhanced CT was 79.6%. The PPV was 91.3% when more than one deep neck space was involved but only 50.0% in patients with isolated retropharyngeal abscesses. In the false-positive group, cellulitis was the most common final result, followed by cystic degeneration of cervical metastases. Five specimens taken intra-operatively revealed malignancy and four of these were not infected. There are some limitations affecting the differentiation of abscesses and cellulitis, particularly in the retropharyngeal space. A central necrotic cervical metastatic lymph node may sometimes also mimic a simple pyogenic deep neck abscess on both clinical pictures and CT images. Routine biopsy of the tissue must be performed during surgical drainage.

  11. A Deep Convolutional Neural Network for segmenting and classifying epithelial and stromal regions in histopathological images

    PubMed Central

    Xu, Jun; Luo, Xiaofei; Wang, Guanhao; Gilmore, Hannah; Madabhushi, Anant

    2016-01-01

    Epithelial (EP) and stromal (ST) are two types of tissues in histological images. Automated segmentation or classification of EP and ST tissues is important when developing computerized system for analyzing the tumor microenvironment. In this paper, a Deep Convolutional Neural Networks (DCNN) based feature learning is presented to automatically segment or classify EP and ST regions from digitized tumor tissue microarrays (TMAs). Current approaches are based on handcraft feature representation, such as color, texture, and Local Binary Patterns (LBP) in classifying two regions. Compared to handcrafted feature based approaches, which involve task dependent representation, DCNN is an end-to-end feature extractor that may be directly learned from the raw pixel intensity value of EP and ST tissues in a data driven fashion. These high-level features contribute to the construction of a supervised classifier for discriminating the two types of tissues. In this work we compare DCNN based models with three handcraft feature extraction based approaches on two different datasets which consist of 157 Hematoxylin and Eosin (H&E) stained images of breast cancer and 1376 immunohistological (IHC) stained images of colorectal cancer, respectively. The DCNN based feature learning approach was shown to have a F1 classification score of 85%, 89%, and 100%, accuracy (ACC) of 84%, 88%, and 100%, and Matthews Correlation Coefficient (MCC) of 86%, 77%, and 100% on two H&E stained (NKI and VGH) and IHC stained data, respectively. Our DNN based approach was shown to outperform three handcraft feature extraction based approaches in terms of the classification of EP and ST regions. PMID:28154470

  12. A Deep Convolutional Neural Network for segmenting and classifying epithelial and stromal regions in histopathological images.

    PubMed

    Xu, Jun; Luo, Xiaofei; Wang, Guanhao; Gilmore, Hannah; Madabhushi, Anant

    2016-05-26

    Epithelial (EP) and stromal (ST) are two types of tissues in histological images. Automated segmentation or classification of EP and ST tissues is important when developing computerized system for analyzing the tumor microenvironment. In this paper, a Deep Convolutional Neural Networks (DCNN) based feature learning is presented to automatically segment or classify EP and ST regions from digitized tumor tissue microarrays (TMAs). Current approaches are based on handcraft feature representation, such as color, texture, and Local Binary Patterns (LBP) in classifying two regions. Compared to handcrafted feature based approaches, which involve task dependent representation, DCNN is an end-to-end feature extractor that may be directly learned from the raw pixel intensity value of EP and ST tissues in a data driven fashion. These high-level features contribute to the construction of a supervised classifier for discriminating the two types of tissues. In this work we compare DCNN based models with three handcraft feature extraction based approaches on two different datasets which consist of 157 Hematoxylin and Eosin (H&E) stained images of breast cancer and 1376 immunohistological (IHC) stained images of colorectal cancer, respectively. The DCNN based feature learning approach was shown to have a F1 classification score of 85%, 89%, and 100%, accuracy (ACC) of 84%, 88%, and 100%, and Matthews Correlation Coefficient (MCC) of 86%, 77%, and 100% on two H&E stained (NKI and VGH) and IHC stained data, respectively. Our DNN based approach was shown to outperform three handcraft feature extraction based approaches in terms of the classification of EP and ST regions.

  13. Deep learning for tumor classification in imaging mass spectrometry.

    PubMed

    Behrmann, Jens; Etmann, Christian; Boskamp, Tobias; Casadonte, Rita; Kriegsmann, Jörg; Maaß, Peter

    2018-04-01

    Tumor classification using imaging mass spectrometry (IMS) data has a high potential for future applications in pathology. Due to the complexity and size of the data, automated feature extraction and classification steps are required to fully process the data. Since mass spectra exhibit certain structural similarities to image data, deep learning may offer a promising strategy for classification of IMS data as it has been successfully applied to image classification. Methodologically, we propose an adapted architecture based on deep convolutional networks to handle the characteristics of mass spectrometry data, as well as a strategy to interpret the learned model in the spectral domain based on a sensitivity analysis. The proposed methods are evaluated on two algorithmically challenging tumor classification tasks and compared to a baseline approach. Competitiveness of the proposed methods is shown on both tasks by studying the performance via cross-validation. Moreover, the learned models are analyzed by the proposed sensitivity analysis revealing biologically plausible effects as well as confounding factors of the considered tasks. Thus, this study may serve as a starting point for further development of deep learning approaches in IMS classification tasks. https://gitlab.informatik.uni-bremen.de/digipath/Deep_Learning_for_Tumor_Classification_in_IMS. jbehrmann@uni-bremen.de or christianetmann@uni-bremen.de. Supplementary data are available at Bioinformatics online.

  14. Accelerating Spaceborne SAR Imaging Using Multiple CPU/GPU Deep Collaborative Computing.

    PubMed

    Zhang, Fan; Li, Guojun; Li, Wei; Hu, Wei; Hu, Yuxin

    2016-04-07

    With the development of synthetic aperture radar (SAR) technologies in recent years, the huge amount of remote sensing data brings challenges for real-time imaging processing. Therefore, high performance computing (HPC) methods have been presented to accelerate SAR imaging, especially the GPU based methods. In the classical GPU based imaging algorithm, GPU is employed to accelerate image processing by massive parallel computing, and CPU is only used to perform the auxiliary work such as data input/output (IO). However, the computing capability of CPU is ignored and underestimated. In this work, a new deep collaborative SAR imaging method based on multiple CPU/GPU is proposed to achieve real-time SAR imaging. Through the proposed tasks partitioning and scheduling strategy, the whole image can be generated with deep collaborative multiple CPU/GPU computing. In the part of CPU parallel imaging, the advanced vector extension (AVX) method is firstly introduced into the multi-core CPU parallel method for higher efficiency. As for the GPU parallel imaging, not only the bottlenecks of memory limitation and frequent data transferring are broken, but also kinds of optimized strategies are applied, such as streaming, parallel pipeline and so on. Experimental results demonstrate that the deep CPU/GPU collaborative imaging method enhances the efficiency of SAR imaging on single-core CPU by 270 times and realizes the real-time imaging in that the imaging rate outperforms the raw data generation rate.

  15. Accelerating Spaceborne SAR Imaging Using Multiple CPU/GPU Deep Collaborative Computing

    PubMed Central

    Zhang, Fan; Li, Guojun; Li, Wei; Hu, Wei; Hu, Yuxin

    2016-01-01

    With the development of synthetic aperture radar (SAR) technologies in recent years, the huge amount of remote sensing data brings challenges for real-time imaging processing. Therefore, high performance computing (HPC) methods have been presented to accelerate SAR imaging, especially the GPU based methods. In the classical GPU based imaging algorithm, GPU is employed to accelerate image processing by massive parallel computing, and CPU is only used to perform the auxiliary work such as data input/output (IO). However, the computing capability of CPU is ignored and underestimated. In this work, a new deep collaborative SAR imaging method based on multiple CPU/GPU is proposed to achieve real-time SAR imaging. Through the proposed tasks partitioning and scheduling strategy, the whole image can be generated with deep collaborative multiple CPU/GPU computing. In the part of CPU parallel imaging, the advanced vector extension (AVX) method is firstly introduced into the multi-core CPU parallel method for higher efficiency. As for the GPU parallel imaging, not only the bottlenecks of memory limitation and frequent data transferring are broken, but also kinds of optimized strategies are applied, such as streaming, parallel pipeline and so on. Experimental results demonstrate that the deep CPU/GPU collaborative imaging method enhances the efficiency of SAR imaging on single-core CPU by 270 times and realizes the real-time imaging in that the imaging rate outperforms the raw data generation rate. PMID:27070606

  16. Comparison of mechanisms involved in image enhancement of Tissue Harmonic Imaging

    NASA Astrophysics Data System (ADS)

    Cleveland, Robin O.; Jing, Yuan

    2006-05-01

    Processes that have been suggested as responsible for the improved imaging in Tissue Harmonic Imaging (THI) include: 1) reduced sensitivity to reverberation, 2) reduced sensitivity to aberration, and 3) reduction in the amplitude of diffraction side lobes. A three-dimensional model of the forward propagation of nonlinear sound beams in media with arbitrary spatial properties (a generalized KZK equation) was developed and solved using a time-domain code. The numerical simulations were validated through experiments with tissue mimicking phantoms. The impact of aberration from tissue-like media was determined through simulations using three-dimensional maps of tissue properties derived from datasets available through the Visible Female Project. The experiments and simulations demonstrated that second harmonic imaging suffers less clutter from reverberation and side-lobes but is not immune to aberration effects. The results indicate that side lobe suppression is the most significant reason for the improvement of second harmonic imaging.

  17. Imaging the hard/soft tissue interface.

    PubMed

    Bannerman, Alistair; Paxton, Jennifer Z; Grover, Liam M

    2014-03-01

    Interfaces between different tissues play an essential role in the biomechanics of native tissues and their recapitulation is now recognized as critical to function. As a consequence, imaging the hard/soft tissue interface has become increasingly important in the area of tissue engineering. Particularly as several biotechnology based products have made it onto the market or are close to human trials and an understanding of their function and development is essential. A range of imaging modalities have been developed that allow a wealth of information on the morphological and physical properties of samples to be obtained non-destructively in vivo or via destructive means. This review summarizes the use of a selection of imaging modalities on interfaces to date considering the strengths and weaknesses of each. We will also consider techniques which have not yet been utilized to their full potential or are likely to play a role in future work in the area.

  18. Earth-from-Luna Limb Imager (ELLI) for Deep Space Gateway

    NASA Astrophysics Data System (ADS)

    Gorkavyi, N.; DeLand, M.

    2018-02-01

    The new type of limb imager with a high-frequency imaging proposed for Deep Space Gateway. Each day this CubeSat' scale imager will generate the global 3D model of the aerosol component of the Earth's atmosphere and Polar Mesospheric Clouds.

  19. Application of Deep Learning in Automated Analysis of Molecular Images in Cancer: A Survey

    PubMed Central

    Xue, Yong; Chen, Shihui; Liu, Yong

    2017-01-01

    Molecular imaging enables the visualization and quantitative analysis of the alterations of biological procedures at molecular and/or cellular level, which is of great significance for early detection of cancer. In recent years, deep leaning has been widely used in medical imaging analysis, as it overcomes the limitations of visual assessment and traditional machine learning techniques by extracting hierarchical features with powerful representation capability. Research on cancer molecular images using deep learning techniques is also increasing dynamically. Hence, in this paper, we review the applications of deep learning in molecular imaging in terms of tumor lesion segmentation, tumor classification, and survival prediction. We also outline some future directions in which researchers may develop more powerful deep learning models for better performance in the applications in cancer molecular imaging. PMID:29114182

  20. ST Spot Detector: a web-based application for automatic spot and tissue detection for spatial Transcriptomics image datasets.

    PubMed

    Wong, Kim; Navarro, José Fernández; Bergenstråhle, Ludvig; Ståhl, Patrik L; Lundeberg, Joakim

    2018-06-01

    Spatial Transcriptomics (ST) is a method which combines high resolution tissue imaging with high troughput transcriptome sequencing data. This data must be aligned with the images for correct visualization, a process that involves several manual steps. Here we present ST Spot Detector, a web tool that automates and facilitates this alignment through a user friendly interface. jose.fernandez.navarro@scilifelab.se. Supplementary data are available at Bioinformatics online.

  1. Correlative Imaging of Fluorescent Proteins in Resin-Embedded Plant Material1

    PubMed Central

    Bell, Karen; Mitchell, Steve; Paultre, Danae; Posch, Markus; Oparka, Karl

    2013-01-01

    Fluorescent proteins (FPs) were developed for live-cell imaging and have revolutionized cell biology. However, not all plant tissues are accessible to live imaging using confocal microscopy, necessitating alternative approaches for protein localization. An example is the phloem, a tissue embedded deep within plant organs and sensitive to damage. To facilitate accurate localization of FPs within recalcitrant tissues, we developed a simple method for retaining FPs after resin embedding. This method is based on low-temperature fixation and dehydration, followed by embedding in London Resin White, and avoids the need for cryosections. We show that a palette of FPs can be localized in plant tissues while retaining good structural cell preservation, and that the polymerized block face can be counterstained with cell wall probes. Using this method we have been able to image green fluorescent protein-labeled plasmodesmata to a depth of more than 40 μm beneath the resin surface. Using correlative light and electron microscopy of the phloem, we were able to locate the same FP-labeled sieve elements in semithin and ultrathin sections. Sections were amenable to antibody labeling, and allowed a combination of confocal and superresolution imaging (three-dimensional-structured illumination microscopy) on the same cells. These correlative imaging methods should find several uses in plant cell biology. PMID:23457228

  2. Bioluminescence-Activated Deep-Tissue Photodynamic Therapy of Cancer

    PubMed Central

    Kim, Yi Rang; Kim, Seonghoon; Choi, Jin Woo; Choi, Sung Yong; Lee, Sang-Hee; Kim, Homin; Hahn, Sei Kwang; Koh, Gou Young; Yun, Seok Hyun

    2015-01-01

    Optical energy can trigger a variety of photochemical processes useful for therapies. Owing to the shallow penetration of light in tissues, however, the clinical applications of light-activated therapies have been limited. Bioluminescence resonant energy transfer (BRET) may provide a new way of inducing photochemical activation. Here, we show that efficient bioluminescence energy-induced photodynamic therapy (PDT) of macroscopic tumors and metastases in deep tissue. For monolayer cell culture in vitro incubated with Chlorin e6, BRET energy of about 1 nJ per cell generated as strong cytotoxicity as red laser light irradiation at 2.2 mW/cm2 for 180 s. Regional delivery of bioluminescence agents via draining lymphatic vessels killed tumor cells spread to the sentinel and secondary lymph nodes, reduced distant metastases in the lung and improved animal survival. Our results show the promising potential of novel bioluminescence-activated PDT. PMID:26000054

  3. Optical metabolic imaging of live tissue cultures

    NASA Astrophysics Data System (ADS)

    Walsh, Alex J.; Cook, Rebecca S.; Arteaga, Carlos L.; Skala, Melissa C.

    2013-02-01

    The fluorescence properties, both intensity and fluorescence lifetime, of NADH and FAD, two coenzymes of metabolism, are sensitive, high resolution measures of cellular metabolism. However, often in vivo measurements of tissue are not feasible. In this study, we investigate the stability over time of two-photon auto-fluorescence imaging of NADH and FAD in live-cultured tissues. Our results demonstrate that cultured tissues remain viable for at least several days post excision. Furthermore, the optical redox ratio, NADH fluorescence lifetime, and FAD fluorescence lifetime do not significantly change in the cultured tissues over time. With these findings, we demonstrate the potential of sustained tissue culture techniques for optical metabolic imaging.

  4. Monitoring of tissue modification with optical coherence tomography

    NASA Astrophysics Data System (ADS)

    Zhang, Wei; Luo, Qingming; Yao, Lei; Cheng, Haiying; Zeng, Shaoqun

    2002-04-01

    An experimental monitoring of tissue modification of in vitro and in vivo rabbit dura mater with administration of osmotical agents, 40% glucose solution and glycerol, using optical coherence tomography was presented. The preliminary results of experimental study of influence of osmotical liquids (glucose solutions, glycerol) of rabbit dura mater were reported. The significant decreasing of the light from surface and increasing of the light from the deep of dura mater under action of osmotical solutions and the increasing of OCT imaging depth were demonstrated. Experiments showed that administration of osmolytes to dura mater allowed for effective and temporary control of its optical characteristics, which made dura mater more transparent, increased the ability of light penetrating the tissue, and consequently improved the optical imaging depth. It is a significant study, which can improve penetration of optical imaging of cerebral function and acquire more information of the deep brain tissue.

  5. Deformation and reperfusion damages and their accumulation in subcutaneous tissues during loading and unloading: a theoretical modeling of deep tissue injuries.

    PubMed

    Mak, Arthur F T; Yu, Yanyan; Kwan, Linda P C; Sun, Lei; Tam, Eric W C

    2011-11-21

    Deep tissue injuries (DTI) involve damages in the subcutaneous tissues under intact skin incurred by prolonged excessive epidermal loadings. This paper presents a new theoretical model for the development of DTI, broadly based on the experimental evidence in the literatures. The model covers the loading damages implicitly inclusive of both the direct mechanical and ischemic injuries, and the additional reperfusion damages and the competing healing processes during the unloading phase. Given the damage accumulated at the end of the loading period, the relative strength of the reperfusion and the healing capacity of the involved tissues system, the model provides a description of the subsequent damage evolution during unloading. The model is used to study parametrically the scenario when reperfusion damage dominates over healing upon unloading and the opposite scenario when the loading and subsequent reperfusion damages remain small relative to the healing capacity of the tissues system. The theoretical model provides an integrated understanding of how tissue damage may further build-up paradoxically even with unloading, how long it would take for the loading and reperfusion damages in the tissues to become fully recovered, and how such loading and reperfusion damages, if not given sufficient time for recovery, may accumulate over multiple loading and unloading cycles, leading to clinical deep tissues ulceration. Copyright © 2011 Elsevier Ltd. All rights reserved.

  6. Three-dimensional fuse deposition modeling of tissue-simulating phantom for biomedical optical imaging

    NASA Astrophysics Data System (ADS)

    Dong, Erbao; Zhao, Zuhua; Wang, Minjie; Xie, Yanjun; Li, Shidi; Shao, Pengfei; Cheng, Liuquan; Xu, Ronald X.

    2015-12-01

    Biomedical optical devices are widely used for clinical detection of various tissue anomalies. However, optical measurements have limited accuracy and traceability, partially owing to the lack of effective calibration methods that simulate the actual tissue conditions. To facilitate standardized calibration and performance evaluation of medical optical devices, we develop a three-dimensional fuse deposition modeling (FDM) technique for freeform fabrication of tissue-simulating phantoms. The FDM system uses transparent gel wax as the base material, titanium dioxide (TiO2) powder as the scattering ingredient, and graphite powder as the absorption ingredient. The ingredients are preheated, mixed, and deposited at the designated ratios layer-by-layer to simulate tissue structural and optical heterogeneities. By printing the sections of human brain model based on magnetic resonance images, we demonstrate the capability for simulating tissue structural heterogeneities. By measuring optical properties of multilayered phantoms and comparing with numerical simulation, we demonstrate the feasibility for simulating tissue optical properties. By creating a rat head phantom with embedded vasculature, we demonstrate the potential for mimicking physiologic processes of a living system.

  7. Potential for Imaging Engineered Tissues with X-Ray Phase Contrast

    PubMed Central

    Appel, Alyssa; Anastasio, Mark A.

    2011-01-01

    As the field of tissue engineering advances, it is crucial to develop imaging methods capable of providing detailed three-dimensional information on tissue structure. X-ray imaging techniques based on phase-contrast (PC) have great potential for a number of biomedical applications due to their ability to provide information about soft tissue structure without exogenous contrast agents. X-ray PC techniques retain the excellent spatial resolution, tissue penetration, and calcified tissue contrast of conventional X-ray techniques while providing drastically improved imaging of soft tissue and biomaterials. This suggests that X-ray PC techniques are very promising for evaluation of engineered tissues. In this review, four different implementations of X-ray PC imaging are described and applications to tissues of relevance to tissue engineering reviewed. In addition, recent applications of X-ray PC to the evaluation of biomaterial scaffolds and engineered tissues are presented and areas for further development and application of these techniques are discussed. Imaging techniques based on X-ray PC have significant potential for improving our ability to image and characterize engineered tissues, and their continued development and optimization could have significant impact on the field of tissue engineering. PMID:21682604

  8. Evaluation of multimodality imaging using image fusion with ultrasound tissue elasticity imaging in an experimental animal model.

    PubMed

    Paprottka, P M; Zengel, P; Cyran, C C; Ingrisch, M; Nikolaou, K; Reiser, M F; Clevert, D A

    2014-01-01

    To evaluate the ultrasound tissue elasticity imaging by comparison to multimodality imaging using image fusion with Magnetic Resonance Imaging (MRI) and conventional grey scale imaging with additional elasticity-ultrasound in an experimental small-animal-squamous-cell carcinoma-model for the assessment of tissue morphology. Human hypopharynx carcinoma cells were subcutaneously injected into the left flank of 12 female athymic nude rats. After 10 days (SD ± 2) of subcutaneous tumor growth, sonographic grey scale including elasticity imaging and MRI measurements were performed using a high-end ultrasound system and a 3T MR. For image fusion the contrast-enhanced MRI DICOM data set was uploaded in the ultrasonic device which has a magnetic field generator, a linear array transducer (6-15 MHz) and a dedicated software package (GE Logic E9), that can detect transducers by means of a positioning system. Conventional grey scale and elasticity imaging were integrated in the image fusion examination. After successful registration and image fusion the registered MR-images were simultaneously shown with the respective ultrasound sectional plane. Data evaluation was performed using the digitally stored video sequence data sets by two experienced radiologist using a modified Tsukuba Elasticity score. The colors "red and green" are assigned for an area of soft tissue, "blue" indicates hard tissue. In all cases a successful image fusion and plan registration with MRI and ultrasound imaging including grey scale and elasticity imaging was possible. The mean tumor volume based on caliper measurements in 3 dimensions was ~323 mm3. 4/12 rats were evaluated with Score I, 5/12 rates were evaluated with Score II, 3/12 rates were evaluated with Score III. There was a close correlation in the fused MRI with existing small necrosis in the tumor. None of the scored II or III lesions was visible by conventional grey scale. The comparison of ultrasound tissue elasticity imaging enables a

  9. Deep Learning for Image-Based Cassava Disease Detection.

    PubMed

    Ramcharan, Amanda; Baranowski, Kelsee; McCloskey, Peter; Ahmed, Babuali; Legg, James; Hughes, David P

    2017-01-01

    Cassava is the third largest source of carbohydrates for human food in the world but is vulnerable to virus diseases, which threaten to destabilize food security in sub-Saharan Africa. Novel methods of cassava disease detection are needed to support improved control which will prevent this crisis. Image recognition offers both a cost effective and scalable technology for disease detection. New deep learning models offer an avenue for this technology to be easily deployed on mobile devices. Using a dataset of cassava disease images taken in the field in Tanzania, we applied transfer learning to train a deep convolutional neural network to identify three diseases and two types of pest damage (or lack thereof). The best trained model accuracies were 98% for brown leaf spot (BLS), 96% for red mite damage (RMD), 95% for green mite damage (GMD), 98% for cassava brown streak disease (CBSD), and 96% for cassava mosaic disease (CMD). The best model achieved an overall accuracy of 93% for data not used in the training process. Our results show that the transfer learning approach for image recognition of field images offers a fast, affordable, and easily deployable strategy for digital plant disease detection.

  10. Sensitive red protein calcium indicators for imaging neural activity

    PubMed Central

    Dana, Hod; Mohar, Boaz; Sun, Yi; Narayan, Sujatha; Gordus, Andrew; Hasseman, Jeremy P; Tsegaye, Getahun; Holt, Graham T; Hu, Amy; Walpita, Deepika; Patel, Ronak; Macklin, John J; Bargmann, Cornelia I; Ahrens, Misha B; Schreiter, Eric R; Jayaraman, Vivek; Looger, Loren L; Svoboda, Karel; Kim, Douglas S

    2016-01-01

    Genetically encoded calcium indicators (GECIs) allow measurement of activity in large populations of neurons and in small neuronal compartments, over times of milliseconds to months. Although GFP-based GECIs are widely used for in vivo neurophysiology, GECIs with red-shifted excitation and emission spectra have advantages for in vivo imaging because of reduced scattering and absorption in tissue, and a consequent reduction in phototoxicity. However, current red GECIs are inferior to the state-of-the-art GFP-based GCaMP6 indicators for detecting and quantifying neural activity. Here we present improved red GECIs based on mRuby (jRCaMP1a, b) and mApple (jRGECO1a), with sensitivity comparable to GCaMP6. We characterized the performance of the new red GECIs in cultured neurons and in mouse, Drosophila, zebrafish and C. elegans in vivo. Red GECIs facilitate deep-tissue imaging, dual-color imaging together with GFP-based reporters, and the use of optogenetics in combination with calcium imaging. DOI: http://dx.doi.org/10.7554/eLife.12727.001 PMID:27011354

  11. Deep Learning in Gastrointestinal Endoscopy.

    PubMed

    Patel, Vivek; Armstrong, David; Ganguli, Malika; Roopra, Sandeep; Kantipudi, Neha; Albashir, Siwar; Kamath, Markad V

    2016-01-01

    Gastrointestinal (GI) endoscopy is used to inspect the lumen or interior of the GI tract for several purposes, including, (1) making a clinical diagnosis, in real time, based on the visual appearances; (2) taking targeted tissue samples for subsequent histopathological examination; and (3) in some cases, performing therapeutic interventions targeted at specific lesions. GI endoscopy is therefore predicated on the assumption that the operator-the endoscopist-is able to identify and characterize abnormalities or lesions accurately and reproducibly. However, as in other areas of clinical medicine, such as histopathology and radiology, many studies have documented marked interobserver and intraobserver variability in lesion recognition. Thus, there is a clear need and opportunity for techniques or methodologies that will enhance the quality of lesion recognition and diagnosis and improve the outcomes of GI endoscopy. Deep learning models provide a basis to make better clinical decisions in medical image analysis. Biomedical image segmentation, classification, and registration can be improved with deep learning. Recent evidence suggests that the application of deep learning methods to medical image analysis can contribute significantly to computer-aided diagnosis. Deep learning models are usually considered to be more flexible and provide reliable solutions for image analysis problems compared to conventional computer vision models. The use of fast computers offers the possibility of real-time support that is important for endoscopic diagnosis, which has to be made in real time. Advanced graphics processing units and cloud computing have also favored the use of machine learning, and more particularly, deep learning for patient care. This paper reviews the rapidly evolving literature on the feasibility of applying deep learning algorithms to endoscopic imaging.

  12. Deep learning for medical image segmentation - using the IBM TrueNorth neurosynaptic system

    NASA Astrophysics Data System (ADS)

    Moran, Steven; Gaonkar, Bilwaj; Whitehead, William; Wolk, Aidan; Macyszyn, Luke; Iyer, Subramanian S.

    2018-03-01

    Deep convolutional neural networks have found success in semantic image segmentation tasks in computer vision and medical imaging. These algorithms are executed on conventional von Neumann processor architectures or GPUs. This is suboptimal. Neuromorphic processors that replicate the structure of the brain are better-suited to train and execute deep learning models for image segmentation by relying on massively-parallel processing. However, given that they closely emulate the human brain, on-chip hardware and digital memory limitations also constrain them. Adapting deep learning models to execute image segmentation tasks on such chips, requires specialized training and validation. In this work, we demonstrate for the first-time, spinal image segmentation performed using a deep learning network implemented on neuromorphic hardware of the IBM TrueNorth Neurosynaptic System and validate the performance of our network by comparing it to human-generated segmentations of spinal vertebrae and disks. To achieve this on neuromorphic hardware, the training model constrains the coefficients of individual neurons to {-1,0,1} using the Energy Efficient Deep Neuromorphic (EEDN)1 networks training algorithm. Given the 1 million neurons and 256 million synapses, the scale and size of the neural network implemented by the IBM TrueNorth allows us to execute the requisite mapping between segmented images and non-uniform intensity MR images >20 times faster than on a GPU-accelerated network and using <0.1 W. This speed and efficiency implies that a trained neuromorphic chip can be deployed in intra-operative environments where real-time medical image segmentation is necessary.

  13. Imaging challenges in biomaterials and tissue engineering

    PubMed Central

    Appel, Alyssa A.; Anastasio, Mark A.; Larson, Jeffery C.; Brey, Eric M.

    2013-01-01

    Biomaterials are employed in the fields of tissue engineering and regenerative medicine (TERM) in order to enhance the regeneration or replacement of tissue function and/or structure. The unique environments resulting from the presence of biomaterials, cells, and tissues result in distinct challenges in regards to monitoring and assessing the results of these interventions. Imaging technologies for three-dimensional (3D) analysis have been identified as a strategic priority in TERM research. Traditionally, histological and immunohistochemical techniques have been used to evaluate engineered tissues. However, these methods do not allow for an accurate volume assessment, are invasive, and do not provide information on functional status. Imaging techniques are needed that enable non-destructive, longitudinal, quantitative, and three-dimensional analysis of TERM strategies. This review focuses on evaluating the application of available imaging modalities for assessment of biomaterials and tissue in TERM applications. Included is a discussion of limitations of these techniques and identification of areas for further development. PMID:23768903

  14. Particle velocity measurements with macroscopic fluorescence imaging in lymph tissue mimicking microfluidic phantoms

    NASA Astrophysics Data System (ADS)

    Hennessy, Ricky; Koo, Chiwan; Ton, Phuc; Han, Arum; Righetti, Raffaella; Maitland, Kristen C.

    2011-03-01

    Ultrasound poroelastography can quantify structural and mechanical properties of tissues such as stiffness, compressibility, and fluid flow rate. This novel ultrasound technique is being explored to detect tissue changes associated with lymphatic disease. We have constructed a macroscopic fluorescence imaging system to validate ultrasonic fluid flow measurements and to provide high resolution imaging of microfluidic phantoms. The optical imaging system is composed of a white light source, excitation and emission filters, and a camera with a zoom lens. The field of view can be adjusted from 100 mm x 75 mm to 10 mm x 7.5 mm. The microfluidic device is made of polydimethylsiloxane (PDMS) and has 9 channels, each 40 μm deep with widths ranging from 30 μm to 200 μm. A syringe pump was used to propel water containing 15 μm diameter fluorescent microspheres through the microchannels, with flow rates ranging from 0.5 μl/min to 10 μl/min. Video was captured at a rate of 25 frames/sec. The velocity of the microspheres in the microchannels was calculated using an algorithm that tracked the movement of the fluorescent microspheres. The imaging system was able to measure particle velocities ranging from 0.2 mm/sec to 10 mm/sec. The range of flow velocities of interest in lymph vessels is between 1 mm/sec to 10 mm/sec; therefore our imaging system is sufficient to measure particle velocity in phantoms modeling lymphatic flow.

  15. Deep image mining for diabetic retinopathy screening.

    PubMed

    Quellec, Gwenolé; Charrière, Katia; Boudi, Yassine; Cochener, Béatrice; Lamard, Mathieu

    2017-07-01

    Deep learning is quickly becoming the leading methodology for medical image analysis. Given a large medical archive, where each image is associated with a diagnosis, efficient pathology detectors or classifiers can be trained with virtually no expert knowledge about the target pathologies. However, deep learning algorithms, including the popular ConvNets, are black boxes: little is known about the local patterns analyzed by ConvNets to make a decision at the image level. A solution is proposed in this paper to create heatmaps showing which pixels in images play a role in the image-level predictions. In other words, a ConvNet trained for image-level classification can be used to detect lesions as well. A generalization of the backpropagation method is proposed in order to train ConvNets that produce high-quality heatmaps. The proposed solution is applied to diabetic retinopathy (DR) screening in a dataset of almost 90,000 fundus photographs from the 2015 Kaggle Diabetic Retinopathy competition and a private dataset of almost 110,000 photographs (e-ophtha). For the task of detecting referable DR, very good detection performance was achieved: A z =0.954 in Kaggle's dataset and A z =0.949 in e-ophtha. Performance was also evaluated at the image level and at the lesion level in the DiaretDB1 dataset, where four types of lesions are manually segmented: microaneurysms, hemorrhages, exudates and cotton-wool spots. For the task of detecting images containing these four lesion types, the proposed detector, which was trained to detect referable DR, outperforms recent algorithms trained to detect those lesions specifically, with pixel-level supervision. At the lesion level, the proposed detector outperforms heatmap generation algorithms for ConvNets. This detector is part of the Messidor® system for mobile eye pathology screening. Because it does not rely on expert knowledge or manual segmentation for detecting relevant patterns, the proposed solution is a promising image

  16. Scalable High Performance Image Registration Framework by Unsupervised Deep Feature Representations Learning

    PubMed Central

    Wu, Guorong; Kim, Minjeong; Wang, Qian; Munsell, Brent C.

    2015-01-01

    Feature selection is a critical step in deformable image registration. In particular, selecting the most discriminative features that accurately and concisely describe complex morphological patterns in image patches improves correspondence detection, which in turn improves image registration accuracy. Furthermore, since more and more imaging modalities are being invented to better identify morphological changes in medical imaging data,, the development of deformable image registration method that scales well to new image modalities or new image applications with little to no human intervention would have a significant impact on the medical image analysis community. To address these concerns, a learning-based image registration framework is proposed that uses deep learning to discover compact and highly discriminative features upon observed imaging data. Specifically, the proposed feature selection method uses a convolutional stacked auto-encoder to identify intrinsic deep feature representations in image patches. Since deep learning is an unsupervised learning method, no ground truth label knowledge is required. This makes the proposed feature selection method more flexible to new imaging modalities since feature representations can be directly learned from the observed imaging data in a very short amount of time. Using the LONI and ADNI imaging datasets, image registration performance was compared to two existing state-of-the-art deformable image registration methods that use handcrafted features. To demonstrate the scalability of the proposed image registration framework image registration experiments were conducted on 7.0-tesla brain MR images. In all experiments, the results showed the new image registration framework consistently demonstrated more accurate registration results when compared to state-of-the-art. PMID:26552069

  17. Scalable High-Performance Image Registration Framework by Unsupervised Deep Feature Representations Learning.

    PubMed

    Wu, Guorong; Kim, Minjeong; Wang, Qian; Munsell, Brent C; Shen, Dinggang

    2016-07-01

    Feature selection is a critical step in deformable image registration. In particular, selecting the most discriminative features that accurately and concisely describe complex morphological patterns in image patches improves correspondence detection, which in turn improves image registration accuracy. Furthermore, since more and more imaging modalities are being invented to better identify morphological changes in medical imaging data, the development of deformable image registration method that scales well to new image modalities or new image applications with little to no human intervention would have a significant impact on the medical image analysis community. To address these concerns, a learning-based image registration framework is proposed that uses deep learning to discover compact and highly discriminative features upon observed imaging data. Specifically, the proposed feature selection method uses a convolutional stacked autoencoder to identify intrinsic deep feature representations in image patches. Since deep learning is an unsupervised learning method, no ground truth label knowledge is required. This makes the proposed feature selection method more flexible to new imaging modalities since feature representations can be directly learned from the observed imaging data in a very short amount of time. Using the LONI and ADNI imaging datasets, image registration performance was compared to two existing state-of-the-art deformable image registration methods that use handcrafted features. To demonstrate the scalability of the proposed image registration framework, image registration experiments were conducted on 7.0-T brain MR images. In all experiments, the results showed that the new image registration framework consistently demonstrated more accurate registration results when compared to state of the art.

  18. Cascaded deep decision networks for classification of endoscopic images

    NASA Astrophysics Data System (ADS)

    Murthy, Venkatesh N.; Singh, Vivek; Sun, Shanhui; Bhattacharya, Subhabrata; Chen, Terrence; Comaniciu, Dorin

    2017-02-01

    Both traditional and wireless capsule endoscopes can generate tens of thousands of images for each patient. It is desirable to have the majority of irrelevant images filtered out by automatic algorithms during an offline review process or to have automatic indication for highly suspicious areas during an online guidance. This also applies to the newly invented endomicroscopy, where online indication of tumor classification plays a significant role. Image classification is a standard pattern recognition problem and is well studied in the literature. However, performance on the challenging endoscopic images still has room for improvement. In this paper, we present a novel Cascaded Deep Decision Network (CDDN) to improve image classification performance over standard Deep neural network based methods. During the learning phase, CDDN automatically builds a network which discards samples that are classified with high confidence scores by a previously trained network and concentrates only on the challenging samples which would be handled by the subsequent expert shallow networks. We validate CDDN using two different types of endoscopic imaging, which includes a polyp classification dataset and a tumor classification dataset. From both datasets we show that CDDN can outperform other methods by about 10%. In addition, CDDN can also be applied to other image classification problems.

  19. Applying Deep Learning in Medical Images: The Case of Bone Age Estimation.

    PubMed

    Lee, Jang Hyung; Kim, Kwang Gi

    2018-01-01

    A diagnostic need often arises to estimate bone age from X-ray images of the hand of a subject during the growth period. Together with measured physical height, such information may be used as indicators for the height growth prognosis of the subject. We present a way to apply the deep learning technique to medical image analysis using hand bone age estimation as an example. Age estimation was formulated as a regression problem with hand X-ray images as input and estimated age as output. A set of hand X-ray images was used to form a training set with which a regression model was trained. An image preprocessing procedure is described which reduces image variations across data instances that are unrelated to age-wise variation. The use of Caffe, a deep learning tool is demonstrated. A rather simple deep learning network was adopted and trained for tutorial purpose. A test set distinct from the training set was formed to assess the validity of the approach. The measured mean absolute difference value was 18.9 months, and the concordance correlation coefficient was 0.78. It is shown that the proposed deep learning-based neural network can be used to estimate a subject's age from hand X-ray images, which eliminates the need for tedious atlas look-ups in clinical environments and should improve the time and cost efficiency of the estimation process.

  20. Multiplexed aberration measurement for deep tissue imaging in vivo

    PubMed Central

    Wang, Chen; Liu, Rui; Milkie, Daniel E.; Sun, Wenzhi; Tan, Zhongchao; Kerlin, Aaron; Chen, Tsai-Wen; Kim, Douglas S.; Ji, Na

    2014-01-01

    We describe a multiplexed aberration measurement method that modulates the intensity or phase of light rays at multiple pupil segments in parallel to determine their phase gradients. Applicable to fluorescent-protein-labeled structures of arbitrary complexity, it allows us to obtain diffraction-limited resolution in various samples in vivo. For the strongly scattering mouse brain, a single aberration correction improves structural and functional imaging of fine neuronal processes over a large imaging volume. PMID:25128976

  1. Polymer dots enable deep in vivo multiphoton fluorescence imaging of cerebrovascular architecture

    NASA Astrophysics Data System (ADS)

    Hassan, Ahmed M.; Wu, Xu; Jarrett, Jeremy W.; Xu, Shihan; Miller, David R.; Yu, Jiangbo; Perillo, Evan P.; Liu, Yen-Liang; Chiu, Daniel T.; Yeh, Hsin-Chih; Dunn, Andrew K.

    2018-02-01

    Deep in vivo imaging of vasculature requires small, bright, and photostable fluorophores suitable for multiphoton microscopy (MPM). Although semiconducting polymer dots (pdots) are an emerging class of highly fluorescent contrast agents with favorable advantages for the next generation of in vivo imaging, their use for deep multiphoton imaging has never before been demonstrated. Here we characterize the multiphoton properties of three pdot variants (CNPPV, PFBT, and PFPV) and demonstrate deep imaging of cortical microvasculature in C57 mice. Specifically, we measure the two- versus three-photon power dependence of these pdots and observe a clear three-photon excitation signature at wavelengths longer than 1300 nm, and a transition from two-photon to three-photon excitation within a 1060 - 1300 nm excitation range. Furthermore, we show that pdots enable in vivo two-photon imaging of cerebrovascular architecture in mice up to 850 μm beneath the pial surface using 800 nm excitation. In contrast with traditional multiphoton probes, we also demonstrate that the broad multiphoton absorption spectrum of pdots permits imaging at longer wavelengths (λex = 1,060 and 1225 nm). These wavelengths approach an ideal biological imaging wavelength near 1,300 nm and confer compatibility with a high-power ytterbium-fiber laser and a high pulse energy optical parametric amplifier, resulting in substantial improvements in signal-to-background ratio (>3.5-fold) and greater cortical imaging depths of 900 μm and 1300 μm. Ultimately, pdots are a versatile tool for MPM due to their extraordinary brightness and broad absorption, which will undoubtedly unlock the ability to interrogate deep structures in vivo.

  2. Phase contrast imaging of buccal mucosa tissues-Feasibility study

    NASA Astrophysics Data System (ADS)

    Fatima, A.; Tripathi, S.; Shripathi, T.; Kulkarni, V. K.; Banda, N. R.; Agrawal, A. K.; Sarkar, P. S.; Kashyap, Y.; Sinha, A.

    2015-06-01

    Phase Contrast Imaging (PCI) technique has been used to interpret physical parameters obtained from the image taken on the normal buccal mucosa tissue extracted from cheek of a patient. The advantages of this method over the conventional imaging techniques are discussed. PCI technique uses the X-ray phase shift at the edges differentiated by very minute density differences and the edge enhanced high contrast images reveal details of soft tissues. The contrast in the images produced is related to changes in the X-ray refractive index of the tissues resulting in higher clarity compared with conventional absorption based X-ray imaging. The results show that this type of imaging has better ability to visualize microstructures of biological soft tissues with good contrast, which can lead to the diagnosis of lesions at an early stage of the diseases.

  3. Objective breast tissue image classification using Quantitative Transmission ultrasound tomography

    NASA Astrophysics Data System (ADS)

    Malik, Bilal; Klock, John; Wiskin, James; Lenox, Mark

    2016-12-01

    Quantitative Transmission Ultrasound (QT) is a powerful and emerging imaging paradigm which has the potential to perform true three-dimensional image reconstruction of biological tissue. Breast imaging is an important application of QT and allows non-invasive, non-ionizing imaging of whole breasts in vivo. Here, we report the first demonstration of breast tissue image classification in QT imaging. We systematically assess the ability of the QT images’ features to differentiate between normal breast tissue types. The three QT features were used in Support Vector Machines (SVM) classifiers, and classification of breast tissue as either skin, fat, glands, ducts or connective tissue was demonstrated with an overall accuracy of greater than 90%. Finally, the classifier was validated on whole breast image volumes to provide a color-coded breast tissue volume. This study serves as a first step towards a computer-aided detection/diagnosis platform for QT.

  4. Research on image retrieval using deep convolutional neural network combining L1 regularization and PRelu activation function

    NASA Astrophysics Data System (ADS)

    QingJie, Wei; WenBin, Wang

    2017-06-01

    In this paper, the image retrieval using deep convolutional neural network combined with regularization and PRelu activation function is studied, and improves image retrieval accuracy. Deep convolutional neural network can not only simulate the process of human brain to receive and transmit information, but also contains a convolution operation, which is very suitable for processing images. Using deep convolutional neural network is better than direct extraction of image visual features for image retrieval. However, the structure of deep convolutional neural network is complex, and it is easy to over-fitting and reduces the accuracy of image retrieval. In this paper, we combine L1 regularization and PRelu activation function to construct a deep convolutional neural network to prevent over-fitting of the network and improve the accuracy of image retrieval

  5. Research on simulated infrared image utility evaluation using deep representation

    NASA Astrophysics Data System (ADS)

    Zhang, Ruiheng; Mu, Chengpo; Yang, Yu; Xu, Lixin

    2018-01-01

    Infrared (IR) image simulation is an important data source for various target recognition systems. However, whether simulated IR images could be used as training data for classifiers depends on the features of fidelity and authenticity of simulated IR images. For evaluation of IR image features, a deep-representation-based algorithm is proposed. Being different from conventional methods, which usually adopt a priori knowledge or manually designed feature, the proposed method can extract essential features and quantitatively evaluate the utility of simulated IR images. First, for data preparation, we employ our IR image simulation system to generate large amounts of IR images. Then, we present the evaluation model of simulated IR image, for which an end-to-end IR feature extraction and target detection model based on deep convolutional neural network is designed. At last, the experiments illustrate that our proposed method outperforms other verification algorithms in evaluating simulated IR images. Cross-validation, variable proportion mixed data validation, and simulation process contrast experiments are carried out to evaluate the utility and objectivity of the images generated by our simulation system. The optimum mixing ratio between simulated and real data is 0.2≤γ≤0.3, which is an effective data augmentation method for real IR images.

  6. Two-dimensional and 3-D images of thick tissue using time-constrained times-of-flight and absorbance spectrophotometry

    NASA Astrophysics Data System (ADS)

    Benaron, David A.; Lennox, M.; Stevenson, David K.

    1992-05-01

    Reconstructing deep-tissue images in real time using spectrophotometric data from optically diffusing thick tissues has been problematic. Continuous wave applications (e.g., pulse oximetry, regional cerebral saturation) ignore both the multiple paths traveled by the photons through the tissue and the effects of scattering, allowing scalar measurements but only under limited conditions; interferometry works poorly in thick, highly-scattering media; frequency- modulated approaches may not allow full deconvolution of scattering and absorbance; and pulsed-light techniques allow for preservation of information regarding the multiple paths taken by light through the tissue, but reconstruction is both computation intensive and limited by the relative surface area available for detection of photons. We have developed a picosecond times-of-flight and absorbance (TOFA) optical system, time-constrained to measure only photons with a narrow range of path lengths and arriving within a narrow angel of the emitter-detector axis. The delay until arrival of the earliest arriving photons is a function of both the scattering and absorbance of the tissues in a direct line between the emitter and detector, reducing the influence of surrounding tissues. Measurement using a variety of emitter and detector locations produces spatial information which can be analyzed in a standard 2-D grid, or subject to computer reconstruction to produce tomographic images representing 3-D structure. Using such a technique, we have been able to demonstrate the principles of tc-TOFA, detect and localize diffusive and/or absorptive objects suspended in highly scattering media (such as blood admixed with yeast), and perform simple 3-D reconstructions using phantom objects. We are now attempting to obtain images in vivo. Potential future applications include use as a research tool, and as a continuous, noninvasive, nondestructive monitor in diagnostic imaging, fetal monitoring, neurologic and cardiac

  7. Improving face image extraction by using deep learning technique

    NASA Astrophysics Data System (ADS)

    Xue, Zhiyun; Antani, Sameer; Long, L. R.; Demner-Fushman, Dina; Thoma, George R.

    2016-03-01

    The National Library of Medicine (NLM) has made a collection of over a 1.2 million research articles containing 3.2 million figure images searchable using the Open-iSM multimodal (text+image) search engine. Many images are visible light photographs, some of which are images containing faces ("face images"). Some of these face images are acquired in unconstrained settings, while others are studio photos. To extract the face regions in the images, we first applied one of the most widely-used face detectors, a pre-trained Viola-Jones detector implemented in Matlab and OpenCV. The Viola-Jones detector was trained for unconstrained face image detection, but the results for the NLM database included many false positives, which resulted in a very low precision. To improve this performance, we applied a deep learning technique, which reduced the number of false positives and as a result, the detection precision was improved significantly. (For example, the classification accuracy for identifying whether the face regions output by this Viola- Jones detector are true positives or not in a test set is about 96%.) By combining these two techniques (Viola-Jones and deep learning) we were able to increase the system precision considerably, while avoiding the need to manually construct a large training set by manual delineation of the face regions.

  8. Propionibacterium acnes in shoulder surgery: true infection, contamination, or commensal of the deep tissue?

    PubMed

    Hudek, Robert; Sommer, Frank; Kerwat, Martina; Abdelkawi, Ayman F; Loos, Franziska; Gohlke, Frank

    2014-12-01

    Propionibacterium acnes has been linked to chronic infections in shoulder surgery. Whether the bacterium is a contaminant or commensal of the deep tissue is unclear. We aimed to assess P. acnes in intraoperative samples of different tissue layers in patients undergoing first-time shoulder surgery. In 118 consecutive patients (mean age, 59.2 years; 75 men, 43 women), intraoperative samples were correlated to preoperative subacromial injection, the type of surgical approach, and gender. One skin, one superficial, one deep tissue, and one test sample were cultured for each patient. The cultures were positive for P. acnes in 36.4% (n = 43) of cases. Subacromial injection was not associated with bacterial growth rates (P = .88 for P. acnes; P = .20 for bacteria other than P. acnes; P = .85 for the anterolateral approach; P = .92 for the deltopectoral approach; P = .56 for men; P = .51 for women). Skin samples were positive for P. acnes in 8.5% (n = 10), superficial samples were positive in 7.6% (n = 9), deep samples were positive in 13.6% (n = 16), and both samples (superficial and deep) were positive in 15.3% (n = 18) of cases (P < .0001). P. acnes was detected in the anterolateral approach in 27.1% (n = 32) of cases and in the deltopectoral approach in 9.3% (n = 11) of cases (P = .01; relative risk, 1.93; 95% confidence interval, 1.08-3.43). Thirty-five of the P. acnes-positive patients were men (81.4%), and 8 patients were women (18.6%; P = .001; relative risk, 2.51; 95% confidence interval, 1.28-4.90). P. acnes was detected in more than one third of patients undergoing first-time shoulder surgery. Preoperative subacromial injection was not associated with bacterial growth. P. acnes was observed more frequently in the deep tissues than in the superficial tissues. The relative risk for obtaining a positive P. acnes culture was 2-fold greater for the anterolateral approach than for the deltopectoral approach, and the risk was 2.5-fold greater

  9. Formulation design facilitates magnetic nanoparticle delivery to diseased cells and tissues

    PubMed Central

    Singh, Dhirender; McMillan, JoEllyn M; Liu, Xin-Ming; Vishwasrao, Hemant M; Kabanov, Alexander V; Sokolsky-Papkov, Marina; Gendelman, Howard E

    2015-01-01

    Magnetic nanoparticles (MNPs) accumulate at disease sites with the aid of magnetic fields; biodegradable MNPs can be designed to facilitate drug delivery, influence disease diagnostics, facilitate tissue regeneration and permit protein purification. Because of their limited toxicity, MNPs are widely used in theranostics, simultaneously facilitating diagnostics and therapeutics. To realize therapeutic end points, iron oxide nanoparticle cores (5–30 nm) are encapsulated in a biocompatible polymer shell with drug cargos. Although limited, the toxic potential of MNPs parallels magnetite composition, along with shape, size and surface chemistry. Clearance is hastened by the reticuloendothelial system. To surmount translational barriers, the crystal structure, particle surface and magnetic properties of MNPs need to be optimized. With this in mind, we provide a comprehensive evaluation of advancements in MNP synthesis, functionalization and design, with an eye towards bench-to-bedside translation. PMID:24646020

  10. Pre-operative CT angiography and three-dimensional image post processing for deep inferior epigastric perforator flap breast reconstructive surgery.

    PubMed

    Lam, D L; Mitsumori, L M; Neligan, P C; Warren, B H; Shuman, W P; Dubinsky, T J

    2012-12-01

    Autologous breast reconstructive surgery with deep inferior epigastric artery (DIEA) perforator flaps has become the mainstay for breast reconstructive surgery. CT angiography and three-dimensional image post processing can depict the number, size, course and location of the DIEA perforating arteries for the pre-operative selection of the best artery to use for the tissue flap. Knowledge of the location and selection of the optimal perforating artery shortens operative times and decreases patient morbidity.

  11. Mesoporous composite nanoparticles for dual-modality ultrasound/magnetic resonance imaging and synergistic chemo-/thermotherapy against deep tumors.

    PubMed

    Zhang, Nan; Wang, Ronghui; Hao, Junnian; Yang, Yang; Zou, Hongmi; Wang, Zhigang

    2017-01-01

    High-intensity focused ultrasound (HIFU) is a promising and noninvasive treatment for solid tumors, which has been explored for potential clinical applications. However, the clinical applications of HIFU for large and deep tumors such as hepatocellular carcinoma (HCC) are severely limited by unsatisfactory imaging guidance, long therapeutic times, and damage to normal tissue around the tumor due to the high power applied. In this study, we developed doxorubicin/perfluorohexane-encapsulated hollow mesoporous Prussian blue nanoparticles (HMPBs-DOX/PFH) as theranostic agents, which can effectively guide HIFU therapy and enhance its therapeutic effects in combination with chemotherapy, by decreasing the cavitation threshold. We investigated the effects of this agent on ultrasound and magnetic resonance imaging in vitro and in vivo. In addition, we showed a highly efficient HIFU therapeutic effect against HCC tumors, as well as controlled drug release, owing to the phase-transitional performance of the PFH. We therefore conclude that HMPB-DOX/PFH is a safe and efficient nanoplatform, which holds significant promise for cancer theranostics against deep tumors in clinical settings.

  12. Mesoporous composite nanoparticles for dual-modality ultrasound/magnetic resonance imaging and synergistic chemo-/thermotherapy against deep tumors

    PubMed Central

    Zhang, Nan; Wang, Ronghui; Hao, Junnian; Yang, Yang; Zou, Hongmi; Wang, Zhigang

    2017-01-01

    High-intensity focused ultrasound (HIFU) is a promising and noninvasive treatment for solid tumors, which has been explored for potential clinical applications. However, the clinical applications of HIFU for large and deep tumors such as hepatocellular carcinoma (HCC) are severely limited by unsatisfactory imaging guidance, long therapeutic times, and damage to normal tissue around the tumor due to the high power applied. In this study, we developed doxorubicin/perfluorohexane-encapsulated hollow mesoporous Prussian blue nanoparticles (HMPBs-DOX/PFH) as theranostic agents, which can effectively guide HIFU therapy and enhance its therapeutic effects in combination with chemotherapy, by decreasing the cavitation threshold. We investigated the effects of this agent on ultrasound and magnetic resonance imaging in vitro and in vivo. In addition, we showed a highly efficient HIFU therapeutic effect against HCC tumors, as well as controlled drug release, owing to the phase-transitional performance of the PFH. We therefore conclude that HMPB-DOX/PFH is a safe and efficient nanoplatform, which holds significant promise for cancer theranostics against deep tumors in clinical settings. PMID:29042775

  13. "Deep-media culture condition" promoted lumen formation of endothelial cells within engineered three-dimensional tissues in vitro.

    PubMed

    Sekiya, Sachiko; Shimizu, Tatsuya; Yamato, Masayuki; Okano, Teruo

    2011-03-01

    In the field of tissue engineering, the induction of microvessels into tissues is an important task because of the need to overcome diffusion limitations of oxygen and nutrients within tissues. Powerful methods to create vessels in engineered tissues are needed for creating real living tissues. In this study, we utilized three-dimensional (3D) highly cell dense tissues fabricated by cell sheet technology. The 3D tissue constructs are close to living-cell dense tissue in vivo. Additionally, creating an endothelial cell (EC) network within tissues promoted neovascularization promptly within the tissue after transplantation in vivo. Compared to the conditions in vivo, however, common in vitro cell culture conditions provide a poor environment for creating lumens within 3D tissue constructs. Therefore, for determining adequate conditions for vascularizing engineered tissue in vitro, our 3D tissue constructs were cultured under a "deep-media culture conditions." Compared to the control conditions, the morphology of ECs showed a visibly strained cytoskeleton, and the density of lumen formation within tissues increased under hydrostatic pressure conditions. Moreover, the increasing expression of vascular endothelial cadherin in the lumens suggested that the vessels were stabilized in the stimulated tissues compared with the control. These findings suggested that deep-media culture conditions improved lumen formation in engineered tissues in vitro.

  14. Tissue feature-based intra-fractional motion tracking for stereoscopic x-ray image guided radiotherapy

    NASA Astrophysics Data System (ADS)

    Xie, Yaoqin; Xing, Lei; Gu, Jia; Liu, Wu

    2013-06-01

    Real-time knowledge of tumor position during radiation therapy is essential to overcome the adverse effect of intra-fractional organ motion. The goal of this work is to develop a tumor tracking strategy by effectively utilizing the inherent image features of stereoscopic x-ray images acquired during dose delivery. In stereoscopic x-ray image guided radiation delivery, two orthogonal x-ray images are acquired either simultaneously or sequentially. The essence of markerless tumor tracking is the reliable identification of inherent points with distinct tissue features on each projection image and their association between two images. The identification of the feature points on a planar x-ray image is realized by searching for points with high intensity gradient. The feature points are associated by using the scale invariance features transform descriptor. The performance of the proposed technique is evaluated by using images of a motion phantom and four archived clinical cases acquired using either a CyberKnife equipped with a stereoscopic x-ray imaging system, or a LINAC equipped with an onboard kV imager and an electronic portal imaging device. In the phantom study, the results obtained using the proposed method agree with the measurements to within 2 mm in all three directions. In the clinical study, the mean error is 0.48 ± 0.46 mm for four patient data with 144 sequential images. In this work, a tissue feature-based tracking method for stereoscopic x-ray image guided radiation therapy is developed. The technique avoids the invasive procedure of fiducial implantation and may greatly facilitate the clinical workflow.

  15. Tissue microarrays and quantitative tissue-based image analysis as a tool for oncology biomarker and diagnostic development.

    PubMed

    Dolled-Filhart, Marisa P; Gustavson, Mark D

    2012-11-01

    Translational oncology has been improved by using tissue microarrays (TMAs), which facilitate biomarker analysis of large cohorts on a single slide. This has allowed for rapid analysis and validation of potential biomarkers for prognostic and predictive value, as well as for evaluation of biomarker prevalence. Coupled with quantitative analysis of immunohistochemical (IHC) staining, objective and standardized biomarker data from tumor samples can further advance companion diagnostic approaches for the identification of drug-responsive or resistant patient subpopulations. This review covers the advantages, disadvantages and applications of TMAs for biomarker research. Research literature and reviews of TMAs and quantitative image analysis methodology have been surveyed for this review (with an AQUA® analysis focus). Applications such as multi-marker diagnostic development and pathway-based biomarker subpopulation analyses are described. Tissue microarrays are a useful tool for biomarker analyses including prevalence surveys, disease progression assessment and addressing potential prognostic or predictive value. By combining quantitative image analysis with TMAs, analyses will be more objective and reproducible, allowing for more robust IHC-based diagnostic test development. Quantitative multi-biomarker IHC diagnostic tests that can predict drug response will allow for greater success of clinical trials for targeted therapies and provide more personalized clinical decision making.

  16. Multimodal tissue perfusion imaging using multi-spectral and thermographic imaging systems applied on clinical data

    NASA Astrophysics Data System (ADS)

    Klaessens, John H. G. M.; Nelisse, Martin; Verdaasdonk, Rudolf M.; Noordmans, Herke Jan

    2013-03-01

    Clinical interventions can cause changes in tissue perfusion, oxygenation or temperature. Real-time imaging of these phenomena could be useful for surgical strategy or understanding of physiological regulation mechanisms. Two noncontact imaging techniques were applied for imaging of large tissue areas: LED based multispectral imaging (MSI, 17 different wavelengths 370 nm-880 nm) and thermal imaging (7.5 to 13.5 μm). Oxygenation concentration changes were calculated using different analyzing methods. The advantages of these methods are presented for stationary and dynamic applications. Concentration calculations of chromophores in tissue require right choices of wavelengths The effects of different wavelength choices for hemoglobin concentration calculations were studied in laboratory conditions and consequently applied in clinical studies. Corrections for interferences during the clinical registrations (ambient light fluctuations, tissue movements) were performed. The wavelength dependency of the algorithms were studied and wavelength sets with the best results will be presented. The multispectral and thermal imaging systems were applied during clinical intervention studies: reperfusion of tissue flap transplantation (ENT), effectiveness of local anesthetic block and during open brain surgery in patients with epileptic seizures. The LED multispectral imaging system successfully imaged the perfusion and oxygenation changes during clinical interventions. The thermal images show local heat distributions over tissue areas as a result of changes in tissue perfusion. Multispectral imaging and thermal imaging provide complementary information and are promising techniques for real-time diagnostics of physiological processes in medicine.

  17. Classify epithelium-stroma in histopathological images based on deep transferable network.

    PubMed

    Yu, X; Zheng, H; Liu, C; Huang, Y; Ding, X

    2018-04-20

    Recently, the deep learning methods have received more attention in histopathological image analysis. However, the traditional deep learning methods assume that training data and test data have the same distributions, which causes certain limitations in real-world histopathological applications. However, it is costly to recollect a large amount of labeled histology data to train a new neural network for each specified image acquisition procedure even for similar tasks. In this paper, an unsupervised domain adaptation is introduced into a typical deep convolutional neural network (CNN) model to mitigate the repeating of the labels. The unsupervised domain adaptation is implemented by adding two regularisation terms, namely the feature-based adaptation and entropy minimisation, to the object function of a widely used CNN model called the AlexNet. Three independent public epithelium-stroma datasets were used to verify the proposed method. The experimental results have demonstrated that in the epithelium-stroma classification, the proposed method can achieve better performance than the commonly used deep learning methods and some existing deep domain adaptation methods. Therefore, the proposed method can be considered as a better option for the real-world applications of histopathological image analysis because there is no requirement for recollection of large-scale labeled data for every specified domain. © 2018 The Authors Journal of Microscopy © 2018 Royal Microscopical Society.

  18. Two-Stage Approach to Image Classification by Deep Neural Networks

    NASA Astrophysics Data System (ADS)

    Ososkov, Gennady; Goncharov, Pavel

    2018-02-01

    The paper demonstrates the advantages of the deep learning networks over the ordinary neural networks on their comparative applications to image classifying. An autoassociative neural network is used as a standalone autoencoder for prior extraction of the most informative features of the input data for neural networks to be compared further as classifiers. The main efforts to deal with deep learning networks are spent for a quite painstaking work of optimizing the structures of those networks and their components, as activation functions, weights, as well as the procedures of minimizing their loss function to improve their performances and speed up their learning time. It is also shown that the deep autoencoders develop the remarkable ability for denoising images after being specially trained. Convolutional Neural Networks are also used to solve a quite actual problem of protein genetics on the example of the durum wheat classification. Results of our comparative study demonstrate the undoubted advantage of the deep networks, as well as the denoising power of the autoencoders. In our work we use both GPU and cloud services to speed up the calculations.

  19. Integrated Raman spectroscopy and trimodal wide-field imaging techniques for real-time in vivo tissue Raman measurements at endoscopy.

    PubMed

    Huang, Zhiwei; Teh, Seng Khoon; Zheng, Wei; Mo, Jianhua; Lin, Kan; Shao, Xiaozhuo; Ho, Khek Yu; Teh, Ming; Yeoh, Khay Guan

    2009-03-15

    We report an integrated Raman spectroscopy and trimodal (white-light reflectance, autofluorescence, and narrow-band) imaging techniques for real-time in vivo tissue Raman measurements at endoscopy. A special 1.8 mm endoscopic Raman probe with filtering modules is developed, permitting effective elimination of interference of fluorescence background and silica Raman in fibers while maximizing tissue Raman collections. We demonstrate that high-quality in vivo Raman spectra of upper gastrointestinal tract can be acquired within 1 s or subseconds under the guidance of wide-field endoscopic imaging modalities, greatly facilitating the adoption of Raman spectroscopy into clinical research and practice during routine endoscopic inspections.

  20. Probing neural tissue with airy light-sheet microscopy: investigation of imaging performance at depth within turbid media

    NASA Astrophysics Data System (ADS)

    Nylk, Jonathan; McCluskey, Kaley; Aggarwal, Sanya; Tello, Javier A.; Dholakia, Kishan

    2017-02-01

    Light-sheet microscopy (LSM) has received great interest for fluorescent imaging applications in biomedicine as it facilitates three-dimensional visualisation of large sample volumes with high spatiotemporal resolution whilst minimising irradiation of, and photo-damage to the specimen. Despite these advantages, LSM can only visualize superficial layers of turbid tissues, such as mammalian neural tissue. Propagation-invariant light modes have played a key role in the development of high-resolution LSM techniques as they overcome the natural divergence of a Gaussian beam, enabling uniform and thin light-sheets over large distances. Most notably, Bessel and Airy beam-based light-sheet imaging modalities have been demonstrated. In the single-photon excitation regime and in lightly scattering specimens, Airy-LSM has given competitive performance with advanced Bessel-LSM techniques. Airy and Bessel beams share the property of self-healing, the ability of the beam to regenerate its transverse beam profile after propagation around an obstacle. Bessel-LSM techniques have been shown to increase the penetration-depth of the illumination into turbid specimens but this effect has been understudied in biologically relevant tissues, particularly for Airy beams. It is expected that Airy-LSM will give a similar enhancement over Gaussian-LSM. In this paper, we report on the comparison of Airy-LSM and Gaussian-LSM imaging modalities within cleared and non-cleared mouse brain tissue. In particular, we examine image quality versus tissue depth by quantitative spatial Fourier analysis of neural structures in virally transduced fluorescent tissue sections, showing a three-fold enhancement at 50 μm depth into non-cleared tissue with Airy-LSM. Complimentary analysis is performed by resolution measurements in bead-injected tissue sections.

  1. Assessing microscope image focus quality with deep learning.

    PubMed

    Yang, Samuel J; Berndl, Marc; Michael Ando, D; Barch, Mariya; Narayanaswamy, Arunachalam; Christiansen, Eric; Hoyer, Stephan; Roat, Chris; Hung, Jane; Rueden, Curtis T; Shankar, Asim; Finkbeiner, Steven; Nelson, Philip

    2018-03-15

    Large image datasets acquired on automated microscopes typically have some fraction of low quality, out-of-focus images, despite the use of hardware autofocus systems. Identification of these images using automated image analysis with high accuracy is important for obtaining a clean, unbiased image dataset. Complicating this task is the fact that image focus quality is only well-defined in foreground regions of images, and as a result, most previous approaches only enable a computation of the relative difference in quality between two or more images, rather than an absolute measure of quality. We present a deep neural network model capable of predicting an absolute measure of image focus on a single image in isolation, without any user-specified parameters. The model operates at the image-patch level, and also outputs a measure of prediction certainty, enabling interpretable predictions. The model was trained on only 384 in-focus Hoechst (nuclei) stain images of U2OS cells, which were synthetically defocused to one of 11 absolute defocus levels during training. The trained model can generalize on previously unseen real Hoechst stain images, identifying the absolute image focus to within one defocus level (approximately 3 pixel blur diameter difference) with 95% accuracy. On a simpler binary in/out-of-focus classification task, the trained model outperforms previous approaches on both Hoechst and Phalloidin (actin) stain images (F-scores of 0.89 and 0.86, respectively over 0.84 and 0.83), despite only having been presented Hoechst stain images during training. Lastly, we observe qualitatively that the model generalizes to two additional stains, Hoechst and Tubulin, of an unseen cell type (Human MCF-7) acquired on a different instrument. Our deep neural network enables classification of out-of-focus microscope images with both higher accuracy and greater precision than previous approaches via interpretable patch-level focus and certainty predictions. The use of

  2. An analysis of image storage systems for scalable training of deep neural networks

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Lim, Seung-Hwan; Young, Steven R; Patton, Robert M

    This study presents a principled empirical evaluation of image storage systems for training deep neural networks. We employ the Caffe deep learning framework to train neural network models for three different data sets, MNIST, CIFAR-10, and ImageNet. While training the models, we evaluate five different options to retrieve training image data: (1) PNG-formatted image files on local file system; (2) pushing pixel arrays from image files into a single HDF5 file on local file system; (3) in-memory arrays to hold the pixel arrays in Python and C++; (4) loading the training data into LevelDB, a log-structured merge tree based key-valuemore » storage; and (5) loading the training data into LMDB, a B+tree based key-value storage. The experimental results quantitatively highlight the disadvantage of using normal image files on local file systems to train deep neural networks and demonstrate reliable performance with key-value storage based storage systems. When training a model on the ImageNet dataset, the image file option was more than 17 times slower than the key-value storage option. Along with measurements on training time, this study provides in-depth analysis on the cause of performance advantages/disadvantages of each back-end to train deep neural networks. We envision the provided measurements and analysis will shed light on the optimal way to architect systems for training neural networks in a scalable manner.« less

  3. A novel end-to-end classifier using domain transferred deep convolutional neural networks for biomedical images.

    PubMed

    Pang, Shuchao; Yu, Zhezhou; Orgun, Mehmet A

    2017-03-01

    Highly accurate classification of biomedical images is an essential task in the clinical diagnosis of numerous medical diseases identified from those images. Traditional image classification methods combined with hand-crafted image feature descriptors and various classifiers are not able to effectively improve the accuracy rate and meet the high requirements of classification of biomedical images. The same also holds true for artificial neural network models directly trained with limited biomedical images used as training data or directly used as a black box to extract the deep features based on another distant dataset. In this study, we propose a highly reliable and accurate end-to-end classifier for all kinds of biomedical images via deep learning and transfer learning. We first apply domain transferred deep convolutional neural network for building a deep model; and then develop an overall deep learning architecture based on the raw pixels of original biomedical images using supervised training. In our model, we do not need the manual design of the feature space, seek an effective feature vector classifier or segment specific detection object and image patches, which are the main technological difficulties in the adoption of traditional image classification methods. Moreover, we do not need to be concerned with whether there are large training sets of annotated biomedical images, affordable parallel computing resources featuring GPUs or long times to wait for training a perfect deep model, which are the main problems to train deep neural networks for biomedical image classification as observed in recent works. With the utilization of a simple data augmentation method and fast convergence speed, our algorithm can achieve the best accuracy rate and outstanding classification ability for biomedical images. We have evaluated our classifier on several well-known public biomedical datasets and compared it with several state-of-the-art approaches. We propose a robust

  4. Image-based metrology of porous tissue engineering scaffolds

    NASA Astrophysics Data System (ADS)

    Rajagopalan, Srinivasan; Robb, Richard A.

    2006-03-01

    Tissue engineering is an interdisciplinary effort aimed at the repair and regeneration of biological tissues through the application and control of cells, porous scaffolds and growth factors. The regeneration of specific tissues guided by tissue analogous substrates is dependent on diverse scaffold architectural indices that can be derived quantitatively from the microCT and microMR images of the scaffolds. However, the randomness of pore-solid distributions in conventional stochastic scaffolds presents unique computational challenges. As a result, image-based characterization of scaffolds has been predominantly qualitative. In this paper, we discuss quantitative image-based techniques that can be used to compute the metrological indices of porous tissue engineering scaffolds. While bulk averaged quantities such as porosity and surface are derived directly from the optimal pore-solid delineations, the spatially distributed geometric indices are derived from the medial axis representations of the pore network. The computational framework proposed (to the best of our knowledge for the first time in tissue engineering) in this paper might have profound implications towards unraveling the symbiotic structure-function relationship of porous tissue engineering scaffolds.

  5. Stokes polarimetry imaging of dog prostate tissue

    NASA Astrophysics Data System (ADS)

    Kim, Jihoon; Johnston, William K., III; Walsh, Joseph T., Jr.

    2010-02-01

    Prostate cancer is the second leading cause of death in the United States in 2009. Radical prostatectomy (complete removal of the prostate) is the most common treatment for prostate cancer, however, differentiating prostate tissue from adjacent bladder, nerves, and muscle is difficult. Improved visualization could improve oncologic outcomes and decrease damage to adjacent nerves and muscle important for preservation of potency and continence. A novel Stokes polarimetry imaging (SPI) system was developed and evaluated using a dog prostate specimen in order to examine the feasibility of the system to differentiate prostate from bladder. The degree of linear polarization (DOLP) image maps from linearly polarized light illumination at different visible wavelengths (475, 510, and 650 nm) were constructed. The SPI system used the polarization property of the prostate tissue. The DOLP images allowed advanced differentiation by distinguishing glandular tissue of prostate from the muscular-stromal tissue in the bladder. The DOLP image at 650 nm effectively differentiated prostate and bladder by strong DOLP in bladder. SPI system has the potential to improve surgical outcomes in open or robotic-assisted laparoscopic removal of the prostate. Further in vivo testing is warranted.

  6. Sensitivity and accuracy of hybrid fluorescence-mediated tomography in deep tissue regions.

    PubMed

    Rosenhain, Stefanie; Al Rawashdeh, Wa'el; Kiessling, Fabian; Gremse, Felix

    2017-09-01

    Fluorescence-mediated tomography (FMT) enables noninvasive assessment of the three-dimensional distribution of near-infrared fluorescence in mice. The combination with micro-computed tomography (µCT) provides anatomical data, enabling improved fluorescence reconstruction and image analysis. The aim of our study was to assess sensitivity and accuracy of µCT-FMT under realistic in vivo conditions in deeply-seated regions. Accordingly, we acquired fluorescence reflectance images (FRI) and µCT-FMT scans of mice which were prepared with rectal insertions with different amounts of fluorescent dye. Default and high-sensitivity scans were acquired and background signal was analyzed for three FMT channels (670 nm, 745 nm, and 790 nm). Analysis was performed for the original and an improved FMT reconstruction using the µCT data. While FRI and the original FMT reconstruction could detect 100 pmol, the improved FMT reconstruction could detect 10 pmol and significantly improved signal localization. By using a finer sampling grid and increasing the exposure time, the sensitivity could be further improved to detect 0.5 pmol. Background signal was highest in the 670 nm channel and most prominent in the gastro-intestinal tract and in organs with high relative amounts of blood. In conclusion, we show that µCT-FMT allows sensitive and accurate assessment of fluorescence in deep tissue regions. © 2017 Wiley-VCH Verlag GmbH & Co. KGaA, Weinheim.

  7. Referred pain and cutaneous responses from deep tissue electrical pain stimulation in the groin.

    PubMed

    Aasvang, E K; Werner, M U; Kehlet, H

    2015-08-01

    Persistent postherniotomy pain is located around the scar and external inguinal ring and is often described as deep rather than cutaneous, with frequent complaints of pain in adjacent areas. Whether this pain is due to local pathology or referred/projected pain is unknown, hindering mechanism-based treatment. Deep tissue electrical pain stimulation by needle electrodes in the right groin (rectus muscle, ilioinguinal/iliohypogastric nerve and perispermatic cord) was combined with assessment of referred/projected pain and the cutaneous heat pain threshold (HPT) at three prespecified areas (both groins and the lower right arm) in 19 healthy subjects. The assessment was repeated 10 days later to assess the reproducibility of individual responses. Deep electrical stimulation elicited pain at the stimulation site in all subjects, and in 15 subjects, pain from areas outside the stimulation area was reported, with 90-100% having the same response on both days, depending on the location. Deep pain stimulation significantly increased the cutaneous HPT (P<0.014). Individual HPT responses before and during deep electrical pain stimulation were significantly correlated (ρ>0.474, P≤0.040) at the two test days for the majority of test areas. Our results corroborate a systematic relationship between deep pain and changes in cutaneous nociception. The individual referred/projected pain patterns and cutaneous responses are variable, but reproducible, supporting individual differences in anatomy and sensory processing. Future studies investigating the responses to deep tissue electrical stimulation in persistent postherniotomy pain patients may advance our understanding of underlying pathophysiological mechanisms and strategies for treatment and prevention. ClinicalTrials.gov (NCT01701427). © The Author 2015. Published by Oxford University Press on behalf of the British Journal of Anaesthesia. All rights reserved. For Permissions, please email: journals.permissions@oup.com.

  8. Deep Multimodal Distance Metric Learning Using Click Constraints for Image Ranking.

    PubMed

    Yu, Jun; Yang, Xiaokang; Gao, Fei; Tao, Dacheng

    2017-12-01

    How do we retrieve images accurately? Also, how do we rank a group of images precisely and efficiently for specific queries? These problems are critical for researchers and engineers to generate a novel image searching engine. First, it is important to obtain an appropriate description that effectively represent the images. In this paper, multimodal features are considered for describing images. The images unique properties are reflected by visual features, which are correlated to each other. However, semantic gaps always exist between images visual features and semantics. Therefore, we utilize click feature to reduce the semantic gap. The second key issue is learning an appropriate distance metric to combine these multimodal features. This paper develops a novel deep multimodal distance metric learning (Deep-MDML) method. A structured ranking model is adopted to utilize both visual and click features in distance metric learning (DML). Specifically, images and their related ranking results are first collected to form the training set. Multimodal features, including click and visual features, are collected with these images. Next, a group of autoencoders is applied to obtain initially a distance metric in different visual spaces, and an MDML method is used to assign optimal weights for different modalities. Next, we conduct alternating optimization to train the ranking model, which is used for the ranking of new queries with click features. Compared with existing image ranking methods, the proposed method adopts a new ranking model to use multimodal features, including click features and visual features in DML. We operated experiments to analyze the proposed Deep-MDML in two benchmark data sets, and the results validate the effects of the method.

  9. Classification of microscopic images of breast tissue

    NASA Astrophysics Data System (ADS)

    Ballerini, Lucia; Franzen, Lennart

    2004-05-01

    Breast cancer is the most common form of cancer among women. The diagnosis is usually performed by the pathologist, that subjectively evaluates tissue samples. The aim of our research is to develop techniques for the automatic classification of cancerous tissue, by analyzing histological samples of intact tissue taken with a biopsy. In our study, we considered 200 images presenting four different conditions: normal tissue, fibroadenosis, ductal cancer and lobular cancer. Methods to extract features have been investigated and described. One method is based on granulometries, which are size-shape descriptors widely used in mathematical morphology. Applications of granulometries lead to distribution functions whose moments are used as features. A second method is based on fractal geometry, that seems very suitable to quantify biological structures. The fractal dimension of binary images has been computed using the euclidean distance mapping. Image classification has then been performed using the extracted features as input of a back-propagation neural network. A new method that combines genetic algorithms and morphological filters has been also investigated. In this case, the classification is based on a correlation measure. Very encouraging results have been obtained with pilot experiments using a small subset of images as training set. Experimental results indicate the effectiveness of the proposed methods. Cancerous tissue was correctly classified in 92.5% of the cases.

  10. Contrast-enhanced photoacoustic imaging with an optical wavelength of 1064 nm

    NASA Astrophysics Data System (ADS)

    Kim, Jeesu; Park, Sara; Park, Gyeong Bae; Choi, Wonseok; Jeong, Unyong; Kim, Chulhong

    2018-02-01

    Photoacoustic (PA) imaging is a biomedical imaging method that can provide both structural and functional information of living tissues beyond the optical diffusion limit by combining the concepts of conventional optical and ultrasound imaging methods. Although endogenous chromophores can be utilized to acquire PA images of biological tissues, exogenous contrast agents that absorb near-infrared (NIR) lights have been extensively explored to improve the contrast and penetration depth of PA images. Here, we demonstrate Bi2Se3 nanoplates, that strongly absorbs NIR lights, as a contrast agent for PA imaging. In particularly, the Bi2Se3 nanoplates produce relatively strong PA signals with an optical wavelength of 1064 nm, which has several advantages for deep tissue imaging including: (1) relatively low absorption by other intrinsic chromophores, (2) cost-effective light source using Nd:YAG laser, and (3) higher available energy than other NIR lights according to American National Standards Institute (ANSI) safety limit. We have investigated deep tissue imaging capability of the Bi2Se3 nanoplates by acquiring in vitro PA images of microtubes under chicken breast tissues. We have also acquired in vivo PA images of bladders, gastrointestinal tracts, and sentinel lymph nodes in mice after injection of the Bi2Se3 nanoplates to verify their applicability to a variety of biomedical research. The results show the promising potential of the Bi2Se3 nanoplates as a PA contrast agent for deep tissue imaging with an optical wavelength of 1064 nm.

  11. Adipose Tissue Quantification by Imaging Methods: A Proposed Classification

    PubMed Central

    Shen, Wei; Wang, ZiMian; Punyanita, Mark; Lei, Jianbo; Sinav, Ahmet; Kral, John G.; Imielinska, Celina; Ross, Robert; Heymsfield, Steven B.

    2007-01-01

    Recent advances in imaging techniques and understanding of differences in the molecular biology of adipose tissue has rendered classical anatomy obsolete, requiring a new classification of the topography of adipose tissue. Adipose tissue is one of the largest body compartments, yet a classification that defines specific adipose tissue depots based on their anatomic location and related functions is lacking. The absence of an accepted taxonomy poses problems for investigators studying adipose tissue topography and its functional correlates. The aim of this review was to critically examine the literature on imaging of whole body and regional adipose tissue and to create the first systematic classification of adipose tissue topography. Adipose tissue terminology was examined in over 100 original publications. Our analysis revealed inconsistencies in the use of specific definitions, especially for the compartment termed “visceral” adipose tissue. This analysis leads us to propose an updated classification of total body and regional adipose tissue, providing a well-defined basis for correlating imaging studies of specific adipose tissue depots with molecular processes. PMID:12529479

  12. Lesion Detection in CT Images Using Deep Learning Semantic Segmentation Technique

    NASA Astrophysics Data System (ADS)

    Kalinovsky, A.; Liauchuk, V.; Tarasau, A.

    2017-05-01

    In this paper, the problem of automatic detection of tuberculosis lesion on 3D lung CT images is considered as a benchmark for testing out algorithms based on a modern concept of Deep Learning. For training and testing of the algorithms a domestic dataset of 338 3D CT scans of tuberculosis patients with manually labelled lesions was used. The algorithms which are based on using Deep Convolutional Networks were implemented and applied in three different ways including slice-wise lesion detection in 2D images using semantic segmentation, slice-wise lesion detection in 2D images using sliding window technique as well as straightforward detection of lesions via semantic segmentation in whole 3D CT scans. The algorithms demonstrate superior performance compared to algorithms based on conventional image analysis methods.

  13. [Advantages and Application Prospects of Deep Learning in Image Recognition and Bone Age Assessment].

    PubMed

    Hu, T H; Wan, L; Liu, T A; Wang, M W; Chen, T; Wang, Y H

    2017-12-01

    Deep learning and neural network models have been new research directions and hot issues in the fields of machine learning and artificial intelligence in recent years. Deep learning has made a breakthrough in the applications of image and speech recognitions, and also has been extensively used in the fields of face recognition and information retrieval because of its special superiority. Bone X-ray images express different variations in black-white-gray gradations, which have image features of black and white contrasts and level differences. Based on these advantages of deep learning in image recognition, we combine it with the research of bone age assessment to provide basic datum for constructing a forensic automatic system of bone age assessment. This paper reviews the basic concept and network architectures of deep learning, and describes its recent research progress on image recognition in different research fields at home and abroad, and explores its advantages and application prospects in bone age assessment. Copyright© by the Editorial Department of Journal of Forensic Medicine.

  14. Classification of time-series images using deep convolutional neural networks

    NASA Astrophysics Data System (ADS)

    Hatami, Nima; Gavet, Yann; Debayle, Johan

    2018-04-01

    Convolutional Neural Networks (CNN) has achieved a great success in image recognition task by automatically learning a hierarchical feature representation from raw data. While the majority of Time-Series Classification (TSC) literature is focused on 1D signals, this paper uses Recurrence Plots (RP) to transform time-series into 2D texture images and then take advantage of the deep CNN classifier. Image representation of time-series introduces different feature types that are not available for 1D signals, and therefore TSC can be treated as texture image recognition task. CNN model also allows learning different levels of representations together with a classifier, jointly and automatically. Therefore, using RP and CNN in a unified framework is expected to boost the recognition rate of TSC. Experimental results on the UCR time-series classification archive demonstrate competitive accuracy of the proposed approach, compared not only to the existing deep architectures, but also to the state-of-the art TSC algorithms.

  15. Pre-operative intra-articular deep tissue sampling with novel retrograde forceps improves the diagnostics in periprosthetic joint infection.

    PubMed

    Wimmer, Matthias D; Ploeger, Milena M; Friedrich, Max J; Hügle, Thomas; Gravius, Sascha; Randau, Thomas M

    2017-07-01

    Histopathological tissue analysis is a key parameter within the diagnostic algorithm for suspected periprosthetic joint infections (PJIs), conventionally acquired in open surgery. In 2014, Hügle and co-workers introduced novel retrograde forceps for retrograde synovial biopsy with simultaneous fluid aspiration of the knee joint. We hypothesised that tissue samples acquired by retrograde synovial biopsy are equal to intra-operatively acquired deep representative tissue samples regarding bacterial detection and differentiation of periprosthetic infectious membranes. Thirty patients (male n = 15, 50%; female n = 15, 50%) with 30 suspected PJIs in painful total hip arthroplasties (THAs) were included in this prospective, controlled, non-blinded trial. The results were compared with intra-operatively obtained representative deep tissue samples. In summary, 27 out of 30 patients were diagnosed correctly as infected (17/17) or non-infected (10/13). The sensitivity to predict a PJI using the Retroforce® sampling forceps in addition to standard diagnostics was 85%, the specificity 100%. Retrograde synovial biopsy is a new and rapid diagnostic procedure under local anaesthesia in patients with painful THAs with similar histological results compared to deep tissue sampling.

  16. Focusing light through biological tissue and tissue-mimicking phantoms up to 9.6 cm in thickness with digital optical phase conjugation

    NASA Astrophysics Data System (ADS)

    Shen, Yuecheng; Liu, Yan; Ma, Cheng; Wang, Lihong V.

    2016-08-01

    Optical phase conjugation (OPC)-based wavefront shaping techniques focus light through or within scattering media, which is critically important for deep-tissue optical imaging, manipulation, and therapy. However, to date, the sample thickness in OPC experiments has been limited to only a few millimeters. Here, by using a laser with a long coherence length and an optimized digital OPC system that can safely deliver more light power, we focused 532-nm light through tissue-mimicking phantoms up to 9.6 cm thick, as well as through ex vivo chicken breast tissue up to 2.5 cm thick. Our results demonstrate that OPC can be achieved even when photons have experienced on average 1000 scattering events. The demonstrated penetration of nearly 10 cm (˜100 transport mean free paths) has never been achieved before by any optical focusing technique, and it shows the promise of OPC for deep-tissue noninvasive optical imaging, manipulation, and therapy.

  17. Deep learning methods for CT image-domain metal artifact reduction

    NASA Astrophysics Data System (ADS)

    Gjesteby, Lars; Yang, Qingsong; Xi, Yan; Shan, Hongming; Claus, Bernhard; Jin, Yannan; De Man, Bruno; Wang, Ge

    2017-09-01

    Artifacts resulting from metal objects have been a persistent problem in CT images over the last four decades. A common approach to overcome their effects is to replace corrupt projection data with values synthesized from an interpolation scheme or by reprojection of a prior image. State-of-the-art correction methods, such as the interpolation- and normalization-based algorithm NMAR, often do not produce clinically satisfactory results. Residual image artifacts remain in challenging cases and even new artifacts can be introduced by the interpolation scheme. Metal artifacts continue to be a major impediment, particularly in radiation and proton therapy planning as well as orthopedic imaging. A new solution to the long-standing metal artifact reduction (MAR) problem is deep learning, which has been successfully applied to medical image processing and analysis tasks. In this study, we combine a convolutional neural network (CNN) with the state-of-the-art NMAR algorithm to reduce metal streaks in critical image regions. Training data was synthesized from CT simulation scans of a phantom derived from real patient images. The CNN is able to map metal-corrupted images to artifact-free monoenergetic images to achieve additional correction on top of NMAR for improved image quality. Our results indicate that deep learning is a novel tool to address CT reconstruction challenges, and may enable more accurate tumor volume estimation for radiation therapy planning.

  18. Image-guided urologic surgery: intraoperative optical imaging and tissue interrogation (Conference Presentation)

    NASA Astrophysics Data System (ADS)

    Liao, Joseph C.

    2017-02-01

    Emerging optical imaging technologies can be integrated in the operating room environment during minimally invasive and open urologic surgery, including oncologic surgery of the bladder, prostate, and kidney. These technologies include macroscopic fluorescence imaging that provides contrast enhancement between normal and diseased tissue and microscopic imaging that provides tissue characterization. Optical imaging technologies that have reached the clinical arena in urologic surgery are reviewed, including photodynamic diagnosis, near infrared fluorescence imaging, optical coherence tomography, and confocal laser endomicroscopy. Molecular imaging represents an exciting future arena in conjugating cancer-specific contrast agents to fluorophores to improve the specificity of disease detection. Ongoing efforts are underway to translate optimal targeting agents and imaging modalities, with the goal to improve cancer-specific and functional outcomes.

  19. Functional imaging of small tissue volumes with diffuse optical tomography

    NASA Astrophysics Data System (ADS)

    Klose, Alexander D.; Hielscher, Andreas H.

    2006-03-01

    Imaging of dynamic changes in blood parameters, functional brain imaging, and tumor imaging are the most advanced application areas of diffuse optical tomography (DOT). When dealing with the image reconstruction problem one is faced with the fact that near-infrared photons, unlike X-rays, are highly scattered when they traverse biological tissue. Image reconstruction schemes are required that model the light propagation inside biological tissue and predict measurements on the tissue surface. By iteratively changing the tissue-parameters until the predictions agree with the real measurements, a spatial distribution of optical properties inside the tissue is found. The optical properties can be related to the tissue oxygenation, inflammation, or to the fluorophore concentration of a biochemical marker. If the model of light propagation is inaccurate, the reconstruction process will lead to an inaccurate result as well. Here, we focus on difficulties that are encountered when DOT is employed for functional imaging of small tissue volumes, for example, in cancer studies involving small animals, or human finger joints for early diagnosis of rheumatoid arthritis. Most of the currently employed image reconstruction methods rely on the diffusion theory that is an approximation to the equation of radiative transfer. But, in the cases of small tissue volumes and tissues that contain low scattering regions diffusion theory has been shown to be of limited applicability Therefore, we employ a light propagation model that is based on the equation of radiative transfer, which promises to overcome the limitations.

  20. Classifying magnetic resonance image modalities with convolutional neural networks

    NASA Astrophysics Data System (ADS)

    Remedios, Samuel; Pham, Dzung L.; Butman, John A.; Roy, Snehashis

    2018-02-01

    Magnetic Resonance (MR) imaging allows the acquisition of images with different contrast properties depending on the acquisition protocol and the magnetic properties of tissues. Many MR brain image processing techniques, such as tissue segmentation, require multiple MR contrasts as inputs, and each contrast is treated differently. Thus it is advantageous to automate the identification of image contrasts for various purposes, such as facilitating image processing pipelines, and managing and maintaining large databases via content-based image retrieval (CBIR). Most automated CBIR techniques focus on a two-step process: extracting features from data and classifying the image based on these features. We present a novel 3D deep convolutional neural network (CNN)- based method for MR image contrast classification. The proposed CNN automatically identifies the MR contrast of an input brain image volume. Specifically, we explored three classification problems: (1) identify T1-weighted (T1-w), T2-weighted (T2-w), and fluid-attenuated inversion recovery (FLAIR) contrasts, (2) identify pre vs postcontrast T1, (3) identify pre vs post-contrast FLAIR. A total of 3418 image volumes acquired from multiple sites and multiple scanners were used. To evaluate each task, the proposed model was trained on 2137 images and tested on the remaining 1281 images. Results showed that image volumes were correctly classified with 97.57% accuracy.

  1. Introducing Convective Cloud Microphysics to a Deep Convection Parameterization Facilitating Aerosol Indirect Effects

    NASA Astrophysics Data System (ADS)

    Alapaty, K.; Zhang, G. J.; Song, X.; Kain, J. S.; Herwehe, J. A.

    2012-12-01

    Short lived pollutants such as aerosols play an important role in modulating not only the radiative balance but also cloud microphysical properties and precipitation rates. In the past, to understand the interactions of aerosols with clouds, several cloud-resolving modeling studies were conducted. These studies indicated that in the presence of anthropogenic aerosols, single-phase deep convection precipitation is reduced or suppressed. On the other hand, anthropogenic aerosol pollution led to enhanced precipitation for mixed-phase deep convective clouds. To date, there have not been many efforts to incorporate such aerosol indirect effects (AIE) in mesoscale models or global models that use parameterization schemes for deep convection. Thus, the objective of this work is to implement a diagnostic cloud microphysical scheme directly into a deep convection parameterization facilitating aerosol indirect effects in the WRF-CMAQ integrated modeling systems. Major research issues addressed in this study are: What is the sensitivity of a deep convection scheme to cloud microphysical processes represented by a bulk double-moment scheme? How close are the simulated cloud water paths as compared to observations? Does increased aerosol pollution lead to increased precipitation for mixed-phase clouds? These research questions are addressed by performing several WRF simulations using the Kain-Fritsch convection parameterization and a diagnostic cloud microphysical scheme. In the first set of simulations (control simulations) the WRF model is used to simulate two scenarios of deep convection over the continental U.S. during two summer periods at 36 km grid resolution. In the second set, these simulations are repeated after incorporating a diagnostic cloud microphysical scheme to study the impacts of inclusion of cloud microphysical processes. Finally, in the third set, aerosol concentrations simulated by the CMAQ modeling system are supplied to the embedded cloud microphysical

  2. Combined spectroscopic imaging and chemometric approach for automatically partitioning tissue types in human prostate tissue biopsies

    NASA Astrophysics Data System (ADS)

    Haka, Abigail S.; Kidder, Linda H.; Lewis, E. Neil

    2001-07-01

    We have applied Fourier transform infrared (FTIR) spectroscopic imaging, coupling a mercury cadmium telluride (MCT) focal plane array detector (FPA) and a Michelson step scan interferometer, to the investigation of various states of malignant human prostate tissue. The MCT FPA used consists of 64x64 pixels, each 61 micrometers 2, and has a spectral range of 2-10.5 microns. Each imaging data set was collected at 16-1 resolution, resulting in 512 image planes and a total of 4096 interferograms. In this article we describe a method for separating different tissue types contained within FTIR spectroscopic imaging data sets of human prostate tissue biopsies. We present images, generated by the Fuzzy C-Means clustering algorithm, which demonstrate the successful partitioning of distinct tissue type domains. Additionally, analysis of differences in the centroid spectra corresponding to different tissue types provides an insight into their biochemical composition. Lastly, we demonstrate the ability to partition tissue type regions in a different data set using centroid spectra calculated from the original data set. This has implications for the use of the Fuzzy C-Means algorithm as an automated technique for the separation and examination of tissue domains in biopsy samples.

  3. Global Magnetospheric Imaging from the Deep Space Gateway in Lunar Orbit

    NASA Astrophysics Data System (ADS)

    Chua, D. H.; Socker, D. G.; Englert, C. R.; Carter, M. T.; Plunkett, S. P.; Korendyke, C. M.; Meier, R. R.

    2018-02-01

    We propose to use the Deep Space Gateway as an observing platform for a magnetospheric imager that will capture the first direct global images of the interface between the incident solar wind and the Earth's magnetosphere.

  4. Confocal Imaging of the Embryonic Heart: How Deep?

    NASA Astrophysics Data System (ADS)

    Miller, Christine E.; Thompson, Robert P.; Bigelow, Michael R.; Gittinger, George; Trusk, Thomas C.; Sedmera, David

    2005-06-01

    Confocal microscopy allows for optical sectioning of tissues, thus obviating the need for physical sectioning and subsequent registration to obtain a three-dimensional representation of tissue architecture. However, practicalities such as tissue opacity, light penetration, and detector sensitivity have usually limited the available depth of imaging to 200 [mu]m. With the emergence of newer, more powerful systems, we attempted to push these limits to those dictated by the working distance of the objective. We used whole-mount immunohistochemical staining followed by clearing with benzyl alcohol-benzyl benzoate (BABB) to visualize three-dimensional myocardial architecture. Confocal imaging of entire chick embryonic hearts up to a depth of 1.5 mm with voxel dimensions of 3 [mu]m was achieved with a 10× dry objective. For the purpose of screening for congenital heart defects, we used endocardial painting with fluorescently labeled poly-L-lysine and imaged BABB-cleared hearts with a 5× objective up to a depth of 2 mm. Two-photon imaging of whole-mount specimens stained with Hoechst nuclear dye produced clear images all the way through stage 29 hearts without significant signal attenuation. Thus, currently available systems allow confocal imaging of fixed samples to previously unattainable depths, the current limiting factors being objective working distance, antibody penetration, specimen autofluorescence, and incomplete clearing.

  5. Large-Scale Image Analytics Using Deep Learning

    NASA Astrophysics Data System (ADS)

    Ganguly, S.; Nemani, R. R.; Basu, S.; Mukhopadhyay, S.; Michaelis, A.; Votava, P.

    2014-12-01

    High resolution land cover classification maps are needed to increase the accuracy of current Land ecosystem and climate model outputs. Limited studies are in place that demonstrates the state-of-the-art in deriving very high resolution (VHR) land cover products. In addition, most methods heavily rely on commercial softwares that are difficult to scale given the region of study (e.g. continents to globe). Complexities in present approaches relate to (a) scalability of the algorithm, (b) large image data processing (compute and memory intensive), (c) computational cost, (d) massively parallel architecture, and (e) machine learning automation. In addition, VHR satellite datasets are of the order of terabytes and features extracted from these datasets are of the order of petabytes. In our present study, we have acquired the National Agricultural Imaging Program (NAIP) dataset for the Continental United States at a spatial resolution of 1-m. This data comes as image tiles (a total of quarter million image scenes with ~60 million pixels) and has a total size of ~100 terabytes for a single acquisition. Features extracted from the entire dataset would amount to ~8-10 petabytes. In our proposed approach, we have implemented a novel semi-automated machine learning algorithm rooted on the principles of "deep learning" to delineate the percentage of tree cover. In order to perform image analytics in such a granular system, it is mandatory to devise an intelligent archiving and query system for image retrieval, file structuring, metadata processing and filtering of all available image scenes. Using the Open NASA Earth Exchange (NEX) initiative, which is a partnership with Amazon Web Services (AWS), we have developed an end-to-end architecture for designing the database and the deep belief network (following the distbelief computing model) to solve a grand challenge of scaling this process across quarter million NAIP tiles that cover the entire Continental United States. The

  6. Deep Convolutional Neural Networks Enable Discrimination of Heterogeneous Digital Pathology Images.

    PubMed

    Khosravi, Pegah; Kazemi, Ehsan; Imielinski, Marcin; Elemento, Olivier; Hajirasouliha, Iman

    2018-01-01

    Pathological evaluation of tumor tissue is pivotal for diagnosis in cancer patients and automated image analysis approaches have great potential to increase precision of diagnosis and help reduce human error. In this study, we utilize several computational methods based on convolutional neural networks (CNN) and build a stand-alone pipeline to effectively classify different histopathology images across different types of cancer. In particular, we demonstrate the utility of our pipeline to discriminate between two subtypes of lung cancer, four biomarkers of bladder cancer, and five biomarkers of breast cancer. In addition, we apply our pipeline to discriminate among four immunohistochemistry (IHC) staining scores of bladder and breast cancers. Our classification pipeline includes a basic CNN architecture, Google's Inceptions with three training strategies, and an ensemble of two state-of-the-art algorithms, Inception and ResNet. Training strategies include training the last layer of Google's Inceptions, training the network from scratch, and fine-tunning the parameters for our data using two pre-trained version of Google's Inception architectures, Inception-V1 and Inception-V3. We demonstrate the power of deep learning approaches for identifying cancer subtypes, and the robustness of Google's Inceptions even in presence of extensive tumor heterogeneity. On average, our pipeline achieved accuracies of 100%, 92%, 95%, and 69% for discrimination of various cancer tissues, subtypes, biomarkers, and scores, respectively. Our pipeline and related documentation is freely available at https://github.com/ih-_lab/CNN_Smoothie. Copyright © 2017 The Authors. Published by Elsevier B.V. All rights reserved.

  7. High-Resolution Photoacoustic Imaging of Ocular Tissues

    PubMed Central

    Silverman, Ronald H.; Kong, Fanting; Chen, Y.C.; Lloyd, Harriet O.; Kim, Hyung Ham; Cannata, Jonathan M.; Shung, K. Kirk; Coleman, D Jackson

    2010-01-01

    Optical coherence tomography (OCT) and ultrasound (US) are methods widely used for diagnostic imaging of the eye. These techniques detect discontinuities in optical refractive index and acoustic impedance respectively. Because these both relate to variations in tissue density or composition, OCT and US images share a qualitatively similar appearance. In photoacoustic imaging (PAI), short light pulses are directed at tissues, pressure is generated due to a rapid energy deposition in the tissue volume, and thermoelastic expansion results in generation of broadband US. PAI thus depicts optical absorption, which is independent of the tissue characteristics imaged by OCT or US. Our aim was to demonstrate the application of PAI in ocular tissues and to do so with lateral resolution comparable to OCT. We developed two PAI assemblies, both of which used single-element US transducers and lasers sharing a common focus. The first assembly had optical and 35-MHz US axes offset by a 30° angle. The second assembly consisted of a 20-MHz ring transducer with a coaxial optics. The laser emitted 5-ns pulses at either 532-nm or 1064-nm, with spot sizes at the focus of 35-μm for the angled probe and 20-μm for the coaxial probe. We compared lateral resolution by scanning 12.5-μm diameter wire targets with pulse/echo US and PAI at each wavelength. We then imaged the anterior segment in whole ex vivo pig eyes and the choroid and ciliary body region in sectioned eyes. PAI data obtained at 1064 nm in the near infrared had higher penetration but reduced signal amplitude compared to that obtained using the 532-nm green wavelength. Images were obtained of the iris, choroid and ciliary processes. The zonules and anterior cornea and lens surfaces were seen at 532 nm. Because the laser spot size was significantly smaller than the US beamwidth at the focus, PAI images had superior resolution than those obtained using conventional US. PMID:20420969

  8. High-resolution photoacoustic imaging of ocular tissues.

    PubMed

    Silverman, Ronald H; Kong, Fanting; Chen, Y C; Lloyd, Harriet O; Kim, Hyung Ham; Cannata, Jonathan M; Shung, K Kirk; Coleman, D Jackson

    2010-05-01

    Optical coherence tomography (OCT) and ultrasound (US) are methods widely used for diagnostic imaging of the eye. These techniques detect discontinuities in optical refractive index and acoustic impedance, respectively. Because these both relate to variations in tissue density or composition, OCT and US images share a qualitatively similar appearance. In photoacoustic imaging (PAI), short light pulses are directed at tissues, pressure is generated due to a rapid energy deposition in the tissue volume and thermoelastic expansion results in generation of broadband US. PAI thus depicts optical absorption, which is independent of the tissue characteristics imaged by OCT or US. Our aim was to demonstrate the application of PAI in ocular tissues and to do so with lateral resolution comparable to OCT. We developed two PAI assemblies, both of which used single-element US transducers and lasers sharing a common focus. The first assembly had optical and 35-MHz US axes offset by a 30 degrees angle. The second assembly consisted of a 20-MHz ring transducer with a coaxial optics. The laser emitted 5-ns pulses at either 532 nm or 1064 nm, with spot sizes at the focus of 35 microm for the angled probe and 20 microm for the coaxial probe. We compared lateral resolution by scanning 12.5 microm diameter wire targets with pulse/echo US and PAI at each wavelength. We then imaged the anterior segment in whole ex vivo pig eyes and the choroid and ciliary body region in sectioned eyes. PAI data obtained at 1064 nm in the near infrared had higher penetration but reduced signal amplitude compared to that obtained using the 532 nm green wavelength. Images were obtained of the iris, choroid and ciliary processes. The zonules and anterior cornea and lens surfaces were seen at 532 nm. Because the laser spot size was significantly smaller than the US beamwidth at the focus, PAI images had superior resolution than those obtained using conventional US. Copyright 2010 World Federation for

  9. Fine-grained leukocyte classification with deep residual learning for microscopic images.

    PubMed

    Qin, Feiwei; Gao, Nannan; Peng, Yong; Wu, Zizhao; Shen, Shuying; Grudtsin, Artur

    2018-08-01

    Leukocyte classification and cytometry have wide applications in medical domain, previous researches usually exploit machine learning techniques to classify leukocytes automatically. However, constrained by the past development of machine learning techniques, for example, extracting distinctive features from raw microscopic images are difficult, the widely used SVM classifier only has relative few parameters to tune, these methods cannot efficiently handle fine-grained classification cases when the white blood cells have up to 40 categories. Based on deep learning theory, a systematic study is conducted on finer leukocyte classification in this paper. A deep residual neural network based leukocyte classifier is constructed at first, which can imitate the domain expert's cell recognition process, and extract salient features robustly and automatically. Then the deep neural network classifier's topology is adjusted according to the prior knowledge of white blood cell test. After that the microscopic image dataset with almost one hundred thousand labeled leukocytes belonging to 40 categories is built, and combined training strategies are adopted to make the designed classifier has good generalization ability. The proposed deep residual neural network based classifier was tested on microscopic image dataset with 40 leukocyte categories. It achieves top-1 accuracy of 77.80%, top-5 accuracy of 98.75% during the training procedure. The average accuracy on the test set is nearly 76.84%. This paper presents a fine-grained leukocyte classification method for microscopic images, based on deep residual learning theory and medical domain knowledge. Experimental results validate the feasibility and effectiveness of our approach. Extended experiments support that the fine-grained leukocyte classifier could be used in real medical applications, assist doctors in diagnosing diseases, reduce human power significantly. Copyright © 2018 Elsevier B.V. All rights reserved.

  10. Mapping tissue oxygen in vivo by photoacoustic lifetime imaging

    NASA Astrophysics Data System (ADS)

    Shao, Qi; Morgounova, Ekaterina; Choi, Jeung-Hwan; Jiang, Chunlan; Bischof, John; Ashkenazi, Shai

    2013-03-01

    Oxygen plays a key role in the energy metabolism of living organisms. Any imbalance in the oxygen levels will affect the metabolic homeostasis and lead to pathophysiological diseases. Hypoxia, a status of low tissue oxygen, is a key factor in tumor biology as it is highly prominent in tumor tissues. However, clinical tools for assessing tissue oxygenation are limited. The gold standard is polarographic needle electrode which is invasive and not capable of mapping (imaging) the oxygen content in tissue. We applied the method of photoacoustic lifetime imaging (PALI) of oxygen-sensitive dye to small animal tissue hypoxia research. PALI is new technology for direct, non-invasive imaging of oxygen. The technique is based on mapping the oxygen-dependent transient optical absorption of Methylene Blue (MB) by pump-probe photoacoustic imaging. Our studies show the feasibility of imaging of dissolved oxygen distribution in phantoms. In vivo experiments demonstrate that the hypoxia region is consistent with the site of subcutaneously xenografted prostate tumor in mice with adequate spatial resolution and penetration depth.

  11. Interactive classification and content-based retrieval of tissue images

    NASA Astrophysics Data System (ADS)

    Aksoy, Selim; Marchisio, Giovanni B.; Tusk, Carsten; Koperski, Krzysztof

    2002-11-01

    We describe a system for interactive classification and retrieval of microscopic tissue images. Our system models tissues in pixel, region and image levels. Pixel level features are generated using unsupervised clustering of color and texture values. Region level features include shape information and statistics of pixel level feature values. Image level features include statistics and spatial relationships of regions. To reduce the gap between low-level features and high-level expert knowledge, we define the concept of prototype regions. The system learns the prototype regions in an image collection using model-based clustering and density estimation. Different tissue types are modeled using spatial relationships of these regions. Spatial relationships are represented by fuzzy membership functions. The system automatically selects significant relationships from training data and builds models which can also be updated using user relevance feedback. A Bayesian framework is used to classify tissues based on these models. Preliminary experiments show that the spatial relationship models we developed provide a flexible and powerful framework for classification and retrieval of tissue images.

  12. 3D reconstruction of synapses with deep learning based on EM Images

    NASA Astrophysics Data System (ADS)

    Xiao, Chi; Rao, Qiang; Zhang, Dandan; Chen, Xi; Han, Hua; Xie, Qiwei

    2017-03-01

    Recently, due to the rapid development of electron microscope (EM) with its high resolution, stacks delivered by EM can be used to analyze a variety of components that are critical to understand brain function. Since synaptic study is essential in neurobiology and can be analyzed by EM stacks, the automated routines for reconstruction of synapses based on EM Images can become a very useful tool for analyzing large volumes of brain tissue and providing the ability to understand the mechanism of brain. In this article, we propose a novel automated method to realize 3D reconstruction of synapses for Automated Tapecollecting Ultra Microtome Scanning Electron Microscopy (ATUM-SEM) with deep learning. Being different from other reconstruction algorithms, which employ classifier to segment synaptic clefts directly. We utilize deep learning method and segmentation algorithm to obtain synaptic clefts as well as promote the accuracy of reconstruction. The proposed method contains five parts: (1) using modified Moving Least Square (MLS) deformation algorithm and Scale Invariant Feature Transform (SIFT) features to register adjacent sections, (2) adopting Faster Region Convolutional Neural Networks (Faster R-CNN) algorithm to detect synapses, (3) utilizing screening method which takes context cues of synapses into consideration to reduce the false positive rate, (4) combining a practical morphology algorithm with a suitable fitting function to segment synaptic clefts and optimize the shape of them, (5) applying the plugin in FIJI to show the final 3D visualization of synapses. Experimental results on ATUM-SEM images demonstrate the effectiveness of our proposed method.

  13. Polarimetric signature imaging of anisotropic bio-medical tissues

    NASA Astrophysics Data System (ADS)

    Wu, Stewart H.; Yang, De-Ming; Chiou, Arthur; Nee, Soe-Mie F.; Nee, Tsu-Wei

    2010-02-01

    Polarimetric imaging of Stokes vector (I, Q, U, V) can provide 4 independent signatures showing the linear and circular polarizations of biological tissues and cells. Using a recently developed Stokes digital imaging system, we measured the Stokes vector images of tissue samples from sections of rat livers containing normal portions and hematomas. The derived Mueller matrix elements can quantitatively provide multi-signature data of the bio-sample. This polarimetric optical technology is a new option of biosensing technology to inspect the structures of tissue samples, particularly for discriminating tumor and non-tumor biopsy. This technology is useful for critical disease discrimination and medical diagnostics applications.

  14. Back-scattered electron imaging of skeletal tissues.

    PubMed

    Boyde, A; Jones, S J

    The use of solid-state back-scattered electron (BSE) detectors in the scanning electron microscopic study of skeletal tissues has been investigated. To minimize the topographic element in the image, flat samples and a ring detector configuration with the sample at normal incidence to the beam and the detector are used. Very flat samples are prepared by diamond micromilling or diamond polishing plastic-embedded tissue. Density discrimination in the image is so good that different density phases within mineralized bone can be imaged. For unembedded spongy bone, cut surfaces can be discriminated from natural surfaces by a topographic contrast mechanism. BSE imaging also presents advantages for unembedded samples with rough topography, such as anorganic preparations of the mineralization zone in cartilage, which give rise to severe charging problems with conventional secondary electron imaging.

  15. Online quantitative analysis of multispectral images of human body tissues

    NASA Astrophysics Data System (ADS)

    Lisenko, S. A.

    2013-08-01

    A method is developed for online monitoring of structural and morphological parameters of biological tissues (haemoglobin concentration, degree of blood oxygenation, average diameter of capillaries and the parameter characterising the average size of tissue scatterers), which involves multispectral tissue imaging, image normalisation to one of its spectral layers and determination of unknown parameters based on their stable regression relation with the spectral characteristics of the normalised image. Regression is obtained by simulating numerically the diffuse reflectance spectrum of the tissue by the Monte Carlo method at a wide variation of model parameters. The correctness of the model calculations is confirmed by the good agreement with the experimental data. The error of the method is estimated under conditions of general variability of structural and morphological parameters of the tissue. The method developed is compared with the traditional methods of interpretation of multispectral images of biological tissues, based on the solution of the inverse problem for each pixel of the image in the approximation of different analytical models.

  16. Characterization of human breast cancer tissues by infrared imaging.

    PubMed

    Verdonck, M; Denayer, A; Delvaux, B; Garaud, S; De Wind, R; Desmedt, C; Sotiriou, C; Willard-Gallo, K; Goormaghtigh, E

    2016-01-21

    Fourier Transform InfraRed (FTIR) spectroscopy coupled to microscopy (IR imaging) has shown unique advantages in detecting morphological and molecular pathologic alterations in biological tissues. The aim of this study was to evaluate the potential of IR imaging as a diagnostic tool to identify characteristics of breast epithelial cells and the stroma. In this study a total of 19 breast tissue samples were obtained from 13 patients. For 6 of the patients, we also obtained Non-Adjacent Non-Tumor tissue samples. Infrared images were recorded on the main cell/tissue types identified in all breast tissue samples. Unsupervised Principal Component Analyses and supervised Partial Least Square Discriminant Analyses (PLS-DA) were used to discriminate spectra. Leave-one-out cross-validation was used to evaluate the performance of PLS-DA models. Our results show that IR imaging coupled with PLS-DA can efficiently identify the main cell types present in FFPE breast tissue sections, i.e. epithelial cells, lymphocytes, connective tissue, vascular tissue and erythrocytes. A second PLS-DA model could distinguish normal and tumor breast epithelial cells in the breast tissue sections. A patient-specific model reached particularly high sensitivity, specificity and MCC rates. Finally, we showed that the stroma located close or at distance from the tumor exhibits distinct spectral characteristics. In conclusion FTIR imaging combined with computational algorithms could be an accurate, rapid and objective tool to identify/quantify breast epithelial cells and differentiate tumor from normal breast tissue as well as normal from tumor-associated stroma, paving the way to the establishment of a potential complementary tool to ensure safe tumor margins.

  17. Deep ultraviolet resonant Raman imaging of a cell

    NASA Astrophysics Data System (ADS)

    Kumamoto, Yasuaki; Taguchi, Atsushi; Smith, Nicholas Isaac; Kawata, Satoshi

    2012-07-01

    We report the first demonstration of deep ultraviolet (DUV) Raman imaging of a cell. Nucleotide distributions in a HeLa cell were observed without any labeling at 257 nm excitation with resonant bands attributable to guanine and adenine. Obtained images represent DNA localization at nucleoli in the nucleus and RNA distribution in the cytoplasm. The presented technique extends the potential of Raman microscopy as a tool to selectively probe nucleic acids in a cell with high sensitivity due to resonance.

  18. Ultrasonically Encoded Photoacoustic Flowgraphy in Biological Tissue

    NASA Astrophysics Data System (ADS)

    Wang, Lidai; Xia, Jun; Yao, Junjie; Maslov, Konstantin I.; Wang, Lihong V.

    2013-11-01

    Blood flow speed is an important functional parameter. Doppler ultrasound flowmetry lacks sufficient sensitivity to slow blood flow (several to tens of millimeters per second) in deep tissue. To address this challenge, we developed ultrasonically encoded photoacoustic flowgraphy combining ultrasonic thermal tagging with photoacoustic imaging. Focused ultrasound generates a confined heat source in acoustically absorptive fluid. Thermal waves propagate with the flow and are directly visualized in pseudo color using photoacoustic computed tomography. The Doppler shift is employed to calculate the flow speed. This method requires only acoustic and optical absorption, and thus is applicable to continuous fluid. A blood flow speed as low as 0.24mm·s-1 was successfully measured. Deep blood flow imaging was experimentally demonstrated under 5-mm-thick chicken breast tissue.

  19. Transfer, Imaging, and Analysis Plate for Facile Handling of 384 Hanging Drop 3D Tissue Spheroids

    PubMed Central

    Cavnar, Stephen P.; Salomonsson, Emma; Luker, Kathryn E.; Luker, Gary D.; Takayama, Shuichi

    2014-01-01

    Three-dimensional culture systems bridge the experimental gap between in vivo and in vitro physiology. However, nonstandardized formation and limited downstream adaptability of 3D cultures have hindered mainstream adoption of these systems for biological applications, especially for low- and moderate-throughput assays commonly used in biomedical research. Here we build on our recent development of a 384-well hanging drop plate for spheroid culture to design a complementary spheroid transfer and imaging (TRIM) plate. The low-aspect ratio wells of the TRIM plate facilitated highfidelity, user-independent, contact-based collection of hanging drop spheroids. Using the TRIM plate, we demonstrated several downstream analyses, including bulk tissue collection for flow cytometry, high-resolution low working-distance immersion imaging, and timely reagent delivery for enzymatic studies. Low working-distance multiphoton imaging revealed a cell type–dependent, macroscopic spheroid structure. Unlike ovarian cancer spheroids, which formed loose, disk-shaped spheroids, human mammary fibroblasts formed tight, spherical, and nutrient-limited spheroids. Beyond the applications we describe here, we expect the hanging drop spheroid plate and complementary TRIM plate to facilitate analyses of spheroids across the spectrum of throughput, particularly for bulk collection of spheroids and high-content imaging. PMID:24051516

  20. Transfer, imaging, and analysis plate for facile handling of 384 hanging drop 3D tissue spheroids.

    PubMed

    Cavnar, Stephen P; Salomonsson, Emma; Luker, Kathryn E; Luker, Gary D; Takayama, Shuichi

    2014-04-01

    Three-dimensional culture systems bridge the experimental gap between in vivo and in vitro physiology. However, nonstandardized formation and limited downstream adaptability of 3D cultures have hindered mainstream adoption of these systems for biological applications, especially for low- and moderate-throughput assays commonly used in biomedical research. Here we build on our recent development of a 384-well hanging drop plate for spheroid culture to design a complementary spheroid transfer and imaging (TRIM) plate. The low-aspect ratio wells of the TRIM plate facilitated high-fidelity, user-independent, contact-based collection of hanging drop spheroids. Using the TRIM plate, we demonstrated several downstream analyses, including bulk tissue collection for flow cytometry, high-resolution low working-distance immersion imaging, and timely reagent delivery for enzymatic studies. Low working-distance multiphoton imaging revealed a cell type-dependent, macroscopic spheroid structure. Unlike ovarian cancer spheroids, which formed loose, disk-shaped spheroids, human mammary fibroblasts formed tight, spherical, and nutrient-limited spheroids. Beyond the applications we describe here, we expect the hanging drop spheroid plate and complementary TRIM plate to facilitate analyses of spheroids across the spectrum of throughput, particularly for bulk collection of spheroids and high-content imaging.

  1. X-ray Phase Contrast Imaging of Calcified Tissue and Biomaterial Structure in Bioreactor Engineered Tissues

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Appel, Alyssa A.; Larson, Jeffery C.; Garson, III, Alfred B.

    2014-11-04

    Tissues engineered in bioreactor systems have been used clinically to replace damaged tissues and organs. In addition, these systems are under continued development for many tissue engineering applications. The ability to quantitatively assess material structure and tissue formation is critical for evaluating bioreactor efficacy and for preimplantation assessment of tissue quality. These techniques allow for the nondestructive and longitudinal monitoring of large engineered tissues within the bioreactor systems and will be essential for the translation of these strategies to viable clinical therapies. X-ray Phase Contrast (XPC) imaging techniques have shown tremendous promise for a number of biomedical applications owing tomore » their ability to provide image contrast based on multiple X-ray properties, including absorption, refraction, and scatter. In this research, mesenchymal stem cell-seeded alginate hydrogels were prepared and cultured under osteogenic conditions in a perfusion bioreactor. The constructs were imaged at various time points using XPC microcomputed tomography (µCT). Imaging was performed with systems using both synchrotron- and tube-based X-ray sources. XPC µCT allowed for simultaneous three-dimensional (3D) quantification of hydrogel size and mineralization, as well as spatial information on hydrogel structure and mineralization. Samples were processed for histological evaluation and XPC showed similar features to histology and quantitative analysis consistent with the histomorphometry. Furthermore, these results provide evidence of the significant potential of techniques based on XPC for noninvasive 3D imaging engineered tissues grown in bioreactors.« less

  2. Distinguishing tracheal and esophageal tissues with hyperspectral imaging and fiber-optic sensing

    NASA Astrophysics Data System (ADS)

    Nawn, Corinne D.; Souhan, Brian E.; Carter, Robert, III; Kneapler, Caitlin; Fell, Nicholas; Ye, Jing Yong

    2016-11-01

    During emergency medical situations, where the patient has an obstructed airway or necessitates respiratory support, endotracheal intubation (ETI) is the medical technique of placing a tube into the trachea in order to facilitate adequate ventilation of the lungs. Complications during ETI, such as repeated attempts, failed intubation, or accidental intubation of the esophagus, can lead to severe consequences or ultimately death. Consequently, a need exists for a feedback mechanism to aid providers in performing successful ETI. Our study examined the spectral reflectance properties of the tracheal and esophageal tissue to determine whether a unique spectral profile exists for either tissue for the purpose of detection. The study began by using a hyperspectral camera to image excised pig tissue samples exposed to white and UV light in order to capture the spectral reflectance properties with high fidelity. After identifying a unique spectral characteristic of the trachea that significantly differed from esophageal tissue, a follow-up investigation used a fiber optic probe to confirm the detectability and consistency of the different reflectance characteristics in a pig model. Our results characterize the unique and consistent spectral reflectance characteristic of tracheal tissue, thereby providing foundational support for exploiting spectral properties to detect the trachea during medical procedures.

  3. Terahertz time-lapse imaging of hydration in physiological tissues

    NASA Astrophysics Data System (ADS)

    Bennett, David B.; Taylor, Zachary D.; Bajwa, Neha; Tewari, Priyamvada; Maccabi, Ashkan; Sung, Shijun; Singh, Rahul S.; Culjat, Martin O.; Grundfest, Warren S.; Brown, Elliott R.

    2011-02-01

    This study describes terahertz (THz) imaging of hydration changes in physiological tissues with high water concentration sensitivity. A fast-scanning, pulsed THz imaging system (centered at 525 GHz; 125 GHz bandwidth) was utilized to acquire a 35 mm x 35 mm field-of-view with 0.5 mm x 0.5 mm pixels in less than two minutes. THz time-lapsed images were taken on three sample systems: (1) a simple binary system of water evaporating from a polypropylene towel, (2) the accumulation of fluid at the site of a sulfuric acid burn on ex vivo porcine skin, and (3) the evaporative dehydration of an ex vivo porcine cornea. The diffusion-regulating behavior of corneal tissue is elucidated, and the correlation of THz reflectivity with tissue hydration is measured using THz spectroscopy on four ex vivo corneas. We conclude that THz imaging can discern small differences in the distribution of water in physiological tissues and is a good candidate for burn and corneal imaging.

  4. Cytopathological image analysis using deep-learning networks in microfluidic microscopy.

    PubMed

    Gopakumar, G; Hari Babu, K; Mishra, Deepak; Gorthi, Sai Siva; Sai Subrahmanyam, Gorthi R K

    2017-01-01

    Cytopathologic testing is one of the most critical steps in the diagnosis of diseases, including cancer. However, the task is laborious and demands skill. Associated high cost and low throughput drew considerable interest in automating the testing process. Several neural network architectures were designed to provide human expertise to machines. In this paper, we explore and propose the feasibility of using deep-learning networks for cytopathologic analysis by performing the classification of three important unlabeled, unstained leukemia cell lines (K562, MOLT, and HL60). The cell images used in the classification are captured using a low-cost, high-throughput cell imaging technique: microfluidics-based imaging flow cytometry. We demonstrate that without any conventional fine segmentation followed by explicit feature extraction, the proposed deep-learning algorithms effectively classify the coarsely localized cell lines. We show that the designed deep belief network as well as the deeply pretrained convolutional neural network outperform the conventionally used decision systems and are important in the medical domain, where the availability of labeled data is limited for training. We hope that our work enables the development of a clinically significant high-throughput microfluidic microscopy-based tool for disease screening/triaging, especially in resource-limited settings.

  5. Magnetoencephalographic imaging of deep corticostriatal network activity during a rewards paradigm.

    PubMed

    Kanal, Eliezer Y; Sun, Mingui; Ozkurt, Tolga E; Jia, Wenyan; Sclabassi, Robert

    2009-01-01

    The human rewards network is a complex system spanning both cortical and subcortical regions. While much is known about the functions of the various components of the network, research on the behavior of the network as a whole has been stymied due to an inability to detect signals at a high enough temporal resolution from both superficial and deep network components simultaneously. In this paper, we describe the application of magnetoencephalographic imaging (MEG) combined with advanced signal processing techniques to this problem. Using data collected while subjects performed a rewards-related gambling paradigm demonstrated to activate the rewards network, we were able to identify neural signals which correspond to deep network activity. We also show that this signal was not observable prior to filtration. These results suggest that MEG imaging may be a viable tool for the detection of deep neural activity.

  6. DeepLab: Semantic Image Segmentation with Deep Convolutional Nets, Atrous Convolution, and Fully Connected CRFs.

    PubMed

    Chen, Liang-Chieh; Papandreou, George; Kokkinos, Iasonas; Murphy, Kevin; Yuille, Alan L

    2018-04-01

    In this work we address the task of semantic image segmentation with Deep Learning and make three main contributions that are experimentally shown to have substantial practical merit. First, we highlight convolution with upsampled filters, or 'atrous convolution', as a powerful tool in dense prediction tasks. Atrous convolution allows us to explicitly control the resolution at which feature responses are computed within Deep Convolutional Neural Networks. It also allows us to effectively enlarge the field of view of filters to incorporate larger context without increasing the number of parameters or the amount of computation. Second, we propose atrous spatial pyramid pooling (ASPP) to robustly segment objects at multiple scales. ASPP probes an incoming convolutional feature layer with filters at multiple sampling rates and effective fields-of-views, thus capturing objects as well as image context at multiple scales. Third, we improve the localization of object boundaries by combining methods from DCNNs and probabilistic graphical models. The commonly deployed combination of max-pooling and downsampling in DCNNs achieves invariance but has a toll on localization accuracy. We overcome this by combining the responses at the final DCNN layer with a fully connected Conditional Random Field (CRF), which is shown both qualitatively and quantitatively to improve localization performance. Our proposed "DeepLab" system sets the new state-of-art at the PASCAL VOC-2012 semantic image segmentation task, reaching 79.7 percent mIOU in the test set, and advances the results on three other datasets: PASCAL-Context, PASCAL-Person-Part, and Cityscapes. All of our code is made publicly available online.

  7. Automated and Adaptable Quantification of Cellular Alignment from Microscopic Images for Tissue Engineering Applications

    PubMed Central

    Xu, Feng; Beyazoglu, Turker; Hefner, Evan; Gurkan, Umut Atakan

    2011-01-01

    Cellular alignment plays a critical role in functional, physical, and biological characteristics of many tissue types, such as muscle, tendon, nerve, and cornea. Current efforts toward regeneration of these tissues include replicating the cellular microenvironment by developing biomaterials that facilitate cellular alignment. To assess the functional effectiveness of the engineered microenvironments, one essential criterion is quantification of cellular alignment. Therefore, there is a need for rapid, accurate, and adaptable methodologies to quantify cellular alignment for tissue engineering applications. To address this need, we developed an automated method, binarization-based extraction of alignment score (BEAS), to determine cell orientation distribution in a wide variety of microscopic images. This method combines a sequenced application of median and band-pass filters, locally adaptive thresholding approaches and image processing techniques. Cellular alignment score is obtained by applying a robust scoring algorithm to the orientation distribution. We validated the BEAS method by comparing the results with the existing approaches reported in literature (i.e., manual, radial fast Fourier transform-radial sum, and gradient based approaches). Validation results indicated that the BEAS method resulted in statistically comparable alignment scores with the manual method (coefficient of determination R2=0.92). Therefore, the BEAS method introduced in this study could enable accurate, convenient, and adaptable evaluation of engineered tissue constructs and biomaterials in terms of cellular alignment and organization. PMID:21370940

  8. A new method for the noninvasive determination of abdominal muscle feedforward activity based on tissue velocity information from tissue Doppler imaging.

    PubMed

    Mannion, A F; Pulkovski, N; Schenk, P; Hodges, P W; Gerber, H; Loupas, T; Gorelick, M; Sprott, H

    2008-04-01

    Rapid arm movements elicit anticipatory activation of the deep-lying abdominal muscles; this appears modified in back pain, but the invasive technique used for its assessment [fine-wire electromyography (EMG)] has precluded its widespread investigation. We examined whether tissue-velocity changes recorded with ultrasound (M-mode) tissue Doppler imaging (TDI) provided a viable noninvasive alternative. Fourteen healthy subjects rapidly flexed, extended, and abducted the shoulder; recordings were made of medial deltoid (MD) surface EMG and of fine-wire EMG and TDI tissue-velocity changes of the contralateral transversus abdominis, obliquus internus, and obliquus externus. Muscle onsets were determined by blinded visual analysis of EMG and TDI data. TDI could not distinguish between the relative activation of the three muscles, so in subsequent analyses only the onset of the earliest abdominal muscle activity was used. The latter occurred <50 ms after the onset of medial deltoid EMG (i.e., was feedforward) and correlated with the corresponding EMG onsets (r = 0.47, P < 0.0001). The mean difference between methods was 20 ms and was likely explained by electromechanical delay; limits of agreement were wide (-40 to +80 ms) but no greater than those typical of repeated measurements using either technique. The between-day standard error of measurement of the TDI onsets (examined in 16 further subjects) was 16 ms. TDI yielded reliable and valid measures of the earliest onset of feedforward activity within the anterolateral abdominal muscle group. The method can be used to assess muscle dysfunction in large groups of back-pain patients and may also be suitable for the noninvasive analysis of other deep-lying or small/thin muscles.

  9. Deep Keck u-Band Imaging of the Hubble Ultra Deep Field: A Catalog of z ~ 3 Lyman Break Galaxies

    NASA Astrophysics Data System (ADS)

    Rafelski, Marc; Wolfe, Arthur M.; Cooke, Jeff; Chen, Hsiao-Wen; Armandroff, Taft E.; Wirth, Gregory D.

    2009-10-01

    We present a sample of 407 z ~ 3 Lyman break galaxies (LBGs) to a limiting isophotal u-band magnitude of 27.6 mag in the Hubble Ultra Deep Field. The LBGs are selected using a combination of photometric redshifts and the u-band drop-out technique enabled by the introduction of an extremely deep u-band image obtained with the Keck I telescope and the blue channel of the Low Resolution Imaging Spectrometer. The Keck u-band image, totaling 9 hr of integration time, has a 1σ depth of 30.7 mag arcsec-2, making it one of the most sensitive u-band images ever obtained. The u-band image also substantially improves the accuracy of photometric redshift measurements of ~50% of the z ~ 3 LBGs, significantly reducing the traditional degeneracy of colors between z ~ 3 and z ~ 0.2 galaxies. This sample provides the most sensitive, high-resolution multi-filter imaging of reliably identified z ~ 3 LBGs for morphological studies of galaxy formation and evolution and the star formation efficiency of gas at high redshift.

  10. Detection of Thermal Erosion Gullies from High-Resolution Images Using Deep Learning

    NASA Astrophysics Data System (ADS)

    Huang, L.; Liu, L.; Jiang, L.; Zhang, T.; Sun, Y.

    2017-12-01

    Thermal erosion gullies, one type of thermokarst landforms, develop due to thawing of ice-rich permafrost. Mapping the location and extent of thermal erosion gullies can help understand the spatial distribution of thermokarst landforms and their temporal evolution. Remote sensing images provide an effective way for mapping thermokarst landforms, especially thermokarst lakes. However, thermal erosion gullies are challenging to map from remote sensing images due to their small sizes and significant variations in geometric/radiometric properties. It is feasible to manually identify these features, as a few previous studies have carried out. However manual methods are labor-intensive, therefore, cannot be used for a large study area. In this work, we conduct automatic mapping of thermal erosion gullies from high-resolution images by using Deep Learning. Our study area is located in Eboling Mountain (Qinghai, China). Within a 6 km2 peatland area underlain by ice-rich permafrost, at least 20 thermal erosional gullies are well developed. The image used is a 15-cm-resolution Digital Orthophoto Map (DOM) generated in July 2016. First, we extracted 14 gully patches and ten non-gully patches as training data. And we performed image augmentation. Next, we fine-tuned the pre-trained model of DeepLab, a deep-learning algorithm for semantic image segmentation based on Deep Convolutional Neural Networks. Then, we performed inference on the whole DOM and obtained intermediate results in forms of polygons for all identified gullies. At last, we removed misidentified polygons based on a few pre-set criteria on the size and shape of each polygon. Our final results include 42 polygons. Validated against field measurements using GPS, most of the gullies are detected correctly. There are 20 false detections due to the small number and low quality of training images. We also found three new gullies that missed in the field observations. This study shows that (1) despite a challenging

  11. Automatic bladder segmentation from CT images using deep CNN and 3D fully connected CRF-RNN.

    PubMed

    Xu, Xuanang; Zhou, Fugen; Liu, Bo

    2018-03-19

    Automatic approach for bladder segmentation from computed tomography (CT) images is highly desirable in clinical practice. It is a challenging task since the bladder usually suffers large variations of appearance and low soft-tissue contrast in CT images. In this study, we present a deep learning-based approach which involves a convolutional neural network (CNN) and a 3D fully connected conditional random fields recurrent neural network (CRF-RNN) to perform accurate bladder segmentation. We also propose a novel preprocessing method, called dual-channel preprocessing, to further advance the segmentation performance of our approach. The presented approach works as following: first, we apply our proposed preprocessing method on the input CT image and obtain a dual-channel image which consists of the CT image and an enhanced bladder density map. Second, we exploit a CNN to predict a coarse voxel-wise bladder score map on this dual-channel image. Finally, a 3D fully connected CRF-RNN refines the coarse bladder score map and produce final fine-localized segmentation result. We compare our approach to the state-of-the-art V-net on a clinical dataset. Results show that our approach achieves superior segmentation accuracy, outperforming the V-net by a significant margin. The Dice Similarity Coefficient of our approach (92.24%) is 8.12% higher than that of the V-net. Moreover, the bladder probability maps performed by our approach present sharper boundaries and more accurate localizations compared with that of the V-net. Our approach achieves higher segmentation accuracy than the state-of-the-art method on clinical data. Both the dual-channel processing and the 3D fully connected CRF-RNN contribute to this improvement. The united deep network composed of the CNN and 3D CRF-RNN also outperforms a system where the CRF model acts as a post-processing method disconnected from the CNN.

  12. Deep linear autoencoder and patch clustering-based unified one-dimensional coding of image and video

    NASA Astrophysics Data System (ADS)

    Li, Honggui

    2017-09-01

    This paper proposes a unified one-dimensional (1-D) coding framework of image and video, which depends on deep learning neural network and image patch clustering. First, an improved K-means clustering algorithm for image patches is employed to obtain the compact inputs of deep artificial neural network. Second, for the purpose of best reconstructing original image patches, deep linear autoencoder (DLA), a linear version of the classical deep nonlinear autoencoder, is introduced to achieve the 1-D representation of image blocks. Under the circumstances of 1-D representation, DLA is capable of attaining zero reconstruction error, which is impossible for the classical nonlinear dimensionality reduction methods. Third, a unified 1-D coding infrastructure for image, intraframe, interframe, multiview video, three-dimensional (3-D) video, and multiview 3-D video is built by incorporating different categories of videos into the inputs of patch clustering algorithm. Finally, it is shown in the results of simulation experiments that the proposed methods can simultaneously gain higher compression ratio and peak signal-to-noise ratio than those of the state-of-the-art methods in the situation of low bitrate transmission.

  13. Wide-field high spatial frequency domain imaging of tissue microstructure

    NASA Astrophysics Data System (ADS)

    Lin, Weihao; Zeng, Bixin; Cao, Zili; Zhu, Danfeng; Xu, M.

    2018-02-01

    Wide-field tissue imaging is usually not capable of resolving tissue microstructure. We present High Spatial Frequency Domain Imaging (HSFDI) - a noncontact imaging modality that spatially maps the tissue microscopic scattering structures over a large field of view. Based on an analytical reflectance model of sub-diffusive light from forward-peaked highly scattering media, HSFDI quantifies the spatially-resolved parameters of the light scattering phase function from the reflectance of structured light modulated at high spatial frequencies. We have demonstrated with ex vivo cancerous tissue to validate the robustness of HSFDI in significant contrast and differentiation of the microstructutral parameters between different types and disease states of tissue.

  14. A deep ALMA image of the Hubble Ultra Deep Field

    NASA Astrophysics Data System (ADS)

    Dunlop, J. S.; McLure, R. J.; Biggs, A. D.; Geach, J. E.; Michałowski, M. J.; Ivison, R. J.; Rujopakarn, W.; van Kampen, E.; Kirkpatrick, A.; Pope, A.; Scott, D.; Swinbank, A. M.; Targett, T. A.; Aretxaga, I.; Austermann, J. E.; Best, P. N.; Bruce, V. A.; Chapin, E. L.; Charlot, S.; Cirasuolo, M.; Coppin, K.; Ellis, R. S.; Finkelstein, S. L.; Hayward, C. C.; Hughes, D. H.; Ibar, E.; Jagannathan, P.; Khochfar, S.; Koprowski, M. P.; Narayanan, D.; Nyland, K.; Papovich, C.; Peacock, J. A.; Rieke, G. H.; Robertson, B.; Vernstrom, T.; Werf, P. P. van der; Wilson, G. W.; Yun, M.

    2017-04-01

    We present the results of the first, deep Atacama Large Millimeter Array (ALMA) imaging covering the full ≃4.5 arcmin2 of the Hubble Ultra Deep Field (HUDF) imaged with Wide Field Camera 3/IR on HST. Using a 45-pointing mosaic, we have obtained a homogeneous 1.3-mm image reaching σ1.3 ≃ 35 μJy, at a resolution of ≃0.7 arcsec. From an initial list of ≃50 > 3.5σ peaks, a rigorous analysis confirms 16 sources with S1.3 > 120 μJy. All of these have secure galaxy counterparts with robust redshifts ( = 2.15). Due to the unparalleled supporting data, the physical properties of the ALMA sources are well constrained, including their stellar masses (M*) and UV+FIR star formation rates (SFR). Our results show that stellar mass is the best predictor of SFR in the high-redshift Universe; indeed at z ≥ 2 our ALMA sample contains seven of the nine galaxies in the HUDF with M* ≥ 2 × 1010 M⊙, and we detect only one galaxy at z > 3.5, reflecting the rapid drop-off of high-mass galaxies with increasing redshift. The detections, coupled with stacking, allow us to probe the redshift/mass distribution of the 1.3-mm background down to S1.3 ≃ 10 μJy. We find strong evidence for a steep star-forming 'main sequence' at z ≃ 2, with SFR ∝M* and a mean specific SFR ≃ 2.2 Gyr-1. Moreover, we find that ≃85 per cent of total star formation at z ≃ 2 is enshrouded in dust, with ≃65 per cent of all star formation at this epoch occurring in high-mass galaxies (M* > 2 × 1010 M⊙), for which the average obscured:unobscured SF ratio is ≃200. Finally, we revisit the cosmic evolution of SFR density; we find this peaks at z ≃ 2.5, and that the star-forming Universe transits from primarily unobscured to primarily obscured at z ≃ 4.

  15. Dual-mode imaging with radiolabeled gold nanorods

    NASA Astrophysics Data System (ADS)

    Agarwal, Ashish; Shao, Xia; Rajian, Justin R.; Zhang, Huanan; Chamberland, David L.; Kotov, Nicholas A.; Wang, Xueding

    2011-05-01

    Many nanoparticle contrast agents have difficulties with deep tissue and near-bone imaging due to limited penetration of visible photons in the body and mineralized tissues. We are looking into the possibility of mediating this problem while retaining the capabilities of the high spatial resolution associated with optical imaging. As such, the potential combination of emerging photoacoustic imaging and nuclear imaging in monitoring of antirheumatic drug delivery by using a newly developed dual-modality contrast agent is investigated. The contrast agent is composed of gold nanorods (GNRs) conjugated to the tumor necrosis factor (TNF-α) antibody and is subsequently radiolabeled by 125I. ELISA experiments designed to test TNF-α binding are performed to prove the specificity and biological activity of the radiolabeled conjugated contrast agent. Photoacoustic and nuclear imaging are performed to visualize the distribution of GNRs in articular tissues of the rat tail joints in situ. Findings from the two imaging modalities correspond well with each other in all experiments. Our system can image GNRs down to a concentration of 10 pM in biological tissues and with a radioactive label of 5 μCi. This study demonstrates the potential of combining photoacoustic and nuclear imaging modalities through one targeted contrast agent for noninvasive monitoring of drug delivery as well as deep and mineralized tissue imaging.

  16. A high-resolution optical imaging system for obtaining the serial transverse section images of biologic tissue

    NASA Astrophysics Data System (ADS)

    Wu, Li; Zhang, Bin; Wu, Ping; Liu, Qian; Gong, Hui

    2007-05-01

    A high-resolution optical imaging system was designed and developed to obtain the serial transverse section images of the biologic tissue, such as the mouse brain, in which new knife-edge imaging technology, high-speed and high-sensitive line-scan CCD and linear air bearing stages were adopted and incorporated with an OLYMPUS microscope. The section images on the tip of the knife-edge were synchronously captured by the reflection imaging in the microscope while cutting the biologic tissue. The biologic tissue can be sectioned at interval of 250 nm with the same resolution of the transverse section images obtained in x and y plane. And the cutting job can be automatically finished based on the control program wrote specially in advance, so we save the mass labor of the registration of the vast images data. In addition, by using this system a larger sample can be cut than conventional ultramicrotome so as to avoid the loss of the tissue structure information because of splitting the tissue sample to meet the size request of the ultramicrotome.

  17. Classifying Acute Ischemic Stroke Onset Time using Deep Imaging Features

    PubMed Central

    Ho, King Chung; Speier, William; El-Saden, Suzie; Arnold, Corey W.

    2017-01-01

    Models have been developed to predict stroke outcomes (e.g., mortality) in attempt to provide better guidance for stroke treatment. However, there is little work in developing classification models for the problem of unknown time-since-stroke (TSS), which determines a patient’s treatment eligibility based on a clinical defined cutoff time point (i.e., <4.5hrs). In this paper, we construct and compare machine learning methods to classify TSS<4.5hrs using magnetic resonance (MR) imaging features. We also propose a deep learning model to extract hidden representations from the MR perfusion-weighted images and demonstrate classification improvement by incorporating these additional imaging features. Finally, we discuss a strategy to visualize the learned features from the proposed deep learning model. The cross-validation results show that our best classifier achieved an area under the curve of 0.68, which improves significantly over current clinical methods (0.58), demonstrating the potential benefit of using advanced machine learning methods in TSS classification. PMID:29854156

  18. Deep Space Gateway Facilitates Exploration of Planetary Crusts: A Human/Robotic Exploration Design Reference Campaign to the Lunar Orientale Basin

    NASA Astrophysics Data System (ADS)

    Head, J. W.; Pieters, C. M.; Scott, D. R.

    2018-02-01

    We outline an Orientale Basin Human/Robotic Architecture that can be facilitated by a Deep Space Gateway International Science Operations Center (DSG-ISOC) (like McMurdo/Antarctica) to address fundamental scientific problems about the Moon and Mars.

  19. Nanoparticle-assisted-multiphoton microscopy for in vivo brain imaging of mice

    NASA Astrophysics Data System (ADS)

    Qian, Jun

    2015-03-01

    Neuro/brain study has attracted much attention during past few years, and many optical methods have been utilized in order to obtain accurate and complete neural information inside the brain. Relying on simultaneous absorption of two or more near-infrared photons by a fluorophore, multiphoton microscopy can achieve deep tissue penetration and efficient light detection noninvasively, which makes it very suitable for thick-tissue and in vivo bioimaging. Nanoparticles possess many unique optical and chemical properties, such as anti-photobleaching, large multiphoton absorption cross-section, and high stability in biological environment, which facilitates their applications in long-term multiphoton microscopy as contrast agents. In this paper, we will introduce several typical nanoparticles (e.g. organic dye doped polymer nanoparticles and gold nanorods) with high multiphoton fluorescence efficiency. We further applied them in two- and three-photon in vivo functional brain imaging of mice, such as brain-microglia imaging, 3D architecture reconstruction of brain blood vessel, and blood velocity measurement.

  20. A survey of clearing techniques for 3D imaging of tissues with special reference to connective tissue.

    PubMed

    Azaripour, Adriano; Lagerweij, Tonny; Scharfbillig, Christina; Jadczak, Anna Elisabeth; Willershausen, Brita; Van Noorden, Cornelis J F

    2016-08-01

    For 3-dimensional (3D) imaging of a tissue, 3 methodological steps are essential and their successful application depends on specific characteristics of the type of tissue. The steps are 1° clearing of the opaque tissue to render it transparent for microscopy, 2° fluorescence labeling of the tissues and 3° 3D imaging. In the past decades, new methodologies were introduced for the clearing steps with their specific advantages and disadvantages. Most clearing techniques have been applied to the central nervous system and other organs that contain relatively low amounts of connective tissue including extracellular matrix. However, tissues that contain large amounts of extracellular matrix such as dermis in skin or gingiva are difficult to clear. The present survey lists methodologies that are available for clearing of tissues for 3D imaging. We report here that the BABB method using a mixture of benzyl alcohol and benzyl benzoate and iDISCO using dibenzylether (DBE) are the most successful methods for clearing connective tissue-rich gingiva and dermis of skin for 3D histochemistry and imaging of fluorescence using light-sheet microscopy. Copyright © 2016 The Authors. Published by Elsevier GmbH.. All rights reserved.

  1. Differentiating fatty and non-fatty tissue using photoacoustic imaging

    NASA Astrophysics Data System (ADS)

    Pan, Leo; Rohling, Robert; Abolmaesumi, Purang; Salcudean, Septimiu; Tang, Shuo

    2014-03-01

    In this paper, we demonstrate a temporal-domain intensity-based photoacoustic imaging method that can differentiate between fatty and non-fatty tissues. PA pressure intensity is partly dependent on the tissue's speed of sound, which increases as temperature increases in non-fatty tissue and decreases in fatty tissue. Therefore, by introducing a temperature change in the tissue and subsequently monitoring the change of the PA intensity, it is possible to distinguish between the two types of tissue. A commercial ultrasound system with a 128-element 5-14 MHz linear array transducer and a tunable ND:YAG laser were used to produce PA images. Ex-vivo bovine fat and porcine liver tissues were precooled to below 10°C and then warmed to room-temperature over ~1 hour period. A thermocouple monitored the temperature rise while PA images were acquired at 0.5°C intervals. The averaged intensity of the illuminated tissue region at each temperature interval was plotted and linearly fitted. Liver samples showed a mean increase of 2.82 %/°C, whereas bovine fat had a mean decrease of 6.24 %/°C. These results demonstrate that this method has the potential to perform tissue differentiation in the temporal-domain.

  2. Real-time photoacoustic and ultrasound dual-modality imaging system facilitated with graphics processing unit and code parallel optimization.

    PubMed

    Yuan, Jie; Xu, Guan; Yu, Yao; Zhou, Yu; Carson, Paul L; Wang, Xueding; Liu, Xiaojun

    2013-08-01

    Photoacoustic tomography (PAT) offers structural and functional imaging of living biological tissue with highly sensitive optical absorption contrast and excellent spatial resolution comparable to medical ultrasound (US) imaging. We report the development of a fully integrated PAT and US dual-modality imaging system, which performs signal scanning, image reconstruction, and display for both photoacoustic (PA) and US imaging all in a truly real-time manner. The back-projection (BP) algorithm for PA image reconstruction is optimized to reduce the computational cost and facilitate parallel computation on a state of the art graphics processing unit (GPU) card. For the first time, PAT and US imaging of the same object can be conducted simultaneously and continuously, at a real-time frame rate, presently limited by the laser repetition rate of 10 Hz. Noninvasive PAT and US imaging of human peripheral joints in vivo were achieved, demonstrating the satisfactory image quality realized with this system. Another experiment, simultaneous PAT and US imaging of contrast agent flowing through an artificial vessel, was conducted to verify the performance of this system for imaging fast biological events. The GPU-based image reconstruction software code for this dual-modality system is open source and available for download from http://sourceforge.net/projects/patrealtime.

  3. Aging in deep gray matter and white matter revealed by diffusional kurtosis imaging.

    PubMed

    Gong, Nan-Jie; Wong, Chun-Sing; Chan, Chun-Chung; Leung, Lam-Ming; Chu, Yiu-Ching

    2014-10-01

    Diffusion tensor imaging has already been extensively used to probe microstructural alterations in white matter tracts, and scarcely, in deep gray matter. However, results in literature regarding age-related degenerative mechanisms in white matter tracts and parametric changes in the putamen are inconsistent. Diffusional kurtosis imaging is a mathematical extension of diffusion tensor imaging, which could more comprehensively mirror microstructure, particularly in isotropic tissues such as gray matter. In this study, we used the diffusional kurtosis imaging method and a white-matter model that provided metrics of explicit neurobiological interpretations in healthy participants (58 in total, aged from 25 to 84 years). Tract-based whole-brain analyses and regions-of-interest (anterior and posterior limbs of the internal capsule, cerebral peduncle, fornix, genu and splenium of corpus callosum, globus pallidus, substantia nigra, red nucleus, putamen, caudate nucleus, and thalamus) analyses were performed to examine parametric differences across regions and correlations with age. In white matter tracts, evidence was found supportive for anterior-posterior gradient and not completely supportive for retrogenesis theory. Age-related degenerations appeared to be broadly driven by axonal loss. Demyelination may also be a major driving mechanism, although confined to the anterior brain. In terms of deep gray matter, higher mean kurtosis and fractional anisotropy in the globus pallidus, substantia nigra, and red nucleus reflected higher microstructural complexity and directionality compared with the putamen, caudate nucleus, and thalamus. In particular, the unique age-related positive correlations for fractional anisotropy, mean kurtosis, and radial kurtosis in the putamen opposite to those in other regions call for further investigation of exact underlying mechanisms. In summary, the results suggested that diffusional kurtosis can provide measurements in a new dimension that

  4. Efficient generation of image chips for training deep learning algorithms

    NASA Astrophysics Data System (ADS)

    Han, Sanghui; Fafard, Alex; Kerekes, John; Gartley, Michael; Ientilucci, Emmett; Savakis, Andreas; Law, Charles; Parhan, Jason; Turek, Matt; Fieldhouse, Keith; Rovito, Todd

    2017-05-01

    Training deep convolutional networks for satellite or aerial image analysis often requires a large amount of training data. For a more robust algorithm, training data need to have variations not only in the background and target, but also radiometric variations in the image such as shadowing, illumination changes, atmospheric conditions, and imaging platforms with different collection geometry. Data augmentation is a commonly used approach to generating additional training data. However, this approach is often insufficient in accounting for real world changes in lighting, location or viewpoint outside of the collection geometry. Alternatively, image simulation can be an efficient way to augment training data that incorporates all these variations, such as changing backgrounds, that may be encountered in real data. The Digital Imaging and Remote Sensing Image Image Generation (DIRSIG) model is a tool that produces synthetic imagery using a suite of physics-based radiation propagation modules. DIRSIG can simulate images taken from different sensors with variation in collection geometry, spectral response, solar elevation and angle, atmospheric models, target, and background. Simulation of Urban Mobility (SUMO) is a multi-modal traffic simulation tool that explicitly models vehicles that move through a given road network. The output of the SUMO model was incorporated into DIRSIG to generate scenes with moving vehicles. The same approach was used when using helicopters as targets, but with slight modifications. Using the combination of DIRSIG and SUMO, we quickly generated many small images, with the target at the center with different backgrounds. The simulations generated images with vehicles and helicopters as targets, and corresponding images without targets. Using parallel computing, 120,000 training images were generated in about an hour. Some preliminary results show an improvement in the deep learning algorithm when real image training data are augmented with

  5. Pitfalls in soft tissue sarcoma imaging: chronic expanding hematomas.

    PubMed

    Jahed, Kiarash; Khazai, Behnaz; Umpierrez, Monica; Subhawong, Ty K; Singer, Adam D

    2018-01-01

    Solid or nodular enhancement is typical of soft tissue sarcomas although high grade soft tissue sarcomas and those with internal hemorrhage often appear heterogeneous with areas of nonenhancement and solid or nodular enhancement. These MRI findings often prompt an orthopedic oncology referral, a biopsy or surgery. However, not all masses with these imaging findings are malignant. We report the multimodality imaging findings of two surgically proven chronic expanding hematomas (CEH) with imaging features that mimicked sarcomas. A third case of nonenhancing CEH of the lower extremity is also presented as a comparison. It is important that in the correct clinical scenario with typical imaging findings, the differential diagnosis of a chronic expanding hematoma be included in the workup of these patients. An image-guided biopsy of nodular tissue within such masses that proves to be negative for malignancy should not necessarily be considered discordant. A correct diagnosis may prevent a morbid unnecessary surgery and may indicate the need for a conservative noninvasive follow-up with imaging.

  6. Imaging of single cells and tissue using MeV ions

    NASA Astrophysics Data System (ADS)

    Watt, F.; Bettiol, A. A.; van Kan, J. A.; Ynsa, M. D.; Minqin, Ren; Rajendran, R.; Huifang, Cui; Fwu-Shen, Sheu; Jenner, A. M.

    2009-06-01

    With the attainment of sub-100 nm high energy (MeV) ion beams, comes the opportunity to image cells and tissue at nano-dimensions. The advantage of MeV ion imaging is that the ions will penetrate whole cells, or relatively thick tissue sections, without any significant loss of resolution. In this paper, we demonstrate that whole cells (cultured N2A neuroblastoma cells ATCC) and tissue sections (rabbit pancreas tissue) can be imaged at sub-100 nm resolutions using scanning transmission ion microscopy (STIM), and that sub-cellular structural details can be identified. In addition to STIM imaging we have also demonstrated for the first time, that sub-cellular proton induced fluorescence imaging (on cultured N2A neuroblastoma cells ATCC) can also be carried out at resolutions of 200 nm, compared with 300-400 nm resolutions achieved by conventional optical fluorescence imaging. The combination of both techniques offers a potentially powerful tool in the quest for elucidating cell function, particularly when it should be possible in the near future to image down to sub-50 nm.

  7. Hemodynamic measurements in deep brain tissues of humans by near-infrared time-resolved spectroscopy

    NASA Astrophysics Data System (ADS)

    Suzuki, Hiroaki; Oda, Motoki; Yamaki, Etsuko; Suzuki, Toshihiko; Yamashita, Daisuke; Yoshimoto, Kenji; Homma, Shu; Yamashita, Yutaka

    2014-03-01

    Using near-infrared time-resolved spectroscopy (TRS), we measured the human head in transmittance mode to obtain the optical properties, tissue oxygenation, and hemodynamics of deep brain tissues in 50 healthy adult volunteers. The right ear canal was irradiated with 3-wavelengths of pulsed light (760, 795, and 835nm), and the photons passing through the human head were collected at the left ear canal. Optical signals with sufficient intensity could be obtained from 46 of the 50 volunteers. By analyzing the temporal profiles based on the photon diffusion theory, we successfully obtained absorption coefficients for each wavelength. The levels of oxygenated hemoglobin (HbO2), deoxygenated hemoglobin (Hb), total hemoglobin (tHb), and tissue oxygen saturation (SO2) were then determined by referring to the hemoglobin spectroscopic data. Compared with the SO2 values for the forehead measurements in reflectance mode, the SO2 values of the transmittance measurements of the human head were approximately 10% lower, and tHb values of the transmittance measurements were always lower than those of the forehead reflectance measurements. Moreover, the level of hemoglobin and the SO2 were strongly correlated between the human head measurements in transmittance mode and the forehead measurements in the reflectance mode, respectively. These results demonstrated a potential application of this TRS system in examining deep brain tissues of humans.

  8. Computer-aided classification of lung nodules on computed tomography images via deep learning technique

    PubMed Central

    Hua, Kai-Lung; Hsu, Che-Hao; Hidayati, Shintami Chusnul; Cheng, Wen-Huang; Chen, Yu-Jen

    2015-01-01

    Lung cancer has a poor prognosis when not diagnosed early and unresectable lesions are present. The management of small lung nodules noted on computed tomography scan is controversial due to uncertain tumor characteristics. A conventional computer-aided diagnosis (CAD) scheme requires several image processing and pattern recognition steps to accomplish a quantitative tumor differentiation result. In such an ad hoc image analysis pipeline, every step depends heavily on the performance of the previous step. Accordingly, tuning of classification performance in a conventional CAD scheme is very complicated and arduous. Deep learning techniques, on the other hand, have the intrinsic advantage of an automatic exploitation feature and tuning of performance in a seamless fashion. In this study, we attempted to simplify the image analysis pipeline of conventional CAD with deep learning techniques. Specifically, we introduced models of a deep belief network and a convolutional neural network in the context of nodule classification in computed tomography images. Two baseline methods with feature computing steps were implemented for comparison. The experimental results suggest that deep learning methods could achieve better discriminative results and hold promise in the CAD application domain. PMID:26346558

  9. Computer-aided classification of lung nodules on computed tomography images via deep learning technique.

    PubMed

    Hua, Kai-Lung; Hsu, Che-Hao; Hidayati, Shintami Chusnul; Cheng, Wen-Huang; Chen, Yu-Jen

    2015-01-01

    Lung cancer has a poor prognosis when not diagnosed early and unresectable lesions are present. The management of small lung nodules noted on computed tomography scan is controversial due to uncertain tumor characteristics. A conventional computer-aided diagnosis (CAD) scheme requires several image processing and pattern recognition steps to accomplish a quantitative tumor differentiation result. In such an ad hoc image analysis pipeline, every step depends heavily on the performance of the previous step. Accordingly, tuning of classification performance in a conventional CAD scheme is very complicated and arduous. Deep learning techniques, on the other hand, have the intrinsic advantage of an automatic exploitation feature and tuning of performance in a seamless fashion. In this study, we attempted to simplify the image analysis pipeline of conventional CAD with deep learning techniques. Specifically, we introduced models of a deep belief network and a convolutional neural network in the context of nodule classification in computed tomography images. Two baseline methods with feature computing steps were implemented for comparison. The experimental results suggest that deep learning methods could achieve better discriminative results and hold promise in the CAD application domain.

  10. Pattern Analysis of Dynamic Susceptibility Contrast-enhanced MR Imaging Demonstrates Peritumoral Tissue Heterogeneity

    PubMed Central

    Akbari, Hamed; Macyszyn, Luke; Da, Xiao; Wolf, Ronald L.; Bilello, Michel; Verma, Ragini; O’Rourke, Donald M.

    2014-01-01

    Purpose To augment the analysis of dynamic susceptibility contrast material–enhanced magnetic resonance (MR) images to uncover unique tissue characteristics that could potentially facilitate treatment planning through a better understanding of the peritumoral region in patients with glioblastoma. Materials and Methods Institutional review board approval was obtained for this study, with waiver of informed consent for retrospective review of medical records. Dynamic susceptibility contrast-enhanced MR imaging data were obtained for 79 patients, and principal component analysis was applied to the perfusion signal intensity. The first six principal components were sufficient to characterize more than 99% of variance in the temporal dynamics of blood perfusion in all regions of interest. The principal components were subsequently used in conjunction with a support vector machine classifier to create a map of heterogeneity within the peritumoral region, and the variance of this map served as the heterogeneity score. Results The calculated principal components allowed near-perfect separability of tissue that was likely highly infiltrated with tumor and tissue that was unlikely infiltrated with tumor. The heterogeneity map created by using the principal components showed a clear relationship between voxels judged by the support vector machine to be highly infiltrated and subsequent recurrence. The results demonstrated a significant correlation (r = 0.46, P < .0001) between the heterogeneity score and patient survival. The hazard ratio was 2.23 (95% confidence interval: 1.4, 3.6; P < .01) between patients with high and low heterogeneity scores on the basis of the median heterogeneity score. Conclusion Analysis of dynamic susceptibility contrast-enhanced MR imaging data by using principal component analysis can help identify imaging variables that can be subsequently used to evaluate the peritumoral region in glioblastoma. These variables are potentially indicative of

  11. Deep machine learning provides state-of-the-art performance in image-based plant phenotyping.

    PubMed

    Pound, Michael P; Atkinson, Jonathan A; Townsend, Alexandra J; Wilson, Michael H; Griffiths, Marcus; Jackson, Aaron S; Bulat, Adrian; Tzimiropoulos, Georgios; Wells, Darren M; Murchie, Erik H; Pridmore, Tony P; French, Andrew P

    2017-10-01

    In plant phenotyping, it has become important to be able to measure many features on large image sets in order to aid genetic discovery. The size of the datasets, now often captured robotically, often precludes manual inspection, hence the motivation for finding a fully automated approach. Deep learning is an emerging field that promises unparalleled results on many data analysis problems. Building on artificial neural networks, deep approaches have many more hidden layers in the network, and hence have greater discriminative and predictive power. We demonstrate the use of such approaches as part of a plant phenotyping pipeline. We show the success offered by such techniques when applied to the challenging problem of image-based plant phenotyping and demonstrate state-of-the-art results (>97% accuracy) for root and shoot feature identification and localization. We use fully automated trait identification using deep learning to identify quantitative trait loci in root architecture datasets. The majority (12 out of 14) of manually identified quantitative trait loci were also discovered using our automated approach based on deep learning detection to locate plant features. We have shown deep learning-based phenotyping to have very good detection and localization accuracy in validation and testing image sets. We have shown that such features can be used to derive meaningful biological traits, which in turn can be used in quantitative trait loci discovery pipelines. This process can be completely automated. We predict a paradigm shift in image-based phenotyping bought about by such deep learning approaches, given sufficient training sets. © The Authors 2017. Published by Oxford University Press.

  12. Instrumentation and method for measuring NIR light absorbed in tissue during MR imaging in medical NIRS measurements

    NASA Astrophysics Data System (ADS)

    Myllylä, Teemu S.; Sorvoja, Hannu S. S.; Nikkinen, Juha; Tervonen, Osmo; Kiviniemi, Vesa; Myllylä, Risto A.

    2011-07-01

    Our goal is to provide a cost-effective method for examining human tissue, particularly the brain, by the simultaneous use of functional magnetic resonance imaging (fMRI) and near-infrared spectroscopy (NIRS). Due to its compatibility requirements, MRI poses a demanding challenge for NIRS measurements. This paper focuses particularly on presenting the instrumentation and a method for the non-invasive measurement of NIR light absorbed in human tissue during MR imaging. One practical method to avoid disturbances in MR imaging involves using long fibre bundles to enable conducting the measurements at some distance from the MRI scanner. This setup serves in fact a dual purpose, since also the NIRS device will be less disturbed by the MRI scanner. However, measurements based on long fibre bundles suffer from light attenuation. Furthermore, because one of our primary goals was to make the measuring method as cost-effective as possible, we used high-power light emitting diodes instead of more expensive lasers. The use of LEDs, however, limits the maximum output power which can be extracted to illuminate the tissue. To meet these requirements, we improved methods of emitting light sufficiently deep into tissue. We also show how to measure NIR light of a very small power level that scatters from the tissue in the MRI environment, which is characterized by strong electromagnetic interference. In this paper, we present the implemented instrumentation and measuring method and report on test measurements conducted during MRI scanning. These measurements were performed in MRI operating rooms housing 1.5 Tesla-strength closed MRI scanners (manufactured by GE) in the Dept. of Diagnostic Radiology at the Oulu University Hospital.

  13. AEGIS-X: Deep Chandra Imaging of the Central Groth Strip

    NASA Astrophysics Data System (ADS)

    Nandra, K.; Laird, E. S.; Aird, J. A.; Salvato, M.; Georgakakis, A.; Barro, G.; Perez-Gonzalez, P. G.; Barmby, P.; Chary, R.-R.; Coil, A.; Cooper, M. C.; Davis, M.; Dickinson, M.; Faber, S. M.; Fazio, G. G.; Guhathakurta, P.; Gwyn, S.; Hsu, L.-T.; Huang, J.-S.; Ivison, R. J.; Koo, D. C.; Newman, J. A.; Rangel, C.; Yamada, T.; Willmer, C.

    2015-09-01

    We present the results of deep Chandra imaging of the central region of the Extended Groth Strip, the AEGIS-X Deep (AEGIS-XD) survey. When combined with previous Chandra observations of a wider area of the strip, AEGIS-X Wide (AEGIS-XW), these provide data to a nominal exposure depth of 800 ks in the three central ACIS-I fields, a region of approximately 0.29 deg2. This is currently the third deepest X-ray survey in existence; a factor ∼ 2-3 shallower than the Chandra Deep Fields (CDFs), but over an area ∼3 times greater than each CDF. We present a catalog of 937 point sources detected in the deep Chandra observations, along with identifications of our X-ray sources from deep ground-based, Spitzer, GALEX, and Hubble Space Telescope imaging. Using a likelihood ratio analysis, we associate multiband counterparts for 929/937 of our X-ray sources, with an estimated 95% reliability, making the identification completeness approximately 94% in a statistical sense. Reliable spectroscopic redshifts for 353 of our X-ray sources are available predominantly from Keck (DEEP2/3) and MMT Hectospec, so the current spectroscopic completeness is ∼38%. For the remainder of the X-ray sources, we compute photometric redshifts based on multiband photometry in up to 35 bands from the UV to mid-IR. Particular attention is given to the fact that the vast majority the X-ray sources are active galactic nuclei and require hybrid templates. Our photometric redshifts have mean accuracy of σ =0.04 and an outlier fraction of approximately 5%, reaching σ =0.03 with less than 4% outliers in the area covered by CANDELS . The X-ray, multiwavelength photometry, and redshift catalogs are made publicly available.

  14. Hadamard-Encoded Multipulses for Contrast-Enhanced Ultrasound Imaging.

    PubMed

    Gong, Ping; Song, Pengfei; Chen, Shigao

    2017-11-01

    The development of contrast-enhanced ultrasound (CEUS) imaging offers great opportunities for new ultrasound clinical applications such as myocardial perfusion imaging and abdominal lesion characterization. In CEUS imaging, the contrast agents (i.e., microbubbles) are utilized to improve the contrast between blood and tissue based on their high nonlinearity under low ultrasound pressure. In this paper, we propose a new CEUS pulse sequence by combining Hadamard-encoded multipulses (HEM) with fundamental frequency bandpass filter (i.e., filter centered on transmit frequency). HEM consecutively emits multipulses encoded by a second-order Hadamard matrix in each of the two transmission events (i.e., pulse-echo events), as opposed to conventional CEUS methods which emit individual pulses in two separate transmission events (i.e., pulse inversion (PI), amplitude modulation (AM), and PIAM). In HEM imaging, the microbubble responses can be improved by the longer transmit pulse, and the tissue harmonics can be suppressed by the fundamental frequency filter, leading to significantly improved contrast-to-tissue ratio (CTR) and signal-to-noise ratio (SNR). In addition, the fast polarity change between consecutive coded pulse emissions excites strong nonlinear microbubble echoes, further enhancing the CEUS image quality. The spatial resolution of HEM image is compromised as compared to other microbubble imaging methods due to the longer transmit pulses and the lower imaging frequency (i.e., fundamental frequency). However, the resolution loss was shown to be negligible and could be offset by the significantly enhanced CTR, SNR, and penetration depth. These properties of HEM can potentially facilitate robust CEUS imaging for many clinical applications, especially for deep abdominal organs and heart.

  15. The effect of deep-tissue massage therapy on blood pressure and heart rate.

    PubMed

    Kaye, Alan David; Kaye, Aaron J; Swinford, Jan; Baluch, Amir; Bawcom, Brad A; Lambert, Thomas J; Hoover, Jason M

    2008-03-01

    In the present study, we describe the effects of deep tissue massage on systolic, diastolic, and mean arterial blood pressure. The study involved 263 volunteers (12% males and 88% females), with an average age of 48.5. Overall muscle spasm/muscle strain was described as either moderate or severe for each patient. Baseline blood pressure and heart rate were measured via an automatic blood pressure cuff. Twenty-one (21) different soothing CDs played in the background as the deep tissue massage was performed over the course of the study. The massages were between 45 and 60 minutes in duration. The data were analyzed using analysis of variance with post-hoc Scheffe's F-test. Results of the present study demonstrated an average systolic pressure reduction of 10.4 mm Hg (p<0.06), a diastolic pressure reduction of 5.3 mm Hg (p<0.04), a mean arterial pressure reduction of 7.0 mm Hg (p<0.47), and an average heart rate reduction of 10.8 beats per minute (p<0.0003), respectively. Additional scientific research in this area is warranted.

  16. Blood Flukes Exploit Peyer's Patch Lymphoid Tissue to Facilitate Transmission from the Mammalian Host

    PubMed Central

    Turner, Joseph D.; Narang, Priyanka; Coles, Mark C.; Mountford, Adrian P.

    2012-01-01

    Schistosomes are blood-dwelling parasitic helminths which produce eggs in order to facilitate transmission. Intestinal schistosomes lay eggs in the mesenteries, however, it is unclear how their eggs escape the vasculature to exit the host. Using a murine model of infection, we reveal that Schistosoma mansoni exploits Peyer's Patches (PP) gut lymphoid tissue as a preferential route of egress for their eggs. Egg deposition is favoured within PP as a result of their more abundant vasculature. Moreover, the presence of eggs causes significant vascular remodeling leading to an expanded venule network. Egg deposition results in a decrease in stromal integrity and lymphoid cellularity, including secretory IgA producing lymphocytes, and the focal recruitment of macrophages. In mice lacking PP, egg excretion is significantly impaired, leading to greater numbers of ova being entrapped in tissues and consequently, exacerbated morbidity. Thus, we demonstrate how schistosomes directly facilitate transmission from the host by targeting lymphoid tissue. For the host, PP-dependency of egg egress represents a trade-off, as limiting potentially life-threatening morbidity is balanced by loss of PP structure and perturbed PP IgA production. PMID:23308064

  17. A Dual-Modality System for Both Multi-Color Ultrasound-Switchable Fluorescence and Ultrasound Imaging

    PubMed Central

    Kandukuri, Jayanth; Yu, Shuai; Cheng, Bingbing; Bandi, Venugopal; D’Souza, Francis; Nguyen, Kytai T.; Hong, Yi; Yuan, Baohong

    2017-01-01

    Simultaneous imaging of multiple targets (SIMT) in opaque biological tissues is an important goal for molecular imaging in the future. Multi-color fluorescence imaging in deep tissues is a promising technology to reach this goal. In this work, we developed a dual-modality imaging system by combining our recently developed ultrasound-switchable fluorescence (USF) imaging technology with the conventional ultrasound (US) B-mode imaging. This dual-modality system can simultaneously image tissue acoustic structure information and multi-color fluorophores in centimeter-deep tissue with comparable spatial resolutions. To conduct USF imaging on the same plane (i.e., x-z plane) as US imaging, we adopted two 90°-crossed ultrasound transducers with an overlapped focal region, while the US transducer (the third one) was positioned at the center of these two USF transducers. Thus, the axial resolution of USF is close to the lateral resolution, which allows a point-by-point USF scanning on the same plane as the US imaging. Both multi-color USF and ultrasound imaging of a tissue phantom were demonstrated. PMID:28165390

  18. Deep sub-wavelength ultrasonic imaging

    NASA Astrophysics Data System (ADS)

    Amireddy, Kiran Kumar; Balasubramaniam, Krishnan; Rajagopal, Prabhu

    2018-04-01

    There is much interest in improving the resolution of ultrasonic inspection, which suffers from large wavelengths typically in the range of millimeters, due to low value of speed of sound in solid media. The authors are interested in achieving this through holey structured metamaterial lenses, and have recently demonstrated an experimental subwavelength resolution of λ/25. However the previous work was in through-transmission mode with reception using Laser Doppler Vibrometer (LDV), which may not be suitable for practical applications. This paper discusses the use of optimized holey structured metalens to achieve a deep sub-wavelength imaging up to λ/18 in through-transmission mode, but using commercially available piezoelectric ultrasonic transducers for both generation and reception of ultrasound.

  19. Tractography patterns of subthalamic nucleus deep brain stimulation.

    PubMed

    Vanegas-Arroyave, Nora; Lauro, Peter M; Huang, Ling; Hallett, Mark; Horovitz, Silvina G; Zaghloul, Kareem A; Lungu, Codrin

    2016-04-01

    Deep brain stimulation therapy is an effective symptomatic treatment for Parkinson's disease, yet the precise mechanisms responsible for its therapeutic effects remain unclear. Although the targets of deep brain stimulation are grey matter structures, axonal modulation is known to play an important role in deep brain stimulation's therapeutic mechanism. Several white matter structures in proximity to the subthalamic nucleus have been implicated in the clinical benefits of deep brain stimulation for Parkinson's disease. We assessed the connectivity patterns that characterize clinically beneficial electrodes in Parkinson's disease patients, after deep brain stimulation of the subthalamic nucleus. We evaluated 22 patients with Parkinson's disease (11 females, age 57 ± 9.1 years, disease duration 13.3 ± 6.3 years) who received bilateral deep brain stimulation of the subthalamic nucleus at the National Institutes of Health. During an initial electrode screening session, one month after deep brain stimulation implantation, the clinical benefits of each contact were determined. The electrode was localized by coregistering preoperative magnetic resonance imaging and postoperative computer tomography images and the volume of tissue activated was estimated from stimulation voltage and impedance. Brain connectivity for the volume of tissue activated of deep brain stimulation contacts was assessed using probabilistic tractography with diffusion-tensor data. Areas most frequently connected to clinically effective contacts included the thalamus, substantia nigra, brainstem and superior frontal gyrus. A series of discriminant analyses demonstrated that the strength of connectivity to the superior frontal gyrus and the thalamus were positively associated with clinical effectiveness. The connectivity patterns observed in our study suggest that the modulation of white matter tracts directed to the superior frontal gyrus and the thalamus is associated with favourable clinical

  20. Enhanced Image of Asteroid Braille from Deep Space 1

    NASA Technical Reports Server (NTRS)

    1999-01-01

    This image was created from a composite of two images which were taken 914 seconds and 932 seconds after the recent Deep Space 1 (DS1) encounter with the asteroid 9969 Braille by the Miniature Integrated Camera Spectrometer (MICAS). Interpolated values were then computed for each pixel in the final image based on the neighboring pixels of the composite. The interpolation minimizes the spatial frequency artifacts of the final image. The Sun is illuminating Braille from below.

    Braille (also known as 1992 KD) was discovered on May 27, 1992 by astronomers Eleanor Helin and Kenneth Lawrence using the 46 centimeter (18 inch) Shmidt telescope at Palomar Observatory, while scanning the skies as part of the Palomar Planet-Crossing Asteroid Survey.

    Deep Space 1 was launched into orbit around the Sun on October 24, 1998 at 5:08 a.m. Pacific Daylight Time from Cape Canaveral Air Station, Florida on a Delta 7326, a variant of the Delta II rocket. An ion engine, operating for more than 1800 hours, was used to maneuver the spacecraft for an encounter with Braille. The closest approach of DS1 to the asteroid, at an approximate distance of 15 kilometers, occurred on July 29,1999 at 04:45 Universal Time, July 28 at 9:46 p.m. Pacific Daylight Time.

  1. [Observation on changes of oxygen partial pressure in the deep tissues along the large intestine meridian during acupuncture in healthy subjects].

    PubMed

    Chen, Ming; Hu, Xiang-long; Wu, Zu-xing

    2010-06-01

    To observe changes of the partial oxygen pressure in the deep tissues along the Large Intestine Meridian (LIM) during acupuncture stimulation, so as to reveal the characteristics of energy metabolism in the tissues along the LIM. Thirty-one healthy volunteer subjects were enlisted in the present study. Partial oxygen pressure (POP) in the tissues (at a depth of about 1.5 cm) of acupoints Binao (LI 14), Shouwuli (LI 13), Shousanli (LI 10), 2 non-acupoints [the midpoints between Quchi (LI 11) and LI 14, and between Yangxi (LI 5) and LI 11) of the LIM, and 10 non-meridian points, 1.5-2.0 cm lateral and medial to each of the tested points of the LIM was detected before, during and after electroacupuncture (EA) stimulation of Hegu (LI 4) by using a tissue oxygen tension needle-like sensor. In normal condition, the POP values in the deep tissues along the LIM were significantly higher than those of the non-meridian control points on its bilateral sides. During and after EA of Hegu (LI 4), the POP levels decreased significantly in the deep tissues along the LIM in comparison with pre-EA (P < 0.01), and had no apparent changes in the non-meridian control points (P > 0.05). POP is significantly higher in the deep tissues along the LIM of healthy subjects under normal conditions, which can be downregulated by EA of Hegu (LI 4), suggesting an increase of both the utilization rate of oxygen and energy metabolism after EA.

  2. Two-tier tissue decomposition for histopathological image representation and classification.

    PubMed

    Gultekin, Tunc; Koyuncu, Can Fahrettin; Sokmensuer, Cenk; Gunduz-Demir, Cigdem

    2015-01-01

    In digital pathology, devising effective image representations is crucial to design robust automated diagnosis systems. To this end, many studies have proposed to develop object-based representations, instead of directly using image pixels, since a histopathological image may contain a considerable amount of noise typically at the pixel-level. These previous studies mostly employ color information to define their objects, which approximately represent histological tissue components in an image, and then use the spatial distribution of these objects for image representation and classification. Thus, object definition has a direct effect on the way of representing the image, which in turn affects classification accuracies. In this paper, our aim is to design a classification system for histopathological images. Towards this end, we present a new model for effective representation of these images that will be used by the classification system. The contributions of this model are twofold. First, it introduces a new two-tier tissue decomposition method for defining a set of multityped objects in an image. Different than the previous studies, these objects are defined combining texture, shape, and size information and they may correspond to individual histological tissue components as well as local tissue subregions of different characteristics. As its second contribution, it defines a new metric, which we call dominant blob scale, to characterize the shape and size of an object with a single scalar value. Our experiments on colon tissue images reveal that this new object definition and characterization provides distinguishing representation of normal and cancerous histopathological images, which is effective to obtain more accurate classification results compared to its counterparts.

  3. Psoriasis skin biopsy image segmentation using Deep Convolutional Neural Network.

    PubMed

    Pal, Anabik; Garain, Utpal; Chandra, Aditi; Chatterjee, Raghunath; Senapati, Swapan

    2018-06-01

    Development of machine assisted tools for automatic analysis of psoriasis skin biopsy image plays an important role in clinical assistance. Development of automatic approach for accurate segmentation of psoriasis skin biopsy image is the initial prerequisite for developing such system. However, the complex cellular structure, presence of imaging artifacts, uneven staining variation make the task challenging. This paper presents a pioneering attempt for automatic segmentation of psoriasis skin biopsy images. Several deep neural architectures are tried for segmenting psoriasis skin biopsy images. Deep models are used for classifying the super-pixels generated by Simple Linear Iterative Clustering (SLIC) and the segmentation performance of these architectures is compared with the traditional hand-crafted feature based classifiers built on popularly used classifiers like K-Nearest Neighbor (KNN), Support Vector Machine (SVM) and Random Forest (RF). A U-shaped Fully Convolutional Neural Network (FCN) is also used in an end to end learning fashion where input is the original color image and the output is the segmentation class map for the skin layers. An annotated real psoriasis skin biopsy image data set of ninety (90) images is developed and used for this research. The segmentation performance is evaluated with two metrics namely, Jaccard's Coefficient (JC) and the Ratio of Correct Pixel Classification (RCPC) accuracy. The experimental results show that the CNN based approaches outperform the traditional hand-crafted feature based classification approaches. The present research shows that practical system can be developed for machine assisted analysis of psoriasis disease. Copyright © 2018 Elsevier B.V. All rights reserved.

  4. Tissue oxygen monitoring by photoacoustic lifetime imaging (PALI) and its application to image-guided photodynamic therapy (PDT)

    NASA Astrophysics Data System (ADS)

    Shao, Qi; Morgounova, Ekaterina; Ashkenazi, Shai

    2015-03-01

    The oxygen partial pressure (pO2), which results from the balance between oxygen delivery and its consumption, is a key component of the physiological state of a tissue. Images of oxygen distribution can provide essential information for identifying hypoxic tissue and optimizing cancer treatment. Previously, we have reported a noninvasive in vivo imaging modality based on photoacoustic lifetime. The technique maps the excited triplet state of oxygen-sensitive dye, thus reflects the spatial and temporal distribution of tissue oxygen. We have applied PALI on tumor on small animals to identify hypoxia area. We also showed that PALI is able monitor changes of tissue oxygen, in an acute ischemia and breathing modulation model. Here we present our work on developing a treatment/imaging modality (PDT-PALI) that integrates PDT and a combined ultrasound/photoacoustic imaging system. The system provides real-time feedback of three essential parameters namely: tissue oxygen, light penetration in tumor location, and distribution of photosensitizer. Tissue oxygen imaging is performed by applying PALI, which relies on photoacoustic probing of oxygen-dependent, excitation lifetime of Methylene Blue (MB) photosensitizer. Lifetime information can also be used to generate image showing the distribution of photosensitizer. The level and penetration depth of PDT illumination can be deduced from photoacoustic imaging at the same wavelength. All images will be combined with ultrasound B-mode images for anatomical reference.

  5. Deep features for efficient multi-biometric recognition with face and ear images

    NASA Astrophysics Data System (ADS)

    Omara, Ibrahim; Xiao, Gang; Amrani, Moussa; Yan, Zifei; Zuo, Wangmeng

    2017-07-01

    Recently, multimodal biometric systems have received considerable research interest in many applications especially in the fields of security. Multimodal systems can increase the resistance to spoof attacks, provide more details and flexibility, and lead to better performance and lower error rate. In this paper, we present a multimodal biometric system based on face and ear, and propose how to exploit the extracted deep features from Convolutional Neural Networks (CNNs) on the face and ear images to introduce more powerful discriminative features and robust representation ability for them. First, the deep features for face and ear images are extracted based on VGG-M Net. Second, the extracted deep features are fused by using a traditional concatenation and a Discriminant Correlation Analysis (DCA) algorithm. Third, multiclass support vector machine is adopted for matching and classification. The experimental results show that the proposed multimodal system based on deep features is efficient and achieves a promising recognition rate up to 100 % by using face and ear. In addition, the results indicate that the fusion based on DCA is superior to traditional fusion.

  6. Deep erosions of the palmar aspect of the navicular bone diagnosed by standing magnetic resonance imaging.

    PubMed

    Sherlock, C; Mair, T; Blunden, T

    2008-11-01

    Erosion of the palmar (flexor) aspect of the navicular bone is difficult to diagnose with conventional imaging techniques. To review the clinical, magnetic resonance (MR) and pathological features of deep erosions of the palmar aspect of the navicular bone. Cases of deep erosions of the palmar aspect of the navicular bone, diagnosed by standing low field MR imaging, were selected. Clinical details, results of diagnostic procedures, MR features and pathological findings were reviewed. Deep erosions of the palmar aspect of the navicular bone were diagnosed in 16 mature horses, 6 of which were bilaterally lame. Sudden onset of lameness was recorded in 63%. Radiography prior to MR imaging showed equivocal changes in 7 horses. The MR features consisted of focal areas of intermediate or high signal intensity on T1-, T2*- and T2-weighted images and STIR images affecting the dorsal aspect of the deep digital flexor tendon, the fibrocartilage of the palmar aspect, subchondral compact bone and medulla of the navicular bone. On follow-up, 7/16 horses (44%) had been subjected to euthanasia and only one was being worked at its previous level. Erosions of the palmar aspect of the navicular bone were confirmed post mortem in 2 horses. Histologically, the lesions were characterised by localised degeneration of fibrocartilage with underlying focal osteonecrosis and fibroplasia. The adjacent deep digital flexor tendon showed fibril formation and fibrocartilaginous metaplasia. Deep erosions of the palmar aspect of the navicular bone are more easily diagnosed by standing low field MR imaging than by conventional radiography. The lesions involve degeneration of the palmar fibrocartilage with underlying osteonecrosis and fibroplasia affecting the subchondral compact bone and medulla, and carry a poor prognosis for return to performance. Diagnosis of shallow erosive lesions of the palmar fibrocartilage may allow therapeutic intervention earlier in the disease process, thereby preventing

  7. Blind source separation of ex-vivo aorta tissue multispectral images

    PubMed Central

    Galeano, July; Perez, Sandra; Montoya, Yonatan; Botina, Deivid; Garzón, Johnson

    2015-01-01

    Blind Source Separation methods (BSS) aim for the decomposition of a given signal in its main components or source signals. Those techniques have been widely used in the literature for the analysis of biomedical images, in order to extract the main components of an organ or tissue under study. The analysis of skin images for the extraction of melanin and hemoglobin is an example of the use of BSS. This paper presents a proof of concept of the use of source separation of ex-vivo aorta tissue multispectral Images. The images are acquired with an interference filter-based imaging system. The images are processed by means of two algorithms: Independent Components analysis and Non-negative Matrix Factorization. In both cases, it is possible to obtain maps that quantify the concentration of the main chromophores present in aortic tissue. Also, the algorithms allow for spectral absorbance of the main tissue components. Those spectral signatures were compared against the theoretical ones by using correlation coefficients. Those coefficients report values close to 0.9, which is a good estimator of the method’s performance. Also, correlation coefficients lead to the identification of the concentration maps according to the evaluated chromophore. The results suggest that Multi/hyper-spectral systems together with image processing techniques is a potential tool for the analysis of cardiovascular tissue. PMID:26137366

  8. Quantitative evaluation of deep and shallow tissue layers' contribution to fNIRS signal using multi-distance optodes and independent component analysis.

    PubMed

    Funane, Tsukasa; Atsumori, Hirokazu; Katura, Takusige; Obata, Akiko N; Sato, Hiroki; Tanikawa, Yukari; Okada, Eiji; Kiguchi, Masashi

    2014-01-15

    To quantify the effect of absorption changes in the deep tissue (cerebral) and shallow tissue (scalp, skin) layers on functional near-infrared spectroscopy (fNIRS) signals, a method using multi-distance (MD) optodes and independent component analysis (ICA), referred to as the MD-ICA method, is proposed. In previous studies, when the signal from the shallow tissue layer (shallow signal) needs to be eliminated, it was often assumed that the shallow signal had no correlation with the signal from the deep tissue layer (deep signal). In this study, no relationship between the waveforms of deep and shallow signals is assumed, and instead, it is assumed that both signals are linear combinations of multiple signal sources, which allows the inclusion of a "shared component" (such as systemic signals) that is contained in both layers. The method also assumes that the partial optical path length of the shallow layer does not change, whereas that of the deep layer linearly increases along with the increase of the source-detector (S-D) distance. Deep- and shallow-layer contribution ratios of each independent component (IC) are calculated using the dependence of the weight of each IC on the S-D distance. Reconstruction of deep- and shallow-layer signals are performed by the sum of ICs weighted by the deep and shallow contribution ratio. Experimental validation of the principle of this technique was conducted using a dynamic phantom with two absorbing layers. Results showed that our method is effective for evaluating deep-layer contributions even if there are high correlations between deep and shallow signals. Next, we applied the method to fNIRS signals obtained on a human head with 5-, 15-, and 30-mm S-D distances during a verbal fluency task, a verbal working memory task (prefrontal area), a finger tapping task (motor area), and a tetrametric visual checker-board task (occipital area) and then estimated the deep-layer contribution ratio. To evaluate the signal separation

  9. Whole slide imaging of unstained tissue using lensfree microscopy

    NASA Astrophysics Data System (ADS)

    Morel, Sophie Nhu An; Hervé, Lionel; Bordy, Thomas; Cioni, Olivier; Delon, Antoine; Fromentin, Catherine; Dinten, Jean-Marc; Allier, Cédric

    2016-04-01

    Pathologist examination of tissue slides provides insightful information about a patient's disease. Traditional analysis of tissue slides is performed under a binocular microscope, which requires staining of the sample and delays the examination. We present a simple cost-effective lensfree imaging method to record 2-4μm resolution wide-field (10 mm2 to 6 cm2) images of unstained tissue slides. The sample processing time is reduced as there is no need for staining. A wide field of view (10 mm2) lensfree hologram is recorded in a single shot and the image is reconstructed in 2s providing a very fast acquisition chain. The acquisition is multispectral, i.e. multiple holograms are recorded simultaneously at three different wavelengths, and a dedicated holographic reconstruction algorithm is used to retrieve both amplitude and phase. Whole tissue slides imaging is obtained by recording 130 holograms with X-Y translation stages and by computing the mosaic of a 25 x 25 mm2 reconstructed image. The reconstructed phase provides a phase-contrast-like image of the unstained specimen, revealing structures of healthy and diseased tissue. Slides from various organs can be reconstructed, e.g. lung, colon, ganglion, etc. To our knowledge, our method is the first technique that enables fast wide-field lensfree imaging of such unlabeled dense samples. This technique is much cheaper and compact than a conventional phase contrast microscope and could be made portable. In sum, we present a new methodology that could quickly provide useful information when a rapid diagnosis is needed, such as tumor margin identification on frozen section biopsies during surgery.

  10. In vivo multiphoton microscopy of deep tissue with gradient index lenses

    NASA Astrophysics Data System (ADS)

    Levene, Michael J.; Dombeck, Daniel A.; Williams, Rebecca M.; Skoch, Jesse; Hickey, Gregory A.; Kasischke, Karl A.; Molloy, Raymond P.; Ingelsson, Martin; Stern, Edward A.; Klucken, Jochen; Bacskai, Brian J.; Zipfel, Warren R.; Hyman, Bradley T.; Webb, Watt W.

    2004-06-01

    Gradient index lenses enable multiphoton microscopy of deep tissues in the intact animal. In order to assess their applicability to clinical research, we present in vivo multiphoton microscopy with gradient index lenses in brain regions associated with Alzheimer's disease and Parkinson's disease in both transgenic and wild-type mice. We also demonstrate microscopy of ovary in wild type mouse using only intrinsic fluorescence and second harmonic generation, signal sources which may prove useful for both the study and diagnosis of cancer.

  11. Beam and tissue factors affecting Cherenkov image intensity for quantitative entrance and exit dosimetry on human tissue

    PubMed Central

    Zhang, Rongxiao; Glaser, Adam K.; Andreozzi, Jacqueline; Jiang, Shudong; Jarvis, Lesley A.; Gladstone, David J.; Pogue, Brian W.

    2017-01-01

    This study’s goal was to determine how Cherenkov radiation emission observed in radiotherapy is affected by predictable factors expected in patient imaging. Factors such as tissue optical properties, radiation beam properties, thickness of tissues, entrance/exit geometry, curved surface effects, curvature and imaging angles were investigated through Monte Carlo simulations. The largest physical cause of variation of the correlation factor between of Cherenkov emission and dose was the entrance/exit geometry (~50%). The largest human tissue effect was from different optical properties (~45%). Beyond these, clinical beam energy varies the correlation factor significantly (~20% for x-ray beams), followed by curved surfaces (~15% for x-ray beams and ~8% for electron beams), and finally, the effect of field size (~5% for x-ray beams). Other investigated factors which caused variations less than 5% were tissue thicknesses and source to surface distance. The effect of non-Lambertian emission was negligible for imaging angles smaller than 60 degrees. The spectrum of Cherenkov emission tends to blue-shift along the curved surface. A simple normalization approach based on the reflectance image was experimentally validated by imaging a range of tissue phantoms, as a first order correction for different tissue optical properties. PMID:27507213

  12. Sustained deep-tissue pain alters functional brain connectivity.

    PubMed

    Kim, Jieun; Loggia, Marco L; Edwards, Robert R; Wasan, Ajay D; Gollub, Randy L; Napadow, Vitaly

    2013-08-01

    Recent functional brain connectivity studies have contributed to our understanding of the neurocircuitry supporting pain perception. However, evoked-pain connectivity studies have employed cutaneous and/or brief stimuli, which induce sensations that differ appreciably from the clinical pain experience. Sustained myofascial pain evoked by pressure cuff affords an excellent opportunity to evaluate functional connectivity change to more clinically relevant sustained deep-tissue pain. Connectivity in specific networks known to be modulated by evoked pain (sensorimotor, salience, dorsal attention, frontoparietal control, and default mode networks: SMN, SLN, DAN, FCN, and DMN) was evaluated with functional-connectivity magnetic resonance imaging, both at rest and during a sustained (6-minute) pain state in healthy adults. We found that pain was stable, with no significant changes of subjects' pain ratings over the stimulation period. Sustained pain reduced connectivity between the SMN and the contralateral leg primary sensorimotor (S1/M1) representation. Such SMN-S1/M1 connectivity decreases were also accompanied by and correlated with increased SLN-S1/M1 connectivity, suggesting recruitment of activated S1/M1 from SMN to SLN. Sustained pain also increased DAN connectivity to pain processing regions such as mid-cingulate cortex, posterior insula, and putamen. Moreover, greater connectivity during pain between contralateral S1/M1 and posterior insula, thalamus, putamen, and amygdala was associated with lower cuff pressures needed to reach the targeted pain sensation. These results demonstrate that sustained pain disrupts resting S1/M1 connectivity by shifting it to a network known to process stimulus salience. Furthermore, increased connectivity between S1/M1 and both sensory and affective processing areas may be an important contribution to interindividual differences in pain sensitivity. Copyright © 2013 International Association for the Study of Pain. Published by

  13. Finite difference time domain (FDTD) modeling of implanted deep brain stimulation electrodes and brain tissue.

    PubMed

    Gabran, S R I; Saad, J H; Salama, M M A; Mansour, R R

    2009-01-01

    This paper demonstrates the electromagnetic modeling and simulation of an implanted Medtronic deep brain stimulation (DBS) electrode using finite difference time domain (FDTD). The model is developed using Empire XCcel and represents the electrode surrounded with brain tissue assuming homogenous and isotropic medium. The model is created to study the parameters influencing the electric field distribution within the tissue in order to provide reference and benchmarking data for DBS and intra-cortical electrode development.

  14. Three-Dimensional Optical Mapping of Nanoparticle Distribution in Intact Tissues.

    PubMed

    Sindhwani, Shrey; Syed, Abdullah Muhammad; Wilhelm, Stefan; Glancy, Dylan R; Chen, Yih Yang; Dobosz, Michael; Chan, Warren C W

    2016-05-24

    The role of tissue architecture in mediating nanoparticle transport, targeting, and biological effects is unknown due to the lack of tools for imaging nanomaterials in whole organs. Here, we developed a rapid optical mapping technique to image nanomaterials in intact organs ex vivo and in three-dimensions (3D). We engineered a high-throughput electrophoretic flow device to simultaneously transform up to 48 tissues into optically transparent structures, allowing subcellular imaging of nanomaterials more than 1 mm deep into tissues which is 25-fold greater than current techniques. A key finding is that nanomaterials can be retained in the processed tissue by chemical cross-linking of surface adsorbed serum proteins to the tissue matrix, which enables nanomaterials to be imaged with respect to cells, blood vessels, and other structures. We developed a computational algorithm to analyze and quantitatively map nanomaterial distribution. This method can be universally applied to visualize the distribution and interactions of materials in whole tissues and animals including such applications as the imaging of nanomaterials, tissue engineered constructs, and biosensors within their intact biological environment.

  15. Joint deep shape and appearance learning: application to optic pathway glioma segmentation

    NASA Astrophysics Data System (ADS)

    Mansoor, Awais; Li, Ien; Packer, Roger J.; Avery, Robert A.; Linguraru, Marius George

    2017-03-01

    Automated tissue characterization is one of the major applications of computer-aided diagnosis systems. Deep learning techniques have recently demonstrated impressive performance for the image patch-based tissue characterization. However, existing patch-based tissue classification techniques struggle to exploit the useful shape information. Local and global shape knowledge such as the regional boundary changes, diameter, and volumetrics can be useful in classifying the tissues especially in scenarios where the appearance signature does not provide significant classification information. In this work, we present a deep neural network-based method for the automated segmentation of the tumors referred to as optic pathway gliomas (OPG) located within the anterior visual pathway (AVP; optic nerve, chiasm or tracts) using joint shape and appearance learning. Voxel intensity values of commonly used MRI sequences are generally not indicative of OPG. To be considered an OPG, current clinical practice dictates that some portion of AVP must demonstrate shape enlargement. The method proposed in this work integrates multiple sequence magnetic resonance image (T1, T2, and FLAIR) along with local boundary changes to train a deep neural network. For training and evaluation purposes, we used a dataset of multiple sequence MRI obtained from 20 subjects (10 controls, 10 NF1+OPG). To our best knowledge, this is the first deep representation learning-based approach designed to merge shape and multi-channel appearance data for the glioma detection. In our experiments, mean misclassification errors of 2:39% and 0:48% were observed respectively for glioma and control patches extracted from the AVP. Moreover, an overall dice similarity coefficient of 0:87+/-0:13 (0:93+/-0:06 for healthy tissue, 0:78+/-0:18 for glioma tissue) demonstrates the potential of the proposed method in the accurate localization and early detection of OPG.

  16. Cell segmentation in histopathological images with deep learning algorithms by utilizing spatial relationships.

    PubMed

    Hatipoglu, Nuh; Bilgin, Gokhan

    2017-10-01

    In many computerized methods for cell detection, segmentation, and classification in digital histopathology that have recently emerged, the task of cell segmentation remains a chief problem for image processing in designing computer-aided diagnosis (CAD) systems. In research and diagnostic studies on cancer, pathologists can use CAD systems as second readers to analyze high-resolution histopathological images. Since cell detection and segmentation are critical for cancer grade assessments, cellular and extracellular structures should primarily be extracted from histopathological images. In response, we sought to identify a useful cell segmentation approach with histopathological images that uses not only prominent deep learning algorithms (i.e., convolutional neural networks, stacked autoencoders, and deep belief networks), but also spatial relationships, information of which is critical for achieving better cell segmentation results. To that end, we collected cellular and extracellular samples from histopathological images by windowing in small patches with various sizes. In experiments, the segmentation accuracies of the methods used improved as the window sizes increased due to the addition of local spatial and contextual information. Once we compared the effects of training sample size and influence of window size, results revealed that the deep learning algorithms, especially convolutional neural networks and partly stacked autoencoders, performed better than conventional methods in cell segmentation.

  17. Epithelium percentage estimation facilitates epithelial quantitative protein measurement in tissue specimens.

    PubMed

    Chen, Jing; Toghi Eshghi, Shadi; Bova, George Steven; Li, Qing Kay; Li, Xingde; Zhang, Hui

    2013-12-01

    The rapid advancement of high-throughput tools for quantitative measurement of proteins has demonstrated the potential for the identification of proteins associated with cancer. However, the quantitative results on cancer tissue specimens are usually confounded by tissue heterogeneity, e.g. regions with cancer usually have significantly higher epithelium content yet lower stromal content. It is therefore necessary to develop a tool to facilitate the interpretation of the results of protein measurements in tissue specimens. Epithelial cell adhesion molecule (EpCAM) and cathepsin L (CTSL) are two epithelial proteins whose expressions in normal and tumorous prostate tissues were confirmed by measuring staining intensity with immunohistochemical staining (IHC). The expressions of these proteins were measured by ELISA in protein extracts from OCT embedded frozen prostate tissues. To eliminate the influence of tissue heterogeneity on epithelial protein quantification measured by ELISA, a color-based segmentation method was developed in-house for estimation of epithelium content using H&E histology slides from the same prostate tissues and the estimated epithelium percentage was used to normalize the ELISA results. The epithelium contents of the same slides were also estimated by a pathologist and used to normalize the ELISA results. The computer based results were compared with the pathologist's reading. We found that both EpCAM and CTSL levels, measured by ELISA assays itself, were greatly affected by epithelium content in the tissue specimens. Without adjusting for epithelium percentage, both EpCAM and CTSL levels appeared significantly higher in tumor tissues than normal tissues with a p value less than 0.001. However, after normalization by the epithelium percentage, ELISA measurements of both EpCAM and CTSL were in agreement with IHC staining results, showing a significant increase only in EpCAM with no difference in CTSL expression in cancer tissues. These results

  18. Performance evaluation of 2D and 3D deep learning approaches for automatic segmentation of multiple organs on CT images

    NASA Astrophysics Data System (ADS)

    Zhou, Xiangrong; Yamada, Kazuma; Kojima, Takuya; Takayama, Ryosuke; Wang, Song; Zhou, Xinxin; Hara, Takeshi; Fujita, Hiroshi

    2018-02-01

    The purpose of this study is to evaluate and compare the performance of modern deep learning techniques for automatically recognizing and segmenting multiple organ regions on 3D CT images. CT image segmentation is one of the important task in medical image analysis and is still very challenging. Deep learning approaches have demonstrated the capability of scene recognition and semantic segmentation on nature images and have been used to address segmentation problems of medical images. Although several works showed promising results of CT image segmentation by using deep learning approaches, there is no comprehensive evaluation of segmentation performance of the deep learning on segmenting multiple organs on different portions of CT scans. In this paper, we evaluated and compared the segmentation performance of two different deep learning approaches that used 2D- and 3D deep convolutional neural networks (CNN) without- and with a pre-processing step. A conventional approach that presents the state-of-the-art performance of CT image segmentation without deep learning was also used for comparison. A dataset that includes 240 CT images scanned on different portions of human bodies was used for performance evaluation. The maximum number of 17 types of organ regions in each CT scan were segmented automatically and compared to the human annotations by using ratio of intersection over union (IU) as the criterion. The experimental results demonstrated the IUs of the segmentation results had a mean value of 79% and 67% by averaging 17 types of organs that segmented by a 3D- and 2D deep CNN, respectively. All the results of the deep learning approaches showed a better accuracy and robustness than the conventional segmentation method that used probabilistic atlas and graph-cut methods. The effectiveness and the usefulness of deep learning approaches were demonstrated for solving multiple organs segmentation problem on 3D CT images.

  19. Perylene-diimide-based nanoparticles as highly efficient photoacoustic agents for deep brain tumor imaging in living mice

    DOE PAGES

    Fan, Quli; Cheng, Kai; Yang, Zhen; ...

    2014-11-06

    In order to promote preclinical and clinical applications of photoacoustic imaging, novel photoacoustic contrast agents are highly desired for molecular imaging of diseases, especially for deep tumor imaging. In this paper, perylene-3,4,9,10-tetracarboxylic diiimide-based near-infrared-absorptive organic nanoparticles are reported as an efficient agent for photoacoustic imaging of deep brain tumors in living mice with enhanced permeability and retention effect

  20. Photoacoustic lifetime imaging for direct in vivo tissue oxygen monitoring

    PubMed Central

    Shao, Qi; Ashkenazi, Shai

    2015-01-01

    Abstract. Measuring the partial pressure of oxygen (pO2) in tissue may provide physicians with essential information about the physiological state of tissue. However, currently available methods for measuring or imaging tissue pO2 have significant limitations, preventing them from being widely used in clinics. Recently, we have reported a direct and noninvasive in vivo imaging modality based on the photoacoustic lifetime which overcomes certain drawbacks of the existing methods. The technique maps the excited triplet state of oxygen-sensitive dye, thus reflecting the spatial and temporal distributions of tissue oxygen. Here, we present two studies which apply photoacoustic lifetime imaging (PALI) to monitor changes of tissue oxygen induced by external modulations. The first study modulates tissue oxygen by controlling the percentage of oxygen a normal mouse inhales. We demonstrate that PALI is able to reflect the change in oxygen level with respect to normal, oxygen-rich, and oxygen-poor breathing conditions. The second study involves an acute ischemia model using a thin thread tied around the hindlimb of a normal mouse to reduce the blood flow. PALI images were acquired before, during, and after the restriction. The drop of tissue pO2 and recovery from hypoxia due to reperfusion were tracked and observed by PALI. PMID:25748857

  1. MARS spectral molecular imaging of lamb tissue: data collection and image analysis

    NASA Astrophysics Data System (ADS)

    Aamir, R.; Chernoglazov, A.; Bateman, C. J.; Butler, A. P. H.; Butler, P. H.; Anderson, N. G.; Bell, S. T.; Panta, R. K.; Healy, J. L.; Mohr, J. L.; Rajendran, K.; Walsh, M. F.; de Ruiter, N.; Gieseg, S. P.; Woodfield, T.; Renaud, P. F.; Brooke, L.; Abdul-Majid, S.; Clyne, M.; Glendenning, R.; Bones, P. J.; Billinghurst, M.; Bartneck, C.; Mandalika, H.; Grasset, R.; Schleich, N.; Scott, N.; Nik, S. J.; Opie, A.; Janmale, T.; Tang, D. N.; Kim, D.; Doesburg, R. M.; Zainon, R.; Ronaldson, J. P.; Cook, N. J.; Smithies, D. J.; Hodge, K.

    2014-02-01

    Spectral molecular imaging is a new imaging technique able to discriminate and quantify different components of tissue simultaneously at high spatial and high energy resolution. Our MARS scanner is an x-ray based small animal CT system designed to be used in the diagnostic energy range (20-140 keV). In this paper, we demonstrate the use of the MARS scanner, equipped with the Medipix3RX spectroscopic photon-processing detector, to discriminate fat, calcium, and water in tissue. We present data collected from a sample of lamb meat including bone as an illustrative example of human tissue imaging. The data is analyzed using our 3D Algebraic Reconstruction Algorithm (MARS-ART) and by material decomposition based on a constrained linear least squares algorithm. The results presented here clearly show the quantification of lipid-like, water-like and bone-like components of tissue. However, it is also clear to us that better algorithms could extract more information of clinical interest from our data. Because we are one of the first to present data from multi-energy photon-processing small animal CT systems, we make the raw, partial and fully processed data available with the intention that others can analyze it using their familiar routines. The raw, partially processed and fully processed data of lamb tissue along with the phantom calibration data can be found at http://hdl.handle.net/10092/8531.

  2. Double-excitation fluorescence spectral imaging: eliminating tissue auto-fluorescence from in vivo PPIX measurements

    NASA Astrophysics Data System (ADS)

    Torosean, Sason; Flynn, Brendan; Samkoe, Kimberley S.; Davis, Scott C.; Gunn, Jason; Axelsson, Johan; Pogue, Brian W.

    2012-02-01

    An ultrasound coupled handheld-probe-based optical fluorescence molecular tomography (FMT) system has been in development for the purpose of quantifying the production of Protoporphyrin IX (PPIX) in aminolevulinic acid treated (ALA), Basal Cell Carcinoma (BCC) in vivo. The design couples fiber-based spectral sampling of PPIX fluorescence emission with a high frequency ultrasound imaging system, allowing regionally localized fluorescence intensities to be quantified [1]. The optical data are obtained by sequential excitation of the tissue with a 633nm laser, at four source locations and five parallel detections at each of the five interspersed detection locations. This method of acquisition permits fluorescence detection for both superficial and deep locations in ultrasound field. The optical boundary data, tissue layers segmented from ultrasound image and diffusion theory are used to estimate the fluorescence in tissue layers. To improve the recovery of the fluorescence signal of PPIX, eliminating tissue autofluorescence is of great importance. Here the approach was to utilize measurements which straddled the steep Qband excitation peak of PPIX, via the integration of an additional laser source, exciting at 637 nm; a wavelength with a 2 fold lower PPIX excitation value than 633nm.The auto-fluorescence spectrum acquired from the 637 nm laser is then used to spectrally decouple the fluorescence data and produce an accurate fluorescence emission signal, because the two wavelengths have very similar auto-fluorescence but substantially different PPIX excitation levels. The accuracy of this method, using a single source detector pair setup, is verified through animal tumor model experiments, and the result is compared to different methods of fluorescence signal recovery.

  3. Fluorescent imaging of cancerous tissues for targeted surgery

    PubMed Central

    Bu, Lihong; Shen, Baozhong; Cheng, Zhen

    2014-01-01

    To maximize tumor excision and minimize collateral damage is the primary goal of cancer surgery. Emerging molecular imaging techniques have to “image-guided surgery” developing into “molecular imaging-guided surgery”, which is termed “targeted surgery” in this review. Consequently, the precision of surgery can be advanced from tissue-scale to molecule-scale, enabling “targeted surgery” to be a component of “targeted therapy”. Evidence from numerous experimental and clinical studies has demonstrated significant benefits of fluorescent imaging in targeted surgery with preoperative molecular diagnostic screening. Fluorescent imaging can help to improve intraoperative staging and enable more radical cytoreduction, detect obscure tumor lesions in special organs, highlight tumor margins, better map lymph node metastases, and identify important normal structures intraoperatively. Though limited tissue penetration of fluorescent imaging and tumor heterogeneity are two major hurdles for current targeted surgery, multimodality imaging and multiplex imaging may provide potential solutions to overcome these issues, respectively. Moreover, though many fluorescent imaging techniques and probes have been investigated, targeted surgery remains at a proof-of-principle stage. The impact of fluorescent imaging on cancer surgery will likely be realized through persistent interdisciplinary amalgamation of research in diverse fields. PMID:25064553

  4. Automatic tissue segmentation of breast biopsies imaged by QPI

    NASA Astrophysics Data System (ADS)

    Majeed, Hassaan; Nguyen, Tan; Kandel, Mikhail; Marcias, Virgilia; Do, Minh; Tangella, Krishnarao; Balla, Andre; Popescu, Gabriel

    2016-03-01

    The current tissue evaluation method for breast cancer would greatly benefit from higher throughput and less inter-observer variation. Since quantitative phase imaging (QPI) measures physical parameters of tissue, it can be used to find quantitative markers, eliminating observer subjectivity. Furthermore, since the pixel values in QPI remain the same regardless of the instrument used, classifiers can be built to segment various tissue components without need for color calibration. In this work we use a texton-based approach to segment QPI images of breast tissue into various tissue components (epithelium, stroma or lumen). A tissue microarray comprising of 900 unstained cores from 400 different patients was imaged using Spatial Light Interference Microscopy. The training data were generated by manually segmenting the images for 36 cores and labelling each pixel (epithelium, stroma or lumen.). For each pixel in the data, a response vector was generated by the Leung-Malik (LM) filter bank and these responses were clustered using the k-means algorithm to find the centers (called textons). A random forest classifier was then trained to find the relationship between a pixel's label and the histogram of these textons in that pixel's neighborhood. The segmentation was carried out on the validation set by calculating the texton histogram in a pixel's neighborhood and generating a label based on the model learnt during training. Segmentation of the tissue into various components is an important step toward efficiently computing parameters that are markers of disease. Automated segmentation, followed by diagnosis, can improve the accuracy and speed of analysis leading to better health outcomes.

  5. Myocardial Tissue Characterization by Magnetic Resonance Imaging

    PubMed Central

    Ferreira, Vanessa M.; Piechnik, Stefan K.; Robson, Matthew D.; Neubauer, Stefan

    2014-01-01

    Cardiac magnetic resonance (CMR) imaging is a well-established noninvasive imaging modality in clinical cardiology. Its unsurpassed accuracy in defining cardiac morphology and function and its ability to provide tissue characterization make it well suited for the study of patients with cardiac diseases. Late gadolinium enhancement was a major advancement in the development of tissue characterization techniques, allowing the unique ability of CMR to differentiate ischemic heart disease from nonischemic cardiomyopathies. Using T2-weighted techniques, areas of edema and inflammation can be identified in the myocardium. A new generation of myocardial mapping techniques are emerging, enabling direct quantitative assessment of myocardial tissue properties in absolute terms. This review will summarize recent developments involving T1-mapping and T2-mapping techniques and focus on the clinical applications and future potential of these evolving CMR methodologies. PMID:24576837

  6. Photoacoustic imaging in both soft and hard biological tissue

    NASA Astrophysics Data System (ADS)

    Li, T.; Dewhurst, R. J.

    2010-03-01

    To date, most Photoacoustic (PA) imaging results have been from soft biotissues. In this study, a PA imaging system with a near-infrared pulsed laser source has been applied to obtain 2-D and 3-D images from both soft tissue and post-mortem dental samples. Imaging results showed that the PA technique has the potential to image human oral disease, such as early-stage teeth decay. For non-invasive photoacoustic imaging, the induced temperature and pressure rises within biotissues should not cause physical damage to the tissue. Several simulations based on the thermoelastic effect have been applied to predict initial temperature and pressure fields within a tooth sample. Predicted initial temperature and pressure rises are below corresponding safety limits.

  7. Big-deep-smart data in imaging for guiding materials design.

    PubMed

    Kalinin, Sergei V; Sumpter, Bobby G; Archibald, Richard K

    2015-10-01

    Harnessing big data, deep data, and smart data from state-of-the-art imaging might accelerate the design and realization of advanced functional materials. Here we discuss new opportunities in materials design enabled by the availability of big data in imaging and data analytics approaches, including their limitations, in material systems of practical interest. We specifically focus on how these tools might help realize new discoveries in a timely manner. Such methodologies are particularly appropriate to explore in light of continued improvements in atomistic imaging, modelling and data analytics methods.

  8. Big-deep-smart data in imaging for guiding materials design

    NASA Astrophysics Data System (ADS)

    Kalinin, Sergei V.; Sumpter, Bobby G.; Archibald, Richard K.

    2015-10-01

    Harnessing big data, deep data, and smart data from state-of-the-art imaging might accelerate the design and realization of advanced functional materials. Here we discuss new opportunities in materials design enabled by the availability of big data in imaging and data analytics approaches, including their limitations, in material systems of practical interest. We specifically focus on how these tools might help realize new discoveries in a timely manner. Such methodologies are particularly appropriate to explore in light of continued improvements in atomistic imaging, modelling and data analytics methods.

  9. Cell Migration in Tissues: Explant Culture and Live Imaging.

    PubMed

    Staneva, Ralitza; Barbazan, Jorge; Simon, Anthony; Vignjevic, Danijela Matic; Krndija, Denis

    2018-01-01

    Cell migration is a process that ensures correct cell localization and function in development and homeostasis. In disease such as cancer, cells acquire an upregulated migratory capacity that leads to their dissemination throughout the body. Live imaging of cell migration allows for better understanding of cell behaviors in development, adult tissue homeostasis and disease. We have optimized live imaging procedures to track cell migration in adult murine tissue explants derived from: (1) healthy gut; (2) primary intestinal carcinoma; and (3) the liver, a common metastatic site. To track epithelial cell migration in the gut, we generated an inducible fluorescent reporter mouse, enabling us to visualize and track individual cells in unperturbed gut epithelium. To image intratumoral cancer cells, we use a spontaneous intestinal cancer model based on the activation of Notch1 and deletion of p53 in the mouse intestinal epithelium, which gives rise to aggressive carcinoma. Interaction of cancer cells with a metastatic niche, the mouse liver, is addressed using a liver colonization model. In summary, we describe a method for long-term 3D imaging of tissue explants by two-photon excitation microscopy. Explant culturing and imaging can help understand dynamic behavior of cells in homeostasis and disease, and would be applicable to various tissues.

  10. Mueller matrix microscope: a quantitative tool to facilitate detections and fibrosis scorings of liver cirrhosis and cancer tissues.

    PubMed

    Wang, Ye; He, Honghui; Chang, Jintao; He, Chao; Liu, Shaoxiong; Li, Migao; Zeng, Nan; Wu, Jian; Ma, Hui

    2016-07-01

    Today the increasing cancer incidence rate is becoming one of the biggest threats to human health.Among all types of cancers, liver cancer ranks in the top five in both frequency and mortality rate all over the world. During the development of liver cancer, fibrosis often evolves as part of a healing process in response to liver damage, resulting in cirrhosis of liver tissues. In a previous study, we applied the Mueller matrix microscope to pathological liver tissue samples and found that both the Mueller matrix polar decomposition (MMPD) and Mueller matrix transformation (MMT) parameters are closely related to the fibrous microstructures. In this paper,we take this one step further to quantitatively facilitate the fibrosis detections and scorings of pathological liver tissue samples in different stages from cirrhosis to cancer using the Mueller matrix microscope. The experimental results of MMPD and MMT parameters for the fibrotic liver tissue samples in different stages are measured and analyzed. We also conduct Monte Carlo simulations based on the sphere birefringence model to examine in detail the influence of structural changes in different fibrosis stages on the imaging parameters. Both the experimental and simulated results indicate that the polarized light microscope and transformed Mueller matrix parameter scan provide additional quantitative information helpful for fibrosis detections and scorings of liver cirrhosis and cancers. Therefore, the polarized light microscope and transformed Mueller matrix parameters have a good application prospect in liver cancer diagnosis.

  11. Mueller matrix microscope: a quantitative tool to facilitate detections and fibrosis scorings of liver cirrhosis and cancer tissues

    NASA Astrophysics Data System (ADS)

    Wang, Ye; He, Honghui; Chang, Jintao; He, Chao; Liu, Shaoxiong; Li, Migao; Zeng, Nan; Wu, Jian; Ma, Hui

    2016-07-01

    Today the increasing cancer incidence rate is becoming one of the biggest threats to human health. Among all types of cancers, liver cancer ranks in the top five in both frequency and mortality rate all over the world. During the development of liver cancer, fibrosis often evolves as part of a healing process in response to liver damage, resulting in cirrhosis of liver tissues. In a previous study, we applied the Mueller matrix microscope to pathological liver tissue samples and found that both the Mueller matrix polar decomposition (MMPD) and Mueller matrix transformation (MMT) parameters are closely related to the fibrous microstructures. In this paper, we take this one step further to quantitatively facilitate the fibrosis detections and scorings of pathological liver tissue samples in different stages from cirrhosis to cancer using the Mueller matrix microscope. The experimental results of MMPD and MMT parameters for the fibrotic liver tissue samples in different stages are measured and analyzed. We also conduct Monte Carlo simulations based on the sphere birefringence model to examine in detail the influence of structural changes in different fibrosis stages on the imaging parameters. Both the experimental and simulated results indicate that the polarized light microscope and transformed Mueller matrix parameters can provide additional quantitative information helpful for fibrosis detections and scorings of liver cirrhosis and cancers. Therefore, the polarized light microscope and transformed Mueller matrix parameters have a good application prospect in liver cancer diagnosis.

  12. Constrained Deep Weak Supervision for Histopathology Image Segmentation.

    PubMed

    Jia, Zhipeng; Huang, Xingyi; Chang, Eric I-Chao; Xu, Yan

    2017-11-01

    In this paper, we develop a new weakly supervised learning algorithm to learn to segment cancerous regions in histopathology images. This paper is under a multiple instance learning (MIL) framework with a new formulation, deep weak supervision (DWS); we also propose an effective way to introduce constraints to our neural networks to assist the learning process. The contributions of our algorithm are threefold: 1) we build an end-to-end learning system that segments cancerous regions with fully convolutional networks (FCNs) in which image-to-image weakly-supervised learning is performed; 2) we develop a DWS formulation to exploit multi-scale learning under weak supervision within FCNs; and 3) constraints about positive instances are introduced in our approach to effectively explore additional weakly supervised information that is easy to obtain and enjoy a significant boost to the learning process. The proposed algorithm, abbreviated as DWS-MIL, is easy to implement and can be trained efficiently. Our system demonstrates the state-of-the-art results on large-scale histopathology image data sets and can be applied to various applications in medical imaging beyond histopathology images, such as MRI, CT, and ultrasound images.

  13. Web-based tool for visualization of electric field distribution in deep-seated body structures and planning of electroporation-based treatments.

    PubMed

    Marčan, Marija; Pavliha, Denis; Kos, Bor; Forjanič, Tadeja; Miklavčič, Damijan

    2015-01-01

    Treatments based on electroporation are a new and promising approach to treating tumors, especially non-resectable ones. The success of the treatment is, however, heavily dependent on coverage of the entire tumor volume with a sufficiently high electric field. Ensuring complete coverage in the case of deep-seated tumors is not trivial and can in best way be ensured by patient-specific treatment planning. The basis of the treatment planning process consists of two complex tasks: medical image segmentation, and numerical modeling and optimization. In addition to previously developed segmentation algorithms for several tissues (human liver, hepatic vessels, bone tissue and canine brain) and the algorithms for numerical modeling and optimization of treatment parameters, we developed a web-based tool to facilitate the translation of the algorithms and their application in the clinic. The developed web-based tool automatically builds a 3D model of the target tissue from the medical images uploaded by the user and then uses this 3D model to optimize treatment parameters. The tool enables the user to validate the results of the automatic segmentation and make corrections if necessary before delivering the final treatment plan. Evaluation of the tool was performed by five independent experts from four different institutions. During the evaluation, we gathered data concerning user experience and measured performance times for different components of the tool. Both user reports and performance times show significant reduction in treatment-planning complexity and time-consumption from 1-2 days to a few hours. The presented web-based tool is intended to facilitate the treatment planning process and reduce the time needed for it. It is crucial for facilitating expansion of electroporation-based treatments in the clinic and ensuring reliable treatment for the patients. The additional value of the tool is the possibility of easy upgrade and integration of modules with new

  14. Deep machine learning based Image classification in hard disk drive manufacturing (Conference Presentation)

    NASA Astrophysics Data System (ADS)

    Rana, Narender; Chien, Chester

    2018-03-01

    A key sensor element in a Hard Disk Drive (HDD) is the read-write head device. The device is complex 3D shape and its fabrication requires over thousand process steps with many of them being various types of image inspection and critical dimension (CD) metrology steps. In order to have high yield of devices across a wafer, very tight inspection and metrology specifications are implemented. Many images are collected on a wafer and inspected for various types of defects and in CD metrology the quality of image impacts the CD measurements. Metrology noise need to be minimized in CD metrology to get better estimate of the process related variations for implementing robust process controls. Though there are specialized tools available for defect inspection and review allowing classification and statistics. However, due to unavailability of such advanced tools or other reasons, many times images need to be manually inspected. SEM Image inspection and CD-SEM metrology tools are different tools differing in software as well. SEM Image inspection and CD-SEM metrology tools are separate tools differing in software and purpose. There have been cases where a significant numbers of CD-SEM images are blurred or have some artefact and there is a need for image inspection along with the CD measurement. Tool may not report a practical metric highlighting the quality of image. Not filtering CD from these blurred images will add metrology noise to the CD measurement. An image classifier can be helpful here for filtering such data. This paper presents the use of artificial intelligence in classifying the SEM images. Deep machine learning is used to train a neural network which is then used to classify the new images as blurred and not blurred. Figure 1 shows the image blur artefact and contingency table of classification results from the trained deep neural network. Prediction accuracy of 94.9 % was achieved in the first model. Paper covers other such applications of the deep neural

  15. Deep tissue optical focusing and optogenetic modulation with time-reversed ultrasonically encoded light

    PubMed Central

    Ruan, Haowen; Brake, Joshua; Robinson, J. Elliott; Liu, Yan; Jang, Mooseok; Xiao, Cheng; Zhou, Chunyi; Gradinaru, Viviana; Yang, Changhuei

    2017-01-01

    Noninvasive light focusing deep inside living biological tissue has long been a goal in biomedical optics. However, the optical scattering of biological tissue prevents conventional optical systems from tightly focusing visible light beyond several hundred micrometers. The recently developed wavefront shaping technique time-reversed ultrasonically encoded (TRUE) focusing enables noninvasive light delivery to targeted locations beyond the optical diffusion limit. However, until now, TRUE focusing has only been demonstrated inside nonliving tissue samples. We present the first example of TRUE focusing in 2-mm-thick living brain tissue and demonstrate its application for optogenetic modulation of neural activity in 800-μm-thick acute mouse brain slices at a wavelength of 532 nm. We found that TRUE focusing enabled precise control of neuron firing and increased the spatial resolution of neuronal excitation fourfold when compared to conventional lens focusing. This work is an important step in the application of TRUE focusing for practical biomedical uses. PMID:29226248

  16. Direct tissue oxygen monitoring by in vivo photoacoustic lifetime imaging (PALI)

    NASA Astrophysics Data System (ADS)

    Shao, Qi; Morgounova, Ekaterina; Ashkenazi, Shai

    2014-03-01

    Tissue oxygen plays a critical role in maintaining tissue viability and in various diseases, including response to therapy. Images of oxygen distribution provide the history of tissue hypoxia and evidence of oxygen availability in the circulatory system. Currently available methods of direct measuring or imaging tissue oxygen all have significant limitations. Previously, we have reported a non-invasive in vivo imaging modality based on photoacoustic lifetime. The technique maps the excited triplet state of oxygen-sensitive dye, thus reflects the spatial and temporal distribution of tissue oxygen. We have applied PALI on tumor hypoxia in small animals, and the hypoxic region imaged by PALI is consistent with the site of the tumor imaged by ultrasound. Here, we present two studies of applying PALI to monitor changes of tissue oxygen by modulations. The first study involves an acute ischemia model using a thin thread tied around the hind limb of a normal mouse to reduce the blood flow. PALI images were acquired before, during, and after the restriction. The drop of muscle pO2 and recovery from hypoxia due to reperfusion were observed by PALI tracking the same region. The second study modulates tissue oxygen by controlling the percentage of oxygen the mouse inhales. We demonstrate that PALI is able to reflect the change of oxygen level with respect to both hyperbaric and hypobaric conditions. We expect this technique to be very attractive for a range of clinical applications in which tissue oxygen mapping would improve therapy decision making and treatment planning.

  17. Compact Tissue-equivalent Proportional Counter for Deep Space Human Missions.

    PubMed

    Straume, T; Braby, L A; Borak, T B; Lusby, T; Warner, D W; Perez-Nunez, D

    2015-10-01

    Effects on human health from the complex radiation environment in deep space have not been measured and can only be simulated here on Earth using experimental systems and beams of radiations produced by accelerators, usually one beam at a time. This makes it particularly important to develop instruments that can be used on deep-space missions to measure quantities that are known to be relatable to the biological effectiveness of space radiation. Tissue-equivalent proportional counters (TEPCs) are such instruments. Unfortunately, present TEPCs are too large and power intensive to be used beyond low Earth orbit (LEO). Here, the authors describe a prototype of a compact TEPC designed for deep space applications with the capability to detect both ambient galactic cosmic rays and intense solar particle event radiation. The device employs an approach that permits real-time determination of yD (and thus quality factor) using a single detector. This was accomplished by assigning sequential sampling intervals as detectors “1” and “2” and requiring the intervals to be brief compared to the change in dose rate. Tests with g rays show that the prototype instrument maintains linear response over the wide dose-rate range expected in space with an accuracy of better than 5% for dose rates above 3 mGy h(-1). Measurements of yD for 200 MeV n(-1) carbon ions were better than 10%. Limited tests with fission spectrum neutrons show absorbed dose-rate accuracy better than 15%.

  18. Compact Tissue-equivalent Proportional Counter for Deep Space Human Missions

    PubMed Central

    Straume, T.; Braby, L.A.; Borak, T.B.; Lusby, T.; Warner, D.W.; Perez-Nunez, D.

    2015-01-01

    Abstract Effects on human health from the complex radiation environment in deep space have not been measured and can only be simulated here on Earth using experimental systems and beams of radiations produced by accelerators, usually one beam at a time. This makes it particularly important to develop instruments that can be used on deep-space missions to measure quantities that are known to be relatable to the biological effectiveness of space radiation. Tissue-equivalent proportional counters (TEPCs) are such instruments. Unfortunately, present TEPCs are too large and power intensive to be used beyond low Earth orbit (LEO). Here, the authors describe a prototype of a compact TEPC designed for deep space applications with the capability to detect both ambient galactic cosmic rays and intense solar particle event radiation. The device employs an approach that permits real-time determination of (and thus quality factor) using a single detector. This was accomplished by assigning sequential sampling intervals as detectors “1” and “2” and requiring the intervals to be brief compared to the change in dose rate. Tests with γ rays show that the prototype instrument maintains linear response over the wide dose-rate range expected in space with an accuracy of better than 5% for dose rates above 3 mGy h−1. Measurements of for 200 MeV n−1 carbon ions were better than 10%. Limited tests with fission spectrum neutrons show absorbed dose-rate accuracy better than 15%. PMID:26313585

  19. New approaches in renal microscopy: volumetric imaging and superresolution microscopy.

    PubMed

    Kim, Alfred H J; Suleiman, Hani; Shaw, Andrey S

    2016-05-01

    Histologic and electron microscopic analysis of the kidney has provided tremendous insight into structures such as the glomerulus and nephron. Recent advances in imaging, such as deep volumetric approaches and superresolution microscopy, have the capacity to dramatically enhance our current understanding of the structure and function of the kidney. Volumetric imaging can generate images millimeters below the surface of the intact kidney. Superresolution microscopy breaks the diffraction barrier inherent in traditional light microscopy, enabling the visualization of fine structures. Here, we describe new approaches to deep volumetric and superresolution microscopy of the kidney. Rapid advances in lasers, microscopic objectives, and tissue preparation have transformed our ability to deep volumetric image the kidney. Innovations in sample preparation have allowed for superresolution imaging with electron microscopy correlation, providing unprecedented insight into the structures within the glomerulus. Technological advances in imaging have revolutionized our capacity to image both large volumes of tissue and the finest structural details of a cell. These new advances have the potential to provide additional profound observations into the normal and pathologic functions of the kidney.

  20. A polarization sensitive hyperspectral imaging system for detection of differences in tissue properties

    NASA Astrophysics Data System (ADS)

    Peller, Joseph A.; Ceja, Nancy K.; Wawak, Amanda J.; Trammell, Susan R.

    2018-02-01

    Polarized light imaging and optical spectroscopy can be used to distinguish between healthy and diseased tissue. In this study, the design and testing of a single-pixel hyperspectral imaging system that uses differences in the polarization of light reflected from tissue to differentiate between healthy and thermally damaged tissue is discussed. Thermal lesions were created in porcine skin (n = 8) samples using an IR laser. The damaged regions were clearly visible in the polarized light hyperspectral images. Reflectance hyperspectral and white light imaging was also obtained for all tissue samples. Sizes of the thermally damaged regions as measured via polarized light hyperspectral imaging are compared to sizes of these regions as measured in the reflectance hyperspectral images and white light images. Good agreement between the sizes measured by all three imaging modalities was found. Hyperspectral polarized light imaging can differentiate between healthy and damaged tissue. Possible applications of this imaging system include determination of tumor margins during cancer surgery or pre-surgical biopsy.

  1. Imaging of cochlear tissue with a grating interferometer and hard X-rays

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Richter, Claus-Peter; Shintani-Smith, Stephanie; Fishman, Andrew

    This article addresses an important current development in medical and biological imaging: the possibility of imaging soft tissue at resolutions in the micron range using hard X-rays. Challenging environments, including the cochlea, require the imaging of soft tissue structure surrounded by bone. We demonstrate that cochlear soft tissue structures can be imaged with hard X-ray phase contrast. Furthermore, we show that only a thin slice of the tissue is required to introduce a large phase shift. It is likely that the phase contrast image of the soft tissue structures is sufficient to image the structures even if surrounded by bone.more » For the present set of experiments, structures with low-absorption contrast have been visualized using in-line phase contrast imaging and a grating interferometer. The experiments have been performed at the Advanced Photon Source at Argonne National Laboratories, a third generation source of synchrotron radiation. The source provides highly coherent X-ray radiation with high-photon flux (>10{sup 12} photons/s) at high-photon energies (5-70 keV). Radiographic and light microscopy images of the gerbil cochlear slice samples were compared. It has been determined that a 20-{micro}m thick tissue slice induces a phase shift between 1/3{pi} and 2/3{pi}.« less

  2. Fast blood flow monitoring in deep tissues with real-time software correlators

    PubMed Central

    Wang, Detian; Parthasarathy, Ashwin B.; Baker, Wesley B.; Gannon, Kimberly; Kavuri, Venki; Ko, Tiffany; Schenkel, Steven; Li, Zhe; Li, Zeren; Mullen, Michael T.; Detre, John A.; Yodh, Arjun G.

    2016-01-01

    We introduce, validate and demonstrate a new software correlator for high-speed measurement of blood flow in deep tissues based on diffuse correlation spectroscopy (DCS). The software correlator scheme employs standard PC-based data acquisition boards to measure temporal intensity autocorrelation functions continuously at 50 – 100 Hz, the fastest blood flow measurements reported with DCS to date. The data streams, obtained in vivo for typical source-detector separations of 2.5 cm, easily resolve pulsatile heart-beat fluctuations in blood flow which were previously considered to be noise. We employ the device to separate tissue blood flow from tissue absorption/scattering dynamics and thereby show that the origin of the pulsatile DCS signal is primarily flow, and we monitor cerebral autoregulation dynamics in healthy volunteers more accurately than with traditional instrumentation as a result of increased data acquisition rates. Finally, we characterize measurement signal-to-noise ratio and identify count rate and averaging parameters needed for optimal performance. PMID:27231588

  3. All-near-infrared multiphoton microscopy interrogates intact tissues at deeper imaging depths than conventional single- and two-photon near-infrared excitation microscopes

    PubMed Central

    Sarder, Pinaki; Yazdanfar, Siavash; Akers, Walter J.; Tang, Rui; Sudlow, Gail P.; Egbulefu, Christopher

    2013-01-01

    Abstract. The era of molecular medicine has ushered in the development of microscopic methods that can report molecular processes in thick tissues with high spatial resolution. A commonality in deep-tissue microscopy is the use of near-infrared (NIR) lasers with single- or multiphoton excitations. However, the relationship between different NIR excitation microscopic techniques and the imaging depths in tissue has not been established. We compared such depth limits for three NIR excitation techniques: NIR single-photon confocal microscopy (NIR SPCM), NIR multiphoton excitation with visible detection (NIR/VIS MPM), and all-NIR multiphoton excitation with NIR detection (NIR/NIR MPM). Homologous cyanine dyes provided the fluorescence. Intact kidneys were harvested after administration of kidney-clearing cyanine dyes in mice. NIR SPCM and NIR/VIS MPM achieved similar maximum imaging depth of ∼100  μm. The NIR/NIR MPM enabled greater than fivefold imaging depth (>500  μm) using the harvested kidneys. Although the NIR/NIR MPM used 1550-nm excitation where water absorption is relatively high, cell viability and histology studies demonstrate that the laser did not induce photothermal damage at the low laser powers used for the kidney imaging. This study provides guidance on the imaging depth capabilities of NIR excitation-based microscopic techniques and reveals the potential to multiplex information using these platforms. PMID:24150231

  4. Soft-Tissue Infections and Their Imaging Mimics: From Cellulitis to Necrotizing Fasciitis.

    PubMed

    Hayeri, Mohammad Reza; Ziai, Pouya; Shehata, Monda L; Teytelboym, Oleg M; Huang, Brady K

    2016-10-01

    Infection of the musculoskeletal system can be associated with high mortality and morbidity if not promptly and accurately diagnosed. These infections are generally diagnosed and managed clinically; however, clinical and laboratory findings sometimes lack sensitivity and specificity, and a definite diagnosis may not be possible. In uncertain situations, imaging is frequently performed to confirm the diagnosis, evaluate the extent of the disease, and aid in treatment planning. In particular, cross-sectional imaging, including computed tomography and magnetic resonance imaging, provides detailed anatomic information in the evaluation of soft tissues due to their inherent high spatial and contrast resolution. Imaging findings of soft-tissue infections can be nonspecific and can have different appearances depending on the depth and anatomic extent of tissue involvement. Although many imaging features of infectious disease can overlap with noninfectious processes, imaging can help establish the diagnosis when combined with the clinical history and laboratory findings. Radiologists should be familiar with the spectrum of imaging findings of soft-tissue infections to better aid the referring physician in managing these patients. The aim of this article is to review the spectrum of soft-tissue infections using a systematic anatomic compartment approach. We discuss the clinical features of soft-tissue infections, their imaging findings with emphasis on cross-sectional imaging, their potential mimics, and clinical management. © RSNA, 2016.

  5. Enhancing SDO/HMI images using deep learning

    NASA Astrophysics Data System (ADS)

    Baso, C. J. Díaz; Ramos, A. Asensio

    2018-06-01

    Context. The Helioseismic and Magnetic Imager (HMI) provides continuum images and magnetograms with a cadence better than one per minute. It has been continuously observing the Sun 24 h a day for the past 7 yr. The trade-off between full disk observations and spatial resolution means that HMI is not adequate for analyzing the smallest-scale events in the solar atmosphere. Aims: Our aim is to develop a new method to enhance HMI data, simultaneously deconvolving and super-resolving images and magnetograms. The resulting images will mimic observations with a diffraction-limited telescope twice the diameter of HMI. Methods: Our method, which we call Enhance, is based on two deep, fully convolutional neural networks that input patches of HMI observations and output deconvolved and super-resolved data. The neural networks are trained on synthetic data obtained from simulations of the emergence of solar active regions. Results: We have obtained deconvolved and super-resolved HMI images. To solve this ill-defined problem with infinite solutions we have used a neural network approach to add prior information from the simulations. We test Enhance against Hinode data that has been degraded to a 28 cm diameter telescope showing very good consistency. The code is open source.

  6. Robust Single Image Super-Resolution via Deep Networks With Sparse Prior.

    PubMed

    Liu, Ding; Wang, Zhaowen; Wen, Bihan; Yang, Jianchao; Han, Wei; Huang, Thomas S

    2016-07-01

    Single image super-resolution (SR) is an ill-posed problem, which tries to recover a high-resolution image from its low-resolution observation. To regularize the solution of the problem, previous methods have focused on designing good priors for natural images, such as sparse representation, or directly learning the priors from a large data set with models, such as deep neural networks. In this paper, we argue that domain expertise from the conventional sparse coding model can be combined with the key ingredients of deep learning to achieve further improved results. We demonstrate that a sparse coding model particularly designed for SR can be incarnated as a neural network with the merit of end-to-end optimization over training data. The network has a cascaded structure, which boosts the SR performance for both fixed and incremental scaling factors. The proposed training and testing schemes can be extended for robust handling of images with additional degradation, such as noise and blurring. A subjective assessment is conducted and analyzed in order to thoroughly evaluate various SR techniques. Our proposed model is tested on a wide range of images, and it significantly outperforms the existing state-of-the-art methods for various scaling factors both quantitatively and perceptually.

  7. In vivo multiphoton tomography and fluorescence lifetime imaging of human brain tumor tissue.

    PubMed

    Kantelhardt, Sven R; Kalasauskas, Darius; König, Karsten; Kim, Ella; Weinigel, Martin; Uchugonova, Aisada; Giese, Alf

    2016-05-01

    High resolution multiphoton tomography and fluorescence lifetime imaging differentiates glioma from adjacent brain in native tissue samples ex vivo. Presently, multiphoton tomography is applied in clinical dermatology and experimentally. We here present the first application of multiphoton and fluorescence lifetime imaging for in vivo imaging on humans during a neurosurgical procedure. We used a MPTflex™ Multiphoton Laser Tomograph (JenLab, Germany). We examined cultured glioma cells in an orthotopic mouse tumor model and native human tissue samples. Finally the multiphoton tomograph was applied to provide optical biopsies during resection of a clinical case of glioblastoma. All tissues imaged by multiphoton tomography were sampled and processed for conventional histopathology. The multiphoton tomograph allowed fluorescence intensity- and fluorescence lifetime imaging with submicron spatial resolution and 200 picosecond temporal resolution. Morphological fluorescence intensity imaging and fluorescence lifetime imaging of tumor-bearing mouse brains and native human tissue samples clearly differentiated tumor and adjacent brain tissue. Intraoperative imaging was found to be technically feasible. Intraoperative image quality was comparable to ex vivo examinations. To our knowledge we here present the first intraoperative application of high resolution multiphoton tomography and fluorescence lifetime imaging of human brain tumors in situ. It allowed in vivo identification and determination of cell density of tumor tissue on a cellular and subcellular level within seconds. The technology shows the potential of rapid intraoperative identification of native glioma tissue without need for tissue processing or staining.

  8. Automated Melanoma Recognition in Dermoscopy Images via Very Deep Residual Networks.

    PubMed

    Yu, Lequan; Chen, Hao; Dou, Qi; Qin, Jing; Heng, Pheng-Ann

    2017-04-01

    Automated melanoma recognition in dermoscopy images is a very challenging task due to the low contrast of skin lesions, the huge intraclass variation of melanomas, the high degree of visual similarity between melanoma and non-melanoma lesions, and the existence of many artifacts in the image. In order to meet these challenges, we propose a novel method for melanoma recognition by leveraging very deep convolutional neural networks (CNNs). Compared with existing methods employing either low-level hand-crafted features or CNNs with shallower architectures, our substantially deeper networks (more than 50 layers) can acquire richer and more discriminative features for more accurate recognition. To take full advantage of very deep networks, we propose a set of schemes to ensure effective training and learning under limited training data. First, we apply the residual learning to cope with the degradation and overfitting problems when a network goes deeper. This technique can ensure that our networks benefit from the performance gains achieved by increasing network depth. Then, we construct a fully convolutional residual network (FCRN) for accurate skin lesion segmentation, and further enhance its capability by incorporating a multi-scale contextual information integration scheme. Finally, we seamlessly integrate the proposed FCRN (for segmentation) and other very deep residual networks (for classification) to form a two-stage framework. This framework enables the classification network to extract more representative and specific features based on segmented results instead of the whole dermoscopy images, further alleviating the insufficiency of training data. The proposed framework is extensively evaluated on ISBI 2016 Skin Lesion Analysis Towards Melanoma Detection Challenge dataset. Experimental results demonstrate the significant performance gains of the proposed framework, ranking the first in classification and the second in segmentation among 25 teams and 28 teams

  9. Anatomical and Functional Images of in vitro and in vivo Tissues by NIR Time-domain Diffuse Optical Tomography

    NASA Astrophysics Data System (ADS)

    Zhao, Huijuan; Gao, Feng; Tanikawa, Yukari; Homma, Kazuhiro; Onodera, Yoichi; Yamada, Yukio

    Near infra-red (NIR) diffuse optical tomography (DOT) has gained much attention and it will be clinically applied to imaging breast, neonatal head, and the hemodynamics of the brain because of its noninvasiveness and deep penetration in biological tissue. Prior to achieving the imaging of infant brain using DOT, the developed methodologies need to be experimentally justified by imaging some real organs with simpler structures. Here we report our results of an in vitro chicken leg and an in vivo exercising human forearm from the data measured by a multi-channel time-resolved NIR system. Tomographic images were reconstructed by a two-dimensional image reconstruction algorithm based on a modified generalized pulse spectrum technique for simultaneous reconstruction of the µa and µs´. The absolute µa- and µs´-images revealed the inner structures of the chicken leg and the forearm, where the bones were clearly distinguished from the muscle. The Δµa-images showed the blood volume changes during the forearm exercise, proving that the system and the image reconstruction algorithm could potentially be used for imaging not only the anatomic structure but also the hemodynamics in neonatal heads.

  10. Deformable image registration for tissues with large displacements

    PubMed Central

    Huang, Xishi; Ren, Jing; Green, Mark

    2017-01-01

    Abstract. Image registration for internal organs and soft tissues is considered extremely challenging due to organ shifts and tissue deformation caused by patients’ movements such as respiration and repositioning. In our previous work, we proposed a fast registration method for deformable tissues with small rotations. We extend our method to deformable registration of soft tissues with large displacements. We analyzed the deformation field of the liver by decomposing the deformation into shift, rotation, and pure deformation components and concluded that in many clinical cases, the liver deformation contains large rotations and small deformations. This analysis justified the use of linear elastic theory in our image registration method. We also proposed a region-based neuro-fuzzy transformation model to seamlessly stitch together local affine and local rigid models in different regions. We have performed the experiments on a liver MRI image set and showed the effectiveness of the proposed registration method. We have also compared the performance of the proposed method with the previous method on tissues with large rotations and showed that the proposed method outperformed the previous method when dealing with the combination of pure deformation and large rotations. Validation results show that we can achieve a target registration error of 1.87±0.87  mm and an average centerline distance error of 1.28±0.78  mm. The proposed technique has the potential to significantly improve registration capabilities and the quality of intraoperative image guidance. To the best of our knowledge, this is the first time that the complex displacement of the liver is explicitly separated into local pure deformation and rigid motion. PMID:28149924

  11. Tissue Pulsatility Imaging of Cerebral Vasoreactivity during Hyperventilation

    PubMed Central

    Kucewicz, John C.; Dunmire, Barbrina; Giardino, Nicholas D.; Leotta, Daniel F.; Paun, Marla; Dager, Stephen R.; Beach, Kirk W.

    2008-01-01

    Tissue Pulsatility Imaging (TPI) is an ultrasonic technique that is being developed at the University of Washington to measure tissue displacement or strain due to blood flow over the cardiac and respiratory cycles. This technique is based in principle on plethysmography, an older non-ultrasound technology for measuring expansion of a whole limb or body part due to perfusion. TPI adapts tissue Doppler signal processing methods to measure the “plethysmographic” signal from hundreds or thousands of sample volumes in an ultrasound image plane. This paper presents a feasibility study to determine if TPI can be used to assess cerebral vasoreactivity. Ultrasound data were collected transcranially through the temporal acoustic window from four subjects before, during, and after voluntary hyperventilation. In each subject, decreases in tissue pulsatility during hyperventilation were observed that were statistically correlated with the subject’s end-tidal CO2 measurements. PMID:18336991

  12. Deep learning in the small sample size setting: cascaded feed forward neural networks for medical image segmentation

    NASA Astrophysics Data System (ADS)

    Gaonkar, Bilwaj; Hovda, David; Martin, Neil; Macyszyn, Luke

    2016-03-01

    Deep Learning, refers to large set of neural network based algorithms, have emerged as promising machine- learning tools in the general imaging and computer vision domains. Convolutional neural networks (CNNs), a specific class of deep learning algorithms, have been extremely effective in object recognition and localization in natural images. A characteristic feature of CNNs, is the use of a locally connected multi layer topology that is inspired by the animal visual cortex (the most powerful vision system in existence). While CNNs, perform admirably in object identification and localization tasks, typically require training on extremely large datasets. Unfortunately, in medical image analysis, large datasets are either unavailable or are extremely expensive to obtain. Further, the primary tasks in medical imaging are organ identification and segmentation from 3D scans, which are different from the standard computer vision tasks of object recognition. Thus, in order to translate the advantages of deep learning to medical image analysis, there is a need to develop deep network topologies and training methodologies, that are geared towards medical imaging related tasks and can work in a setting where dataset sizes are relatively small. In this paper, we present a technique for stacked supervised training of deep feed forward neural networks for segmenting organs from medical scans. Each `neural network layer' in the stack is trained to identify a sub region of the original image, that contains the organ of interest. By layering several such stacks together a very deep neural network is constructed. Such a network can be used to identify extremely small regions of interest in extremely large images, inspite of a lack of clear contrast in the signal or easily identifiable shape characteristics. What is even more intriguing is that the network stack achieves accurate segmentation even when it is trained on a single image with manually labelled ground truth. We validate

  13. Precursors to radiopharmaceutical agents for tissue imaging

    DOEpatents

    Srivastava, Prem C.; Knapp, Jr., Furn F.

    1988-01-01

    A class of radiolabeled compounds to be used in tissue imaging that exhibits rapid brain uptake, good brain:blood radioactivity ratios, and long retention times. The imaging agents are more specifically radioiodinated aromatic amines attached to dihydropyridine carriers, that exhibit heart as well as brain specificity. In addition to the radiolabeled compounds, classes of compounds are also described that are used as precursors and intermediates in the preparation of the imaging agents.

  14. Deep learning of joint myelin and T1w MRI features in normal-appearing brain tissue to distinguish between multiple sclerosis patients and healthy controls.

    PubMed

    Yoo, Youngjin; Tang, Lisa Y W; Brosch, Tom; Li, David K B; Kolind, Shannon; Vavasour, Irene; Rauscher, Alexander; MacKay, Alex L; Traboulsee, Anthony; Tam, Roger C

    2018-01-01

    Myelin imaging is a form of quantitative magnetic resonance imaging (MRI) that measures myelin content and can potentially allow demyelinating diseases such as multiple sclerosis (MS) to be detected earlier. Although focal lesions are the most visible signs of MS pathology on conventional MRI, it has been shown that even tissues that appear normal may exhibit decreased myelin content as revealed by myelin-specific images (i.e., myelin maps). Current methods for analyzing myelin maps typically use global or regional mean myelin measurements to detect abnormalities, but ignore finer spatial patterns that may be characteristic of MS. In this paper, we present a machine learning method to automatically learn, from multimodal MR images, latent spatial features that can potentially improve the detection of MS pathology at early stage. More specifically, 3D image patches are extracted from myelin maps and the corresponding T1-weighted (T1w) MRIs, and are used to learn a latent joint myelin-T1w feature representation via unsupervised deep learning. Using a data set of images from MS patients and healthy controls, a common set of patches are selected via a voxel-wise t -test performed between the two groups. In each MS image, any patches overlapping with focal lesions are excluded, and a feature imputation method is used to fill in the missing values. A feature selection process (LASSO) is then utilized to construct a sparse representation. The resulting normal-appearing features are used to train a random forest classifier. Using the myelin and T1w images of 55 relapse-remitting MS patients and 44 healthy controls in an 11-fold cross-validation experiment, the proposed method achieved an average classification accuracy of 87.9% (SD = 8.4%), which is higher and more consistent across folds than those attained by regional mean myelin (73.7%, SD = 13.7%) and T1w measurements (66.7%, SD = 10.6%), or deep-learned features in either the myelin (83.8%, SD = 11.0%) or T1w

  15. Chromatic confocal microscopy for multi-depth imaging of epithelial tissue

    PubMed Central

    Olsovsky, Cory; Shelton, Ryan; Carrasco-Zevallos, Oscar; Applegate, Brian E.; Maitland, Kristen C.

    2013-01-01

    We present a novel chromatic confocal microscope capable of volumetric reflectance imaging of microstructure in non-transparent tissue. Our design takes advantage of the chromatic aberration of aspheric lenses that are otherwise well corrected. Strong chromatic aberration, generated by multiple aspheres, longitudinally disperses supercontinuum light onto the sample. The backscattered light detected with a spectrometer is therefore wavelength encoded and each spectrum corresponds to a line image. This approach obviates the need for traditional axial mechanical scanning techniques that are difficult to implement for endoscopy and susceptible to motion artifact. A wavelength range of 590-775 nm yielded a >150 µm imaging depth with ~3 µm axial resolution. The system was further demonstrated by capturing volumetric images of buccal mucosa. We believe these represent the first microstructural images in non-transparent biological tissue using chromatic confocal microscopy that exhibit long imaging depth while maintaining acceptable resolution for resolving cell morphology. Miniaturization of this optical system could bring enhanced speed and accuracy to endomicroscopic in vivo volumetric imaging of epithelial tissue. PMID:23667789

  16. Fully automated, deep learning segmentation of oxygen-induced retinopathy images

    PubMed Central

    Xiao, Sa; Bucher, Felicitas; Wu, Yue; Rokem, Ariel; Lee, Cecilia S.; Marra, Kyle V.; Fallon, Regis; Diaz-Aguilar, Sophia; Aguilar, Edith; Friedlander, Martin; Lee, Aaron Y.

    2017-01-01

    Oxygen-induced retinopathy (OIR) is a widely used model to study ischemia-driven neovascularization (NV) in the retina and to serve in proof-of-concept studies in evaluating antiangiogenic drugs for ocular, as well as nonocular, diseases. The primary parameters that are analyzed in this mouse model include the percentage of retina with vaso-obliteration (VO) and NV areas. However, quantification of these two key variables comes with a great challenge due to the requirement of human experts to read the images. Human readers are costly, time-consuming, and subject to bias. Using recent advances in machine learning and computer vision, we trained deep learning neural networks using over a thousand segmentations to fully automate segmentation in OIR images. While determining the percentage area of VO, our algorithm achieved a similar range of correlation coefficients to that of expert inter-human correlation coefficients. In addition, our algorithm achieved a higher range of correlation coefficients compared with inter-expert correlation coefficients for quantification of the percentage area of neovascular tufts. In summary, we have created an open-source, fully automated pipeline for the quantification of key values of OIR images using deep learning neural networks. PMID:29263301

  17. Compact, multi-exposure speckle contrast optical spectroscopy (SCOS) device for measuring deep tissue blood flow

    PubMed Central

    Dragojević, Tanja; Hollmann, Joseph L.; Tamborini, Davide; Portaluppi, Davide; Buttafava, Mauro; Culver, Joseph P.; Villa, Federica; Durduran, Turgut

    2017-01-01

    Speckle contrast optical spectroscopy (SCOS) measures absolute blood flow in deep tissue, by taking advantage of multi-distance (previously reported in the literature) or multi-exposure (reported here) approach. This method promises to use inexpensive detectors to obtain good signal-to-noise ratio, but it has not yet been implemented in a suitable manner for a mass production. Here we present a new, compact, low power consumption, 32 by 2 single photon avalanche diode (SPAD) array that has no readout noise, low dead time and has high sensitivity in low light conditions, such as in vivo measurements. To demonstrate the capability to measure blood flow in deep tissue, healthy volunteers were measured, showing no significant differences from the diffuse correlation spectroscopy. In the future, this array can be miniaturized to a low-cost, robust, battery operated wireless device paving the way for measuring blood flow in a wide-range of applications from sport injury recovery and training to, on-field concussion detection to wearables. PMID:29359106

  18. Deep multi-spectral ensemble learning for electronic cleansing in dual-energy CT colonography

    NASA Astrophysics Data System (ADS)

    Tachibana, Rie; Näppi, Janne J.; Hironaka, Toru; Kim, Se Hyung; Yoshida, Hiroyuki

    2017-03-01

    We developed a novel electronic cleansing (EC) method for dual-energy CT colonography (DE-CTC) based on an ensemble deep convolution neural network (DCNN) and multi-spectral multi-slice image patches. In the method, an ensemble DCNN is used to classify each voxel of a DE-CTC image volume into five classes: luminal air, soft tissue, tagged fecal materials, and partial-volume boundaries between air and tagging and those between soft tissue and tagging. Each DCNN acts as a voxel classifier, where an input image patch centered at the voxel is generated as input to the DCNNs. An image patch has three channels that are mapped from a region-of-interest containing the image plane of the voxel and the two adjacent image planes. Six different types of spectral input image datasets were derived using two dual-energy CT images, two virtual monochromatic images, and two material images. An ensemble DCNN was constructed by use of a meta-classifier that combines the output of multiple DCNNs, each of which was trained with a different type of multi-spectral image patches. The electronically cleansed CTC images were calculated by removal of regions classified as other than soft tissue, followed by a colon surface reconstruction. For pilot evaluation, 359 volumes of interest (VOIs) representing sources of subtraction artifacts observed in current EC schemes were sampled from 30 clinical CTC cases. Preliminary results showed that the ensemble DCNN can yield high accuracy in labeling of the VOIs, indicating that deep learning of multi-spectral EC with multi-slice imaging could accurately remove residual fecal materials from CTC images without generating major EC artifacts.

  19. In Vivo Optical Imaging for Targeted Drug Kinetics and Localization for Oral Surgery and Super-Resolution, Facilitated by Printed Phantoms

    NASA Astrophysics Data System (ADS)

    Bentz, Brian Z.

    Many human cancer cell types over-express folate receptors, and this provides an opportunity to develop targeted anti-cancer drugs. For these drugs to be effective, their kinetics must be well understood in vivo and in deep tissue where tumors occur. We demonstrate a method for imaging these parameters by incorporating a kinetic compartment model and fluorescence into optical diffusion tomography (ODT). The kinetics were imaged in a live mouse, and found to be in agreement with previous in vitro studies, demonstrating the validity of the method and its feasibility as an effective tool in preclinical drug development studies. Progress in developing optical imaging for biomedical applications requires customizable and often complex objects known as "phantoms" for testing and evaluation. We present new optical phantoms fabricated using inexpensive 3D printing methods with multiple materials, allowing for the placement of complex inhomogeneities in heterogeneous or anatomically realistic geometries, as opposed to previous phantoms which were limited to simple shapes formed by molds or machining. Furthermore, we show that Mie theory can be used to design the optical properties to match a target tissue. The phantom fabrication methods are versatile, can be applied to optical imaging methods besides diffusive imaging, and can be used in the calibration of live animal imaging data. Applications of diffuse optical imaging in the operating theater have been limited in part due to computational burden. We present an approach for the fast localization of arteries in the roof of the mouth that has the potential to reduce complications. Furthermore, we use the extracted position information to fabricate a custom surgical guide using 3D printing that could protect the arteries during surgery. The resolution of ODT is severely limited by the attenuation of high spatial frequencies. We present a super-resolution method achieved through the point localization of fluorescent

  20. Imaging of connective tissue diseases of the head and neck

    PubMed Central

    2016-01-01

    We review the imaging appearance of connective tissue diseases of the head and neck. Bilateral sialadenitis and dacryoadenitis are seen in Sjögren’s syndrome; ankylosis of the temporo-mandibular joint with sclerosis of the crico-arytenoid joint are reported in rheumatoid arthritis and lupus panniculitis with atypical infection are reported in patients with systemic lupus erythematosus. Relapsing polychondritis shows subglottic stenosis, prominent ear and saddle nose; progressive systemic sclerosis shows osteolysis of the mandible, fibrosis of the masseter muscle with calcinosis of the subcutaneous tissue and dermatomyositis/polymyositis shows condylar erosions and autoimmune thyroiditis. Vascular thrombosis is reported in antiphospholipid antibodies syndrome; cervical lymphadenopathy is seen in adult-onset Still’s disease, and neuropathy with thyroiditis reported in mixed connective tissue disorder. Imaging is important to detect associated malignancy with connective tissue disorders. Correlation of the imaging findings with demographic data and clinical findings are important for the diagnosis of connective tissue disorders. PMID:26988082

  1. Tissue classification for laparoscopic image understanding based on multispectral texture analysis

    NASA Astrophysics Data System (ADS)

    Zhang, Yan; Wirkert, Sebastian J.; Iszatt, Justin; Kenngott, Hannes; Wagner, Martin; Mayer, Benjamin; Stock, Christian; Clancy, Neil T.; Elson, Daniel S.; Maier-Hein, Lena

    2016-03-01

    Intra-operative tissue classification is one of the prerequisites for providing context-aware visualization in computer-assisted minimally invasive surgeries. As many anatomical structures are difficult to differentiate in conventional RGB medical images, we propose a classification method based on multispectral image patches. In a comprehensive ex vivo study we show (1) that multispectral imaging data is superior to RGB data for organ tissue classification when used in conjunction with widely applied feature descriptors and (2) that combining the tissue texture with the reflectance spectrum improves the classification performance. Multispectral tissue analysis could thus evolve as a key enabling technique in computer-assisted laparoscopy.

  2. Three-dimensional high-resolution second-harmonic generation imaging of endogenous structural proteins in biological tissues.

    PubMed Central

    Campagnola, Paul J; Millard, Andrew C; Terasaki, Mark; Hoppe, Pamela E; Malone, Christian J; Mohler, William A

    2002-01-01

    We find that several key endogenous protein structures give rise to intense second-harmonic generation (SHG)-nonabsorptive frequency doubling of an excitation laser line. Second-harmonic imaging microscopy (SHIM) on a laser-scanning system proves, therefore, to be a powerful and unique tool for high-resolution, high-contrast, three-dimensional studies of live cell and tissue architecture. Unlike fluorescence, SHG suffers no inherent photobleaching or toxicity and does not require exogenous labels. Unlike polarization microscopy, SHIM provides intrinsic confocality and deep sectioning in complex tissues. In this study, we demonstrate the clarity of SHIM optical sectioning within unfixed, unstained thick specimens. SHIM and two-photon excited fluorescence (TPEF) were combined in a dual-mode nonlinear microscopy to elucidate the molecular sources of SHG in live cells and tissues. SHG arose not only from coiled-coil complexes within connective tissues and muscle thick filaments, but also from microtubule arrays within interphase and mitotic cells. Both polarization dependence and a local symmetry cancellation effect of SHG allowed the signal from species generating the second harmonic to be decoded, by ratiometric correlation with TPEF, to yield information on local structure below optical resolution. The physical origin of SHG within these tissues is addressed and is attributed to the laser interaction with dipolar protein structures that is enhanced by the intrinsic chirality of the protein helices. PMID:11751336

  3. Study on the Classification of GAOFEN-3 Polarimetric SAR Images Using Deep Neural Network

    NASA Astrophysics Data System (ADS)

    Zhang, J.; Zhang, J.; Zhao, Z.

    2018-04-01

    Polarimetric Synthetic Aperture Radar (POLSAR) imaging principle determines that the image quality will be affected by speckle noise. So the recognition accuracy of traditional image classification methods will be reduced by the effect of this interference. Since the date of submission, Deep Convolutional Neural Network impacts on the traditional image processing methods and brings the field of computer vision to a new stage with the advantages of a strong ability to learn deep features and excellent ability to fit large datasets. Based on the basic characteristics of polarimetric SAR images, the paper studied the types of the surface cover by using the method of Deep Learning. We used the fully polarimetric SAR features of different scales to fuse RGB images to the GoogLeNet model based on convolution neural network Iterative training, and then use the trained model to test the classification of data validation.First of all, referring to the optical image, we mark the surface coverage type of GF-3 POLSAR image with 8m resolution, and then collect the samples according to different categories. To meet the GoogLeNet model requirements of 256 × 256 pixel image input and taking into account the lack of full-resolution SAR resolution, the original image should be pre-processed in the process of resampling. In this paper, POLSAR image slice samples of different scales with sampling intervals of 2 m and 1 m to be trained separately and validated by the verification dataset. Among them, the training accuracy of GoogLeNet model trained with resampled 2-m polarimetric SAR image is 94.89 %, and that of the trained SAR image with resampled 1 m is 92.65 %.

  4. The use of infrared thermal imaging in the diagnosis of deep vein thrombosis

    NASA Astrophysics Data System (ADS)

    Kacmaz, Seydi; Ercelebi, Ergun; Zengin, Suat; Cindoruk, Sener

    2017-11-01

    The diagnosis of Deep Vein Thrombosis is of vital importance, especially in emergency situations where there is a lack of time and the patient's condition is critical. Late diagnosis causes cost increase, long waiting time, and improper treatment. Today, with the rapidly developing technology, the cost of thermal cameras is gradually decreasing day by day. Studies have shown that many diseases are associated with heat. As a result, infrared images are thought to be a tool for diagnosing various diseases. In this study, it has been shown that infrared thermal imaging can be used as a pre-screening test in the diagnosis of Deep Vein Thrombosis with the developed computer aided software. In addition, a sample combination is shown for applications that utilize emergency services to perform diagnosis and treatment of Deep Vein Thrombosis as soon as possible.

  5. First cosmic-ray images of bone and soft tissue

    NASA Astrophysics Data System (ADS)

    Mrdja, Dusan; Bikit, Istvan; Bikit, Kristina; Slivka, Jaroslav; Hansman, Jan; Oláh, László; Varga, Dezső

    2016-11-01

    More than 120 years after Roentgen's first X-ray image, the first cosmic-ray muon images of bone and soft tissue are created. The pictures, shown in the present paper, represent the first radiographies of structures of organic origin ever recorded by cosmic rays. This result is achieved by a uniquely designed, simple and versatile cosmic-ray muon-imaging system, which consists of four plastic scintillation detectors and a muon tracker. This system does not use scattering or absorption of muons in order to deduct image information, but takes advantage of the production rate of secondaries in the target materials, detected in coincidence with muons. The 2D image slices of cow femur bone are obtained at several depths along the bone axis, together with the corresponding 3D image. Real organic soft tissue, polymethyl methacrylate and water, never seen before by any other muon imaging techniques, are also registered in the images. Thus, similar imaging systems, placed around structures of organic or inorganic origin, can be used for tomographic imaging using only the omnipresent cosmic radiation.

  6. Evaluation of Deep Learning Based Stereo Matching Methods: from Ground to Aerial Images

    NASA Astrophysics Data System (ADS)

    Liu, J.; Ji, S.; Zhang, C.; Qin, Z.

    2018-05-01

    Dense stereo matching has been extensively studied in photogrammetry and computer vision. In this paper we evaluate the application of deep learning based stereo methods, which were raised from 2016 and rapidly spread, on aerial stereos other than ground images that are commonly used in computer vision community. Two popular methods are evaluated. One learns matching cost with a convolutional neural network (known as MC-CNN); the other produces a disparity map in an end-to-end manner by utilizing both geometry and context (known as GC-net). First, we evaluate the performance of the deep learning based methods for aerial stereo images by a direct model reuse. The models pre-trained on KITTI 2012, KITTI 2015 and Driving datasets separately, are directly applied to three aerial datasets. We also give the results of direct training on target aerial datasets. Second, the deep learning based methods are compared to the classic stereo matching method, Semi-Global Matching(SGM), and a photogrammetric software, SURE, on the same aerial datasets. Third, transfer learning strategy is introduced to aerial image matching based on the assumption of a few target samples available for model fine tuning. It experimentally proved that the conventional methods and the deep learning based methods performed similarly, and the latter had greater potential to be explored.

  7. Segmentation and classification of colon glands with deep convolutional neural networks and total variation regularization.

    PubMed

    Kainz, Philipp; Pfeiffer, Michael; Urschler, Martin

    2017-01-01

    Segmentation of histopathology sections is a necessary preprocessing step for digital pathology. Due to the large variability of biological tissue, machine learning techniques have shown superior performance over conventional image processing methods. Here we present our deep neural network-based approach for segmentation and classification of glands in tissue of benign and malignant colorectal cancer, which was developed to participate in the GlaS@MICCAI2015 colon gland segmentation challenge. We use two distinct deep convolutional neural networks (CNN) for pixel-wise classification of Hematoxylin-Eosin stained images. While the first classifier separates glands from background, the second classifier identifies gland-separating structures. In a subsequent step, a figure-ground segmentation based on weighted total variation produces the final segmentation result by regularizing the CNN predictions. We present both quantitative and qualitative segmentation results on the recently released and publicly available Warwick-QU colon adenocarcinoma dataset associated with the GlaS@MICCAI2015 challenge and compare our approach to the simultaneously developed other approaches that participated in the same challenge. On two test sets, we demonstrate our segmentation performance and show that we achieve a tissue classification accuracy of 98% and 95%, making use of the inherent capability of our system to distinguish between benign and malignant tissue. Our results show that deep learning approaches can yield highly accurate and reproducible results for biomedical image analysis, with the potential to significantly improve the quality and speed of medical diagnoses.

  8. Segmentation and classification of colon glands with deep convolutional neural networks and total variation regularization

    PubMed Central

    Kainz, Philipp; Pfeiffer, Michael

    2017-01-01

    Segmentation of histopathology sections is a necessary preprocessing step for digital pathology. Due to the large variability of biological tissue, machine learning techniques have shown superior performance over conventional image processing methods. Here we present our deep neural network-based approach for segmentation and classification of glands in tissue of benign and malignant colorectal cancer, which was developed to participate in the GlaS@MICCAI2015 colon gland segmentation challenge. We use two distinct deep convolutional neural networks (CNN) for pixel-wise classification of Hematoxylin-Eosin stained images. While the first classifier separates glands from background, the second classifier identifies gland-separating structures. In a subsequent step, a figure-ground segmentation based on weighted total variation produces the final segmentation result by regularizing the CNN predictions. We present both quantitative and qualitative segmentation results on the recently released and publicly available Warwick-QU colon adenocarcinoma dataset associated with the GlaS@MICCAI2015 challenge and compare our approach to the simultaneously developed other approaches that participated in the same challenge. On two test sets, we demonstrate our segmentation performance and show that we achieve a tissue classification accuracy of 98% and 95%, making use of the inherent capability of our system to distinguish between benign and malignant tissue. Our results show that deep learning approaches can yield highly accurate and reproducible results for biomedical image analysis, with the potential to significantly improve the quality and speed of medical diagnoses. PMID:29018612

  9. A comparative study for chest radiograph image retrieval using binary texture and deep learning classification.

    PubMed

    Anavi, Yaron; Kogan, Ilya; Gelbart, Elad; Geva, Ofer; Greenspan, Hayit

    2015-08-01

    In this work various approaches are investigated for X-ray image retrieval and specifically chest pathology retrieval. Given a query image taken from a data set of 443 images, the objective is to rank images according to similarity. Different features, including binary features, texture features, and deep learning (CNN) features are examined. In addition, two approaches are investigated for the retrieval task. One approach is based on the distance of image descriptors using the above features (hereon termed the "descriptor"-based approach); the second approach ("classification"-based approach) is based on a probability descriptor, generated by a pair-wise classification of each two classes (pathologies) and their decision values using an SVM classifier. Best results are achieved using deep learning features in a classification scheme.

  10. Optical-thermal light-tissue interactions during photoacoustic imaging

    NASA Astrophysics Data System (ADS)

    Gould, Taylor; Wang, Quanzeng; Pfefer, T. Joshua

    2014-03-01

    Photoacoustic imaging (PAI) has grown rapidly as a biomedical imaging technique in recent years, with key applications in cancer diagnosis and oximetry. In spite of these advances, the literature provides little insight into thermal tissue interactions involved in PAI. To elucidate these basic phenomena, we have developed, validated, and implemented a three-dimensional numerical model of tissue photothermal (PT) response to repetitive laser pulses. The model calculates energy deposition, fluence distributions, transient temperature and damage profiles in breast tissue with blood vessels and generalized perfusion. A parametric evaluation of these outputs vs. vessel diameter and depth, optical beam diameter, wavelength, and irradiance, was performed. For a constant radiant exposure level, increasing beam diameter led to a significant increase in subsurface heat generation rate. Increasing vessel diameter resulted in two competing effects - reduced mean energy deposition in the vessel due to light attenuation and greater thermal superpositioning due to reduced thermal relaxation. Maximum temperatures occurred either at the surface or in subsurface regions of the dermis, depending on vessel geometry and position. Results are discussed in terms of established exposure limits and levels used in prior studies. While additional experimental and numerical study is needed, numerical modeling represents a powerful tool for elucidating the effect of PA imaging devices on biological tissue.

  11. Interpreting CARS images of tissue within the C-H-stretching region

    NASA Astrophysics Data System (ADS)

    Dietzek, Benjamin; Meyer, Tobias; Medyukhina, Anna; Bergner, Norbert; Krafft, Christoph; Romeike, Bernd F. M.; Reichart, Rupert; Kalff, Rolf; Schmitt, Michael; Popp, Jürgen

    2014-03-01

    Single band coherent anti-Stokes Raman scattering (CARS) microscopy within the CH-stretching region is applied to detect individual cells and nuclei of human brain tissue and brain tumors - an information which allows for histopathologic grading of the tissue. The CARS image contrast within the C-H-stretching region correlated to the tissue composition. Based on the specific application example of identifying nuclei within (coherent) Raman images of neurotissue sections, we shall derive general design parameters for lasers optimally suited to serve in a clinical environment and discuss the potential of recently developed methods to analyze spectrally resolved CARS images and image segmentation algorithms.

  12. Magnetic Resonance Imaging of Adipose Tissue in Metabolic Dysfunction.

    PubMed

    Franz, Daniela; Syväri, Jan; Weidlich, Dominik; Baum, Thomas; Rummeny, Ernst J; Karampinos, Dimitrios C

    2018-06-06

     Adipose tissue has become an increasingly important tissue target in medicine. It plays a central role in the storage and release of energy throughout the human body and has recently gained interest for its endocrinologic function. Magnetic resonance imaging (MRI) is an established method for quantitative direct evaluation of adipose tissue distribution, and is used increasingly as the modality of choice for metabolic phenotyping. The purpose of this review was the identification and presentation of the currently available literature on MRI of adipose tissue in metabolic dysfunction.  A PubMed (http://www.ncbi.nlm.nih.gov/pubmed) keyword search up to August 2017 without starting date limitation was performed and reference lists of relevant articles were searched.  MRI provides excellent tools for the evaluation of adipose tissue distribution and further characterization of the tissue. Standard as well as newly developed MRI techniques allow a risk stratification for the development of metabolic dysfunction and enable monitoring without the use of ionizing radiation or contrast material.   · Different types of adipose tissue play a crucial role in various types of metabolic dysfunction.. · Magnetic resonance imaging (MRI) is an excellent tool for noninvasive adipose tissue evaluation with respect to distribution, composition and metabolic activity.. · Both standard and newly developed MRI techniques can be used for risk stratification for the development of metabolic dysfunction and allow monitoring without the use of ionizing radiation or contrast material.. · Franz D, Syväri J, Weidlich D et al. Magnetic Resonance Imaging of Adipose Tissue in Metabolic Dysfunction. Fortschr Röntgenstr 2018; DOI: 10.1055/a-0612-8006. © Georg Thieme Verlag KG Stuttgart · New York.

  13. Altering the architecture of tissue engineered hypertrophic cartilaginous grafts facilitates vascularisation and accelerates mineralisation.

    PubMed

    Sheehy, Eamon J; Vinardell, Tatiana; Toner, Mary E; Buckley, Conor T; Kelly, Daniel J

    2014-01-01

    Cartilaginous tissues engineered using mesenchymal stem cells (MSCs) can be leveraged to generate bone in vivo by executing an endochondral program, leading to increased interest in the use of such hypertrophic grafts for the regeneration of osseous defects. During normal skeletogenesis, canals within the developing hypertrophic cartilage play a key role in facilitating endochondral ossification. Inspired by this developmental feature, the objective of this study was to promote endochondral ossification of an engineered cartilaginous construct through modification of scaffold architecture. Our hypothesis was that the introduction of channels into MSC-seeded hydrogels would firstly facilitate the in vitro development of scaled-up hypertrophic cartilaginous tissues, and secondly would accelerate vascularisation and mineralisation of the graft in vivo. MSCs were encapsulated into hydrogels containing either an array of micro-channels, or into non-channelled 'solid' controls, and maintained in culture conditions known to promote a hypertrophic cartilaginous phenotype. Solid constructs accumulated significantly more sGAG and collagen in vitro, while channelled constructs accumulated significantly more calcium. In vivo, the channels acted as conduits for vascularisation and accelerated mineralisation of the engineered graft. Cartilaginous tissue within the channels underwent endochondral ossification, producing lamellar bone surrounding a hematopoietic marrow component. This study highlights the potential of utilising engineering methodologies, inspired by developmental skeletal processes, in order to enhance endochondral bone regeneration strategies.

  14. Altering the Architecture of Tissue Engineered Hypertrophic Cartilaginous Grafts Facilitates Vascularisation and Accelerates Mineralisation

    PubMed Central

    Sheehy, Eamon J.; Vinardell, Tatiana; Toner, Mary E.; Buckley, Conor T.; Kelly, Daniel J.

    2014-01-01

    Cartilaginous tissues engineered using mesenchymal stem cells (MSCs) can be leveraged to generate bone in vivo by executing an endochondral program, leading to increased interest in the use of such hypertrophic grafts for the regeneration of osseous defects. During normal skeletogenesis, canals within the developing hypertrophic cartilage play a key role in facilitating endochondral ossification. Inspired by this developmental feature, the objective of this study was to promote endochondral ossification of an engineered cartilaginous construct through modification of scaffold architecture. Our hypothesis was that the introduction of channels into MSC-seeded hydrogels would firstly facilitate the in vitro development of scaled-up hypertrophic cartilaginous tissues, and secondly would accelerate vascularisation and mineralisation of the graft in vivo. MSCs were encapsulated into hydrogels containing either an array of micro-channels, or into non-channelled ‘solid’ controls, and maintained in culture conditions known to promote a hypertrophic cartilaginous phenotype. Solid constructs accumulated significantly more sGAG and collagen in vitro, while channelled constructs accumulated significantly more calcium. In vivo, the channels acted as conduits for vascularisation and accelerated mineralisation of the engineered graft. Cartilaginous tissue within the channels underwent endochondral ossification, producing lamellar bone surrounding a hematopoietic marrow component. This study highlights the potential of utilising engineering methodologies, inspired by developmental skeletal processes, in order to enhance endochondral bone regeneration strategies. PMID:24595316

  15. In vivo multimodal nonlinear optical imaging of mucosal tissue

    NASA Astrophysics Data System (ADS)

    Sun, Ju; Shilagard, Tuya; Bell, Brent; Motamedi, Massoud; Vargas, Gracie

    2004-05-01

    We present a multimodal nonlinear imaging approach to elucidate microstructures and spectroscopic features of oral mucosa and submucosa in vivo. The hamster buccal pouch was imaged using 3-D high resolution multiphoton and second harmonic generation microscopy. The multimodal imaging approach enables colocalization and differentiation of prominent known spectroscopic and structural features such as keratin, epithelial cells, and submucosal collagen at various depths in tissue. Visualization of cellular morphology and epithelial thickness are in excellent agreement with histological observations. These results suggest that multimodal nonlinear optical microscopy can be an effective tool for studying the physiology and pathology of mucosal tissue.

  16. A New Image: Online Communities to Facilitate Teacher Professional Development

    ERIC Educational Resources Information Center

    Lock, Jennifer V.

    2006-01-01

    Realizing the potential of online or virtual communities to facilitate teacher professional development requires educators to change their current perceptions of professional development. This calls for educators to develop new images of ongoing opportunities for professional development, based on their needs within an online community of learners…

  17. The national DBS brain tissue network pilot study: need for more tissue and more standardization.

    PubMed

    Vedam-Mai, V; Krock, N; Ullman, M; Foote, K D; Shain, W; Smith, K; Yachnis, A T; Steindler, D; Reynolds, B; Merritt, S; Pagan, F; Marjama-Lyons, J; Hogarth, P; Resnick, A S; Zeilman, P; Okun, M S

    2011-08-01

    Over 70,000 DBS devices have been implanted worldwide; however, there remains a paucity of well-characterized post-mortem DBS brains available to researchers. We propose that the overall understanding of DBS can be improved through the establishment of a Deep Brain Stimulation-Brain Tissue Network (DBS-BTN), which will further our understanding of DBS and brain function. The objectives of the tissue bank are twofold: (a) to provide a complete (clinical, imaging and pathological) database for DBS brain tissue samples, and (b) to make available DBS tissue samples to researchers, which will help our understanding of disease and underlying brain circuitry. Standard operating procedures for processing DBS brains were developed as part of the pilot project. Complete data files were created for individual patients and included demographic information, clinical information, imaging data, pathology, and DBS lead locations/settings. 19 DBS brains were collected from 11 geographically dispersed centers from across the U.S. The average age at the time of death was 69.3 years (51-92, with a standard deviation or SD of 10.13). The male:female ratio was almost 3:1. Average post-mortem interval from death to brain collection was 10.6 h (SD of 7.17). The DBS targets included: subthalamic nucleus, globus pallidus interna, and ventralis intermedius nucleus of the thalamus. In 16.7% of cases the clinical diagnosis failed to match the pathological diagnosis. We provide neuropathological findings from the cohort, and perilead responses to DBS. One of the most important observations made in this pilot study was the missing data, which was approximately 25% of all available data fields. Preliminary results demonstrated the feasibility and utility of creating a National DBS-BTN resource for the scientific community. We plan to improve our techniques to remedy omitted clinical/research data, and expand the Network to include a larger donor pool. We will enhance sample preparation to

  18. Confocal microscopy imaging of solid tissue

    EPA Science Inventory

    Confocal laser scanning microscopy (CLSM) is a technique that is capable of generating serial sections of whole-mount tissue and then reassembling the computer acquired images as a virtual 3-dimensional structure. In many ways CLSM offers an alternative to traditional sectioning ...

  19. [Registration and 3D rendering of serial tissue section images].

    PubMed

    Liu, Zhexing; Jiang, Guiping; Dong, Wu; Zhang, Yu; Xie, Xiaomian; Hao, Liwei; Wang, Zhiyuan; Li, Shuxiang

    2002-12-01

    It is an important morphological research method to reconstruct the 3D imaging from serial section tissue images. Registration of serial images is a key step to 3D reconstruction. Firstly, an introduction to the segmentation-counting registration algorithm is presented, which is based on the joint histogram. After thresholding of the two images to be registered, the criterion function is defined as counting in a specific region of the joint histogram, which greatly speeds up the alignment process. Then, the method is used to conduct the serial tissue image matching task, and lies a solid foundation for 3D rendering. Finally, preliminary surface rendering results are presented.

  20. Excitation-scanning hyperspectral imaging as a means to discriminate various tissues types

    NASA Astrophysics Data System (ADS)

    Deal, Joshua; Favreau, Peter F.; Lopez, Carmen; Lall, Malvika; Weber, David S.; Rich, Thomas C.; Leavesley, Silas J.

    2017-02-01

    Little is currently known about the fluorescence excitation spectra of disparate tissues and how these spectra change with pathological state. Current imaging diagnostic techniques have limited capacity to investigate fluorescence excitation spectral characteristics. This study utilized excitation-scanning hyperspectral imaging to perform a comprehensive assessment of fluorescence spectral signatures of various tissues. Immediately following tissue harvest, a custom inverted microscope (TE-2000, Nikon Instruments) with Xe arc lamp and thin film tunable filter array (VersaChrome, Semrock, Inc.) were used to acquire hyperspectral image data from each sample. Scans utilized excitation wavelengths from 340 nm to 550 nm in 5 nm increments. Hyperspectral images were analyzed with custom Matlab scripts including linear spectral unmixing (LSU), principal component analysis (PCA), and Gaussian mixture modeling (GMM). Spectra were examined for potential characteristic features such as consistent intensity peaks at specific wavelengths or intensity ratios among significant wavelengths. The resultant spectral features were conserved among tissues of similar molecular composition. Additionally, excitation spectra appear to be a mixture of pure endmembers with commonalities across tissues of varied molecular composition, potentially identifiable through GMM. These results suggest the presence of common autofluorescent molecules in most tissues and that excitationscanning hyperspectral imaging may serve as an approach for characterizing tissue composition as well as pathologic state. Future work will test the feasibility of excitation-scanning hyperspectral imaging as a contrast mode for discriminating normal and pathological tissues.

  1. Deep Convolutional Neural Network-Based Early Automated Detection of Diabetic Retinopathy Using Fundus Image.

    PubMed

    Xu, Kele; Feng, Dawei; Mi, Haibo

    2017-11-23

    The automatic detection of diabetic retinopathy is of vital importance, as it is the main cause of irreversible vision loss in the working-age population in the developed world. The early detection of diabetic retinopathy occurrence can be very helpful for clinical treatment; although several different feature extraction approaches have been proposed, the classification task for retinal images is still tedious even for those trained clinicians. Recently, deep convolutional neural networks have manifested superior performance in image classification compared to previous handcrafted feature-based image classification methods. Thus, in this paper, we explored the use of deep convolutional neural network methodology for the automatic classification of diabetic retinopathy using color fundus image, and obtained an accuracy of 94.5% on our dataset, outperforming the results obtained by using classical approaches.

  2. Near-Infrared II Fluorescence for Imaging Hindlimb Vessel Regeneration with Dynamic Tissue Perfusion Measurement

    PubMed Central

    Hong, Guosong; Lee, Jerry C.; Jha, Arshi; Diao, Shuo; Nakayama, Karina H.; Hou, Luqia; Doyle, Timothy C.; Robinson, Joshua T.; Antaris, Alexander L.; Dai, Hongjie; Cooke, John P.; Huang, Ngan F.

    2014-01-01

    Background Real-time vascular imaging that provides both anatomic and hemodynamic information could greatly facilitate the diagnosis of vascular diseases and provide accurate assessment of therapeutic effects. Here we have developed a novel fluorescence-based all-optical method, named near-infrared II (NIR-II) fluorescence imaging, to image murine hindlimb vasculature and blood flow in an experimental model of peripheral arterial disease, by exploiting fluorescence in the NIR-II region (1000–1400 nm) of photon wavelengths. Methods and Results Owing to the reduced photon scattering of NIR-II fluorescence compared to traditional NIR fluorescence imaging and thus much deeper penetration depth into the body, we demonstrated that the mouse hindlimb vasculature could be imaged with higher spatial resolution than in vivo microCT. Furthermore, imaging over 26 days revealed a significant increase in hindlimb microvascular density in response to experimentally induced ischemia within the first 8 days of the surgery (P < 0.005), which was confirmed by histological analysis of microvascular density. Moreover, the tissue perfusion in the ischemic hindlimb could be quantitatively measured by the dynamic NIR-II method, revealing the temporal kinetics of blood flow recovery that resembled microbead-based blood flowmetry and laser Doppler blood spectroscopy. Conclusions The penetration depth of millimeters, high spatial resolution and fast acquisition rate of NIR-II imaging makes it a useful imaging tool for murine models of vascular disease. PMID:24657826

  3. Near-infrared II fluorescence for imaging hindlimb vessel regeneration with dynamic tissue perfusion measurement.

    PubMed

    Hong, Guosong; Lee, Jerry C; Jha, Arshi; Diao, Shuo; Nakayama, Karina H; Hou, Luqia; Doyle, Timothy C; Robinson, Joshua T; Antaris, Alexander L; Dai, Hongjie; Cooke, John P; Huang, Ngan F

    2014-05-01

    Real-time vascular imaging that provides both anatomic and hemodynamic information could greatly facilitate the diagnosis of vascular diseases and provide accurate assessment of therapeutic effects. Here, we have developed a novel fluorescence-based all-optical method, named near-infrared II (NIR-II) fluorescence imaging, to image murine hindlimb vasculature and blood flow in an experimental model of peripheral arterial disease, by exploiting fluorescence in the NIR-II region (1000-1400 nm) of photon wavelengths. Because of the reduced photon scattering of NIR-II fluorescence compared with traditional NIR fluorescence imaging and thus much deeper penetration depth into the body, we demonstrated that the mouse hindlimb vasculature could be imaged with higher spatial resolution than in vivo microscopic computed tomography. Furthermore, imaging during 26 days revealed a significant increase in hindlimb microvascular density in response to experimentally induced ischemia within the first 8 days of the surgery (P<0.005), which was confirmed by histological analysis of microvascular density. Moreover, the tissue perfusion in the ischemic hindlimb could be quantitatively measured by the dynamic NIR-II method, revealing the temporal kinetics of blood flow recovery that resembled microbead-based blood flowmetry and laser Doppler blood spectroscopy. The penetration depth of millimeters, high spatial resolution, and fast acquisition rate of NIR-II imaging make it a useful imaging tool for murine models of vascular disease. © 2014 American Heart Association, Inc.

  4. Magnetic resonance imaging characteristics of nonthermal irreversible electroporation in vegetable tissue.

    PubMed

    Hjouj, Mohammad; Rubinsky, Boris

    2010-07-01

    We introduce and characterize the use of MRI for studying nonthermal irreversible electroporation (NTIRE) in a vegetative tissue model. NTIRE is a new minimally invasive surgical technique for tissue ablation in which microsecond, high electric-field pulses form nanoscale defects in the cell membrane that lead to cell death. Clinical NTIRE sequences were applied to a potato tuber tissue model. The potato is used for NTIRE studies because cell damage is readily visible with optical means through a natural oxidation process of released intracellular enzymes (polyphenol oxidase) and the formation of brown-black melanins. MRI sequences of the treated area were taken at various times before and after NTIRE and compared with photographic images. A comparison was made between T1W, T2W, FLAIR and STIR MRIs of NTIRE and photographic images. Some MRI sequences show changes in areas treated by irreversible electroporation. T1W and FLAIR produce brighter images of the treated areas. In contrast, the signal was lost from the treated area when a suppression technique, STIR, was used. There was similarity between optical photographic images of the treated tissue and MRIs of the same areas. This is the first study to characterize MRI of NTIRE in vegetative tissue. We find that NTIRE produces changes in vegetative tissue that can be imaged by certain MRI sequences. This could make MRI an effective tool to study the fundamentals of NTIRE in nonanimal tissue.

  5. Optical redox imaging indices discriminate human breast cancer from normal tissues

    NASA Astrophysics Data System (ADS)

    Xu, He N.; Tchou, Julia; Feng, Min; Zhao, Huaqing; Li, Lin Z.

    2016-11-01

    Our long-term goal was to investigate the potential of incorporating redox imaging technique as a breast cancer (BC) diagnosis component to increase the positive predictive value of suspicious imaging finding and to reduce unnecessary biopsies and overdiagnosis. We previously found that precancer and cancer tissues in animal models displayed abnormal mitochondrial redox state. We also revealed abnormal mitochondrial redox state in cancerous specimens from three BC patients. Here, we extend our study to include biopsies of 16 patients. Tissue aliquots were collected from both apparently normal and cancerous tissues from the affected cancer-bearing breasts shortly after surgical resection. All specimens were snap-frozen and scanned with the Chance redox scanner, i.e., the three-dimensional cryogenic NADH/Fp (reduced nicotinamide adenine dinucleotide/oxidized flavoproteins) fluorescence imager. We found both Fp and NADH in the cancerous tissues roughly tripled that in the normal tissues (p<0.05). The redox ratio Fp/(NADH + Fp) was ˜27% higher in the cancerous tissues (p<0.05). Additionally, Fp, or NADH, or the redox ratio alone could predict cancer with reasonable sensitivity and specificity. Our findings suggest that the optical redox imaging technique can provide parameters independent of clinical factors for discriminating cancer from noncancer breast tissues in human patients.

  6. Optical redox imaging indices discriminate human breast cancer from normal tissues

    PubMed Central

    Xu, He N.; Tchou, Julia; Feng, Min; Zhao, Huaqing; Li, Lin Z.

    2016-01-01

    Abstract. Our long-term goal was to investigate the potential of incorporating redox imaging technique as a breast cancer (BC) diagnosis component to increase the positive predictive value of suspicious imaging finding and to reduce unnecessary biopsies and overdiagnosis. We previously found that precancer and cancer tissues in animal models displayed abnormal mitochondrial redox state. We also revealed abnormal mitochondrial redox state in cancerous specimens from three BC patients. Here, we extend our study to include biopsies of 16 patients. Tissue aliquots were collected from both apparently normal and cancerous tissues from the affected cancer-bearing breasts shortly after surgical resection. All specimens were snap-frozen and scanned with the Chance redox scanner, i.e., the three-dimensional cryogenic NADH/Fp (reduced nicotinamide adenine dinucleotide/oxidized flavoproteins) fluorescence imager. We found both Fp and NADH in the cancerous tissues roughly tripled that in the normal tissues (p<0.05). The redox ratio Fp/(NADH + Fp) was ∼27% higher in the cancerous tissues (p<0.05). Additionally, Fp, or NADH, or the redox ratio alone could predict cancer with reasonable sensitivity and specificity. Our findings suggest that the optical redox imaging technique can provide parameters independent of clinical factors for discriminating cancer from noncancer breast tissues in human patients. PMID:27896360

  7. Monitoring Cartilage Tissue Engineering Using Magnetic Resonance Spectroscopy, Imaging, and Elastography

    PubMed Central

    Klatt, Dieter; Magin, Richard L.

    2013-01-01

    A key technical challenge in cartilage tissue engineering is the development of a noninvasive method for monitoring the composition, structure, and function of the tissue at different growth stages. Due to its noninvasive, three-dimensional imaging capabilities and the breadth of available contrast mechanisms, magnetic resonance imaging (MRI) techniques can be expected to play a leading role in assessing engineered cartilage. In this review, we describe the new MR-based tools (spectroscopy, imaging, and elastography) that can provide quantitative biomarkers for cartilage tissue development both in vitro and in vivo. Magnetic resonance spectroscopy can identify the changing molecular structure and alternations in the conformation of major macromolecules (collagen and proteoglycans) using parameters such as chemical shift, relaxation rates, and magnetic spin couplings. MRI provides high-resolution images whose contrast reflects developing tissue microstructure and porosity through changes in local relaxation times and the apparent diffusion coefficient. Magnetic resonance elastography uses low-frequency mechanical vibrations in conjunction with MRI to measure soft tissue mechanical properties (shear modulus and viscosity). When combined, these three techniques provide a noninvasive, multiscale window for characterizing cartilage tissue growth at all stages of tissue development, from the initial cell seeding of scaffolds to the development of the extracellular matrix during construct incubation, and finally, to the postimplantation assessment of tissue integration in animals and patients. PMID:23574498

  8. Evaluating color performance of whole-slide imaging devices by multispectral-imaging of biological tissues

    NASA Astrophysics Data System (ADS)

    Saleheen, Firdous; Badano, Aldo; Cheng, Wei-Chung

    2017-03-01

    The color reproducibility of two whole-slide imaging (WSI) devices was evaluated with biological tissue slides. Three tissue slides (human colon, skin, and kidney) were used to test a modern and a legacy WSI devices. The color truth of the tissue slides was obtained using a multispectral imaging system. The output WSI images were compared with the color truth to calculate the color difference for each pixel. A psychophysical experiment was also conducted to measure the perceptual color reproducibility (PCR) of the same slides with four subjects. The experiment results show that the mean color differences of the modern, legacy, and monochrome WSI devices are 10.94+/-4.19, 22.35+/-8.99, and 42.74+/-2.96 ▵E00, while their mean PCRs are 70.35+/-7.64%, 23.06+/-14.68%, and 0.91+/-1.01%, respectively.

  9. High Resolution X-ray Phase Contrast Imaging with Acoustic Tissue-Selective Contrast Enhancement

    DTIC Science & Technology

    2008-06-01

    Imaging with Acoustic Tissue-Selective Contrast Enhancement PRINCIPAL INVESTIGATOR: Gerald J. Diebold, Ph.D. CONTRACTING... Contrast Imaging with Acoustic Tissue-Selective Contrast Enhancement 5b. GRANT NUMBER W81XWH-04-1-0481 5c. PROGRAM ELEMENT NUMBER 6. AUTHOR(S...additional phase contrast features are visible at the interfaces of soft tissues as slight contrast enhancements . The image sequence in Fig. 2 shows an image

  10. Direct microCT imaging of non-mineralized connective tissues at high resolution.

    PubMed

    Naveh, Gili R S; Brumfeld, Vlad; Dean, Mason; Shahar, Ron; Weiner, Steve

    2014-01-01

    The 3D imaging of soft tissues in their native state is challenging, especially when high resolution is required. An X-ray-based microCT is, to date, the best choice for high resolution 3D imaging of soft tissues. However, since X-ray attenuation of soft tissues is very low, contrasting enhancement using different staining materials is needed. The staining procedure, which also usually involves tissue fixation, causes unwanted and to some extent unknown tissue alterations. Here, we demonstrate that a method that enables 3D imaging of soft tissues without fixing and staining using an X-ray-based bench-top microCT can be applied to a variety of different tissues. With the sample mounted in a custom-made loading device inside a humidity chamber, we obtained soft tissue contrast and generated 3D images of fresh, soft tissues with a resolution of 1 micron voxel size. We identified three critical conditions which make it possible to image soft tissues: humidified environment, mechanical stabilization of the sample and phase enhancement. We demonstrate the capability of the technique using different specimens: an intervertebral disc, the non-mineralized growth plate, stingray tessellated radials (calcified cartilage) and the collagenous network of the periodontal ligament. Since the scanned specimen is fresh an interesting advantage of this technique is the ability to scan a specimen under load and track the changes of the different structures. This method offers a unique opportunity for obtaining valuable insights into 3D structure-function relationships of soft tissues.

  11. Live biospeckle laser imaging of root tissues.

    PubMed

    Braga, Roberto A; Dupuy, L; Pasqual, M; Cardoso, R R

    2009-06-01

    Live imaging is now a central component for the study of plant developmental processes. Currently, most techniques are extremely constraining: they rely on the marking of specific cellular structures which generally apply to model species because they require genetic transformations. The biospeckle laser (BSL) system was evaluated as an instrument to measure biological activity in plant tissues. The system allows collecting biospeckle patterns from roots which are grown in gels. Laser illumination has been optimized to obtain the images without undesirable specular reflections from the glass tube. Data on two different plant species were obtained and the ability of three different methods to analyze the biospeckle patterns are presented. The results showed that the biospeckle could provide quantitative indicators of the molecular activity from roots which are grown in gel substrate in tissue culture. We also presented a particular experimental configuration and the optimal approach to analyze the images. This may serve as a basis to further works on live BSL in order to study root development.

  12. Deep convolutional networks for pancreas segmentation in CT imaging

    NASA Astrophysics Data System (ADS)

    Roth, Holger R.; Farag, Amal; Lu, Le; Turkbey, Evrim B.; Summers, Ronald M.

    2015-03-01

    Automatic organ segmentation is an important prerequisite for many computer-aided diagnosis systems. The high anatomical variability of organs in the abdomen, such as the pancreas, prevents many segmentation methods from achieving high accuracies when compared to state-of-the-art segmentation of organs like the liver, heart or kidneys. Recently, the availability of large annotated training sets and the accessibility of affordable parallel computing resources via GPUs have made it feasible for "deep learning" methods such as convolutional networks (ConvNets) to succeed in image classification tasks. These methods have the advantage that used classification features are trained directly from the imaging data. We present a fully-automated bottom-up method for pancreas segmentation in computed tomography (CT) images of the abdomen. The method is based on hierarchical coarse-to-fine classification of local image regions (superpixels). Superpixels are extracted from the abdominal region using Simple Linear Iterative Clustering (SLIC). An initial probability response map is generated, using patch-level confidences and a two-level cascade of random forest classifiers, from which superpixel regions with probabilities larger 0.5 are retained. These retained superpixels serve as a highly sensitive initial input of the pancreas and its surroundings to a ConvNet that samples a bounding box around each superpixel at different scales (and random non-rigid deformations at training time) in order to assign a more distinct probability of each superpixel region being pancreas or not. We evaluate our method on CT images of 82 patients (60 for training, 2 for validation, and 20 for testing). Using ConvNets we achieve maximum Dice scores of an average 68% +/- 10% (range, 43-80%) in testing. This shows promise for accurate pancreas segmentation, using a deep learning approach and compares favorably to state-of-the-art methods.

  13. Multiview hyperspectral topography of tissue structural and functional characteristics

    NASA Astrophysics Data System (ADS)

    Liu, Peng; Huang, Jiwei; Zhang, Shiwu; Xu, Ronald X.

    2016-01-01

    Accurate and in vivo characterization of structural, functional, and molecular characteristics of biological tissue will facilitate quantitative diagnosis, therapeutic guidance, and outcome assessment in many clinical applications, such as wound healing, cancer surgery, and organ transplantation. We introduced and tested a multiview hyperspectral imaging technique for noninvasive topographic imaging of cutaneous wound oxygenation. The technique integrated a multiview module and a hyperspectral module in a single portable unit. Four plane mirrors were cohered to form a multiview reflective mirror set with a rectangular cross section. The mirror set was placed between a hyperspectral camera and the target biological tissue. For a single image acquisition task, a hyperspectral data cube with five views was obtained. The five-view hyperspectral image consisted of a main objective image and four reflective images. Three-dimensional (3-D) topography of the scene was achieved by correlating the matching pixels between the objective image and the reflective images. 3-D mapping of tissue oxygenation was achieved using a hyperspectral oxygenation algorithm. The multiview hyperspectral imaging technique was validated in a wound model, a tissue-simulating blood phantom, and in vivo biological tissue. The experimental results demonstrated the technical feasibility of using multiview hyperspectral imaging for 3-D topography of tissue functional properties.

  14. Facilitating Facilitators to Facilitate, in Problem or Enquiry Based Learning Sessions

    ERIC Educational Resources Information Center

    Coelho, Catherine

    2014-01-01

    Problem based learning (PBL) has been used in dental education over the past 20 years and uses a patient case scenario to stimulate learning in a small group setting, where a trained facilitator does not teach but guides the group to bring about deep contextualized learning, to be empathetic to each other and to encourage fair and equitable…

  15. Visualization and tissue classification of human breast cancer images using ultrahigh-resolution OCT.

    PubMed

    Yao, Xinwen; Gan, Yu; Chang, Ernest; Hibshoosh, Hanina; Feldman, Sheldon; Hendon, Christine

    2017-03-01

    Breast cancer is one of the most common cancers, and recognized as the third leading cause of mortality in women. Optical coherence tomography (OCT) enables three dimensional visualization of biological tissue with micrometer level resolution at high speed, and can play an important role in early diagnosis and treatment guidance of breast cancer. In particular, ultra-high resolution (UHR) OCT provides images with better histological correlation. This paper compared UHR OCT performance with standard OCT in breast cancer imaging qualitatively and quantitatively. Automatic tissue classification algorithms were used to automatically detect invasive ductal carcinoma in ex vivo human breast tissue. Human breast tissues, including non-neoplastic/normal tissues from breast reduction and tumor samples from mastectomy specimens, were excised from patients at Columbia University Medical Center. The tissue specimens were imaged by two spectral domain OCT systems at different wavelengths: a home-built ultra-high resolution (UHR) OCT system at 800 nm (measured as 2.72 μm axial and 5.52 μm lateral) and a commercial OCT system at 1,300 nm with standard resolution (measured as 6.5 μm axial and 15 μm lateral), and their imaging performances were analyzed qualitatively. Using regional features derived from OCT images produced by the two systems, we developed an automated classification algorithm based on relevance vector machine (RVM) to differentiate hollow-structured adipose tissue against solid tissue. We further developed B-scan based features for RVM to classify invasive ductal carcinoma (IDC) against normal fibrous stroma tissue among OCT datasets produced by the two systems. For adipose classification, 32 UHR OCT B-scans from 9 normal specimens, and 28 standard OCT B-scans from 6 normal and 4 IDC specimens were employed. For IDC classification, 152 UHR OCT B-scans from 6 normal and 13 IDC specimens, and 104 standard OCT B-scans from 5 normal and 8 IDC specimens

  16. Longitudinal in vivo two-photon fluorescence imaging

    PubMed Central

    Crowe, Sarah E.; Ellis-Davies, Graham C.R.

    2014-01-01

    Fluorescence microscopy is an essential technique for the basic sciences, especially biomedical research. Since the invention of laser scanning confocal microscopy in 1980s, that enabled imaging both fixed and living biological tissue with three-dimensional precision, high-resolution fluorescence imaging has revolutionized biological research. Confocal microscopy, by its very nature, has one fundamental limitation. Due to the confocal pinhole, deep tissue fluorescence imaging is not practical. In contrast (no pun intended), two-photon fluorescence microscopy allows, in principle, the collection of all emitted photons from fluorophores in the imaged voxel, dramatically extending our ability to see deep into living tissue. Since the development of transgenic mice with genetically encoded fluorescent protein in neocortical cells in 2000, two-photon imaging has enabled the dynamics of individual synapses to be followed for up to two years. Since the initial landmark contributions to this field in 2002, the technique has been used to understand how neuronal structure are changed by experience, learning and memory and various diseases. Here we provide a basic summary of the crucial elements that are required for such studies, and discuss many applications of longitudinal two-photon fluorescence microscopy that have appeared since 2002. PMID:24214350

  17. Medical ultrasound: imaging of soft tissue strain and elasticity

    PubMed Central

    Wells, Peter N. T.; Liang, Hai-Dong

    2011-01-01

    After X-radiography, ultrasound is now the most common of all the medical imaging technologies. For millennia, manual palpation has been used to assist in diagnosis, but it is subjective and restricted to larger and more superficial structures. Following an introduction to the subject of elasticity, the elasticity of biological soft tissues is discussed and published data are presented. The basic physical principles of pulse-echo and Doppler ultrasonic techniques are explained. The history of ultrasonic imaging of soft tissue strain and elasticity is summarized, together with a brief critique of previously published reviews. The relevant techniques—low-frequency vibration, step, freehand and physiological displacement, and radiation force (displacement, impulse, shear wave and acoustic emission)—are described. Tissue-mimicking materials are indispensible for the assessment of these techniques and their characteristics are reported. Emerging clinical applications in breast disease, cardiology, dermatology, gastroenterology, gynaecology, minimally invasive surgery, musculoskeletal studies, radiotherapy, tissue engineering, urology and vascular disease are critically discussed. It is concluded that ultrasonic imaging of soft tissue strain and elasticity is now sufficiently well developed to have clinical utility. The potential for further research is examined and it is anticipated that the technology will become a powerful mainstream investigative tool. PMID:21680780

  18. Nonlinear interferometric vibrational imaging of biological tissue

    NASA Astrophysics Data System (ADS)

    Jiang, Zhi; Marks, Daniel L.; Geddes, Joseph B., III; Boppart, Stephen A.

    2008-02-01

    We demonstrate imaging with the technique of nonlinear interferometric vibrational imaging (NIVI). Experimental images using this instrumentation and method have been acquired from both phantom and biological tissues. In our system, coherent anti-Stokes Raman scattering (CARS) signals are detected by spectral interferometry, which is able to fully restore high resolution Raman spectrum on each focal spot of a sample covering multiple Raman bands using broadband pump and Stokes laser beams. Spectral-domain detection has been demonstrated and allows for a significant increase in image acquiring speed, in signal-to-noise, and in interferometric signal stability.

  19. Deep UV autofluorescence microscopy for cell biology and tissue histology.

    PubMed

    Jamme, Frédéric; Kascakova, Slavka; Villette, Sandrine; Allouche, Fatma; Pallu, Stéphane; Rouam, Valérie; Réfrégiers, Matthieu

    2013-07-01

    Autofluorescence spectroscopy is a powerful tool for molecular histology and for following metabolic processes in biological samples as it does not require labelling. However, at the microscopic scale, it is mostly limited to visible and near infrared excitation of the samples. Several interesting and naturally occurring fluorophores can be excited in the UV and deep UV (DUV), but cannot be monitored in cellulo nor in vivo due to a lack of available microscopic instruments working in this wavelength range. To fulfil this need, we have developed a synchrotron-coupled DUV microspectrofluorimeter which is operational since 2010. An extended selection of endogenous autofluorescent probes that can be excited in DUV, including their spectral characteristics, is presented. The distribution of the probes in various biological samples, including cultured cells, soft tissues, bone sections and maize stems, is shown to illustrate the possibilities offered by this system. In this work we demonstrate that DUV autofluorescence is a powerful tool for tissue histology and cell biology. To fulfil this need, we have developed a synchrotron-coupled DUV microspectrofluorimeter which is operational since 2010. An extended selection of endogenous autofluorescent probes that can be excited in DUV, including their spectral characteristics, is presented. The distribution of the probes in various biological samples, including cultured cells, soft tissues, bone sections and maize stems, is shown to illustrate the possibilities offered by this system. In this work we demonstrate that DUV autofluorescence is a powerful tool for tissue histology and cell biology. In this work we demonstrate that DUV autofluorescence is a powerful tool for tissue histology and cell biology. © 2013 Société Française des Microscopies and Société de Biologie Cellulaire de France. Published by John Wiley & Sons Ltd.

  20. Imaging of oral pathological tissue using optical coherence tomography

    NASA Astrophysics Data System (ADS)

    Canjau, Silvana; Todea, Carmen; Sinescu, Cosmin; Duma, Virgil-Florin; Topala, Florin I.; Podoleanu, Adrian G.

    2014-01-01

    Oral squamous cell carcinoma (OSCC) constitutes 90% of oral cancer. Early detection is a cornerstone to improve survival. Interaction of light with tissues may highlight changes in tissue structure and metabolism. We propose optical coherence tomography (OCT), as a non-invasive diagnosis method, being a new high-resolution optical technique that permits tri-dimensional (3-D), real-time imaging of near surface abnormalities in complex tissues. In this study half of the excisional biopsy was directed to the pathologist and the other half was assigned for OCT investigation. Histopathology validated the results. Areas of OSCC of the buccal mucosa were identified in the OCT images. The elements obserced included extensive epithelial down-growth, the disruption of the basement membrane, with areas of erosion, an epithelial layer that was highly variable in thickness and invasion into the sub-epithelial layers. Therefore, OCT appears to be a highly promising imaging modality.

  1. Automatically identifying, counting, and describing wild animals in camera-trap images with deep learning.

    PubMed

    Norouzzadeh, Mohammad Sadegh; Nguyen, Anh; Kosmala, Margaret; Swanson, Alexandra; Palmer, Meredith S; Packer, Craig; Clune, Jeff

    2018-06-19

    Having accurate, detailed, and up-to-date information about the location and behavior of animals in the wild would improve our ability to study and conserve ecosystems. We investigate the ability to automatically, accurately, and inexpensively collect such data, which could help catalyze the transformation of many fields of ecology, wildlife biology, zoology, conservation biology, and animal behavior into "big data" sciences. Motion-sensor "camera traps" enable collecting wildlife pictures inexpensively, unobtrusively, and frequently. However, extracting information from these pictures remains an expensive, time-consuming, manual task. We demonstrate that such information can be automatically extracted by deep learning, a cutting-edge type of artificial intelligence. We train deep convolutional neural networks to identify, count, and describe the behaviors of 48 species in the 3.2 million-image Snapshot Serengeti dataset. Our deep neural networks automatically identify animals with >93.8% accuracy, and we expect that number to improve rapidly in years to come. More importantly, if our system classifies only images it is confident about, our system can automate animal identification for 99.3% of the data while still performing at the same 96.6% accuracy as that of crowdsourced teams of human volunteers, saving >8.4 y (i.e., >17,000 h at 40 h/wk) of human labeling effort on this 3.2 million-image dataset. Those efficiency gains highlight the importance of using deep neural networks to automate data extraction from camera-trap images, reducing a roadblock for this widely used technology. Our results suggest that deep learning could enable the inexpensive, unobtrusive, high-volume, and even real-time collection of a wealth of information about vast numbers of animals in the wild. Copyright © 2018 the Author(s). Published by PNAS.

  2. Development of two-photon fluorescence microscopy for quantitative imaging in turbid tissues

    NASA Astrophysics Data System (ADS)

    Coleno, Mariah Lee

    Two-photon laser scanning fluorescence microscopy (TPM) is a high resolution, non-invasive biological imaging technique that can be used to image turbid tissues both in vitro and in vivo at depths of several hundred microns. Although TPM has been widely used to image tissue structures, no one has focused on using TPM to extract quantitative information from turbid tissues at depth. As a result, this thesis addresses the quantitative characterization of two-photon signals in turbid media. Initially, a two-photon microscope system is constructed, and two-photon images that validate system performance are obtained. Then TPM is established as an imaging technique that can be used to validate theoretical observations already listed in the literature. In particular, TPM is found to validate the exponential dependence of the fluorescence intensity decay with depth in turbid tissue model systems. Results from these studies next prompted experimental investigation into whether TPM could be used to determine tissue optical properties. Comparing the exponential dependence of the decay with a Monte Carlo model involving tissue optical properties, TPM is shown to be useful for determining the optical properties (total attenuation coefficient) of thick, turbid tissues on a small spatial scale. Next, a role for TPM for studying and optimizing wound healing is demonstrated. In particular, TPM is used to study the effects of perturbations (growth factors, PDT) on extracellular matrix remodeling in artificially engineered skin tissues. Results from these studies combined with tissue contraction studies are shown to demonstrate ways to modulate tissues to optimize the wound healing immune response and reduce scarring. In the end, TPM is shown to be an extremely important quantitative biological imaging technique that can be used to optimize wound repair.

  3. Marginal Space Deep Learning: Efficient Architecture for Volumetric Image Parsing.

    PubMed

    Ghesu, Florin C; Krubasik, Edward; Georgescu, Bogdan; Singh, Vivek; Yefeng Zheng; Hornegger, Joachim; Comaniciu, Dorin

    2016-05-01

    Robust and fast solutions for anatomical object detection and segmentation support the entire clinical workflow from diagnosis, patient stratification, therapy planning, intervention and follow-up. Current state-of-the-art techniques for parsing volumetric medical image data are typically based on machine learning methods that exploit large annotated image databases. Two main challenges need to be addressed, these are the efficiency in scanning high-dimensional parametric spaces and the need for representative image features which require significant efforts of manual engineering. We propose a pipeline for object detection and segmentation in the context of volumetric image parsing, solving a two-step learning problem: anatomical pose estimation and boundary delineation. For this task we introduce Marginal Space Deep Learning (MSDL), a novel framework exploiting both the strengths of efficient object parametrization in hierarchical marginal spaces and the automated feature design of Deep Learning (DL) network architectures. In the 3D context, the application of deep learning systems is limited by the very high complexity of the parametrization. More specifically 9 parameters are necessary to describe a restricted affine transformation in 3D, resulting in a prohibitive amount of billions of scanning hypotheses. The mechanism of marginal space learning provides excellent run-time performance by learning classifiers in clustered, high-probability regions in spaces of gradually increasing dimensionality. To further increase computational efficiency and robustness, in our system we learn sparse adaptive data sampling patterns that automatically capture the structure of the input. Given the object localization, we propose a DL-based active shape model to estimate the non-rigid object boundary. Experimental results are presented on the aortic valve in ultrasound using an extensive dataset of 2891 volumes from 869 patients, showing significant improvements of up to 45

  4. ReagentTF: a rapid and versatile optical clearing method for biological imaging(Conference Presentation)

    NASA Astrophysics Data System (ADS)

    Yu, Tingting; Zhu, Jingtan; Li, Yusha; Qi, Yisong; Xu, Jianyi; Gong, Hui; Luo, Qingming; Zhu, Dan

    2017-02-01

    The emergence of various optical clearing methods provides a great potential for imaging deep inside tissues by combining with multiple-labelling and microscopic imaging techniques. They were generally developed for specific imaging demand thus presented some non-negligible limitations such as long incubation time, tissue deformation, fluorescence quenching, incompatibility with immunostaining or lipophilic tracers. In this study, we developed a rapid and versatile clearing method, termed ReagentTF, for deep imaging of various fluorescent samples. This method can not only efficiently clear embryos, neonatal whole-brains and adult thick brain sections by simple immersion in aqueous mixtures with minimal volume change, but also can preserve fluorescence of various fluorescent proteins and simultaneously be compatible with immunostaining and lipophilic neuronal dyes. We demonstrate the effectiveness of this method in reconstructing the cell distributions of mouse hippocampus, visualizing the neural projection from CA1 (Cornu Ammonis 1) to HDB (nucleus of the horizontal limb of the diagonal band), and observing the growth of forelimb plexus in whole-mount embryos. These results suggest that ReagentTF is useful for large-volume imaging and will be an option for the deep imaging of biological tissues.

  5. White Matter Tract Injury is Associated with Deep Gray Matter Iron Deposition in Multiple Sclerosis.

    PubMed

    Bergsland, Niels; Tavazzi, Eleonora; Laganà, Maria Marcella; Baglio, Francesca; Cecconi, Pietro; Viotti, Stefano; Zivadinov, Robert; Baselli, Giuseppe; Rovaris, Marco

    2017-01-01

    With respect to healthy controls (HCs), increased iron concentrations in the deep gray matter (GM) and decreased white matter (WM) integrity are common findings in multiple sclerosis (MS) patients. The association between these features of the disease remains poorly understood. We investigated the relationship between deep iron deposition in the deep GM and WM injury in associated fiber tracts in MS patients. Sixty-six MS patients (mean age 50.0 years, median Expanded Disability Status Scale 5.25, mean disease duration 19.1 years) and 29 HCs, group matched for age and sex were imaged on a 1.5T scanner. Susceptibility-weighted imaging and diffusion tensor imaging (DTI) were used for assessing high-pass filtered phase values in the deep GM and normal appearing WM (NAWM) integrity in associated fiber tracts, respectively. Correlation analyses investigated the associations between filtered phase values (suggestive of iron content) and WM damage. Areas indicative of increased iron levels were found in the left and right caudates as well as in the left thalamus. MS patients presented with decreased DTI-derived measures of tissue integrity in the associated WM tracts. Greater mean, axial and radial diffusivities were associated with increased iron levels in all three GM areas (r values .393 to .514 with corresponding P values .003 to <.0001). Global NAWM diffusivity measures were not related to mean filtered phase values within the deep GM. Increased iron concentration in the deep GM is associated with decreased tissue integrity of the connected WM in MS patients. Copyright © 2016 by the American Society of Neuroimaging.

  6. Experimental and numerical investigation of tissue harmonic imaging (THI)

    NASA Astrophysics Data System (ADS)

    Jing, Yuan; Yang, Xinmai; Cleveland, Robin O.

    2003-04-01

    In THI the probing ultrasonic pulse has enough amplitude that it undergoes nonlinear distortion and energy shifts from the fundamental frequency of the pulse into its higher harmonics. Images generated from the second harmonic (SH) have superior quality to the images formed from the fundamental frequency. Experiments with a single element focused ultrasound transducer were used to compare a line target embedded in a tissue phantom using either fundamental or SH imaging. SH imaging showed an improvement in both the axial resolution (0.70 mm vs 0.92 mm) and the lateral resolution (1.02 mm vs 2.70 mm) of the target. In addition, the contrast-to-tissue ratio of the target was 2 dB higher with SH imaging. A three-dimensional model of the forward propagation has been developed to simulate the experimental system. The model is based on a time-domain code for solving the KZK equation and accounts for arbitrary spatial variations in all tissue properties. The code was used to determine the impact of a nearfield layer of fat on the fundamental and second harmonic signals. For a 15 mm thick layer the SH side-lobes remained the same but the fundamental side-lobes increased by 2 dB. [Work supported by the NSF through the Center for Subsurface Sensing and Imaging Systems.

  7. Deep supervised dictionary learning for no-reference image quality assessment

    NASA Astrophysics Data System (ADS)

    Huang, Yuge; Liu, Xuesong; Tian, Xiang; Zhou, Fan; Chen, Yaowu; Jiang, Rongxin

    2018-03-01

    We propose a deep convolutional neural network (CNN) for general no-reference image quality assessment (NR-IQA), i.e., accurate prediction of image quality without a reference image. The proposed model consists of three components such as a local feature extractor that is a fully CNN, an encoding module with an inherent dictionary that aggregates local features to output a fixed-length global quality-aware image representation, and a regression module that maps the representation to an image quality score. Our model can be trained in an end-to-end manner, and all of the parameters, including the weights of the convolutional layers, the dictionary, and the regression weights, are simultaneously learned from the loss function. In addition, the model can predict quality scores for input images of arbitrary sizes in a single step. We tested our method on commonly used image quality databases and showed that its performance is comparable with that of state-of-the-art general-purpose NR-IQA algorithms.

  8. Deep Learning Automates the Quantitative Analysis of Individual Cells in Live-Cell Imaging Experiments.

    PubMed

    Van Valen, David A; Kudo, Takamasa; Lane, Keara M; Macklin, Derek N; Quach, Nicolas T; DeFelice, Mialy M; Maayan, Inbal; Tanouchi, Yu; Ashley, Euan A; Covert, Markus W

    2016-11-01

    Live-cell imaging has opened an exciting window into the role cellular heterogeneity plays in dynamic, living systems. A major critical challenge for this class of experiments is the problem of image segmentation, or determining which parts of a microscope image correspond to which individual cells. Current approaches require many hours of manual curation and depend on approaches that are difficult to share between labs. They are also unable to robustly segment the cytoplasms of mammalian cells. Here, we show that deep convolutional neural networks, a supervised machine learning method, can solve this challenge for multiple cell types across the domains of life. We demonstrate that this approach can robustly segment fluorescent images of cell nuclei as well as phase images of the cytoplasms of individual bacterial and mammalian cells from phase contrast images without the need for a fluorescent cytoplasmic marker. These networks also enable the simultaneous segmentation and identification of different mammalian cell types grown in co-culture. A quantitative comparison with prior methods demonstrates that convolutional neural networks have improved accuracy and lead to a significant reduction in curation time. We relay our experience in designing and optimizing deep convolutional neural networks for this task and outline several design rules that we found led to robust performance. We conclude that deep convolutional neural networks are an accurate method that require less curation time, are generalizable to a multiplicity of cell types, from bacteria to mammalian cells, and expand live-cell imaging capabilities to include multi-cell type systems.

  9. Deep Learning Automates the Quantitative Analysis of Individual Cells in Live-Cell Imaging Experiments

    DOE PAGES

    Van Valen, David A.; Kudo, Takamasa; Lane, Keara M.; ...

    2016-11-04

    Live-cell imaging has opened an exciting window into the role cellular heterogeneity plays in dynamic, living systems. A major critical challenge for this class of experiments is the problem of image segmentation, or determining which parts of a microscope image correspond to which individual cells. Current approaches require many hours of manual curation and depend on approaches that are difficult to share between labs. They are also unable to robustly segment the cytoplasms of mammalian cells. Here, we show that deep convolutional neural networks, a supervised machine learning method, can solve this challenge for multiple cell types across the domainsmore » of life. We demonstrate that this approach can robustly segment fluorescent images of cell nuclei as well as phase images of the cytoplasms of individual bacterial and mammalian cells from phase contrast images without the need for a fluorescent cytoplasmic marker. These networks also enable the simultaneous segmentation and identification of different mammalian cell types grown in co-culture. A quantitative comparison with prior methods demonstrates that convolutional neural networks have improved accuracy and lead to a significant reduction in curation time. We relay our experience in designing and optimizing deep convolutional neural networks for this task and outline several design rules that we found led to robust performance. We conclude that deep convolutional neural networks are an accurate method that require less curation time, are generalizable to a multiplicity of cell types, from bacteria to mammalian cells, and expand live-cell imaging capabilities to include multi-cell type systems.« less

  10. Deep Learning Automates the Quantitative Analysis of Individual Cells in Live-Cell Imaging Experiments

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Van Valen, David A.; Kudo, Takamasa; Lane, Keara M.

    Live-cell imaging has opened an exciting window into the role cellular heterogeneity plays in dynamic, living systems. A major critical challenge for this class of experiments is the problem of image segmentation, or determining which parts of a microscope image correspond to which individual cells. Current approaches require many hours of manual curation and depend on approaches that are difficult to share between labs. They are also unable to robustly segment the cytoplasms of mammalian cells. Here, we show that deep convolutional neural networks, a supervised machine learning method, can solve this challenge for multiple cell types across the domainsmore » of life. We demonstrate that this approach can robustly segment fluorescent images of cell nuclei as well as phase images of the cytoplasms of individual bacterial and mammalian cells from phase contrast images without the need for a fluorescent cytoplasmic marker. These networks also enable the simultaneous segmentation and identification of different mammalian cell types grown in co-culture. A quantitative comparison with prior methods demonstrates that convolutional neural networks have improved accuracy and lead to a significant reduction in curation time. We relay our experience in designing and optimizing deep convolutional neural networks for this task and outline several design rules that we found led to robust performance. We conclude that deep convolutional neural networks are an accurate method that require less curation time, are generalizable to a multiplicity of cell types, from bacteria to mammalian cells, and expand live-cell imaging capabilities to include multi-cell type systems.« less

  11. Deep Learning Automates the Quantitative Analysis of Individual Cells in Live-Cell Imaging Experiments

    PubMed Central

    Van Valen, David A.; Lane, Keara M.; Quach, Nicolas T.; Maayan, Inbal

    2016-01-01

    Live-cell imaging has opened an exciting window into the role cellular heterogeneity plays in dynamic, living systems. A major critical challenge for this class of experiments is the problem of image segmentation, or determining which parts of a microscope image correspond to which individual cells. Current approaches require many hours of manual curation and depend on approaches that are difficult to share between labs. They are also unable to robustly segment the cytoplasms of mammalian cells. Here, we show that deep convolutional neural networks, a supervised machine learning method, can solve this challenge for multiple cell types across the domains of life. We demonstrate that this approach can robustly segment fluorescent images of cell nuclei as well as phase images of the cytoplasms of individual bacterial and mammalian cells from phase contrast images without the need for a fluorescent cytoplasmic marker. These networks also enable the simultaneous segmentation and identification of different mammalian cell types grown in co-culture. A quantitative comparison with prior methods demonstrates that convolutional neural networks have improved accuracy and lead to a significant reduction in curation time. We relay our experience in designing and optimizing deep convolutional neural networks for this task and outline several design rules that we found led to robust performance. We conclude that deep convolutional neural networks are an accurate method that require less curation time, are generalizable to a multiplicity of cell types, from bacteria to mammalian cells, and expand live-cell imaging capabilities to include multi-cell type systems. PMID:27814364

  12. High-resolution imaging of biological tissue with full-field optical coherence tomography

    NASA Astrophysics Data System (ADS)

    Zhu, Yue; Gao, Wanrong

    2015-03-01

    A new full-field optical coherence tomography system with high-resolution has been developed for imaging of cells and tissues. Compared with other FF-OCT (Full-field optical coherence tomography, FF-OCT) systems illuminated with optical fiber bundle, the improved Köhler illumination arrangement with a halogen lamp was used in the proposed FF-OCT system. High numerical aperture microscopic objectives were used for imaging and a piezoelectric ceramic transducer (PZT) was used for phase-shifting. En-face tomographic images can be obtained by applying the five-step phase-shifting algorithm to a series of interferometric images which are recorded by a smart camera. Three-dimensional images can be generated from these tomographic images. Imaging of the chip of Intel Pentium 4 processor demonstrated the ultrahigh resolution of the system (lateral resolution is 0.8μm ), which approaches the theoretical resolution 0.7 μm× 0.5 μm (lateral × axial). En-face images of cells of onion show an excellent performance of the system in generating en-face images of biological tissues. Then, unstained pig stomach was imaged as a tissue and gastric pits could be easily recognized using FF-OCT system. Our study provides evidence for the potential ability of FFOCT in identifying gastric pits from pig stomach tissue. Finally, label-free and unstained ex vivo human liver tissues from both normal and tumor were imaged with this FFOCT system. The results show that the setup has the potential for medical diagnosis applications such liver cancer diagnosis.

  13. A novel near-infrared nanomaterial with high quantum efficiency and its applications in real time in vivo imaging

    NASA Astrophysics Data System (ADS)

    Cui, X. X.; Fan, Q.; Shi, S. J.; Wen, W. H.; Chen, D. F.; Guo, H. T.; Xu, Y. T.; Gao, F.; Nie, R. Z.; Ford, Harold D.; Tang, Gordon H.; Hou, C. Q.; Peng, B.

    2018-05-01

    Fluorescence imaging signal is severely limited by the quantum efficiency and emission wavelength. To overcome these challenges, novel NIR-emitting K5NdLi2F10 nanoparticles under NIR excitation was introduced as fluorescence imaging probe for the first time. The photostability of K5NdLi2F10 nanoparticles in the water, phosphate buffer saline, fetal bovine serum and living mice was investigated. The fluorescence signal was detected with depths of 3.5 and 2.0 cm in phantom and pork tissue, respectively. Fluorescence spectrum with a significant signal-to-background ratio of 10:1 was captured in living mice. Moreover, clear NIR images were virtualized for the living mice after intravenous injection. The imaging ability of nanoparticles in tumor-beard mice were evaluated, the enrichment of K5NdLi2F10 nanoparticles in tumor site due to the enhanced permeability and retention effect was confirmed. The systematic studies of toxicity, bio-distribution and in-vivo dynamic imaging suggest that these materials give high biocompatibility and low toxicity. These NIR-emitting nanoparticles with high quantum efficiency, high penetration and low toxicity might facilitate tumor identification in deep tissues more sensitively.

  14. A novel near-infrared nanomaterial with high quantum efficiency and its applications in real time in vivo imaging.

    PubMed

    Cui, X X; Fan, Q; Shi, S J; Wen, W H; Chen, D F; Guo, H T; Xu, Y T; Gao, F; Nie, R Z; Ford, Harold D; Tang, Gordon H; Hou, C Q; Peng, B

    2018-05-18

    Fluorescence imaging signal is severely limited by the quantum efficiency and emission wavelength. To overcome these challenges, novel NIR-emitting K 5 NdLi 2 F 10 nanoparticles under NIR excitation was introduced as fluorescence imaging probe for the first time. The photostability of K 5 NdLi 2 F 10 nanoparticles in the water, phosphate buffer saline, fetal bovine serum and living mice was investigated. The fluorescence signal was detected with depths of 3.5 and 2.0 cm in phantom and pork tissue, respectively. Fluorescence spectrum with a significant signal-to-background ratio of 10:1 was captured in living mice. Moreover, clear NIR images were virtualized for the living mice after intravenous injection. The imaging ability of nanoparticles in tumor-beard mice were evaluated, the enrichment of K 5 NdLi 2 F 10 nanoparticles in tumor site due to the enhanced permeability and retention effect was confirmed. The systematic studies of toxicity, bio-distribution and in-vivo dynamic imaging suggest that these materials give high biocompatibility and low toxicity. These NIR-emitting nanoparticles with high quantum efficiency, high penetration and low toxicity might facilitate tumor identification in deep tissues more sensitively.

  15. Super-resolution for asymmetric resolution of FIB-SEM 3D imaging using AI with deep learning.

    PubMed

    Hagita, Katsumi; Higuchi, Takeshi; Jinnai, Hiroshi

    2018-04-12

    Scanning electron microscopy equipped with a focused ion beam (FIB-SEM) is a promising three-dimensional (3D) imaging technique for nano- and meso-scale morphologies. In FIB-SEM, the specimen surface is stripped by an ion beam and imaged by an SEM installed orthogonally to the FIB. The lateral resolution is governed by the SEM, while the depth resolution, i.e., the FIB milling direction, is determined by the thickness of the stripped thin layer. In most cases, the lateral resolution is superior to the depth resolution; hence, asymmetric resolution is generated in the 3D image. Here, we propose a new approach based on an image-processing or deep-learning-based method for super-resolution of 3D images with such asymmetric resolution, so as to restore the depth resolution to achieve symmetric resolution. The deep-learning-based method learns from high-resolution sub-images obtained via SEM and recovers low-resolution sub-images parallel to the FIB milling direction. The 3D morphologies of polymeric nano-composites are used as test images, which are subjected to the deep-learning-based method as well as conventional methods. We find that the former yields superior restoration, particularly as the asymmetric resolution is increased. Our super-resolution approach for images having asymmetric resolution enables observation time reduction.

  16. Electrochemical imaging of cells and tissues

    PubMed Central

    Lin, Tzu-En; Rapino, Stefania; Girault, Hubert H.

    2018-01-01

    The technological and experimental progress in electrochemical imaging of biological specimens is discussed with a view on potential applications for skin cancer diagnostics, reproductive medicine and microbial testing. The electrochemical analysis of single cell activity inside cell cultures, 3D cellular aggregates and microtissues is based on the selective detection of electroactive species involved in biological functions. Electrochemical imaging strategies, based on nano/micrometric probes scanning over the sample and sensor array chips, respectively, can be made sensitive and selective without being affected by optical interference as many other microscopy techniques. The recent developments in microfabrication, electronics and cell culturing/tissue engineering have evolved in affordable and fast-sampling electrochemical imaging platforms. We believe that the topics discussed herein demonstrate the applicability of electrochemical imaging devices in many areas related to cellular functions. PMID:29899947

  17. Preliminary study of synthetic aperture tissue harmonic imaging on in-vivo data

    NASA Astrophysics Data System (ADS)

    Rasmussen, Joachim H.; Hemmsen, Martin C.; Madsen, Signe S.; Hansen, Peter M.; Nielsen, Michael B.; Jensen, Jørgen A.

    2013-03-01

    A method for synthetic aperture tissue harmonic imaging is investigated. It combines synthetic aperture sequen- tial beamforming (SASB) with tissue harmonic imaging (THI) to produce an increased and more uniform spatial resolution and improved side lobe reduction compared to conventional B-mode imaging. Synthetic aperture sequential beamforming tissue harmonic imaging (SASB-THI) was implemented on a commercially available BK 2202 Pro Focus UltraView ultrasound system and compared to dynamic receive focused tissue harmonic imag- ing (DRF-THI) in clinical scans. The scan sequence that was implemented on the UltraView system acquires both SASB-THI and DRF-THI simultaneously. Twenty-four simultaneously acquired video sequences of in-vivo abdominal SASB-THI and DRF-THI scans on 3 volunteers of 4 different sections of liver and kidney tissues were created. Videos of the in-vivo scans were presented in double blinded studies to two radiologists for image quality performance scoring. Limitations to the systems transmit stage prevented user defined transmit apodization to be applied. Field II simulations showed that side lobes in SASB could be improved by using Hanning transmit apodization. Results from the image quality study show, that in the current configuration on the UltraView system, where no transmit apodization was applied, SASB-THI and DRF-THI produced equally good images. It is expected that given the use of transmit apodization, SASB-THI could be further improved.

  18. Visualizing deep neural network by alternately image blurring and deblurring.

    PubMed

    Wang, Feng; Liu, Haijun; Cheng, Jian

    2018-01-01

    Visualization from trained deep neural networks has drawn massive public attention in recent. One of the visualization approaches is to train images maximizing the activation of specific neurons. However, directly maximizing the activation would lead to unrecognizable images, which cannot provide any meaningful information. In this paper, we introduce a simple but effective technique to constrain the optimization route of the visualization. By adding two totally inverse transformations, image blurring and deblurring, to the optimization procedure, recognizable images can be created. Our algorithm is good at extracting the details in the images, which are usually filtered by previous methods in the visualizations. Extensive experiments on AlexNet, VGGNet and GoogLeNet illustrate that we can better understand the neural networks utilizing the knowledge obtained by the visualization. Copyright © 2017 Elsevier Ltd. All rights reserved.

  19. Stimulated Raman photoacoustic imaging

    PubMed Central

    Yakovlev, Vladislav V.; Zhang, Hao F.; Noojin, Gary D.; Denton, Michael L.; Thomas, Robert J.; Scully, Marlan O.

    2010-01-01

    Achieving label-free, molecular-specific imaging with high spatial resolution in deep tissue is often considered the grand challenge of optical imaging. To accomplish this goal, significant optical scattering in tissues has to be overcome while achieving molecular specificity without resorting to extrinsic labeling. We demonstrate the feasibility of developing such an optical imaging modality by combining the molecularly specific stimulated Raman excitation with the photoacoustic detection. By employing two ultrashort excitation laser pulses, separated in frequency by the vibrational frequency of a targeted molecule, only the specific vibrational level of the target molecules in the illuminated tissue volume is excited. This targeted optical absorption generates ultrasonic waves (referred to as stimulated Raman photoacoustic waves) which are detected using a traditional ultrasonic transducer to form an image following the design of the established photoacoustic microscopy. PMID:21059930

  20. Multiview hyperspectral topography of tissue structural and functional characteristics

    NASA Astrophysics Data System (ADS)

    Zhang, Shiwu; Liu, Peng; Huang, Jiwei; Xu, Ronald

    2012-12-01

    Accurate and in vivo characterization of structural, functional, and molecular characteristics of biological tissue will facilitate quantitative diagnosis, therapeutic guidance, and outcome assessment in many clinical applications, such as wound healing, cancer surgery, and organ transplantation. However, many clinical imaging systems have limitations and fail to provide noninvasive, real time, and quantitative assessment of biological tissue in an operation room. To overcome these limitations, we developed and tested a multiview hyperspectral imaging system. The multiview hyperspectral imaging system integrated the multiview and the hyperspectral imaging techniques in a single portable unit. Four plane mirrors are cohered together as a multiview reflective mirror set with a rectangular cross section. The multiview reflective mirror set was placed between a hyperspectral camera and the measured biological tissue. For a single image acquisition task, a hyperspectral data cube with five views was obtained. The five-view hyperspectral image consisted of a main objective image and four reflective images. Three-dimensional topography of the scene was achieved by correlating the matching pixels between the objective image and the reflective images. Three-dimensional mapping of tissue oxygenation was achieved using a hyperspectral oxygenation algorithm. The multiview hyperspectral imaging technique is currently under quantitative validation in a wound model, a tissue-simulating blood phantom, and an in vivo biological tissue model. The preliminary results have demonstrated the technical feasibility of using multiview hyperspectral imaging for three-dimensional topography of tissue functional properties.