Sample records for quantitatively analyzed results

  1. Methods for Quantitative Interpretation of Retarding Field Analyzer Data

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Calvey, J.R.; Crittenden, J.A.; Dugan, G.F.

    2011-03-28

    Over the course of the CesrTA program at Cornell, over 30 Retarding Field Analyzers (RFAs) have been installed in the CESR storage ring, and a great deal of data has been taken with them. These devices measure the local electron cloud density and energy distribution, and can be used to evaluate the efficacy of different cloud mitigation techniques. Obtaining a quantitative understanding of RFA data requires use of cloud simulation programs, as well as a detailed model of the detector itself. In a drift region, the RFA can be modeled by postprocessing the output of a simulation code, and onemore » can obtain best fit values for important simulation parameters with a chi-square minimization method.« less

  2. Quantitative Analysis in the General Chemistry Laboratory: Training Students to Analyze Individual Results in the Context of Collective Data

    ERIC Educational Resources Information Center

    Ling, Chris D.; Bridgeman, Adam J.

    2011-01-01

    Titration experiments are ideal for generating large data sets for use in quantitative-analysis activities that are meaningful and transparent to general chemistry students. We report the successful implementation of a sophisticated quantitative exercise in which the students identify a series of unknown acids by determining their molar masses…

  3. Oxygen analyzer

    DOEpatents

    Benner, W.H.

    1984-05-08

    An oxygen analyzer which identifies and classifies microgram quantities of oxygen in ambient particulate matter and for quantitating organic oxygen in solvent extracts of ambient particulate matter. A sample is pyrolyzed in oxygen-free nitrogen gas (N/sub 2/), and the resulting oxygen quantitatively converted to carbon monoxide (CO) by contact with hot granular carbon (C). Two analysis modes are made possible: (1) rapid determination of total pyrolyzable obtained by decomposing the sample at 1135/sup 0/C, or (2) temperature-programmed oxygen thermal analysis obtained by heating the sample from room temperature to 1135/sup 0/C as a function of time. The analyzer basically comprises a pyrolysis tube containing a bed of granular carbon under N/sub 2/, ovens used to heat the carbon and/or decompose the sample, and a non-dispersive infrared CO detector coupled to a mini-computer to quantitate oxygen in the decomposition products and control oven heating.

  4. Oxygen analyzer

    DOEpatents

    Benner, William H.

    1986-01-01

    An oxygen analyzer which identifies and classifies microgram quantities of oxygen in ambient particulate matter and for quantitating organic oxygen in solvent extracts of ambient particulate matter. A sample is pyrolyzed in oxygen-free nitrogen gas (N.sub.2), and the resulting oxygen quantitatively converted to carbon monoxide (CO) by contact with hot granular carbon (C). Two analysis modes are made possible: (1) rapid determination of total pyrolyzable oxygen obtained by decomposing the sample at 1135.degree. C., or (2) temperature-programmed oxygen thermal analysis obtained by heating the sample from room temperature to 1135.degree. C. as a function of time. The analyzer basically comprises a pyrolysis tube containing a bed of granular carbon under N.sub.2, ovens used to heat the carbon and/or decompose the sample, and a non-dispersive infrared CO detector coupled to a mini-computer to quantitate oxygen in the decomposition products and control oven heating.

  5. SDAR 1.0 a New Quantitative Toolkit for Analyze Stratigraphic Data

    NASA Astrophysics Data System (ADS)

    Ortiz, John; Moreno, Carlos; Cardenas, Andres; Jaramillo, Carlos

    2015-04-01

    Since the foundation of stratigraphy geoscientists have recognized that data obtained from stratigraphic columns (SC), two dimensional schemes recording descriptions of both geological and paleontological features (e.g., thickness of rock packages, grain size, fossil and lithological components, and sedimentary structures), are key elements for establishing reliable hypotheses about the distribution in space and time of rock sequences, and ancient sedimentary environmental and paleobiological dynamics. Despite the tremendous advances on the way geoscientists store, plot, and quantitatively analyze sedimentological and paleontological data (e.g., Macrostrat [http://www.macrostrat.org/], Paleobiology Database [http://www.paleodb.org/], respectively), there is still a lack of computational methodologies designed to quantitatively examine data from a highly detailed SCs. Moreover, frequently the stratigraphic information is plotted "manually" using vector graphics editors (e.g., Corel Draw, Illustrator), however, this information although store on a digital format, cannot be used readily for any quantitative analysis. Therefore, any attempt to examine the stratigraphic data in an analytical fashion necessarily takes further steps. Given these issues, we have developed the sofware 'Stratigraphic Data Analysis in R' (SDAR), which stores in a database all sedimentological, stratigraphic, and paleontological information collected from a SC, allowing users to generate high-quality graphic plots (including one or multiple features stored in the database). SDAR also encompasses quantitative analyses helping users to quantify stratigraphic information (e.g. grain size, sorting and rounding, proportion of sand/shale). Finally, given that the SDAR analysis module, has been written in the open-source high-level computer language "R graphics/statistics language" [R Development Core Team, 2014], it is already loaded with many of the crucial features required to accomplish basic and

  6. Guidelines for Reporting Quantitative Methods and Results in Primary Research

    ERIC Educational Resources Information Center

    Norris, John M.; Plonsky, Luke; Ross, Steven J.; Schoonen, Rob

    2015-01-01

    Adequate reporting of quantitative research about language learning involves careful consideration of the logic, rationale, and actions underlying both study designs and the ways in which data are analyzed. These guidelines, commissioned and vetted by the board of directors of "Language Learning," outline the basic expectations for…

  7. Quantitative performance measurements of bent crystal Laue analyzers for X-ray fluorescence spectroscopy.

    PubMed

    Karanfil, C; Bunker, G; Newville, M; Segre, C U; Chapman, D

    2012-05-01

    Third-generation synchrotron radiation sources pose difficult challenges for energy-dispersive detectors for XAFS because of their count rate limitations. One solution to this problem is the bent crystal Laue analyzer (BCLA), which removes most of the undesired scatter and fluorescence before it reaches the detector, effectively eliminating detector saturation due to background. In this paper experimental measurements of BCLA performance in conjunction with a 13-element germanium detector, and a quantitative analysis of the signal-to-noise improvement of BCLAs are presented. The performance of BCLAs are compared with filters and slits.

  8. Solution identification and quantitative analysis of fiber-capacitive drop analyzer based on multivariate statistical methods

    NASA Astrophysics Data System (ADS)

    Chen, Zhe; Qiu, Zurong; Huo, Xinming; Fan, Yuming; Li, Xinghua

    2017-03-01

    A fiber-capacitive drop analyzer is an instrument which monitors a growing droplet to produce a capacitive opto-tensiotrace (COT). Each COT is an integration of fiber light intensity signals and capacitance signals and can reflect the unique physicochemical property of a liquid. In this study, we propose a solution analytical and concentration quantitative method based on multivariate statistical methods. Eight characteristic values are extracted from each COT. A series of COT characteristic values of training solutions at different concentrations compose a data library of this kind of solution. A two-stage linear discriminant analysis is applied to analyze different solution libraries and establish discriminant functions. Test solutions can be discriminated by these functions. After determining the variety of test solutions, Spearman correlation test and principal components analysis are used to filter and reduce dimensions of eight characteristic values, producing a new representative parameter. A cubic spline interpolation function is built between the parameters and concentrations, based on which we can calculate the concentration of the test solution. Methanol, ethanol, n-propanol, and saline solutions are taken as experimental subjects in this paper. For each solution, nine or ten different concentrations are chosen to be the standard library, and the other two concentrations compose the test group. By using the methods mentioned above, all eight test solutions are correctly identified and the average relative error of quantitative analysis is 1.11%. The method proposed is feasible which enlarges the applicable scope of recognizing liquids based on the COT and improves the concentration quantitative precision, as well.

  9. Integrating Quantitative and Qualitative Results in Health Science Mixed Methods Research Through Joint Displays

    PubMed Central

    Guetterman, Timothy C.; Fetters, Michael D.; Creswell, John W.

    2015-01-01

    PURPOSE Mixed methods research is becoming an important methodology to investigate complex health-related topics, yet the meaningful integration of qualitative and quantitative data remains elusive and needs further development. A promising innovation to facilitate integration is the use of visual joint displays that bring data together visually to draw out new insights. The purpose of this study was to identify exemplar joint displays by analyzing the various types of joint displays being used in published articles. METHODS We searched for empirical articles that included joint displays in 3 journals that publish state-of-the-art mixed methods research. We analyzed each of 19 identified joint displays to extract the type of display, mixed methods design, purpose, rationale, qualitative and quantitative data sources, integration approaches, and analytic strategies. Our analysis focused on what each display communicated and its representation of mixed methods analysis. RESULTS The most prevalent types of joint displays were statistics-by-themes and side-by-side comparisons. Innovative joint displays connected findings to theoretical frameworks or recommendations. Researchers used joint displays for convergent, explanatory sequential, exploratory sequential, and intervention designs. We identified exemplars for each of these designs by analyzing the inferences gained through using the joint display. Exemplars represented mixed methods integration, presented integrated results, and yielded new insights. CONCLUSIONS Joint displays appear to provide a structure to discuss the integrated analysis and assist both researchers and readers in understanding how mixed methods provides new insights. We encourage researchers to use joint displays to integrate and represent mixed methods analysis and discuss their value. PMID:26553895

  10. Integrating Quantitative and Qualitative Results in Health Science Mixed Methods Research Through Joint Displays.

    PubMed

    Guetterman, Timothy C; Fetters, Michael D; Creswell, John W

    2015-11-01

    Mixed methods research is becoming an important methodology to investigate complex health-related topics, yet the meaningful integration of qualitative and quantitative data remains elusive and needs further development. A promising innovation to facilitate integration is the use of visual joint displays that bring data together visually to draw out new insights. The purpose of this study was to identify exemplar joint displays by analyzing the various types of joint displays being used in published articles. We searched for empirical articles that included joint displays in 3 journals that publish state-of-the-art mixed methods research. We analyzed each of 19 identified joint displays to extract the type of display, mixed methods design, purpose, rationale, qualitative and quantitative data sources, integration approaches, and analytic strategies. Our analysis focused on what each display communicated and its representation of mixed methods analysis. The most prevalent types of joint displays were statistics-by-themes and side-by-side comparisons. Innovative joint displays connected findings to theoretical frameworks or recommendations. Researchers used joint displays for convergent, explanatory sequential, exploratory sequential, and intervention designs. We identified exemplars for each of these designs by analyzing the inferences gained through using the joint display. Exemplars represented mixed methods integration, presented integrated results, and yielded new insights. Joint displays appear to provide a structure to discuss the integrated analysis and assist both researchers and readers in understanding how mixed methods provides new insights. We encourage researchers to use joint displays to integrate and represent mixed methods analysis and discuss their value. © 2015 Annals of Family Medicine, Inc.

  11. Quantitatively analyzing the mechanism of giant circular dichroism in extrinsic plasmonic chiral nanostructures by tracking the interplay of electric and magnetic dipoles.

    PubMed

    Hu, Li; Tian, Xiaorui; Huang, Yingzhou; Fang, Liang; Fang, Yurui

    2016-02-14

    Plasmonic chirality has drawn much attention because of tunable circular dichroism (CD) and the enhancement for chiral molecule signals. Although various mechanisms have been proposed to explain the plasmonic CD, a quantitative explanation like the ab initio mechanism for chiral molecules, is still unavailable. In this study, a mechanism similar to the mechanisms associated with chiral molecules was analyzed. The giant extrinsic circular dichroism of a plasmonic splitting rectangle ring was quantitatively investigated from a theoretical standpoint. The interplay of the electric and magnetic modes of the meta-structure is proposed to explain the giant CD. We analyzed the interplay using both an analytical coupled electric-magnetic dipole model and a finite element method model. The surface charge distributions showed that the circular current yielded by the splitting rectangle ring causes the ring to behave like a magneton at some resonant modes, which then interact with the electric modes, resulting in a mixing of the two types of modes. The strong interplay of the two mode types is primarily responsible for the giant CD. The analysis of the chiral near-field of the structure shows potential applications for chiral molecule sensing.

  12. Ten Years of LibQual: A Study of Qualitative and Quantitative Survey Results at the University of Mississippi 2001-2010

    ERIC Educational Resources Information Center

    Greenwood, Judy T.; Watson, Alex P.; Dennis, Melissa

    2011-01-01

    This article analyzes quantitative adequacy gap scores and coded qualitative comments from LibQual surveys at the University of Mississippi from 2001 to 2010, looking for relationships between library policy changes and LibQual results and any other trends that emerged. Analysis found no relationship between changes in policy and survey results…

  13. Standard Reference Line Combined with One-Point Calibration-Free Laser-Induced Breakdown Spectroscopy (CF-LIBS) to Quantitatively Analyze Stainless and Heat Resistant Steel.

    PubMed

    Fu, Hongbo; Wang, Huadong; Jia, Junwei; Ni, Zhibo; Dong, Fengzhong

    2018-01-01

    Due to the influence of major elements' self-absorption, scarce observable spectral lines of trace elements, and relative efficiency correction of experimental system, accurate quantitative analysis with calibration-free laser-induced breakdown spectroscopy (CF-LIBS) is in fact not easy. In order to overcome these difficulties, standard reference line (SRL) combined with one-point calibration (OPC) is used to analyze six elements in three stainless-steel and five heat-resistant steel samples. The Stark broadening and Saha - Boltzmann plot of Fe are used to calculate the electron density and the plasma temperature, respectively. In the present work, we tested the original SRL method, the SRL with the OPC method, and intercept with the OPC method. The final calculation results show that the latter two methods can effectively improve the overall accuracy of quantitative analysis and the detection limits of trace elements.

  14. 2L-PCA: a two-level principal component analyzer for quantitative drug design and its applications.

    PubMed

    Du, Qi-Shi; Wang, Shu-Qing; Xie, Neng-Zhong; Wang, Qing-Yan; Huang, Ri-Bo; Chou, Kuo-Chen

    2017-09-19

    A two-level principal component predictor (2L-PCA) was proposed based on the principal component analysis (PCA) approach. It can be used to quantitatively analyze various compounds and peptides about their functions or potentials to become useful drugs. One level is for dealing with the physicochemical properties of drug molecules, while the other level is for dealing with their structural fragments. The predictor has the self-learning and feedback features to automatically improve its accuracy. It is anticipated that 2L-PCA will become a very useful tool for timely providing various useful clues during the process of drug development.

  15. The Pioneer 10 plasma analyzer results at Jupiter

    NASA Technical Reports Server (NTRS)

    Wolfe, J. H.

    1975-01-01

    Results are reported for the Pioneer 10 plasma-analyzer experiment at Jupiter. The analyzer system consisted of dual 90-deg quadrispherical electrostatic analyzers, multiple charged-particle detectors, and attendant electronics; it was capable of determining the incident plasma-distribution parameters over the energy range from 100 to 18,000 eV for protons and from approximately 1 to 500 eV for electrons. Data are presented on the interaction between the solar wind and the Jovian magnetosphere, the interplanetary ion flux, observations of the magnetosheath plasma, and traversals of the bow shock and magnetopause. Values are estimated for the proton isotropic temperature, number density, and bulk velocity within the magnetosheath flow field as well as for the beta parameter, ion number density, and magnetic-energy density of the magnetospheric plasma. It is argued that Jupiter has a reasonably thick magnetosphere somewhat similar to earth's except for the vastly different scale sizes involved.

  16. Connecting qualitative observation and quantitative measurement for enhancing quantitative literacy in plant anatomy course

    NASA Astrophysics Data System (ADS)

    Nuraeni, E.; Rahmat, A.

    2018-05-01

    Forming of cognitive schemes of plant anatomy concepts is performed by processing of qualitative and quantitative data obtained from microscopic observations. To enhancing student’s quantitative literacy, strategy of plant anatomy course was modified by adding the task to analyze quantitative data produced by quantitative measurement of plant anatomy guided by material course. Participant in this study was 24 biology students and 35 biology education students. Quantitative Literacy test, complex thinking in plant anatomy test and questioner used to evaluate the course. Quantitative literacy capability data was collected by quantitative literacy test with the rubric from the Association of American Colleges and Universities, Complex thinking in plant anatomy by test according to Marzano and questioner. Quantitative literacy data are categorized according to modified Rhodes and Finley categories. The results showed that quantitative literacy of biology education students is better than biology students.

  17. Approaches to quantitating the results of differentially dyed cottons

    USDA-ARS?s Scientific Manuscript database

    The differential dyeing (DD) method has served as a subjective method for visually determining immature cotton fibers. In an attempt to quantitate the results of the differential dyeing method, and thus offer an efficient means of elucidating cotton maturity without visual discretion, image analysi...

  18. The Impact of Acquisition Dose on Quantitative Breast Density Estimation with Digital Mammography: Results from ACRIN PA 4006.

    PubMed

    Chen, Lin; Ray, Shonket; Keller, Brad M; Pertuz, Said; McDonald, Elizabeth S; Conant, Emily F; Kontos, Despina

    2016-09-01

    Purpose To investigate the impact of radiation dose on breast density estimation in digital mammography. Materials and Methods With institutional review board approval and Health Insurance Portability and Accountability Act compliance under waiver of consent, a cohort of women from the American College of Radiology Imaging Network Pennsylvania 4006 trial was retrospectively analyzed. All patients underwent breast screening with a combination of dose protocols, including standard full-field digital mammography, low-dose digital mammography, and digital breast tomosynthesis. A total of 5832 images from 486 women were analyzed with previously validated, fully automated software for quantitative estimation of density. Clinical Breast Imaging Reporting and Data System (BI-RADS) density assessment results were also available from the trial reports. The influence of image acquisition radiation dose on quantitative breast density estimation was investigated with analysis of variance and linear regression. Pairwise comparisons of density estimations at different dose levels were performed with Student t test. Agreement of estimation was evaluated with quartile-weighted Cohen kappa values and Bland-Altman limits of agreement. Results Radiation dose of image acquisition did not significantly affect quantitative density measurements (analysis of variance, P = .37 to P = .75), with percent density demonstrating a high overall correlation between protocols (r = 0.88-0.95; weighted κ = 0.83-0.90). However, differences in breast percent density (1.04% and 3.84%, P < .05) were observed within high BI-RADS density categories, although they were significantly correlated across the different acquisition dose levels (r = 0.76-0.92, P < .05). Conclusion Precision and reproducibility of automated breast density measurements with digital mammography are not substantially affected by variations in radiation dose; thus, the use of low-dose techniques for the purpose of density estimation

  19. The Impact of Acquisition Dose on Quantitative Breast Density Estimation with Digital Mammography: Results from ACRIN PA 4006

    PubMed Central

    Chen, Lin; Ray, Shonket; Keller, Brad M.; Pertuz, Said; McDonald, Elizabeth S.; Conant, Emily F.

    2016-01-01

    Purpose To investigate the impact of radiation dose on breast density estimation in digital mammography. Materials and Methods With institutional review board approval and Health Insurance Portability and Accountability Act compliance under waiver of consent, a cohort of women from the American College of Radiology Imaging Network Pennsylvania 4006 trial was retrospectively analyzed. All patients underwent breast screening with a combination of dose protocols, including standard full-field digital mammography, low-dose digital mammography, and digital breast tomosynthesis. A total of 5832 images from 486 women were analyzed with previously validated, fully automated software for quantitative estimation of density. Clinical Breast Imaging Reporting and Data System (BI-RADS) density assessment results were also available from the trial reports. The influence of image acquisition radiation dose on quantitative breast density estimation was investigated with analysis of variance and linear regression. Pairwise comparisons of density estimations at different dose levels were performed with Student t test. Agreement of estimation was evaluated with quartile-weighted Cohen kappa values and Bland-Altman limits of agreement. Results Radiation dose of image acquisition did not significantly affect quantitative density measurements (analysis of variance, P = .37 to P = .75), with percent density demonstrating a high overall correlation between protocols (r = 0.88–0.95; weighted κ = 0.83–0.90). However, differences in breast percent density (1.04% and 3.84%, P < .05) were observed within high BI-RADS density categories, although they were significantly correlated across the different acquisition dose levels (r = 0.76–0.92, P < .05). Conclusion Precision and reproducibility of automated breast density measurements with digital mammography are not substantially affected by variations in radiation dose; thus, the use of low-dose techniques for the purpose of density

  20. A cascading failure model for analyzing railway accident causation

    NASA Astrophysics Data System (ADS)

    Liu, Jin-Tao; Li, Ke-Ping

    2018-01-01

    In this paper, a new cascading failure model is proposed for quantitatively analyzing the railway accident causation. In the model, the loads of nodes are redistributed according to the strength of the causal relationships between the nodes. By analyzing the actual situation of the existing prevention measures, a critical threshold of the load parameter in the model is obtained. To verify the effectiveness of the proposed cascading model, simulation experiments of a train collision accident are performed. The results show that the cascading failure model can describe the cascading process of the railway accident more accurately than the previous models, and can quantitatively analyze the sensitivities and the influence of the causes. In conclusion, this model can assist us to reveal the latent rules of accident causation to reduce the occurrence of railway accidents.

  1. An approach for quantitatively analyzing the genuine tripartite nonlocality of general three-qubit states

    NASA Astrophysics Data System (ADS)

    Su, Zhaofeng; Li, Lvzhou; Ling, Jie

    2018-04-01

    Nonlocality is an important resource for quantum information processing. Genuine tripartite nonlocality, which is sufficiently confirmed by the violation of Svetlichny inequality, is a kind of more precious resource than the standard one. The genuine tripartite nonlocality is usually quantified by the amount of maximal violation of Svetlichny inequality. The problem of detecting and quantifying the genuine tripartite nonlocality of quantum states is of practical significance but still open for the case of general three-qubit quantum states. In this paper, we quantitatively investigate the genuine nonlocality of three-qubit states, which not only include pure states but also include mixed states. Firstly, we derive a simplified formula for the genuine nonlocality of a general three-qubit state, which is a function of the corresponding three correlation matrices. Secondly, we develop three properties of the genuine nonlocality which can help us to analyze the genuine nonlocality of complex states and understand the nature of quantum nonlocality. Further, we get analytical results of genuine nonlocality for two classes of three-qubit states which have special correlation matrices. In particular, the genuine nonlocality of generalized three-qubit GHZ states, which is derived by Ghose et al. (Phys. Rev. Lett. 102, 250404, 2009), and that of three-qubit GHZ-symmetric states, which is derived by Paul et al. (Phys. Rev. A 94, 032101, 2016), can be easily derived by applying the strategy and properties developed in this paper.

  2. Diagnostic accuracy of uriSed automated urine microscopic sediment analyzer and dipstick parameters in predicting urine culture test results.

    PubMed

    Huysal, Kağan; Budak, Yasemin U; Karaca, Ayse Ulusoy; Aydos, Murat; Kahvecioğlu, Serdar; Bulut, Mehtap; Polat, Murat

    2013-01-01

    Urinary tract infection (UTI) is one of the most common types of infection. Currently, diagnosis is primarily based on microbiologic culture, which is time- and labor-consuming. The aim of this study was to assess the diagnostic accuracy of urinalysis results from UriSed (77 Electronica, Budapest, Hungary), an automated microscopic image-based sediment analyzer, in predicting positive urine cultures. We examined a total of 384 urine specimens from hospitalized patients and outpatients attending our hospital on the same day for urinalysis, dipstick tests and semi-quantitative urine culture. The urinalysis results were compared with those of conventional semiquantitative urine culture. Of 384 urinary specimens, 68 were positive for bacteriuria by culture, and were thus considered true positives. Comparison of these results with those obtained from the UriSed analyzer indicated that the analyzer had a specificity of 91.1%, a sensitivity of 47.0%, a positive predictive value (PPV) of 53.3% (95% confidence interval (CI) = 40.8-65.3), and a negative predictive value (NPV) of 88.8% (95% CI = 85.0-91.8%). The accuracy was 83.3% when the urine leukocyte parameter was used, 76.8% when bacteriuria analysis of urinary sediment was used, and 85.1% when the bacteriuria and leukocyturia parameters were combined. The presence of nitrite was the best indicator of culture positivity (99.3% specificity) but had a negative likelihood ratio of 0.7, indicating that it was not a reliable clinical test. Although the specificity of the UriSed analyzer was within acceptable limits, the sensitivity value was low. Thus, UriSed urinalysis resuIts do not accurately predict the outcome of culture.

  3. A quantitative headspace-solid-phase microextraction-gas chromatography-flame ionization detector method to analyze short chain free fatty acids in rat feces.

    PubMed

    Fiorini, Dennis; Boarelli, Maria Chiara; Gabbianelli, Rosita; Ballini, Roberto; Pacetti, Deborah

    2016-09-01

    This study sought to develop and validate a quantitative method to analyze short chain free fatty acids (SCFAs) in rat feces by solid-phase microextraction and gas chromatography (SPME-GC) using the salt mixture ammonium sulfate and sodium dihydrogen phosphate as salting out agent. Conditioning and extraction time, linearity, limits of detection and quantification, repeatability, and recovery were evaluated. The proposed method allows quantification with improved sensitivity as compared with other methods exploiting SPME-GC. The method has been applied to analyze rat fecal samples, quantifying acetic, propionic, isobutyric, butyric, isopentanoic, pentanoic, and hexanoic acids. Copyright © 2016 Elsevier Inc. All rights reserved.

  4. Qualitative versus Quantitative Results: An Experimental Introduction to Data Interpretation.

    ERIC Educational Resources Information Center

    Johnson, Eric R.; Alter, Paula

    1989-01-01

    Described is an experiment in which the student can ascertain the meaning of a negative result from a qualitative test by performing a more sensitive quantitative test on the same sample. Methodology for testing urinary glucose with a spectrophotometer at 630 nm and with commercial assaying glucose strips is presented. (MVL)

  5. Method for analyzing microbial communities

    DOEpatents

    Zhou, Jizhong [Oak Ridge, TN; Wu, Liyou [Oak Ridge, TN

    2010-07-20

    The present invention provides a method for quantitatively analyzing microbial genes, species, or strains in a sample that contains at least two species or strains of microorganisms. The method involves using an isothermal DNA polymerase to randomly and representatively amplify genomic DNA of the microorganisms in the sample, hybridizing the resultant polynucleotide amplification product to a polynucleotide microarray that can differentiate different genes, species, or strains of microorganisms of interest, and measuring hybridization signals on the microarray to quantify the genes, species, or strains of interest.

  6. Quantitative MR imaging in fracture dating--Initial results.

    PubMed

    Baron, Katharina; Neumayer, Bernhard; Widek, Thomas; Schick, Fritz; Scheicher, Sylvia; Hassler, Eva; Scheurer, Eva

    2016-04-01

    For exact age determinations of bone fractures in a forensic context (e.g. in cases of child abuse) improved knowledge of the time course of the healing process and use of non-invasive modern imaging technology is of high importance. To date, fracture dating is based on radiographic methods by determining the callus status and thereby relying on an expert's experience. As a novel approach, this study aims to investigate the applicability of magnetic resonance imaging (MRI) for bone fracture dating by systematically investigating time-resolved changes in quantitative MR characteristics after a fracture event. Prior to investigating fracture healing in children, adults were examined for this study in order to test the methodology for this application. Altogether, 31 MR examinations in 17 subjects (♀: 11 ♂: 6; median age 34 ± 15 y, scanned 1-5 times over a period of up to 200 days after the fracture event) were performed on a clinical 3T MR scanner (TimTrio, Siemens AG, Germany). All subjects were treated conservatively for a fracture in either a long bone or in the collar bone. Both, qualitative and quantitative MR measurements were performed in all subjects. MR sequences for a quantitative measurement of relaxation times T1 and T2 in the fracture gap and musculature were applied. Maps of quantitative MR parameters T1, T2, and magnetisation transfer ratio (MTR) were calculated and evaluated by investigating changes over time in the fractured area by defined ROIs. Additionally, muscle areas were examined as reference regions to validate this approach. Quantitative evaluation of 23 MR data sets (12 test subjects, ♀: 7 ♂: 5) showed an initial peak in T1 values in the fractured area (T1=1895 ± 607 ms), which decreased over time to a value of 1094 ± 182 ms (200 days after the fracture event). T2 values also peaked for early-stage fractures (T2=115 ± 80 ms) and decreased to 73 ± 33 ms within 21 days after the fracture event. After that time point, no

  7. Analyzing the texture changes in the quantitative phase maps of adipocytes

    NASA Astrophysics Data System (ADS)

    Roitshtain, Darina; Sharabani-Yosef, Orna; Gefen, Amit; Shaked, Natan T.

    2016-03-01

    We present a new analysis tool for studying texture changes in the quantitative phase maps of live cells acquired by wide-field interferometry. The sensitivity of wide-field interferometry systems to small changes in refractive index enables visualizing cells and inner cell organelles without the using fluorescent dyes or other cell-invasive approaches, which may affect the measurement and require external labeling. Our label-free texture-analysis tool is based directly on the optical path delay profile of the sample and does not necessitate decoupling refractive index and thickness in the cell quantitative phase profile; thus, relevant parameters can be calculated using a single-frame acquisition. Our experimental system includes low-coherence wide-field interferometer, combined with simultaneous florescence microscopy system for validation. We used this system and analysis tool for studying lipid droplets formation in adipocytes. The latter demonstration is relevant for various cellular functions such as lipid metabolism, protein storage and degradation to viral replication. These processes are functionally linked to several physiological and pathological conditions, including obesity and metabolic diseases. Quantification of these biological phenomena based on the texture changes in the cell phase map has a potential as a new cellular diagnosis tool.

  8. Analyzing the BBOB results by means of benchmarking concepts.

    PubMed

    Mersmann, O; Preuss, M; Trautmann, H; Bischl, B; Weihs, C

    2015-01-01

    We present methods to answer two basic questions that arise when benchmarking optimization algorithms. The first one is: which algorithm is the "best" one? and the second one is: which algorithm should I use for my real-world problem? Both are connected and neither is easy to answer. We present a theoretical framework for designing and analyzing the raw data of such benchmark experiments. This represents a first step in answering the aforementioned questions. The 2009 and 2010 BBOB benchmark results are analyzed by means of this framework and we derive insight regarding the answers to the two questions. Furthermore, we discuss how to properly aggregate rankings from algorithm evaluations on individual problems into a consensus, its theoretical background and which common pitfalls should be avoided. Finally, we address the grouping of test problems into sets with similar optimizer rankings and investigate whether these are reflected by already proposed test problem characteristics, finding that this is not always the case.

  9. Quantitative Hydrocarbon Surface Analysis

    NASA Technical Reports Server (NTRS)

    Douglas, Vonnie M.

    2000-01-01

    The elimination of ozone depleting substances, such as carbon tetrachloride, has resulted in the use of new analytical techniques for cleanliness verification and contamination sampling. The last remaining application at Rocketdyne which required a replacement technique was the quantitative analysis of hydrocarbons by infrared spectrometry. This application, which previously utilized carbon tetrachloride, was successfully modified using the SOC-400, a compact portable FTIR manufactured by Surface Optics Corporation. This instrument can quantitatively measure and identify hydrocarbons from solvent flush of hardware as well as directly analyze the surface of metallic components without the use of ozone depleting chemicals. Several sampling accessories are utilized to perform analysis for various applications.

  10. Quantitative learning strategies based on word networks

    NASA Astrophysics Data System (ADS)

    Zhao, Yue-Tian-Yi; Jia, Zi-Yang; Tang, Yong; Xiong, Jason Jie; Zhang, Yi-Cheng

    2018-02-01

    Learning English requires a considerable effort, but the way that vocabulary is introduced in textbooks is not optimized for learning efficiency. With the increasing population of English learners, learning process optimization will have significant impact and improvement towards English learning and teaching. The recent developments of big data analysis and complex network science provide additional opportunities to design and further investigate the strategies in English learning. In this paper, quantitative English learning strategies based on word network and word usage information are proposed. The strategies integrate the words frequency with topological structural information. By analyzing the influence of connected learned words, the learning weights for the unlearned words and dynamically updating of the network are studied and analyzed. The results suggest that quantitative strategies significantly improve learning efficiency while maintaining effectiveness. Especially, the optimized-weight-first strategy and segmented strategies outperform other strategies. The results provide opportunities for researchers and practitioners to reconsider the way of English teaching and designing vocabularies quantitatively by balancing the efficiency and learning costs based on the word network.

  11. ImatraNMR: Novel software for batch integration and analysis of quantitative NMR spectra

    NASA Astrophysics Data System (ADS)

    Mäkelä, A. V.; Heikkilä, O.; Kilpeläinen, I.; Heikkinen, S.

    2011-08-01

    Quantitative NMR spectroscopy is a useful and important tool for analysis of various mixtures. Recently, in addition of traditional quantitative 1D 1H and 13C NMR methods, a variety of pulse sequences aimed for quantitative or semiquantitative analysis have been developed. To obtain actual usable results from quantitative spectra, they must be processed and analyzed with suitable software. Currently, there are many processing packages available from spectrometer manufacturers and third party developers, and most of them are capable of analyzing and integration of quantitative spectra. However, they are mainly aimed for processing single or few spectra, and are slow and difficult to use when large numbers of spectra and signals are being analyzed, even when using pre-saved integration areas or custom scripting features. In this article, we present a novel software, ImatraNMR, designed for batch analysis of quantitative spectra. In addition to capability of analyzing large number of spectra, it provides results in text and CSV formats, allowing further data-analysis using spreadsheet programs or general analysis programs, such as Matlab. The software is written with Java, and thus it should run in any platform capable of providing Java Runtime Environment version 1.6 or newer, however, currently it has only been tested with Windows and Linux (Ubuntu 10.04). The software is free for non-commercial use, and is provided with source code upon request.

  12. Comprehensive Quantitative Analysis on Privacy Leak Behavior

    PubMed Central

    Fan, Lejun; Wang, Yuanzhuo; Jin, Xiaolong; Li, Jingyuan; Cheng, Xueqi; Jin, Shuyuan

    2013-01-01

    Privacy information is prone to be leaked by illegal software providers with various motivations. Privacy leak behavior has thus become an important research issue of cyber security. However, existing approaches can only qualitatively analyze privacy leak behavior of software applications. No quantitative approach, to the best of our knowledge, has been developed in the open literature. To fill this gap, in this paper we propose for the first time four quantitative metrics, namely, possibility, severity, crypticity, and manipulability, for privacy leak behavior analysis based on Privacy Petri Net (PPN). In order to compare the privacy leak behavior among different software, we further propose a comprehensive metric, namely, overall leak degree, based on these four metrics. Finally, we validate the effectiveness of the proposed approach using real-world software applications. The experimental results demonstrate that our approach can quantitatively analyze the privacy leak behaviors of various software types and reveal their characteristics from different aspects. PMID:24066046

  13. Comprehensive quantitative analysis on privacy leak behavior.

    PubMed

    Fan, Lejun; Wang, Yuanzhuo; Jin, Xiaolong; Li, Jingyuan; Cheng, Xueqi; Jin, Shuyuan

    2013-01-01

    Privacy information is prone to be leaked by illegal software providers with various motivations. Privacy leak behavior has thus become an important research issue of cyber security. However, existing approaches can only qualitatively analyze privacy leak behavior of software applications. No quantitative approach, to the best of our knowledge, has been developed in the open literature. To fill this gap, in this paper we propose for the first time four quantitative metrics, namely, possibility, severity, crypticity, and manipulability, for privacy leak behavior analysis based on Privacy Petri Net (PPN). In order to compare the privacy leak behavior among different software, we further propose a comprehensive metric, namely, overall leak degree, based on these four metrics. Finally, we validate the effectiveness of the proposed approach using real-world software applications. The experimental results demonstrate that our approach can quantitatively analyze the privacy leak behaviors of various software types and reveal their characteristics from different aspects.

  14. Comparison of avian biochemical test results with Abaxis VetScan and Hitachi 911 analyzers.

    PubMed

    Greenacre, Cheryl B; Flatland, Bente; Souza, Marcy J; Fry, Michael M

    2008-12-01

    To compare results of clinical biochemical analysis using an Abaxis VetScan bench-top analyzer with reagents specifically marketed for avian use and a Hitachi 911 analyzer, plasma (both methods) and whole blood (VetScan method) samples from 20 clinically healthy Hispaniolan Amazon parrots (Amazona ventralis) were analyzed. Correlation between methods was very high (r = 0.9-1.0) for aspartate aminotransferase (AST), calcium, glucose, and uric acid; high (r = 0.7-0.89) for creatine kinase (CK), phosphorus, potassium, and total protein; moderate (r = 0.5-0.69) for globulin; and low (r = 0.3-0.49) for albumin and sodium. VetScan analyzer results for globulin, sodium, and uric acid had a constant negative bias (values below those from the Hitachi method). Based on difference plot analysis, results for AST, calcium, CK, and glucose are comparable. Because 16 of 20 values fell below the lower detection limit of the VetScan analyzer, bile acid data were excluded from analysis. By using a relatively small sample size (0.1 ml whole blood or plasma), the VetScan analyzer offers rapid in-house results, compact size, and ease of operation. For 4 of the most clinically relevant biochemical analytes used in avian medicine (AST, calcium, CK, glucose), it offers reliable values. For an additional 4 analytes (phosphorous, potassium, total protein, uric acid), establishing analyzer-specific reference intervals is recommended. Neither the VetScan nor the Hitachi method is recommended to assess albumin and globulin concentrations.

  15. MISSE-7 MESA Miniaturized Electrostatic Analyzer - Ion Spectra Analysis Preliminary Results

    NASA Astrophysics Data System (ADS)

    Enloe, C. L.; Balthazor, R. L.; McHarg, M. G.; Clark, A. L.; Waite, D.; Wallerstein, A. J.; Wilson, K. A.

    2011-12-01

    The 7th Materials on the International Space Station Experiment (MISSE-7) was launched in November 2009 and retrieved on STS-134 in April 2011. One of the onboard experiments, the Miniaturized Electrostatic Analyzer (MESA), is a small low-cost low-size/weight/power ion and electron spectrometer that was pointed into ram during the majority of the time onboard. Over 800 Mb of data has been obtained by taking spectra every three minutes on-orbit. The data has been analyzed with a novel "parameterizing the parameters" method suitable for on-orbit data analysis using low-cost microcontrollers. Preliminary results are shown.

  16. ImatraNMR: novel software for batch integration and analysis of quantitative NMR spectra.

    PubMed

    Mäkelä, A V; Heikkilä, O; Kilpeläinen, I; Heikkinen, S

    2011-08-01

    Quantitative NMR spectroscopy is a useful and important tool for analysis of various mixtures. Recently, in addition of traditional quantitative 1D (1)H and (13)C NMR methods, a variety of pulse sequences aimed for quantitative or semiquantitative analysis have been developed. To obtain actual usable results from quantitative spectra, they must be processed and analyzed with suitable software. Currently, there are many processing packages available from spectrometer manufacturers and third party developers, and most of them are capable of analyzing and integration of quantitative spectra. However, they are mainly aimed for processing single or few spectra, and are slow and difficult to use when large numbers of spectra and signals are being analyzed, even when using pre-saved integration areas or custom scripting features. In this article, we present a novel software, ImatraNMR, designed for batch analysis of quantitative spectra. In addition to capability of analyzing large number of spectra, it provides results in text and CSV formats, allowing further data-analysis using spreadsheet programs or general analysis programs, such as Matlab. The software is written with Java, and thus it should run in any platform capable of providing Java Runtime Environment version 1.6 or newer, however, currently it has only been tested with Windows and Linux (Ubuntu 10.04). The software is free for non-commercial use, and is provided with source code upon request. Copyright © 2011 Elsevier Inc. All rights reserved.

  17. Closed loop models for analyzing engineering requirements for simulators

    NASA Technical Reports Server (NTRS)

    Baron, S.; Muralidharan, R.; Kleinman, D.

    1980-01-01

    A closed loop analytic model, incorporating a model for the human pilot, (namely, the optimal control model) that would allow certain simulation design tradeoffs to be evaluated quantitatively was developed. This model was applied to a realistic flight control problem. The resulting model is used to analyze both overall simulation effects and the effects of individual elements. The results show that, as compared to an ideal continuous simulation, the discrete simulation can result in significant performance and/or workload penalties.

  18. Applying Knowledge of Quantitative Design and Analysis

    ERIC Educational Resources Information Center

    Baskas, Richard S.

    2011-01-01

    This study compared and contrasted two quantitative scholarly articles in relation to their research designs. Their designs were analyzed by the comparison of research references and research specific vocabulary to describe how various research methods were used. When researching and analyzing quantitative scholarly articles, it is imperative to…

  19. Quantitative Data Analysis--In the Graduate Curriculum

    ERIC Educational Resources Information Center

    Albers, Michael J.

    2017-01-01

    A quantitative research study collects numerical data that must be analyzed to help draw the study's conclusions. Teaching quantitative data analysis is not teaching number crunching, but teaching a way of critical thinking for how to analyze the data. The goal of data analysis is to reveal the underlying patterns, trends, and relationships of a…

  20. Quantitative coronary plaque analysis predicts high-risk plaque morphology on coronary computed tomography angiography: results from the ROMICAT II trial.

    PubMed

    Liu, Ting; Maurovich-Horvat, Pál; Mayrhofer, Thomas; Puchner, Stefan B; Lu, Michael T; Ghemigian, Khristine; Kitslaar, Pieter H; Broersen, Alexander; Pursnani, Amit; Hoffmann, Udo; Ferencik, Maros

    2018-02-01

    Semi-automated software can provide quantitative assessment of atherosclerotic plaques on coronary CT angiography (CTA). The relationship between established qualitative high-risk plaque features and quantitative plaque measurements has not been studied. We analyzed the association between quantitative plaque measurements and qualitative high-risk plaque features on coronary CTA. We included 260 patients with plaque who underwent coronary CTA in the Rule Out Myocardial Infarction/Ischemia Using Computer Assisted Tomography (ROMICAT) II trial. Quantitative plaque assessment and qualitative plaque characterization were performed on a per coronary segment basis. Quantitative coronary plaque measurements included plaque volume, plaque burden, remodeling index, and diameter stenosis. In qualitative analysis, high-risk plaque was present if positive remodeling, low CT attenuation plaque, napkin-ring sign or spotty calcium were detected. Univariable and multivariable logistic regression analyses were performed to assess the association between quantitative and qualitative high-risk plaque assessment. Among 888 segments with coronary plaque, high-risk plaque was present in 391 (44.0%) segments by qualitative analysis. In quantitative analysis, segments with high-risk plaque had higher total plaque volume, low CT attenuation plaque volume, plaque burden and remodeling index. Quantitatively assessed low CT attenuation plaque volume (odds ratio 1.12 per 1 mm 3 , 95% CI 1.04-1.21), positive remodeling (odds ratio 1.25 per 0.1, 95% CI 1.10-1.41) and plaque burden (odds ratio 1.53 per 0.1, 95% CI 1.08-2.16) were associated with high-risk plaque. Quantitative coronary plaque characteristics (low CT attenuation plaque volume, positive remodeling and plaque burden) measured by semi-automated software correlated with qualitative assessment of high-risk plaque features.

  1. Analyzing force concept inventory with item response theory

    NASA Astrophysics Data System (ADS)

    Wang, Jing; Bao, Lei

    2010-10-01

    Item response theory is a popular assessment method used in education. It rests on the assumption of a probability framework that relates students' innate ability and their performance on test questions. Item response theory transforms students' raw test scores into a scaled proficiency score, which can be used to compare results obtained with different test questions. The scaled score also addresses the issues of ceiling effects and guessing, which commonly exist in quantitative assessment. We used item response theory to analyze the force concept inventory (FCI). Our results show that item response theory can be useful for analyzing physics concept surveys such as the FCI and produces results about the individual questions and student performance that are beyond the capability of classical statistics. The theory yields detailed measurement parameters regarding the difficulty, discrimination features, and probability of correct guess for each of the FCI questions.

  2. What Really Happens in Quantitative Group Research? Results of a Content Analysis of Recent Quantitative Research in "JSGW"

    ERIC Educational Resources Information Center

    Boyle, Lauren H.; Whittaker, Tiffany A.; Eyal, Maytal; McCarthy, Christopher J.

    2017-01-01

    The authors conducted a content analysis on quantitative studies published in "The Journal for Specialists in Group Work" ("JSGW") between 2012 and 2015. This brief report provides a general overview of the current practices of quantitative group research in counseling. The following study characteristics are reported and…

  3. Quantitative Muscle Ultrasonography in Carpal Tunnel Syndrome.

    PubMed

    Lee, Hyewon; Jee, Sungju; Park, Soo Ho; Ahn, Seung-Chan; Im, Juneho; Sohn, Min Kyun

    2016-12-01

    To assess the reliability of quantitative muscle ultrasonography (US) in healthy subjects and to evaluate the correlation between quantitative muscle US findings and electrodiagnostic study results in patients with carpal tunnel syndrome (CTS). The clinical significance of quantitative muscle US in CTS was also assessed. Twenty patients with CTS and 20 age-matched healthy volunteers were recruited. All control and CTS subjects underwent a bilateral median and ulnar nerve conduction study (NCS) and quantitative muscle US. Transverse US images of the abductor pollicis brevis (APB) and abductor digiti minimi (ADM) were obtained to measure muscle cross-sectional area (CSA), thickness, and echo intensity (EI). EI was determined using computer-assisted, grayscale analysis. Inter-rater and intra-rater reliability for quantitative muscle US in control subjects, and differences in muscle thickness, CSA, and EI between the CTS patient and control groups were analyzed. Relationships between quantitative US parameters and electrodiagnostic study results were evaluated. Quantitative muscle US had high inter-rater and intra-rater reliability in the control group. Muscle thickness and CSA were significantly decreased, and EI was significantly increased in the APB of the CTS group (all p<0.05). EI demonstrated a significant positive correlation with latency of the median motor and sensory NCS in CTS patients (p<0.05). These findings suggest that quantitative muscle US parameters may be useful for detecting muscle changes in CTS. Further study involving patients with other neuromuscular diseases is needed to evaluate peripheral muscle change using quantitative muscle US.

  4. Biotransformation of dichlorodiphenyltrichloroethane in the benthic polychaete, Nereis succinea: quantitative estimation by analyzing the partitioning of chemicals between gut fluid and lipid.

    PubMed

    Wang, Fei; Pei, Yuan-yuan; You, Jing

    2015-02-01

    Biotransformation plays an important role in the bioaccumulation and toxicity of a chemical in biota. Dichlorodiphenyltrichloroethane (DDT) commonly co-occurs with its metabolites (dichlorodiphenyldichloroethane [DDD] and dichlorodiphenyldichloroethylene [DDE]), in the environment; thus it is a challenge to accurately quantify the biotransformation rates of DDT and distinguish the sources of the accumulated metabolites in an organism. The present study describes a method developed to quantitatively analyze the biotransformation of p,p'-DDT in the benthic polychaete, Nereis succinea. The lugworms were exposed to sediments spiked with DDT at various concentrations for 28 d. Degradation of DDT to DDD and DDE occurred in sediments during the aging period, and approximately two-thirds of the DDT remained in the sediment. To calculate the biotransformation rates, residues of individual compounds measured in the bioaccumulation testing (after biotransformation) were compared with residues predicted by analyzing the partitioning of the parent and metabolite compounds between gut fluid and tissue lipid (before biotransformation). The results suggest that sediment ingestion rates decreased when DDT concentrations in sediment increased. Extensive biotransformation of DDT occurred in N. succinea, with 86% of DDT being metabolized to DDD and <2% being transformed to DDE. Of the DDD that accumulated in the lugworms, approximately 70% was the result of DDT biotransformation, and the remaining 30% was from direct uptake of sediment-associated DDD. In addition, the biotransformation was not dependent on bulk sediment concentrations, but rather on bioaccessible concentrations of the chemicals in sediment, which were quantified by gut fluid extraction. The newly established method improved the accuracy of prediction of the bioaccumulation and toxicity of DDTs. © 2014 SETAC.

  5. Dynamic and quantitative method of analyzing service consistency evolution based on extended hierarchical finite state automata.

    PubMed

    Fan, Linjun; Tang, Jun; Ling, Yunxiang; Li, Benxian

    2014-01-01

    This paper is concerned with the dynamic evolution analysis and quantitative measurement of primary factors that cause service inconsistency in service-oriented distributed simulation applications (SODSA). Traditional methods are mostly qualitative and empirical, and they do not consider the dynamic disturbances among factors in service's evolution behaviors such as producing, publishing, calling, and maintenance. Moreover, SODSA are rapidly evolving in terms of large-scale, reusable, compositional, pervasive, and flexible features, which presents difficulties in the usage of traditional analysis methods. To resolve these problems, a novel dynamic evolution model extended hierarchical service-finite state automata (EHS-FSA) is constructed based on finite state automata (FSA), which formally depict overall changing processes of service consistency states. And also the service consistency evolution algorithms (SCEAs) based on EHS-FSA are developed to quantitatively assess these impact factors. Experimental results show that the bad reusability (17.93% on average) is the biggest influential factor, the noncomposition of atomic services (13.12%) is the second biggest one, and the service version's confusion (1.2%) is the smallest one. Compared with previous qualitative analysis, SCEAs present good effectiveness and feasibility. This research can guide the engineers of service consistency technologies toward obtaining a higher level of consistency in SODSA.

  6. Dynamic and Quantitative Method of Analyzing Service Consistency Evolution Based on Extended Hierarchical Finite State Automata

    PubMed Central

    Fan, Linjun; Tang, Jun; Ling, Yunxiang; Li, Benxian

    2014-01-01

    This paper is concerned with the dynamic evolution analysis and quantitative measurement of primary factors that cause service inconsistency in service-oriented distributed simulation applications (SODSA). Traditional methods are mostly qualitative and empirical, and they do not consider the dynamic disturbances among factors in service's evolution behaviors such as producing, publishing, calling, and maintenance. Moreover, SODSA are rapidly evolving in terms of large-scale, reusable, compositional, pervasive, and flexible features, which presents difficulties in the usage of traditional analysis methods. To resolve these problems, a novel dynamic evolution model extended hierarchical service-finite state automata (EHS-FSA) is constructed based on finite state automata (FSA), which formally depict overall changing processes of service consistency states. And also the service consistency evolution algorithms (SCEAs) based on EHS-FSA are developed to quantitatively assess these impact factors. Experimental results show that the bad reusability (17.93% on average) is the biggest influential factor, the noncomposition of atomic services (13.12%) is the second biggest one, and the service version's confusion (1.2%) is the smallest one. Compared with previous qualitative analysis, SCEAs present good effectiveness and feasibility. This research can guide the engineers of service consistency technologies toward obtaining a higher level of consistency in SODSA. PMID:24772033

  7. Quantitative influence of risk factors on blood glucose level.

    PubMed

    Chen, Songjing; Luo, Senlin; Pan, Limin; Zhang, Tiemei; Han, Longfei; Zhao, Haixiu

    2014-01-01

    The aim of this study is to quantitatively analyze the influence of risk factors on the blood glucose level, and to provide theory basis for understanding the characteristics of blood glucose change and confirming the intervention index for type 2 diabetes. The quantitative method is proposed to analyze the influence of risk factors on blood glucose using back propagation (BP) neural network. Ten risk factors are screened first. Then the cohort is divided into nine groups by gender and age. According to the minimum error principle, nine BP models are trained respectively. The quantitative values of the influence of different risk factors on the blood glucose change can be obtained by sensitivity calculation. The experiment results indicate that weight is the leading cause of blood glucose change (0.2449). The second factors are cholesterol, age and triglyceride. The total ratio of these four factors reaches to 77% of the nine screened risk factors. And the sensitivity sequences can provide judgment method for individual intervention. This method can be applied to risk factors quantitative analysis of other diseases and potentially used for clinical practitioners to identify high risk populations for type 2 diabetes as well as other disease.

  8. Convergent and sequential synthesis designs: implications for conducting and reporting systematic reviews of qualitative and quantitative evidence.

    PubMed

    Hong, Quan Nha; Pluye, Pierre; Bujold, Mathieu; Wassef, Maggy

    2017-03-23

    Systematic reviews of qualitative and quantitative evidence can provide a rich understanding of complex phenomena. This type of review is increasingly popular, has been used to provide a landscape of existing knowledge, and addresses the types of questions not usually covered in reviews relying solely on either quantitative or qualitative evidence. Although several typologies of synthesis designs have been developed, none have been tested on a large sample of reviews. The aim of this review of reviews was to identify and develop a typology of synthesis designs and methods that have been used and to propose strategies for synthesizing qualitative and quantitative evidence. A review of systematic reviews combining qualitative and quantitative evidence was performed. Six databases were searched from inception to December 2014. Reviews were included if they were systematic reviews combining qualitative and quantitative evidence. The included reviews were analyzed according to three concepts of synthesis processes: (a) synthesis methods, (b) sequence of data synthesis, and (c) integration of data and synthesis results. A total of 459 reviews were included. The analysis of this literature highlighted a lack of transparency in reporting how evidence was synthesized and a lack of consistency in the terminology used. Two main types of synthesis designs were identified: convergent and sequential synthesis designs. Within the convergent synthesis design, three subtypes were found: (a) data-based convergent synthesis design, where qualitative and quantitative evidence is analyzed together using the same synthesis method, (b) results-based convergent synthesis design, where qualitative and quantitative evidence is analyzed separately using different synthesis methods and results of both syntheses are integrated during a final synthesis, and (c) parallel-results convergent synthesis design consisting of independent syntheses of qualitative and quantitative evidence and an

  9. CesrTA Retarding Field Analyzer Modeling Results

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Calvey, J.R.; Celata, C.M.; Crittenden, J.A.

    2010-05-23

    Retarding field analyzers (RFAs) provide an effective measure of the local electron cloud density and energy distribution. Proper interpretation of RFA data can yield information about the behavior of the cloud, as well as the surface properties of the instrumented vacuum chamber. However, due to the complex interaction of the cloud with the RFA itself, understanding these measurements can be nontrivial. This paper examines different methods for interpreting RFA data via cloud simulation programs. Techniques include postprocessing the output of a simulation code to predict the RFA response; and incorporating an RFA model into the cloud modeling program itself.

  10. Interlaboratory Comparison of Quantitative PCR Test Results for Dehalococcoides

    EPA Science Inventory

    Quantitative PCR (qPCR) techniques have been widely used to measure Dehalococcoides (Dhc) DNA in the groundwater at field sites for several years. Interpretation of these data may be complicated when different laboratories using alternate methods conduct the analysis. An...

  11. Comparison of 99mTc-MDP SPECT qualitative vs quantitative results in patients with suspected condylar hyperplasia.

    PubMed

    López Buitrago, D F; Ruiz Botero, J; Corral, C M; Carmona, A R; Sabogal, A

    To compare qualitative vs quantitative results of Single Photon Emission Computerised Tomography (SPECT), calculated from percentage of 99m Tc-MDP (methylene diphosphonate) uptake, in condyles of patients with a presumptive clinical diagnosis of condylar hyperplasia. A retrospective, descriptive study was conducted on the 99m Tc-MDP SPECT bone scintigraphy reports from 51 patients, with clinical impression of facial asymmetry related to condylar hyperplasia referred by their specialist in orthodontics or maxillofacial surgery, to a nuclear medicine department in order to take this type of test. Quantitative data from 99m Tc-MDP condylar uptake of each were obtained and compared with qualitative image interpretation reported by a nuclear medicine expert. The concordances between the 51 qualitative and quantitative reports results was established. The total sample included 32 women (63%) and 19 men (37%). The patient age range was 13-45 years (21±8 years). According to qualitative reports, 19 patients were positive for right side condylar hyperplasia, 12 for left side condylar hyperplasia, with 8 bilateral, and 12 negative. The quantitative reports diagnosed 16 positives for right side condylar hyperplasia, 10 for left side condylar hyperplasia, and 25 negatives. Nuclear medicine images are an important diagnostic tool, but the qualitative interpretation of the images is not as reliable as the quantitative calculation. The agreement between the two types of report is low (39.2%, Kappa=0.13; P>.2). The main limitation of quantitative reports is that they do not register bilateral condylar hyperplasia cases. Copyright © 2017 Elsevier España, S.L.U. y SEMNIM. All rights reserved.

  12. Analyzing students' attitudes towards science during inquiry-based lessons

    NASA Astrophysics Data System (ADS)

    Kostenbader, Tracy C.

    Due to the logistics of guided-inquiry lesson, students learn to problem solve and develop critical thinking skills. This mixed-methods study analyzed the students' attitudes towards science during inquiry lessons. My quantitative results from a repeated measures survey showed no significant difference between student attitudes when taught with either structured-inquiry or guided-inquiry lessons. The qualitative results analyzed through a constant-comparative method did show that students generate positive interest, critical thinking and low level stress during guided-inquiry lessons. The qualitative research also gave insight into a teacher's transition to guided-inquiry. This study showed that with my students, their attitudes did not change during this transition according to the qualitative data however, the qualitative data did how high levels of excitement. The results imply that students like guided-inquiry laboratories, even though they require more work, just as much as they like traditional laboratories with less work and less opportunity for creativity.

  13. Improving Middle School Students’ Quantitative Literacy through Inquiry Lab and Group Investigation

    NASA Astrophysics Data System (ADS)

    Aisya, N. S. M.; Supriatno, B.; Saefudin; Anggraeni, S.

    2017-02-01

    The purpose of this study was to analyze the application of metacognitive strategies learning based Vee Diagram through Inquiry Lab and Group Investigation toward students’ quantitative literacy. This study compared two treatments on learning activity in middle school. The metacognitive strategies have applied to the content of environmental pollution at 7th grade. This study used a quantitative approach with quasi-experimental method. The research sample were the 7th grade students, involves 27 students in the experimental through Inquiry Lab and 27 students in the experimental through Group Investigation. The instruments that used in this research were pretest and posttest quantitative literacy skills, learning step observation sheets, and the questionnaire of teachers and students responses. As the result, N-gain average of pretest and posttest increased in both experimental groups. The average of posttest score was 61,11 for the Inquiry Lab and 54,01 to the Group Investigation. The average score of N-gain quantitative literacy skill of Inquiry Lab class was 0,492 and Group Investigation class was 0,426. Both classes of experiments showed an average N-gain in the medium category. The data has been analyzed statistically by using SPSS ver.23 and the results showed that although both the learning model can develop quantitative literacy, but there is not significantly different of improving students’ quantitative literacy between Inquiry Lab and Group Investigation in environmental pollution material.

  14. Solving Quantitative Problems: Guidelines for Teaching Derived from Research.

    ERIC Educational Resources Information Center

    Kramers-Pals, H.; Pilot, A.

    1988-01-01

    Presents four guidelines for teaching quantitative problem-solving based on research results: analyze difficulties of students, develop a system of heuristics, select and map key relations, and design instruction with proper orientation, exercise, and feedback. Discusses the four guidelines and uses flow charts and diagrams to show how the…

  15. Quantitative Method for Analyzing the Allocation of Risks in Transportation Construction

    DOT National Transportation Integrated Search

    1979-04-01

    The report presents a conceptual model of risk that was developed to analyze the impact on owner's cost of alternate allocations of risk among owner and contractor in mass transit construction. A model and analysis procedure are developed, based on d...

  16. A sample-to-result system for blood coagulation tests on a microfluidic disk analyzer

    PubMed Central

    Lin, Chia-Hui; Liu, Cheng-Yuan; Shih, Chih-Hsin; Lu, Chien-Hsing

    2014-01-01

    In this report, we describe in detail a microfluidic analyzer, which is able to conduct blood coagulation tests using whole blood samples. Sample preparation steps, such as whole blood aliquoting and metering, plasma separation, decanting, and mixing with reagents were performed in sequence through microfluidic functions integrated on a disk. Both prothrombin time (PT) and activated partial thromboplastin time (aPTT) were carried out on the same platform and the test results can be reported in 5 min. Fifty clinical samples were tested for both PT and aPTT utilizing the microfluidic disk analyzer and the instrument used in hospitals. The test results showed good correlation and agreement between the two instruments. PMID:25332733

  17. Development of online NIR urine analyzing system based on AOTF

    NASA Astrophysics Data System (ADS)

    Wan, Feng; Sun, Zhendong; Li, Xiaoxia

    2006-09-01

    In this paper, some key techniques on development of on-line MR urine analyzing system based on AOTF (Acousto - Optics Tunable Filter) are introduced. Problems about designing the optical system including collimation of incident light and working distance (the shortest distance for separating incident light and diffracted light) are analyzed and researched. DDS (Direct Digital Synthesizer) controlled by microprocessor is used to realize the wavelength scan. The experiment results show that this MR urine analyzing system based on. AOTF has 10000 - 4000cm -1 wavelength range and O.3ms wavelength transfer rate. Compare with the conventional Fourier Transform NIP. spectrophotometer for analyzing multi-components in urine, this system features low cost, small volume and on-line measurement function. Unscrambler software (multivariate statistical software by CAMO Inc. Norway) is selected as the software for processing the data. This system can realize on line quantitative analysis of protein, urea and creatinine in urine.

  18. A Quantitative Proteomics Approach to Clinical Research with Non-Traditional Samples

    PubMed Central

    Licier, Rígel; Miranda, Eric; Serrano, Horacio

    2016-01-01

    The proper handling of samples to be analyzed by mass spectrometry (MS) can guarantee excellent results and a greater depth of analysis when working in quantitative proteomics. This is critical when trying to assess non-traditional sources such as ear wax, saliva, vitreous humor, aqueous humor, tears, nipple aspirate fluid, breast milk/colostrum, cervical-vaginal fluid, nasal secretions, bronco-alveolar lavage fluid, and stools. We intend to provide the investigator with relevant aspects of quantitative proteomics and to recognize the most recent clinical research work conducted with atypical samples and analyzed by quantitative proteomics. Having as reference the most recent and different approaches used with non-traditional sources allows us to compare new strategies in the development of novel experimental models. On the other hand, these references help us to contribute significantly to the understanding of the proportions of proteins in different proteomes of clinical interest and may lead to potential advances in the emerging field of precision medicine. PMID:28248241

  19. A Quantitative Proteomics Approach to Clinical Research with Non-Traditional Samples.

    PubMed

    Licier, Rígel; Miranda, Eric; Serrano, Horacio

    2016-10-17

    The proper handling of samples to be analyzed by mass spectrometry (MS) can guarantee excellent results and a greater depth of analysis when working in quantitative proteomics. This is critical when trying to assess non-traditional sources such as ear wax, saliva, vitreous humor, aqueous humor, tears, nipple aspirate fluid, breast milk/colostrum, cervical-vaginal fluid, nasal secretions, bronco-alveolar lavage fluid, and stools. We intend to provide the investigator with relevant aspects of quantitative proteomics and to recognize the most recent clinical research work conducted with atypical samples and analyzed by quantitative proteomics. Having as reference the most recent and different approaches used with non-traditional sources allows us to compare new strategies in the development of novel experimental models. On the other hand, these references help us to contribute significantly to the understanding of the proportions of proteins in different proteomes of clinical interest and may lead to potential advances in the emerging field of precision medicine.

  20. Assessing agreement between preclinical magnetic resonance imaging and histology: An evaluation of their image qualities and quantitative results

    PubMed Central

    Elschner, Cindy; Korn, Paula; Hauptstock, Maria; Schulz, Matthias C.; Range, Ursula; Jünger, Diana; Scheler, Ulrich

    2017-01-01

    One consequence of demographic change is the increasing demand for biocompatible materials for use in implants and prostheses. This is accompanied by a growing number of experimental animals because the interactions between new biomaterials and its host tissue have to be investigated. To evaluate novel materials and engineered tissues the use of non-destructive imaging modalities have been identified as a strategic priority. This provides the opportunity for studying interactions repeatedly with individual animals, along with the advantages of reduced biological variability and decreased number of laboratory animals. However, histological techniques are still the golden standard in preclinical biomaterial research. The present article demonstrates a detailed method comparison between histology and magnetic resonance imaging. This includes the presentation of their image qualities as well as the detailed statistical analysis for assessing agreement between quantitative measures. Exemplarily, the bony ingrowth of tissue engineered bone substitutes for treatment of a cleft-like maxillary bone defect has been evaluated. By using a graphical concordance analysis the mean difference between MRI results and histomorphometrical measures has been examined. The analysis revealed a slightly but significant bias in the case of the bone volume (biasHisto−MRI:Bone volume=2.40 %, p<0.005) and a clearly significant deviation for the remaining defect width (biasHisto−MRI:Defect width=−6.73 %, p≪0.005). But the study although showed a considerable effect of the analyzed section position to the quantitative result. It could be proven, that the bias of the data sets was less originated due to the imaging modalities, but mainly on the evaluation of different slice positions. The article demonstrated that method comparisons not always need the use of an independent animal study, additionally. PMID:28666026

  1. Quantitative results of stellar evolution and pulsation theories.

    NASA Technical Reports Server (NTRS)

    Fricke, K.; Stobie, R. S.; Strittmatter, P. A.

    1971-01-01

    The discrepancy between the masses of Cepheid variables deduced from evolution theory and pulsation theory is examined. The effect of input physics on evolutionary tracks is first discussed; in particular, changes in the opacity are considered. The sensitivity of pulsation masses to opacity changes and to the ascribed values of luminosity and effective temperature are then analyzed. The Cepheid mass discrepancy is discussed in the light of the results already obtained. Other astronomical evidence, including the mass-luminosity relation for main sequence stars, the solar neutrino flux, and cluster ages are also considered in an attempt to determine the most likely source of error in the event that substantial mass loss has not occurred.

  2. Quantitative fluorescence angiography for neurosurgical interventions.

    PubMed

    Weichelt, Claudia; Duscha, Philipp; Steinmeier, Ralf; Meyer, Tobias; Kuß, Julia; Cimalla, Peter; Kirsch, Matthias; Sobottka, Stephan B; Koch, Edmund; Schackert, Gabriele; Morgenstern, Ute

    2013-06-01

    Present methods for quantitative measurement of cerebral perfusion during neurosurgical operations require additional technology for measurement, data acquisition, and processing. This study used conventional fluorescence video angiography--as an established method to visualize blood flow in brain vessels--enhanced by a quantifying perfusion software tool. For these purposes, the fluorescence dye indocyanine green is given intravenously, and after activation by a near-infrared light source the fluorescence signal is recorded. Video data are analyzed by software algorithms to allow quantification of the blood flow. Additionally, perfusion is measured intraoperatively by a reference system. Furthermore, comparing reference measurements using a flow phantom were performed to verify the quantitative blood flow results of the software and to validate the software algorithm. Analysis of intraoperative video data provides characteristic biological parameters. These parameters were implemented in the special flow phantom for experimental validation of the developed software algorithms. Furthermore, various factors that influence the determination of perfusion parameters were analyzed by means of mathematical simulation. Comparing patient measurement, phantom experiment, and computer simulation under certain conditions (variable frame rate, vessel diameter, etc.), the results of the software algorithms are within the range of parameter accuracy of the reference methods. Therefore, the software algorithm for calculating cortical perfusion parameters from video data presents a helpful intraoperative tool without complex additional measurement technology.

  3. Classification of cassava genotypes based on qualitative and quantitative data.

    PubMed

    Oliveira, E J; Oliveira Filho, O S; Santos, V S

    2015-02-02

    We evaluated the genetic variation of cassava accessions based on qualitative (binomial and multicategorical) and quantitative traits (continuous). We characterized 95 accessions obtained from the Cassava Germplasm Bank of Embrapa Mandioca e Fruticultura; we evaluated these accessions for 13 continuous, 10 binary, and 25 multicategorical traits. First, we analyzed the accessions based only on quantitative traits; next, we conducted joint analysis (qualitative and quantitative traits) based on the Ward-MLM method, which performs clustering in two stages. According to the pseudo-F, pseudo-t2, and maximum likelihood criteria, we identified five and four groups based on quantitative trait and joint analysis, respectively. The smaller number of groups identified based on joint analysis may be related to the nature of the data. On the other hand, quantitative data are more subject to environmental effects in the phenotype expression; this results in the absence of genetic differences, thereby contributing to greater differentiation among accessions. For most of the accessions, the maximum probability of classification was >0.90, independent of the trait analyzed, indicating a good fit of the clustering method. Differences in clustering according to the type of data implied that analysis of quantitative and qualitative traits in cassava germplasm might explore different genomic regions. On the other hand, when joint analysis was used, the means and ranges of genetic distances were high, indicating that the Ward-MLM method is very useful for clustering genotypes when there are several phenotypic traits, such as in the case of genetic resources and breeding programs.

  4. The effect of different methods and image analyzers on the results of the in vivo comet assay.

    PubMed

    Kyoya, Takahiro; Iwamoto, Rika; Shimanura, Yuko; Terada, Megumi; Masuda, Shuichi

    2018-01-01

    The in vivo comet assay is a widely used genotoxicity test that can detect DNA damage in a range of organs. It is included in the Organisation for Economic Co-operation and Development Guidelines for the Testing of Chemicals. However, various protocols are still used for this assay, and several different image analyzers are used routinely to evaluate the results. Here, we verified a protocol that largely contributes to the equivalence of results, and we assessed the effect on the results when slides made from the same sample were analyzed using two different image analyzers (Comet Assay IV vs Comet Analyzer). Standardizing the agarose concentrations and DNA unwinding and electrophoresis times had a large impact on the equivalence of the results between the different methods used for the in vivo comet assay. In addition, there was some variation in the sensitivity of the two different image analyzers tested; however this variation was considered to be minor and became negligible when the test conditions were standardized between the two different methods. By standardizing the concentrations of low melting agarose and DNA unwinding and electrophoresis times between both methods used in the current study, the sensitivity to detect the genotoxicity of a positive control substance in the in vivo comet assay became generally comparable, independently of the image analyzer used. However, there may still be the possibility that other conditions, except for the three described here, could affect the reproducibility of the in vivo comet assay.

  5. Quantitative magnetic resonance imaging in traumatic brain injury.

    PubMed

    Bigler, E D

    2001-04-01

    Quantitative neuroimaging has now become a well-established method for analyzing magnetic resonance imaging in traumatic brain injury (TBI). A general review of studies that have examined quantitative changes following TBI is presented. The consensus of quantitative neuroimaging studies is that most brain structures demonstrate changes in volume or surface area after injury. The patterns of atrophy are consistent with the generalized nature of brain injury and diffuse axonal injury. Various clinical caveats are provided including how quantitative neuroimaging findings can be used clinically and in predicting rehabilitation outcome. The future of quantitative neuroimaging also is discussed.

  6. X-ray chemical analyzer for field applications

    DOEpatents

    Gamba, Otto O. M.

    1977-01-01

    A self-supporting portable field multichannel X-ray chemical analyzer system comprising a lightweight, flexibly connected, remotely locatable, radioisotope-excited sensing probe utilizing a cryogenically-cooled solid state semi-conductor crystal detector for fast in situ non-destructive, qualitative and quantitative analysis of elements in solid, powder, liquid or slurried form, utilizing an X-ray energy dispersive spectrometry technique.

  7. Quantitative Amyloid Imaging in Autosomal Dominant Alzheimer's Disease: Results from the DIAN Study Group.

    PubMed

    Su, Yi; Blazey, Tyler M; Owen, Christopher J; Christensen, Jon J; Friedrichsen, Karl; Joseph-Mathurin, Nelly; Wang, Qing; Hornbeck, Russ C; Ances, Beau M; Snyder, Abraham Z; Cash, Lisa A; Koeppe, Robert A; Klunk, William E; Galasko, Douglas; Brickman, Adam M; McDade, Eric; Ringman, John M; Thompson, Paul M; Saykin, Andrew J; Ghetti, Bernardino; Sperling, Reisa A; Johnson, Keith A; Salloway, Stephen P; Schofield, Peter R; Masters, Colin L; Villemagne, Victor L; Fox, Nick C; Förster, Stefan; Chen, Kewei; Reiman, Eric M; Xiong, Chengjie; Marcus, Daniel S; Weiner, Michael W; Morris, John C; Bateman, Randall J; Benzinger, Tammie L S

    2016-01-01

    Amyloid imaging plays an important role in the research and diagnosis of dementing disorders. Substantial variation in quantitative methods to measure brain amyloid burden exists in the field. The aim of this work is to investigate the impact of methodological variations to the quantification of amyloid burden using data from the Dominantly Inherited Alzheimer's Network (DIAN), an autosomal dominant Alzheimer's disease population. Cross-sectional and longitudinal [11C]-Pittsburgh Compound B (PiB) PET imaging data from the DIAN study were analyzed. Four candidate reference regions were investigated for estimation of brain amyloid burden. A regional spread function based technique was also investigated for the correction of partial volume effects. Cerebellar cortex, brain-stem, and white matter regions all had stable tracer retention during the course of disease. Partial volume correction consistently improves sensitivity to group differences and longitudinal changes over time. White matter referencing improved statistical power in the detecting longitudinal changes in relative tracer retention; however, the reason for this improvement is unclear and requires further investigation. Full dynamic acquisition and kinetic modeling improved statistical power although it may add cost and time. Several technical variations to amyloid burden quantification were examined in this study. Partial volume correction emerged as the strategy that most consistently improved statistical power for the detection of both longitudinal changes and across-group differences. For the autosomal dominant Alzheimer's disease population with PiB imaging, utilizing brainstem as a reference region with partial volume correction may be optimal for current interventional trials. Further investigation of technical issues in quantitative amyloid imaging in different study populations using different amyloid imaging tracers is warranted.

  8. Introduction to Quantitative Science, a Ninth-Grade Laboratory-Centered Course Stressing Quantitative Observation and Mathematical Analysis of Experimental Results. Final Report.

    ERIC Educational Resources Information Center

    Badar, Lawrence J.

    This report, in the form of a teacher's guide, presents materials for a ninth grade introductory course on Introduction to Quantitative Science (IQS). It is intended to replace a traditional ninth grade general science with a process oriented course that will (1) unify the sciences, and (2) provide a quantitative preparation for the new science…

  9. ELISA-BASE: An Integrated Bioinformatics Tool for Analyzing and Tracking ELISA Microarray Data

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    White, Amanda M.; Collett, James L.; Seurynck-Servoss, Shannon L.

    ELISA-BASE is an open-source database for capturing, organizing and analyzing protein enzyme-linked immunosorbent assay (ELISA) microarray data. ELISA-BASE is an extension of the BioArray Soft-ware Environment (BASE) database system, which was developed for DNA microarrays. In order to make BASE suitable for protein microarray experiments, we developed several plugins for importing and analyzing quantitative ELISA microarray data. Most notably, our Protein Microarray Analysis Tool (ProMAT) for processing quantita-tive ELISA data is now available as a plugin to the database.

  10. A quantitative model for designing keyboard layout.

    PubMed

    Shieh, K K; Lin, C C

    1999-02-01

    This study analyzed the quantitative relationship between keytapping times and ergonomic principles in typewriting skills. Keytapping times and key-operating characteristics of a female subject typing on the Qwerty and Dvorak keyboards for six weeks each were collected and analyzed. The results showed that characteristics of the typed material and the movements of hands and fingers were significantly related to keytapping times. The most significant factors affecting keytapping times were association frequency between letters, consecutive use of the same hand or finger, and the finger used. A regression equation for relating keytapping times to ergonomic principles was fitted to the data. Finally, a protocol for design of computerized keyboard layout based on the regression equation was proposed.

  11. Quantitative variability of renewable energy resources in Norway

    NASA Astrophysics Data System (ADS)

    Christakos, Konstantinos; Varlas, George; Cheliotis, Ioannis; Aalstad, Kristoffer; Papadopoulos, Anastasios; Katsafados, Petros; Steeneveld, Gert-Jan

    2017-04-01

    Based on European Union (EU) targets for 2030, the share of renewable energy (RE) consumption should be increased at 27%. RE resources such as hydropower, wind, wave power and solar power are strongly depending on the chaotic behavior of the weather conditions and climate. Due to this dependency, the prediction of the spatiotemporal variability of the RE resources is more crucial factor than in other energy resources (i.e. carbon based energy). The fluctuation of the RE resources can affect the development of the RE technologies, the energy grid, supply and prices. This study investigates the variability of the potential RE resources in Norway. More specifically, hydropower, wind, wave, and solar power are quantitatively analyzed and correlated with respect to various spatial and temporal scales. In order to analyze the diversities and their interrelationships, reanalysis and observational data of wind, precipitation, wave, and solar radiation are used for a quantitative assessment. The results indicate a high variability of marine RE resources in the North Sea and the Norwegian Sea.

  12. Preliminary pioneer 10 encounter results from the ames research center plasma analyzer experiment.

    PubMed

    Wolfe, J H; Collard, H R; Mihalov, J D; Intriligator, D S

    1974-01-25

    Preliminary results from the Ames Research Center plasma analyzer experiment for the Pioneer 10 Jupiter encounter indicate that Jupiter has a detached bow shock and magnetopause similar to the case at Earth but much larger in spatial extent. In contrast to Earth, Jupiter's outer magnetosphere appears to be highly inflated by thermal plasma and therefore highly responsive in size to changes in solar wind dynamic pressure.

  13. Quantitative Susceptibility Mapping after Sports-Related Concussion.

    PubMed

    Koch, K M; Meier, T B; Karr, R; Nencka, A S; Muftuler, L T; McCrea, M

    2018-06-07

    Quantitative susceptibility mapping using MR imaging can assess changes in brain tissue structure and composition. This report presents preliminary results demonstrating changes in tissue magnetic susceptibility after sports-related concussion. Longitudinal quantitative susceptibility mapping metrics were produced from imaging data acquired from cohorts of concussed and control football athletes. One hundred thirty-six quantitative susceptibility mapping datasets were analyzed across 3 separate visits (24 hours after injury, 8 days postinjury, and 6 months postinjury). Longitudinal quantitative susceptibility mapping group analyses were performed on stability-thresholded brain tissue compartments and selected subregions. Clinical concussion metrics were also measured longitudinally in both cohorts and compared with the measured quantitative susceptibility mapping. Statistically significant increases in white matter susceptibility were identified in the concussed athlete group during the acute (24 hour) and subacute (day 8) period. These effects were most prominent at the 8-day visit but recovered and showed no significant difference from controls at the 6-month visit. The subcortical gray matter showed no statistically significant group differences. Observed susceptibility changes after concussion appeared to outlast self-reported clinical recovery metrics at a group level. At an individual subject level, susceptibility increases within the white matter showed statistically significant correlations with return-to-play durations. The results of this preliminary investigation suggest that sports-related concussion can induce physiologic changes to brain tissue that can be detected using MR imaging-based magnetic susceptibility estimates. In group analyses, the observed tissue changes appear to persist beyond those detected on clinical outcome assessments and were associated with return-to-play duration after sports-related concussion. © 2018 by American Journal of

  14. A quantitative assessment of the Hadoop framework for analyzing massively parallel DNA sequencing data.

    PubMed

    Siretskiy, Alexey; Sundqvist, Tore; Voznesenskiy, Mikhail; Spjuth, Ola

    2015-01-01

    New high-throughput technologies, such as massively parallel sequencing, have transformed the life sciences into a data-intensive field. The most common e-infrastructure for analyzing this data consists of batch systems that are based on high-performance computing resources; however, the bioinformatics software that is built on this platform does not scale well in the general case. Recently, the Hadoop platform has emerged as an interesting option to address the challenges of increasingly large datasets with distributed storage, distributed processing, built-in data locality, fault tolerance, and an appealing programming methodology. In this work we introduce metrics and report on a quantitative comparison between Hadoop and a single node of conventional high-performance computing resources for the tasks of short read mapping and variant calling. We calculate efficiency as a function of data size and observe that the Hadoop platform is more efficient for biologically relevant data sizes in terms of computing hours for both split and un-split data files. We also quantify the advantages of the data locality provided by Hadoop for NGS problems, and show that a classical architecture with network-attached storage will not scale when computing resources increase in numbers. Measurements were performed using ten datasets of different sizes, up to 100 gigabases, using the pipeline implemented in Crossbow. To make a fair comparison, we implemented an improved preprocessor for Hadoop with better performance for splittable data files. For improved usability, we implemented a graphical user interface for Crossbow in a private cloud environment using the CloudGene platform. All of the code and data in this study are freely available as open source in public repositories. From our experiments we can conclude that the improved Hadoop pipeline scales better than the same pipeline on high-performance computing resources, we also conclude that Hadoop is an economically viable

  15. A quantitative evaluation of spurious results in the infrared spectroscopic measurement of CO2 isotope ratios

    NASA Astrophysics Data System (ADS)

    Mansfield, C. D.; Rutt, H. N.

    2002-02-01

    The possible generation of spurious results, arising from the application of infrared spectroscopic techniques to the measurement of carbon isotope ratios in breath, due to coincident absorption bands has been re-examined. An earlier investigation, which approached the problem qualitatively, fulfilled its aspirations in providing an unambiguous assurance that 13C16O2/12C16O2 ratios can be confidently measured for isotopic breath tests using instruments based on infrared absorption. Although this conclusion still stands, subsequent quantitative investigation has revealed an important exception that necessitates a strict adherence to sample collection protocol. The results show that concentrations and decay rates of the coincident breath trace compounds acetonitrile and carbon monoxide, found in the breath sample of a heavy smoker, can produce spurious results. Hence, findings from this investigation justify the concern that breath trace compounds present a risk to the accurate measurement of carbon isotope ratios in breath when using broadband, non-dispersive, ground state absorption infrared spectroscopy. It provides recommendations on the length of smoking abstention required to avoid generation of spurious results and also reaffirms, through quantitative argument, the validity of using infrared absorption spectroscopy to measure CO2 isotope ratios in breath.

  16. Ion-Exclusion Chromatography for Analyzing Organics in Water

    NASA Technical Reports Server (NTRS)

    Sauer, Richard; Rutz, Jeffrey A.; Schultz, John R.

    2006-01-01

    A liquid-chromatography technique has been developed for use in the quantitative analysis of urea (and of other nonvolatile organic compounds typically found with urea) dissolved in water. The technique involves the use of a column that contains an ion-exclusion resin; heretofore, this column has been sold for use in analyzing monosaccharides and food softeners, but not for analyzing water supplies. The prior technique commonly used to analyze water for urea content has been one of high-performance liquid chromatography (HPLC), with reliance on hydrophobic interactions between analytes in a water sample and long-chain alkyl groups bonded to an HPLC column. The prior technique has proven inadequate because of a strong tendency toward co-elution of urea with other compounds. Co-elution often causes the urea and other compounds to be crowded into a narrow region of the chromatogram (see left part of figure), thereby giving rise to low chromatographic resolution and misidentification of compounds. It is possible to quantitate urea or another analyte via ultraviolet- and visible-light absorbance measurements, but in order to perform such measurements, it is necessary to dilute the sample, causing a significant loss of sensitivity. The ion-exclusion resin used in the improved technique is sulfonated polystyrene in the calcium form. Whereas the alkyl-chain column used in the prior technique separates compounds on the basis of polarity only, the ion-exclusion-resin column used in the improved technique separates compounds on the basis of both molecular size and electric charge. As a result, the degree of separation is increased: instead of being crowded together into a single chromatographic peak only about 1 to 2 minutes wide as in the prior technique, the chromatographic peaks of different compounds are now separated from each other and spread out over a range about 33 minutes wide (see right part of figure), and the urea peak can readily be distinguished from the other

  17. Optical analyzer

    DOEpatents

    Hansen, A.D.

    1987-09-28

    An optical analyzer wherein a sample of particulate matter, and particularly of organic matter, which has been collected on a quartz fiber filter is placed in a combustion tube, and light from a light source is passed through the sample. The temperature of the sample is raised at a controlled rate and in a controlled atmosphere. The magnitude of the transmission of light through the sample is detected as the temperature is raised. A data processor, differentiator and a two pen recorder provide a chart of the optical transmission versus temperature and the rate of change of optical transmission versus temperature signatures (T and D) of the sample. These signatures provide information as to physical and chemical processes and a variety of quantitative and qualitative information about the sample. Additional information is obtained by repeating the run in different atmospheres and/or different rates or heating with other samples of the same particulate material collected on other filters. 7 figs.

  18. Analyzing Medical Image Search Behavior: Semantics and Prediction of Query Results.

    PubMed

    De-Arteaga, Maria; Eggel, Ivan; Kahn, Charles E; Müller, Henning

    2015-10-01

    Log files of information retrieval systems that record user behavior have been used to improve the outcomes of retrieval systems, understand user behavior, and predict events. In this article, a log file of the ARRS GoldMiner search engine containing 222,005 consecutive queries is analyzed. Time stamps are available for each query, as well as masked IP addresses, which enables to identify queries from the same person. This article describes the ways in which physicians (or Internet searchers interested in medical images) search and proposes potential improvements by suggesting query modifications. For example, many queries contain only few terms and therefore are not specific; others contain spelling mistakes or non-medical terms that likely lead to poor or empty results. One of the goals of this report is to predict the number of results a query will have since such a model allows search engines to automatically propose query modifications in order to avoid result lists that are empty or too large. This prediction is made based on characteristics of the query terms themselves. Prediction of empty results has an accuracy above 88%, and thus can be used to automatically modify the query to avoid empty result sets for a user. The semantic analysis and data of reformulations done by users in the past can aid the development of better search systems, particularly to improve results for novice users. Therefore, this paper gives important ideas to better understand how people search and how to use this knowledge to improve the performance of specialized medical search engines.

  19. A System Evaluation Theory Analyzing Value and Results Chain for Institutional Accreditation in Oman

    ERIC Educational Resources Information Center

    Paquibut, Rene Ymbong

    2017-01-01

    Purpose: This paper aims to apply the system evaluation theory (SET) to analyze the institutional quality standards of Oman Academic Accreditation Authority using the results chain and value chain tools. Design/methodology/approach: In systems thinking, the institutional standards are connected as input, process, output and feedback and leads to…

  20. Quantitative assessments of arousal by analyzing microsaccade rates and pupil fluctuations prior to slow eye movements.

    PubMed

    Honda, Shogo; Kohama, Takeshi; Tanaka, Tatsuro; Yoshida, Hisashi

    2014-01-01

    It is well known that a decline of arousal level causes of poor performance of movements or judgments. Our previous study indicates that microsaccade (MS) rates and pupil fluctuations change before slow eye movements (SEMs) (Honda et al. 2013). However, SEM detection of this study was obscure and insufficient. In this study, we propose a new SEM detection method and analyze MS rates and pupil fluctuations while subjects maintain their gaze on a target. We modified Shin et al.'s method, which is optimized for EOG (electrooculography) signals, to extract the period of sustaining SEMs using a general eye tracker. After SEM detection, we analyzed MS rates and pupil fluctuations prior to the initiation of SEMs. As a result, we were able to detect SEMs more precisely than in our previous study. Moreover, the results of eye movements and pupil fluctuations analyses show that gradual rise of MS rate and longitudinal miosis are observed prior to the initiation of SEMs, which is consistent with our previous study. These findings suggest that monitoring eye movements and pupil fluctuations may evaluate the arousal level more precisely. Further, we found that these tendencies become more significant when they are restricted to the initial SEMs.

  1. Increasing Literacy in Quantitative Methods: The Key to the Future of Canadian Psychology

    PubMed Central

    Counsell, Alyssa; Cribbie, Robert A.; Harlow, Lisa. L.

    2016-01-01

    Quantitative methods (QM) dominate empirical research in psychology. Unfortunately most researchers in psychology receive inadequate training in QM. This creates a challenge for researchers who require advanced statistical methods to appropriately analyze their data. Many of the recent concerns about research quality, replicability, and reporting practices are directly tied to the problematic use of QM. As such, improving quantitative literacy in psychology is an important step towards eliminating these concerns. The current paper will include two main sections that discuss quantitative challenges and opportunities. The first section discusses training and resources for students and presents descriptive results on the number of quantitative courses required and available to graduate students in Canadian psychology departments. In the second section, we discuss ways of improving quantitative literacy for faculty, researchers, and clinicians. This includes a strong focus on the importance of collaboration. The paper concludes with practical recommendations for improving quantitative skills and literacy for students and researchers in Canada. PMID:28042199

  2. Increasing Literacy in Quantitative Methods: The Key to the Future of Canadian Psychology.

    PubMed

    Counsell, Alyssa; Cribbie, Robert A; Harlow, Lisa L

    2016-08-01

    Quantitative methods (QM) dominate empirical research in psychology. Unfortunately most researchers in psychology receive inadequate training in QM. This creates a challenge for researchers who require advanced statistical methods to appropriately analyze their data. Many of the recent concerns about research quality, replicability, and reporting practices are directly tied to the problematic use of QM. As such, improving quantitative literacy in psychology is an important step towards eliminating these concerns. The current paper will include two main sections that discuss quantitative challenges and opportunities. The first section discusses training and resources for students and presents descriptive results on the number of quantitative courses required and available to graduate students in Canadian psychology departments. In the second section, we discuss ways of improving quantitative literacy for faculty, researchers, and clinicians. This includes a strong focus on the importance of collaboration. The paper concludes with practical recommendations for improving quantitative skills and literacy for students and researchers in Canada.

  3. Quantitative Amyloid Imaging in Autosomal Dominant Alzheimer’s Disease: Results from the DIAN Study Group

    PubMed Central

    Su, Yi; Blazey, Tyler M.; Owen, Christopher J.; Christensen, Jon J.; Friedrichsen, Karl; Joseph-Mathurin, Nelly; Wang, Qing; Hornbeck, Russ C.; Ances, Beau M.; Snyder, Abraham Z.; Cash, Lisa A.; Koeppe, Robert A.; Klunk, William E.; Galasko, Douglas; Brickman, Adam M.; McDade, Eric; Ringman, John M.; Thompson, Paul M.; Saykin, Andrew J.; Ghetti, Bernardino; Sperling, Reisa A.; Johnson, Keith A.; Salloway, Stephen P.; Schofield, Peter R.; Masters, Colin L.; Villemagne, Victor L.; Fox, Nick C.; Förster, Stefan; Chen, Kewei; Reiman, Eric M.; Xiong, Chengjie; Marcus, Daniel S.; Weiner, Michael W.; Morris, John C.; Bateman, Randall J.; Benzinger, Tammie L. S.

    2016-01-01

    Amyloid imaging plays an important role in the research and diagnosis of dementing disorders. Substantial variation in quantitative methods to measure brain amyloid burden exists in the field. The aim of this work is to investigate the impact of methodological variations to the quantification of amyloid burden using data from the Dominantly Inherited Alzheimer’s Network (DIAN), an autosomal dominant Alzheimer’s disease population. Cross-sectional and longitudinal [11C]-Pittsburgh Compound B (PiB) PET imaging data from the DIAN study were analyzed. Four candidate reference regions were investigated for estimation of brain amyloid burden. A regional spread function based technique was also investigated for the correction of partial volume effects. Cerebellar cortex, brain-stem, and white matter regions all had stable tracer retention during the course of disease. Partial volume correction consistently improves sensitivity to group differences and longitudinal changes over time. White matter referencing improved statistical power in the detecting longitudinal changes in relative tracer retention; however, the reason for this improvement is unclear and requires further investigation. Full dynamic acquisition and kinetic modeling improved statistical power although it may add cost and time. Several technical variations to amyloid burden quantification were examined in this study. Partial volume correction emerged as the strategy that most consistently improved statistical power for the detection of both longitudinal changes and across-group differences. For the autosomal dominant Alzheimer’s disease population with PiB imaging, utilizing brainstem as a reference region with partial volume correction may be optimal for current interventional trials. Further investigation of technical issues in quantitative amyloid imaging in different study populations using different amyloid imaging tracers is warranted. PMID:27010959

  4. Diagnostic accuracy of stress perfusion CMR in comparison with quantitative coronary angiography: fully quantitative, semiquantitative, and qualitative assessment.

    PubMed

    Mordini, Federico E; Haddad, Tariq; Hsu, Li-Yueh; Kellman, Peter; Lowrey, Tracy B; Aletras, Anthony H; Bandettini, W Patricia; Arai, Andrew E

    2014-01-01

    This study's primary objective was to determine the sensitivity, specificity, and accuracy of fully quantitative stress perfusion cardiac magnetic resonance (CMR) versus a reference standard of quantitative coronary angiography. We hypothesized that fully quantitative analysis of stress perfusion CMR would have high diagnostic accuracy for identifying significant coronary artery stenosis and exceed the accuracy of semiquantitative measures of perfusion and qualitative interpretation. Relatively few studies apply fully quantitative CMR perfusion measures to patients with coronary disease and comparisons to semiquantitative and qualitative methods are limited. Dual bolus dipyridamole stress perfusion CMR exams were performed in 67 patients with clinical indications for assessment of myocardial ischemia. Stress perfusion images alone were analyzed with a fully quantitative perfusion (QP) method and 3 semiquantitative methods including contrast enhancement ratio, upslope index, and upslope integral. Comprehensive exams (cine imaging, stress/rest perfusion, late gadolinium enhancement) were analyzed qualitatively with 2 methods including the Duke algorithm and standard clinical interpretation. A 70% or greater stenosis by quantitative coronary angiography was considered abnormal. The optimum diagnostic threshold for QP determined by receiver-operating characteristic curve occurred when endocardial flow decreased to <50% of mean epicardial flow, which yielded a sensitivity of 87% and specificity of 93%. The area under the curve for QP was 92%, which was superior to semiquantitative methods: contrast enhancement ratio: 78%; upslope index: 82%; and upslope integral: 75% (p = 0.011, p = 0.019, p = 0.004 vs. QP, respectively). Area under the curve for QP was also superior to qualitative methods: Duke algorithm: 70%; and clinical interpretation: 78% (p < 0.001 and p < 0.001 vs. QP, respectively). Fully quantitative stress perfusion CMR has high diagnostic accuracy for

  5. Quantitative risk assessment system (QRAS)

    NASA Technical Reports Server (NTRS)

    Tan, Zhibin (Inventor); Mosleh, Ali (Inventor); Weinstock, Robert M (Inventor); Smidts, Carol S (Inventor); Chang, Yung-Hsien (Inventor); Groen, Francisco J (Inventor); Swaminathan, Sankaran (Inventor)

    2001-01-01

    A quantitative risk assessment system (QRAS) builds a risk model of a system for which risk of failure is being assessed, then analyzes the risk of the system corresponding to the risk model. The QRAS performs sensitivity analysis of the risk model by altering fundamental components and quantifications built into the risk model, then re-analyzes the risk of the system using the modifications. More particularly, the risk model is built by building a hierarchy, creating a mission timeline, quantifying failure modes, and building/editing event sequence diagrams. Multiplicities, dependencies, and redundancies of the system are included in the risk model. For analysis runs, a fixed baseline is first constructed and stored. This baseline contains the lowest level scenarios, preserved in event tree structure. The analysis runs, at any level of the hierarchy and below, access this baseline for risk quantitative computation as well as ranking of particular risks. A standalone Tool Box capability exists, allowing the user to store application programs within QRAS.

  6. The simultaneous quantitation of ten amino acids in soil extracts by mass fragmentography

    NASA Technical Reports Server (NTRS)

    Pereira, W. E.; Hoyano, Y.; Reynolds, W. E.; Summons, R. E.; Duffield, A. M.

    1972-01-01

    A specific and sensitive method for the identification and simultaneous quantitation by mass fragmentography of ten of the amino acids present in soil was developed. The technique uses a computer driven quadrupole mass spectrometer and a commercial preparation of deuterated amino acids is used as internal standards for purposes of quantitation. The results obtained are comparable with those from an amino acid analyzer. In the quadrupole mass spectrometer-computer system up to 25 pre-selected ions may be monitored sequentially. This allows a maximum of 12 different amino acids (one specific ion in each of the undeuterated and deuterated amino acid spectra) to be quantitated. The method is relatively rapid (analysis time of approximately one hour) and is capable of the quantitation of nanogram quantities of amino acids.

  7. Bridging the Qualitative/Quantitative Software Divide

    PubMed Central

    Annechino, Rachelle; Antin, Tamar M. J.; Lee, Juliet P.

    2011-01-01

    To compare and combine qualitative and quantitative data collected from respondents in a mixed methods study, the research team developed a relational database to merge survey responses stored and analyzed in SPSS and semistructured interview responses stored and analyzed in the qualitative software package ATLAS.ti. The process of developing the database, as well as practical considerations for researchers who may wish to use similar methods, are explored. PMID:22003318

  8. Towards quantitative magnetic particle imaging: A comparison with magnetic particle spectroscopy

    NASA Astrophysics Data System (ADS)

    Paysen, Hendrik; Wells, James; Kosch, Olaf; Steinhoff, Uwe; Trahms, Lutz; Schaeffter, Tobias; Wiekhorst, Frank

    2018-05-01

    Magnetic Particle Imaging (MPI) is a quantitative imaging modality with promising features for several biomedical applications. Here, we study quantitatively the raw data obtained during MPI measurements. We present a method for the calibration of the MPI scanner output using measurements from a magnetic particle spectrometer (MPS) to yield data in units of magnetic moments. The calibration technique is validated in a simplified MPI mode with a 1D excitation field. Using the calibrated results from MPS and MPI, we determine and compare the detection limits for each system. The detection limits were found to be 5.10-12 Am2 for MPS and 3.6.10-10 Am2 for MPI. Finally, the quantitative information contained in a standard MPI measurement with a 3D excitation is analyzed and compared to the previous results, showing a decrease in signal amplitudes of the odd harmonics related to the case of 1D excitation. We propose physical explanations for all acquired results; and discuss the possible benefits for the improvement of MPI technology.

  9. Examining the Role of Numeracy in College STEM Courses: Results from the Quantitative Reasoning for College Science (QuaRCS) Assessment Instrument

    NASA Astrophysics Data System (ADS)

    Follette, Katherine B.; McCarthy, Donald W.; Dokter, Erin F.; Buxner, Sanlyn; Prather, Edward E.

    2016-01-01

    Is quantitative literacy a prerequisite for science literacy? Can students become discerning voters, savvy consumers and educated citizens without it? Should college science courses for nonmajors be focused on "science appreciation", or should they engage students in the messy quantitative realities of modern science? We will present results from the recently developed and validated Quantitative Reasoning for College Science (QuaRCS) Assessment, which probes both quantitative reasoning skills and attitudes toward mathematics. Based on data from nearly two thousand students enrolled in nineteen general education science courses, we show that students in these courses did not demonstrate significant skill or attitude improvements over the course of a single semester, but find encouraging evidence for longer term trends.

  10. Effects of two types of medical contrast media on routine chemistry results by three automated chemistry analyzers.

    PubMed

    Park, Yu Jin; Rim, John Hoon; Yim, Jisook; Lee, Sang-Guk; Kim, Jeong-Ho

    2017-08-01

    The use of iodinated contrast media has grown in popularity in the past two decades, but relatively little attention has been paid to the possible interferential effects of contrast media on laboratory test results. Herein, we investigate medical contrast media interference with routine chemistry results obtained by three automated chemistry analyzers. Ten levels of pooled serum were used in the study. Two types of medical contrast media [Iopamiro (iopamidol) and Omnipaque (iohexol)] were evaluated. To evaluate the dose-dependent effects of the contrast media, iopamidol and iohexol were spiked separately into aliquots of serum for final concentrations of 1.8%, 3.6%, 5.5%, 7.3%, and 9.1%. The 28 analytes included in the routine chemistry panel were measured by using Hitachi 7600, AU5800, and Cobas c702 analyzers. We calculated the delta percentage difference (DPD) between the samples and the control, and examined dose-dependent trends. When the mean DPD values were compared with the reference cut-off criteria, the only uniformly interferential effect observed for all analyzers was in total protein with iopamidol. Two additional analytes that showed trends toward interferential effects only in few analyzers and exceeded the limits of the allowable error were the serum iron and the total CO 2 . The other combinations of analyzer and contrast showed no consistent dose-dependent propensity for change in any analyte level. Our study suggests that many of the analytes included in routine chemistry results, except total protein and serum iron, are not significantly affected by iopamidol and iohexol. These results suggest that it would be beneficial to apply a flexible medical evaluation process for patients requiring both laboratory tests and imaging studies, minimizing the need for strict regulations for sequential tests. Copyright © 2017 The Canadian Society of Clinical Chemists. Published by Elsevier Inc. All rights reserved.

  11. [Study on ethnic medicine quantitative reference herb,Tibetan medicine fruits of Capsicum frutescens as a case].

    PubMed

    Zan, Ke; Cui, Gan; Guo, Li-Nong; Ma, Shuang-Cheng; Zheng, Jian

    2018-05-01

    High price and difficult to get of reference substance have become obstacles to HPLC assay of ethnic medicine. A new method based on quantitative reference herb (QRH) was proposed. Specific chromatograms in fruits of Capsicum frutescens were employed to determine peak positions, and HPLC quantitative reference herb was prepared from fruits of C. frutescens. The content of capsaicin and dihydrocapsaicin in the quantitative control herb was determined by HPLC. Eleven batches of fruits of C. frutescens were analyzed with quantitative reference herb and reference substance respectively. The results showed no difference. The present method is feasible for quality control of ethnic medicines and quantitative reference herb is suitable to replace reference substances in assay. Copyright© by the Chinese Pharmaceutical Association.

  12. Reproducibility and quantitation of amplicon sequencing-based detection

    PubMed Central

    Zhou, Jizhong; Wu, Liyou; Deng, Ye; Zhi, Xiaoyang; Jiang, Yi-Huei; Tu, Qichao; Xie, Jianping; Van Nostrand, Joy D; He, Zhili; Yang, Yunfeng

    2011-01-01

    To determine the reproducibility and quantitation of the amplicon sequencing-based detection approach for analyzing microbial community structure, a total of 24 microbial communities from a long-term global change experimental site were examined. Genomic DNA obtained from each community was used to amplify 16S rRNA genes with two or three barcode tags as technical replicates in the presence of a small quantity (0.1% wt/wt) of genomic DNA from Shewanella oneidensis MR-1 as the control. The technical reproducibility of the amplicon sequencing-based detection approach is quite low, with an average operational taxonomic unit (OTU) overlap of 17.2%±2.3% between two technical replicates, and 8.2%±2.3% among three technical replicates, which is most likely due to problems associated with random sampling processes. Such variations in technical replicates could have substantial effects on estimating β-diversity but less on α-diversity. A high variation was also observed in the control across different samples (for example, 66.7-fold for the forward primer), suggesting that the amplicon sequencing-based detection approach could not be quantitative. In addition, various strategies were examined to improve the comparability of amplicon sequencing data, such as increasing biological replicates, and removing singleton sequences and less-representative OTUs across biological replicates. Finally, as expected, various statistical analyses with preprocessed experimental data revealed clear differences in the composition and structure of microbial communities between warming and non-warming, or between clipping and non-clipping. Taken together, these results suggest that amplicon sequencing-based detection is useful in analyzing microbial community structure even though it is not reproducible and quantitative. However, great caution should be taken in experimental design and data interpretation when the amplicon sequencing-based detection approach is used for quantitative

  13. [Determination of acidity and vitamin C in apples using portable NIR analyzer].

    PubMed

    Yang, Fan; Li, Ya-Ting; Gu, Xuan; Ma, Jiang; Fan, Xing; Wang, Xiao-Xuan; Zhang, Zhuo-Yong

    2011-09-01

    Near infrared (NIR) spectroscopy technology based on a portable NIR analyzer, combined with kernel Isomap algorithm and generalized regression neural network (GRNN) has been applied to establishing quantitative models for prediction of acidity and vitamin C in six kinds of apple samples. The obtained results demonstrated that the fitting and the predictive accuracy of the models with kernel Isomap algorithm were satisfactory. The correlation between actual and predicted values of calibration samples (R(c)) obtained by the acidity model was 0.999 4, and for prediction samples (R(p)) was 0.979 9. The root mean square error of prediction set (RMSEP) was 0.055 8. For the vitamin C model, R(c) was 0.989 1, R(p) was 0.927 2, and RMSEP was 4.043 1. Results proved that the portable NIR analyzer can be a feasible tool for the determination of acidity and vitamin C in apples.

  14. CTG Analyzer: A graphical user interface for cardiotocography.

    PubMed

    Sbrollini, Agnese; Agostinelli, Angela; Burattini, Luca; Morettini, Micaela; Di Nardo, Francesco; Fioretti, Sandro; Burattini, Laura

    2017-07-01

    Cardiotocography (CTG) is the most commonly used test for establishing the good health of the fetus during pregnancy and labor. CTG consists in the recording of fetal heart rate (FHR; bpm) and maternal uterine contractions (UC; mmHg). FHR is characterized by baseline, baseline variability, tachycardia, bradycardia, acceleration and decelerations. Instead, UC signal is characterized by presence of contractions and contractions period. Such parameters are usually evaluated by visual inspection. However, visual analysis of CTG recordings has a well-demonstrated poor reproducibility, due to the complexity of physiological phenomena affecting fetal heart rhythm and being related to clinician's experience. Computerized tools in support of clinicians represents a possible solution for improving correctness in CTG interpretation. This paper proposes CTG Analyzer as a graphical tool for automatic and objective analysis of CTG tracings. CTG Analyzer was developed under MATLAB®; it is a very intuitive and user friendly graphical user interface. FHR time series and UC signal are represented one under the other, on a grid with reference lines, as usually done for CTG reports printed on paper. Colors help identification of FHR and UC features. Automatic analysis is based on some unchangeable features definitions provided by the FIGO guidelines, and other arbitrary settings whose default values can be changed by the user. Eventually, CTG Analyzer provides a report file listing all the quantitative results of the analysis. Thus, CTG Analyzer represents a potentially useful graphical tool for automatic and objective analysis of CTG tracings.

  15. Utility of DWI with quantitative ADC values in ovarian tumors: a meta-analysis of diagnostic test performance.

    PubMed

    Pi, Shan; Cao, Rong; Qiang, Jin Wei; Guo, Yan Hui

    2018-01-01

    Background Diffusion-weighted imaging (DWI) and quantitative apparent diffusion coefficient (ADC) values are widely used in the differential diagnosis of ovarian tumors. Purpose To assess the diagnostic performance of quantitative ADC values in ovarian tumors. Material and Methods PubMed, Embase, the Cochrane Library, and local databases were searched for studies assessing ovarian tumors using quantitative ADC values. We quantitatively analyzed the diagnostic performances for two clinical problems: benign vs. malignant tumors and borderline vs. malignant tumors. We evaluated diagnostic performances by the pooled sensitivity and specificity values and by summary receiver operating characteristic (SROC) curves. Subgroup analyses were used to analyze study heterogeneity. Results From the 742 studies identified in the search results, 16 studies met our inclusion criteria. A total of ten studies evaluated malignant vs. benign ovarian tumors and six studies assessed malignant vs. borderline ovarian tumors. Regarding the diagnostic accuracy of quantitative ADC values for distinguishing between malignant and benign ovarian tumors, the pooled sensitivity and specificity values were 0.91 and 0.91, respectively. The area under the SROC curve (AUC) was 0.96. For differentiating borderline from malignant tumors, the pooled sensitivity and specificity values were 0.89 and 0.79, and the AUC was 0.91. The methodological quality of the included studies was moderate. Conclusion Quantitative ADC values could serve as useful preoperative markers for predicting the nature of ovarian tumors. Nevertheless, prospective trials focused on standardized imaging parameters are needed to evaluate the clinical value of quantitative ADC values in ovarian tumors.

  16. Theoretical framework for analyzing structural compliance properties of proteins.

    PubMed

    Arikawa, Keisuke

    2018-01-01

    We propose methods for directly analyzing structural compliance (SC) properties of elastic network models of proteins, and we also propose methods for extracting information about motion properties from the SC properties. The analysis of SC properties involves describing the relationships between the applied forces and the deformations. When decomposing the motion according to the magnitude of SC (SC mode decomposition), we can obtain information about the motion properties under the assumption that the lower SC mode motions or the softer motions occur easily. For practical applications, the methods are formulated in a general form. The parts where forces are applied and those where deformations are evaluated are separated from each other for enabling the analyses of allosteric interactions between the specified parts. The parts are specified not only by the points but also by the groups of points (the groups are treated as flexible bodies). In addition, we propose methods for quantitatively evaluating the properties based on the screw theory and the considerations of the algebraic structures of the basic equations expressing the SC properties. These methods enable quantitative discussions about the relationships between the SC mode motions and the motions estimated from two different conformations; they also help identify the key parts that play important roles for the motions by comparing the SC properties with those of partially constrained models. As application examples, lactoferrin and ATCase are analyzed. The results show that we can understand their motion properties through their lower SC mode motions or the softer motions.

  17. Theoretical framework for analyzing structural compliance properties of proteins

    PubMed Central

    2018-01-01

    We propose methods for directly analyzing structural compliance (SC) properties of elastic network models of proteins, and we also propose methods for extracting information about motion properties from the SC properties. The analysis of SC properties involves describing the relationships between the applied forces and the deformations. When decomposing the motion according to the magnitude of SC (SC mode decomposition), we can obtain information about the motion properties under the assumption that the lower SC mode motions or the softer motions occur easily. For practical applications, the methods are formulated in a general form. The parts where forces are applied and those where deformations are evaluated are separated from each other for enabling the analyses of allosteric interactions between the specified parts. The parts are specified not only by the points but also by the groups of points (the groups are treated as flexible bodies). In addition, we propose methods for quantitatively evaluating the properties based on the screw theory and the considerations of the algebraic structures of the basic equations expressing the SC properties. These methods enable quantitative discussions about the relationships between the SC mode motions and the motions estimated from two different conformations; they also help identify the key parts that play important roles for the motions by comparing the SC properties with those of partially constrained models. As application examples, lactoferrin and ATCase are analyzed. The results show that we can understand their motion properties through their lower SC mode motions or the softer motions. PMID:29607281

  18. Optical analyzer

    DOEpatents

    Hansen, Anthony D.

    1989-02-07

    An optical analyzer (10) wherein a sample (19) of particulate matter, and particularly of organic matter, which has been collected on a quartz fiber filter (20) is placed in a combustion tube (11), and light from a light source (14) is passed through the sample (19). The temperature of the sample (19) is raised at a controlled rate and in a controlled atmosphere. The magnitude of the transmission of light through the sample (19) is detected (18) as the temperature is raised. A data processor (23), differentiator (28) and a two pen recorder (24) provide a chart of the optical transmission versus temperature and the rate of change of optical transmission versus temperature signatures (T and D) of the sample (19). These signatures provide information as to physical and chemical processes and a variety of quantitative and qualitative information about the sample (19). Additional information is obtained by repeating the run in different atmospheres and/or different rates of heating with other samples of the same particulate material collected on other filters.

  19. Optical analyzer

    DOEpatents

    Hansen, Anthony D.

    1989-01-01

    An optical analyzer (10) wherein a sample (19) of particulate matter, and particularly of organic matter, which has been collected on a quartz fiber filter (20) is placed in a combustion tube (11), and light from a light source (14) is passed through the sample (19). The temperature of the sample (19) is raised at a controlled rate and in a controlled atmosphere. The magnitude of the transmission of light through the sample (19) is detected (18) as the temperature is raised. A data processor (23), differentiator (28) and a two pen recorder (24) provide a chart of the optical transmission versus temperature and the rate of change of optical transmission versus temperature signatures (T and D) of the sample (19). These signatures provide information as to physical and chemical processes and a variety of quantitative and qualitative information about the sample (19). Additional information is obtained by repeating the run in different atmospheres and/or different rates of heating with other samples of the same particulate material collected on other filters.

  20. The Development of Mathematical Knowledge for Teaching for Quantitative Reasoning Using Video-Based Instruction

    NASA Astrophysics Data System (ADS)

    Walters, Charles David

    Quantitative reasoning (P. W. Thompson, 1990, 1994) is a powerful mathematical tool that enables students to engage in rich problem solving across the curriculum. One way to support students' quantitative reasoning is to develop prospective secondary teachers' (PSTs) mathematical knowledge for teaching (MKT; Ball, Thames, & Phelps, 2008) related to quantitative reasoning. However, this may prove challenging, as prior to entering the classroom, PSTs often have few opportunities to develop MKT by examining and reflecting on students' thinking. Videos offer one avenue through which such opportunities are possible. In this study, I report on the design of a mini-course for PSTs that featured a series of videos created as part of a proof-of-concept NSF-funded project. These MathTalk videos highlight the ways in which the quantitative reasoning of two high school students developed over time. Using a mixed approach to grounded theory, I analyzed pre- and postinterviews using an extant coding scheme based on the Silverman and Thompson (2008) framework for the development of MKT. This analysis revealed a shift in participants' affect as well as three distinct shifts in their MKT around quantitative reasoning with distances, including shifts in: (a) quantitative reasoning; (b) point of view (decentering); and (c) orientation toward problem solving. Using the four-part focusing framework (Lobato, Hohensee, & Rhodehamel, 2013), I analyzed classroom data to account for how participants' noticing was linked with the shifts in MKT. Notably, their increased noticing of aspects of MKT around quantitative reasoning with distances, which features prominently in the MathTalk videos, seemed to contribute to the emergence of the shifts in MKT. Results from this study link elements of the learning environment to the development of specific facets of MKT around quantitative reasoning with distances. These connections suggest that vicarious experiences with two students' quantitative

  1. Preliminary research on eddy current bobbin quantitative test for heat exchange tube in nuclear power plant

    NASA Astrophysics Data System (ADS)

    Qi, Pan; Shao, Wenbin; Liao, Shusheng

    2016-02-01

    For quantitative defects detection research on heat transfer tube in nuclear power plants (NPP), two parts of work are carried out based on the crack as the main research objects. (1) Production optimization of calibration tube. Firstly, ASME, RSEM and homemade crack calibration tubes are applied to quantitatively analyze the defects depth on other designed crack test tubes, and then the judgment with quantitative results under crack calibration tube with more accuracy is given. Base on that, weight analysis of influence factors for crack depth quantitative test such as crack orientation, length, volume and so on can be undertaken, which will optimize manufacture technology of calibration tubes. (2) Quantitative optimization of crack depth. Neural network model with multi-calibration curve adopted to optimize natural crack test depth generated in in-service tubes shows preliminary ability to improve quantitative accuracy.

  2. Preliminary Results of Acoustic Radiation Force Impulse Imaging by Combined Qualitative and Quantitative Analyses for Evaluation of Breast Lesions.

    PubMed

    Wang, Lin; Wan, Cai-Feng; Du, Jing; Li, Feng-Hua

    2018-04-15

    The purpose of this study was to evaluate the application of a new elastographic technique, acoustic radiation force impulse (ARFI) imaging, and its diagnostic performance for characterizing breast lesions. One hundred consecutive female patients with 126 breast lesions were enrolled in our study. After routine breast ultrasound examinations, the patients underwent ARFI elasticity imaging. Virtual Touch tissue imaging (VTI) and Virtual Touch tissue quantification (Siemens Medical Solutions, Mountain View, CA) were used to qualitatively and quantitatively analyze the elasticity and hardness of tumors. A receiver operating characteristic curve analysis was performed to evaluate the diagnostic performance of ARFI for discrimination between benign and malignant breast lesions. Pathologic analysis revealed 40 lesions in the malignant group and 86 lesions in the benign group. Different VTI patterns were observed in benign and malignant breast lesions. Eighty lesions (93.0%) of benign group had pattern 1, 2, or 3, whereas all pattern 4b lesions (n = 20 [50.0%]) were malignant. Regarding the quantitative analysis, the mean VTI-to-B-mode area ratio, internal shear wave velocity, and marginal shear wave velocity of benign lesions were statistically significantly lower than those of malignant lesions (all P < .001). The cutoff point for a scoring system constructed to evaluate the diagnostic performance of ARFI was estimated to be between 3 and 4 points for malignancy, with sensitivity of 77.5%, specificity of 96.5%, accuracy of 90.5%, and an area under the curve of 0.933. The application of ARFI technology has shown promising results by noninvasively providing substantial complementary information and could potentially serve as an effective diagnostic tool for differentiation between benign and malignant breast lesions. © 2018 by the American Institute of Ultrasound in Medicine.

  3. High resolution energy analyzer for broad ion beam characterization

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Kanarov, V.; Hayes, A.; Yevtukhov, R.

    2008-09-15

    analyzed IEDF shape or mean energy value. This led us to conclude that the optimized analyzer construction provides an energy resolution considerably narrower than the investigated ion beam energy spectrum full width at half maximum, and the derived energy spectrum is an objective and accurate representation of the analyzed broad ion beam energy distribution characteristics. A quantitative study of the focusing voltage and retarding grid field effects based on the experimental data and modeling results have supported this conclusion.« less

  4. Teaching Fundamental Skills in Microsoft Excel to First-Year Students in Quantitative Analysis

    ERIC Educational Resources Information Center

    Rubin, Samuel J.; Abrams, Binyomin

    2015-01-01

    Despite their technological savvy, most students entering university lack the necessary computer skills to succeed in a quantitative analysis course, in which they are often expected to input, analyze, and plot results of experiments without any previous formal education in Microsoft Excel or similar programs. This lack of formal education results…

  5. Interprofessional collaboration from nurses and physicians – A triangulation of quantitative and qualitative data

    PubMed

    Schärli, Marianne; Müller, Rita; Martin, Jacqueline S; Spichiger, Elisabeth; Spirig, Rebecca

    2017-01-01

    Background: Interprofessional collaboration between nurses and physicians is a recurrent challenge in daily clinical practice. To ameliorate the situation, quantitative or qualitative studies are conducted. However, the results of these studies have often been limited by the methods chosen. Aim: To describe the synthesis of interprofessional collaboration from the nursing perspective by triangulating quantitative and qualitative data. Method: Data triangulation was performed as a sub-project of the interprofessional Sinergia DRG Research program. Initially, quantitative and qualitative data were analyzed separately in a mixed methods design. By means of triangulation a „meta-matrix“ resulted in a four-step process. Results: The „meta-matrix“ displays all relevant quantitative and qualitative results as well as their interrelations on one page. Relevance, influencing factors as well as consequences of interprofessional collaboration for patients, relatives and systems become visible. Conclusion: For the first time, the interprofessional collaboration from the nursing perspective at five Swiss hospitals is shown in a „meta-matrix“. The consequences of insufficient collaboration between nurses and physicians are considerable. This is why it’s necessary to invest in interprofessional concepts. In the „meta-matrix“ the factors which influence the interprofessional collaboration positively or negatively are visible.

  6. Ultrasound introscopic image quantitative characteristics for medical diagnosis

    NASA Astrophysics Data System (ADS)

    Novoselets, Mikhail K.; Sarkisov, Sergey S.; Gridko, Alexander N.; Tcheban, Anatoliy K.

    1993-09-01

    The results on computer aided extraction of quantitative characteristics (QC) of ultrasound introscopic images for medical diagnosis are presented. Thyroid gland (TG) images of Chernobil Accident sufferers are considered. It is shown that TG diseases can be associated with some values of selected QCs of random echo distribution in the image. The possibility of these QCs usage for TG diseases recognition in accordance with calculated values is analyzed. The role of speckle noise elimination in the solution of the problem on TG diagnosis is considered too.

  7. Quantitative proteomic analysis of intact plastids.

    PubMed

    Shiraya, Takeshi; Kaneko, Kentaro; Mitsui, Toshiaki

    2014-01-01

    Plastids are specialized cell organelles in plant cells that are differentiated into various forms including chloroplasts, chromoplasts, and amyloplasts, and fulfill important functions in maintaining the overall cell metabolism and sensing environmental factors such as sunlight. It is therefore important to grasp the mechanisms of differentiation and functional changes of plastids in order to enhance the understanding of vegetality. In this chapter, details of a method for the extraction of intact plastids that makes analysis possible while maintaining the plastid functions are provided; in addition, a quantitative shotgun method for analyzing the composition and changes in the content of proteins in plastids as a result of environmental impacts is described.

  8. Analysis Results for Lunar Soil Simulant Using a Portable X-Ray Fluorescence Analyzer

    NASA Technical Reports Server (NTRS)

    Boothe, R. E.

    2006-01-01

    Lunar soil will potentially be used for oxygen generation, water generation, and as filler for building blocks during habitation missions on the Moon. NASA s in situ fabrication and repair program is evaluating portable technologies that can assess the chemistry of lunar soil and lunar soil simulants. This Technical Memorandum summarizes the results of the JSC 1 lunar soil simulant analysis using the TRACeR III IV handheld x-ray fluorescence analyzer, manufactured by KeyMaster Technologies, Inc. The focus of the evaluation was to determine how well the current instrument configuration would detect and quantify the components of JSC-1.

  9. Visualizing the Critique: Integrating Quantitative Reasoning with the Design Process

    ERIC Educational Resources Information Center

    Weinstein, Kathryn

    2017-01-01

    In the age of "Big Data," information is often quantitative in nature. The ability to analyze information through the sifting of data has been identified as a core competency for success in navigating daily life and participation in the contemporary workforce. This skill, known as Quantitative Reasoning (QR), is characterized by the…

  10. Quantitative T2-Mapping and T2⁎-Mapping Evaluation of Changes in Cartilage Matrix after Acute Anterior Cruciate Ligament Rupture and the Correlation between the Results of Both Methods

    PubMed Central

    Tao, Hongyue; Qiao, Yang; Hu, Yiwen; Xie, Yuxue; Lu, Rong; Yan, Xu

    2018-01-01

    Objectives To quantitatively assess changes in cartilage matrix after acute anterior cruciate ligament (ACL) rupture using T2- and T2⁎-mapping and analyze the correlation between the results of both methods. Methods Twenty-three patients and 23 healthy controls were enrolled and underwent quantitative MRI examination. The knee cartilage was segmented into six compartments, including lateral femur (LF), lateral tibia (LT), medial femur (MF), medial tibia (MT), trochlea (Tr), and patella (Pa). T2 and T2⁎ values were measured in full-thickness as well as superficial and deep layers of each cartilage compartment. Differences of T2 and T2⁎ values between patients and controls were compared using unpaired Student's t-test, and the correlation between their reciprocals was analyzed using Pearson's correlation coefficient. Results ACL-ruptured patients showed higher T2 and T2⁎ values in full-thickness and superficial layers of medial and lateral tibiofemoral joint. Meanwhile, patients exhibited higher T2⁎ values in deep layers of lateral tibiofemoral joint. The elevated percentages of T2 and T2⁎ value in superficial LT were most significant (20.738%, 17.525%). The reciprocal of T2⁎ value was correlated with that of T2 value (r = 0.886, P < 0.001). Conclusion The early degeneration could occur in various knee cartilage compartments after acute ACL rupture, especially in the superficial layer of LT. T2⁎-mapping might be more sensitive in detecting deep layer of cartilage than T2-mapping. PMID:29888279

  11. [Quantitative mineralogical analyzes of kidney stones and diagnosing metabolic disorders in female patients with calcium oxalate urolithiasis].

    PubMed

    Kustov, A V; Moryganov, M A; Strel'nikov, A I; Zhuravleva, N I; Airapetyan, A O

    2016-02-01

    To conduct a complex examination of female patients with calcium oxalate urolithiasis to detect metabolic disorders, leading to stone formation. The study was carried out using complex physical and chemical methods, including quantitative X-ray phase analysis of urinary stones, pH measurement, volumetry, urine and blood spectrophotometry. Quantitative mineralogical composition of stones, daily urine pH profile, daily urinary excretion of ions of calcium, magnesium, oxalate, phosphate, citrate and uric acid were determined in 20 female patients with calcium oxalate stones. We have shown that most of the stones comprised calcium oxalate monohydrate or mixtures of calcium oxalate dihydrate and hydroxyapatite. Among the identified abnormalities, the most frequent were hypocitraturia and hypercalciuria - 90 and 45%, respectively. Our findings revealed that the daily secretion of citrate and oxalate in patients older than 50 years was significantly lower than in younger patients. In conclusion, daily urinary citrate excretion should be measured in female patients with calcium oxalate stones. This is necessary both to determine the causes of stone formation, and to monitor the effectiveness of citrate therapy.

  12. Quantitative analysis of rib movement based on dynamic chest bone images: preliminary results

    NASA Astrophysics Data System (ADS)

    Tanaka, R.; Sanada, S.; Oda, M.; Mitsutaka, M.; Suzuki, K.; Sakuta, K.; Kawashima, H.

    2014-03-01

    Rib movement during respiration is one of the diagnostic criteria in pulmonary impairments. In general, the rib movement is assessed in fluoroscopy. However, the shadows of lung vessels and bronchi overlapping ribs prevent accurate quantitative analysis of rib movement. Recently, an image-processing technique for separating bones from soft tissue in static chest radiographs, called "bone suppression technique", has been developed. Our purpose in this study was to evaluate the usefulness of dynamic bone images created by the bone suppression technique in quantitative analysis of rib movement. Dynamic chest radiographs of 10 patients were obtained using a dynamic flat-panel detector (FPD). Bone suppression technique based on a massive-training artificial neural network (MTANN) was applied to the dynamic chest images to create bone images. Velocity vectors were measured in local areas on the dynamic bone images, which formed a map. The velocity maps obtained with bone and original images for scoliosis and normal cases were compared to assess the advantages of bone images. With dynamic bone images, we were able to quantify and distinguish movements of ribs from those of other lung structures accurately. Limited rib movements of scoliosis patients appeared as reduced rib velocity vectors. Vector maps in all normal cases exhibited left-right symmetric distributions, whereas those in abnormal cases showed nonuniform distributions. In conclusion, dynamic bone images were useful for accurate quantitative analysis of rib movements: Limited rib movements were indicated as a reduction of rib movement and left-right asymmetric distribution on vector maps. Thus, dynamic bone images can be a new diagnostic tool for quantitative analysis of rib movements without additional radiation dose.

  13. Qualitative and Quantitative Detection of Botulinum Neurotoxins from Complex Matrices: Results of the First International Proficiency Test

    PubMed Central

    Worbs, Sylvia; Fiebig, Uwe; Zeleny, Reinhard; Schimmel, Heinz; Rummel, Andreas; Luginbühl, Werner; Dorner, Brigitte G.

    2015-01-01

    In the framework of the EU project EQuATox, a first international proficiency test (PT) on the detection and quantification of botulinum neurotoxins (BoNT) was conducted. Sample materials included BoNT serotypes A, B and E spiked into buffer, milk, meat extract and serum. Different methods were applied by the participants combining different principles of detection, identification and quantification. Based on qualitative assays, 95% of all results reported were correct. Successful strategies for BoNT detection were based on a combination of complementary immunological, MS-based and functional methods or on suitable functional in vivo/in vitro approaches (mouse bioassay, hemidiaphragm assay and Endopep-MS assay). Quantification of BoNT/A, BoNT/B and BoNT/E was performed by 48% of participating laboratories. It turned out that precise quantification of BoNT was difficult, resulting in a substantial scatter of quantitative data. This was especially true for results obtained by the mouse bioassay which is currently considered as “gold standard” for BoNT detection. The results clearly demonstrate the urgent need for certified BoNT reference materials and the development of methods replacing animal testing. In this context, the BoNT PT provided the valuable information that both the Endopep-MS assay and the hemidiaphragm assay delivered quantitative results superior to the mouse bioassay. PMID:26703724

  14. Rigour in quantitative research.

    PubMed

    Claydon, Leica Sarah

    2015-07-22

    This article which forms part of the research series addresses scientific rigour in quantitative research. It explores the basis and use of quantitative research and the nature of scientific rigour. It examines how the reader may determine whether quantitative research results are accurate, the questions that should be asked to determine accuracy and the checklists that may be used in this process. Quantitative research has advantages in nursing, since it can provide numerical data to help answer questions encountered in everyday practice.

  15. Light scattering application for quantitative estimation of apoptosis

    NASA Astrophysics Data System (ADS)

    Bilyy, Rostyslav O.; Stoika, Rostyslav S.; Getman, Vasyl B.; Bilyi, Olexander I.

    2004-05-01

    Estimation of cell proliferation and apoptosis are in focus of instrumental methods used in modern biomedical sciences. Present study concerns monitoring of functional state of cells, specifically the development of their programmed death or apoptosis. The available methods for such purpose are either very expensive, or require time-consuming operations. Their specificity and sensitivity are frequently not sufficient for making conclusions which could be used in diagnostics or treatment monitoring. We propose a novel method for apoptosis measurement based on quantitative determination of cellular functional state taking into account their physical characteristics. This method uses the patented device -- laser microparticle analyser PRM-6 -- for analyzing light scattering by the microparticles, including cells. The method gives an opportunity for quick, quantitative, simple (without complicated preliminary cell processing) and relatively cheap measurement of apoptosis in cellular population. The elaborated method was used for studying apoptosis expression in murine leukemia cells of L1210 line and human lymphoblastic leukemia cells of K562 line. The results obtained by the proposed method permitted measuring cell number in tested sample, detecting and quantitative characterization of functional state of cells, particularly measuring the ratio of the apoptotic cells in suspension.

  16. Analyzing the impacts of global trade and investment on non-communicable diseases and risk factors: a critical review of methodological approaches used in quantitative analyses.

    PubMed

    Cowling, Krycia; Thow, Anne Marie; Pollack Porter, Keshia

    2018-05-24

    A key mechanism through which globalization has impacted health is the liberalization of trade and investment, yet relatively few studies to date have used quantitative methods to investigate the impacts of global trade and investment policies on non-communicable diseases and risk factors. Recent reviews of this literature have found heterogeneity in results and a range of quality across studies, which may be in part attributable to a lack of conceptual clarity and methodological inconsistencies. This study is a critical review of methodological approaches used in the quantitative literature on global trade and investment and diet, tobacco, alcohol, and related health outcomes, with the objective of developing recommendations and providing resources to guide future robust, policy relevant research. A review of reviews, expert review, and reference tracing were employed to identify relevant studies, which were evaluated using a novel quality assessment tool designed for this research. Eight review articles and 34 quantitative studies were identified for inclusion. Important ways to improve this literature were identified and discussed: clearly defining exposures of interest and not conflating trade and investment; exploring mechanisms of broader relationships; increasing the use of individual-level data; ensuring consensus and consistency in key confounding variables; utilizing more sector-specific versus economy-wide trade and investment indicators; testing and adequately adjusting for autocorrelation and endogeneity when using longitudinal data; and presenting results from alternative statistical models and sensitivity analyses. To guide the development of future analyses, recommendations for international data sources for selected trade and investment indicators, as well as key gaps in the literature, are presented. More methodologically rigorous and consistent approaches in future quantitative studies on the impacts of global trade and investment policies on non

  17. Three-dimensional quantitative analysis of adhesive remnants and enamel loss resulting from debonding orthodontic molar tubes

    PubMed Central

    2014-01-01

    Aims Presenting a new method for direct, quantitative analysis of enamel surface. Measurement of adhesive remnants and enamel loss resulting from debonding molar tubes. Material and methods Buccal surfaces of fifteen extracted human molars were directly scanned with an optic blue-light 3D scanner to the nearest 2 μm. After 20 s etching molar tubes were bonded and after 24 h storing in 0.9% saline - debonded. Then 3D scanning was repeated. Superimposition and comparison were proceeded and shape alterations of the entire objects were analyzed using specialized computer software. Residual adhesive heights as well as enamel loss depths have been obtained for the entire buccal surfaces. Residual adhesive volume and enamel loss volume have been calculated for every tooth. Results The maximum height of adhesive remaining on enamel surface was 0.76 mm and the volume on particular teeth ranged from 0.047 mm3 to 4.16 mm3. The median adhesive remnant volume was 0.988 mm3. Mean depths of enamel loss for particular teeth ranged from 0.0076 mm to 0.0416 mm. Highest maximum depth of enamel loss was 0.207 mm. Median volume of enamel loss was 0.104 mm3 and maximum volume was 1.484 mm3. Conclusions Blue-light 3D scanning is able to provide direct precise scans of the enamel surface, which can be superimposed in order to calculate shape alterations. Debonding molar tubes leaves a certain amount of adhesive remnants on the enamel, however the interface fracture pattern varies for particular teeth and areas of enamel loss are present as well. PMID:25208969

  18. Quantitatively differentiating microstructural variations of skeletal muscle tissues by multispectral Mueller matrix imaging

    NASA Astrophysics Data System (ADS)

    Dong, Yang; He, Honghui; He, Chao; Ma, Hui

    2016-10-01

    Polarized light is sensitive to the microstructures of biological tissues and can be used to detect physiological changes. Meanwhile, spectral features of the scattered light can also provide abundant microstructural information of tissues. In this paper, we take the backscattering polarization Mueller matrix images of bovine skeletal muscle tissues during the 24-hour experimental time, and analyze their multispectral behavior using quantitative Mueller matrix parameters. In the processes of rigor mortis and proteolysis of muscle samples, multispectral frequency distribution histograms (FDHs) of the Mueller matrix elements can reveal rich qualitative structural information. In addition, we analyze the temporal variations of the sample using the multispectral Mueller matrix transformation (MMT) parameters. The experimental results indicate that the different stages of rigor mortis and proteolysis for bovine skeletal muscle samples can be judged by these MMT parameters. The results presented in this work show that combining with the multispectral technique, the FDHs and MMT parameters can characterize the microstructural variation features of skeletal muscle tissues. The techniques have the potential to be used as tools for quantitative assessment of meat qualities in food industry.

  19. Quantitative Susceptibility Mapping of the Midbrain in Parkinson’s Disease

    PubMed Central

    Du, Guangwei; Liu, Tian; Lewis, Mechelle M.; Kong, Lan; Wang, Yi; Connor, James; Mailman, Richard B.; Huang, Xuemei

    2017-01-01

    Background Parkinson’s disease (PD) is marked pathologically by dopamine neuron loss and iron overload in the substantia nigra pars compacta. Midbrain iron content is reported to be increased in PD based on magnetic resonance imaging (MRI) R2* changes. Because quantitative susceptibility mapping is a novel MRI approach to measure iron content, we compared it with R2* for assessing midbrain changes in PD. Methods Quantitative susceptibility mapping and R2* maps were obtained from 47 PD patients and 47 healthy controls. Midbrain susceptibility and R2* values were analyzed by using both voxel-based and region-of-interest approaches in normalized space, and analyzed along with clinical data, including disease duration, Unified Parkinson’s Disease Rating Scale (UPDRS) I, II, and III sub-scores, and levodopa-equivalent daily dosage. All studies were done while PD patients were “on drug.” Results Compared with controls, PD patients showed significantly increased susceptibility values in both right (cluster size = 106 mm3) and left (164 mm3) midbrain, located ventrolateral to the red nucleus that corresponded to the substantia nigra pars compacta. Susceptibility values in this region were correlated significantly with disease duration, UPDRS II, and levodopa-equivalent daily dosage. Conversely, R2* was increased significantly only in a much smaller region (62 mm3) of the left lateral substantia nigra pars compacta and was not significantly correlated with clinical parameters. Conclusion The use of quantitative susceptibility mapping demonstrated marked nigral changes that correlated with clinical PD status more sensitively than R2*. These data suggest that quantitative susceptibility mapping may be a superior imaging biomarker to R2* for estimating brain iron levels in PD. PMID:26362242

  20. PRIORITIZING FUTURE RESEACH ON OFF-LABEL PRESCRIBING: RESULTS OF A QUANTITATIVE EVALUATION

    PubMed Central

    Walton, Surrey M.; Schumock, Glen T.; Lee, Ky-Van; Alexander, G. Caleb; Meltzer, David; Stafford, Randall S.

    2015-01-01

    Background Drug use for indications not approved by the Food and Drug Administration exceeds 20% of prescribing. Available compendia indicate that a minority of off-label uses are well supported by evidence. Policy makers, however, lack information to identify where systematic reviews of the evidence or other research would be most valuable. Methods We developed a quantitative model for prioritizing individual drugs for future research on off-label uses. The base model incorporated three key factors, 1) the volume of off-label use with inadequate evidence, 2) safety, and 3) cost and market considerations. Nationally representative prescribing data were used to estimate the number of off-label drug uses by indication from 1/2005 through 6/2007 in the United States, and these indications were then categorized according to the adequacy of scientific support. Black box warnings and safety alerts were used to quantify drug safety. Drug cost, date of market entry, and marketing expenditures were used to quantify cost and market considerations. Each drug was assigned a relative value for each factor, and the factors were then weighted in the final model to produce a priority score. Sensitivity analyses were conducted by varying the weightings and model parameters. Results Drugs that were consistently ranked highly in both our base model and sensitivity analyses included quetiapine, warfarin, escitalopram, risperidone, montelukast, bupropion, sertraline, venlafaxine, celecoxib, lisinopril, duloxetine, trazodone, olanzapine, and epoetin alfa. Conclusion Future research into off-label drug use should focus on drugs used frequently with inadequate supporting evidence, particularly if further concerns are raised by known safety issues, high drug cost, recent market entry, and extensive marketing. Based on quantitative measures of these factors, we have prioritized drugs where targeted research and policy activities have high potential value. PMID:19025425

  1. A Novel Method for Analyzing Extremely Biased Agonism at G Protein–Coupled Receptors

    PubMed Central

    Zhou, Lei; Ehlert, Frederick J.; Bohn, Laura M.

    2015-01-01

    Seven transmembrane receptors were originally named and characterized based on their ability to couple to heterotrimeric G proteins. The assortment of coupling partners for G protein–coupled receptors has subsequently expanded to include other effectors (most notably the βarrestins). This diversity of partners available to the receptor has prompted the pursuit of ligands that selectively activate only a subset of the available partners. A biased or functionally selective ligand may be able to distinguish between different active states of the receptor, and this would result in the preferential activation of one signaling cascade more than another. Although application of the “standard” operational model for analyzing ligand bias is useful and suitable in most cases, there are limitations that arise when the biased agonist fails to induce a significant response in one of the assays being compared. In this article, we describe a quantitative method for measuring ligand bias that is particularly useful for such cases of extreme bias. Using simulations and experimental evidence from several κ opioid receptor agonists, we illustrate a “competitive” model for quantitating the degree and direction of bias. By comparing the results obtained from the competitive model with the standard model, we demonstrate that the competitive model expands the potential for evaluating the bias of very partial agonists. We conclude the competitive model provides a useful mechanism for analyzing the bias of partial agonists that exhibit extreme bias. PMID:25680753

  2. Implementing a Cost Effectiveness Analyzer for Web-Supported Academic Instruction: A Campus Wide Analysis

    ERIC Educational Resources Information Center

    Cohen, Anat; Nachmias, Rafi

    2009-01-01

    This paper describes the implementation of a quantitative cost effectiveness analyzer for Web-supported academic instruction that was developed in Tel Aviv University during a long term study. The paper presents the cost effectiveness analysis of Tel Aviv University campus. Cost and benefit of 3,453 courses were analyzed, exemplifying campus-wide…

  3. Adsorption of monocomponent enzymes in enzyme mixture analyzed quantitatively during hydrolysis of lignocellulose substrates.

    PubMed

    Várnai, Anikó; Viikari, Liisa; Marjamaa, Kaisa; Siika-aho, Matti

    2011-01-01

    The adsorption of purified Trichoderma reesei cellulases (TrCel7A, TrCel6A and TrCel5A) and xylanase TrXyn11 and Aspergillus niger β-glucosidase AnCel3A was studied in enzyme mixture during hydrolysis of two pretreated lignocellulosic materials, steam pretreated and catalytically delignified spruce, along with microcrystalline cellulose (Avicel). The enzyme mixture was compiled to resemble the composition of commercial cellulase preparations. The hydrolysis was carried out at 35 °C to mimic the temperature of the simultaneous saccharification and fermentation (SSF). Enzyme adsorption was followed by analyzing the activity and the protein amount of the individual free enzymes in the hydrolysis supernatant. Most enzymes adsorbed quickly at early stages of the hydrolysis and remained bound throughout the hydrolysis, although the conversion reached was fairly high. Only with the catalytically oxidized spruce samples, the bound enzymes started to be released as the hydrolysis degree reached 80%. The results based on enzyme activities and protein assay were in good accordance. Copyright © 2010 Elsevier Ltd. All rights reserved.

  4. Quantitative Analysis of High-Quality Officer Selection by Commandants Career-Level Education Board

    DTIC Science & Technology

    2017-03-01

    due to Marines being evaluated before the end of their initial service commitment. Our research utilizes quantitative variables to analyze the...not provide detailed information why. B. LIMITATIONS The photograph analysis in this research is strictly limited to a quantitative analysis in...NAVAL POSTGRADUATE SCHOOL MONTEREY, CALIFORNIA THESIS Approved for public release. Distribution is unlimited. QUANTITATIVE

  5. Quantitative PCR for HTLV-1 provirus in adult T-cell leukemia/lymphoma using paraffin tumor sections.

    PubMed

    Kato, Junki; Masaki, Ayako; Fujii, Keiichiro; Takino, Hisashi; Murase, Takayuki; Yonekura, Kentaro; Utsunomiya, Atae; Ishida, Takashi; Iida, Shinsuke; Inagaki, Hiroshi

    2016-11-01

    Detection of HTLV-1 provirus using paraffin tumor sections may assist the diagnosis of adult T-cell leukemia/lymphoma (ATLL). For the detection, non-quantitative PCR assay has been reported, but its usefulness and limitations remain unclear. To our knowledge, quantitative PCR assay using paraffin tumor sections has not been reported. Using paraffin sections from ATLLs and non-ATLL T-cell lymphomas, we first performed non-quantitative PCR for HTLV-1 provirus. Next, we determined tumor ratios and carried out quantitative PCR to obtain provirus copy numbers. The results were analyzed with a simple regression model and a novel criterion, cut-off using 95 % rejection limits. Our quantitative PCR assay showed an excellent association between tumor ratios and the copy numbers (r = 0.89, P < 0.0001). The 95 % rejection limits provided a statistical basis for the range for the determination of HTLV-1 involvement. Its application suggested that results of non-quantitative PCR assay should be interpreted very carefully and that our quantitative PCR assay is useful to estimate the status of HTLV-1 involvement in the tumor cases. In conclusion, our quantitative PCR assay using paraffin tumor sections may be useful for the screening of ATLL cases, especially in HTLV-1 non-endemic areas where easy access to serological testing for HTLV-1 infection is limited. © 2016 Japanese Society of Pathology and John Wiley & Sons Australia, Ltd.

  6. Methodological Options and Their Implications: An Example Using Secondary Data to Analyze Latino Educational Expectations

    ERIC Educational Resources Information Center

    Wells, Ryan S.; Lynch, Cassie M.; Seifert, Tricia A.

    2011-01-01

    A number of studies over decades have examined determinants of educational expectations. However, even among the subset of quantitative studies, there is considerable variation in the methods used to operationally define and analyze expectations. Using a systematic literature review and several regression methods to analyze Latino students'…

  7. Quantitative measurement of feline colonic transit

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Krevsky, B.; Somers, M.B.; Maurer, A.H.

    1988-10-01

    Colonic transit scintigraphy, a method for quantitatively evaluating the movement of the fecal stream in vivo, was employed to evaluate colonic transit in the cat. Scintigraphy was performed in duplicate in five cats and repeated four times in one cat. After instillation of an 111In marker into the cecum through a surgically implanted silicone cecostomy tube, colonic movement of the instillate was quantitated for 24 h using gamma scintigraphy. Antegrade and retrograde motion of radionuclide was observed. The cecum and ascending colon emptied rapidly, with a half-emptying time of 1.68 +/- 0.56 h (mean +/- SE). After 24 h, 25.1more » +/- 5.2% of the activity remained in the transverse colon. The progression of the geometric center was initially rapid, followed later by a delayed phase. Geometric center reproducibility was found to be high when analyzed using simple linear regression (slope = 0.92; r = 0.73; P less than 0.01). Atropine (0.1 mg/kg im) was found to delay cecum and ascending colon emptying and delay progression of the geometric center. These results demonstrate both 1) the ability of colonic transit scintigraphy to detect changes in transit induced by pharmacological manipulation and 2) the fact that muscarinic blockade inhibits antegrade transit of the fecal stream. We conclude that feline colonic transit may be studied in a quantitative and reproducible manner with colonic transit scintigraphy.« less

  8. Sample normalization methods in quantitative metabolomics.

    PubMed

    Wu, Yiman; Li, Liang

    2016-01-22

    To reveal metabolomic changes caused by a biological event in quantitative metabolomics, it is critical to use an analytical tool that can perform accurate and precise quantification to examine the true concentration differences of individual metabolites found in different samples. A number of steps are involved in metabolomic analysis including pre-analytical work (e.g., sample collection and storage), analytical work (e.g., sample analysis) and data analysis (e.g., feature extraction and quantification). Each one of them can influence the quantitative results significantly and thus should be performed with great care. Among them, the total sample amount or concentration of metabolites can be significantly different from one sample to another. Thus, it is critical to reduce or eliminate the effect of total sample amount variation on quantification of individual metabolites. In this review, we describe the importance of sample normalization in the analytical workflow with a focus on mass spectrometry (MS)-based platforms, discuss a number of methods recently reported in the literature and comment on their applicability in real world metabolomics applications. Sample normalization has been sometimes ignored in metabolomics, partially due to the lack of a convenient means of performing sample normalization. We show that several methods are now available and sample normalization should be performed in quantitative metabolomics where the analyzed samples have significant variations in total sample amounts. Copyright © 2015 Elsevier B.V. All rights reserved.

  9. Quantitative evaluation of translational medicine based on scientometric analysis and information extraction.

    PubMed

    Zhang, Yin; Diao, Tianxi; Wang, Lei

    2014-12-01

    Designed to advance the two-way translational process between basic research and clinical practice, translational medicine has become one of the most important areas in biomedicine. The quantitative evaluation of translational medicine is valuable for the decision making of global translational medical research and funding. Using the scientometric analysis and information extraction techniques, this study quantitatively analyzed the scientific articles on translational medicine. The results showed that translational medicine had significant scientific output and impact, specific core field and institute, and outstanding academic status and benefit. While it is not considered in this study, the patent data are another important indicators that should be integrated in the relevant research in the future. © 2014 Wiley Periodicals, Inc.

  10. Quantitative Determination of Citric and Ascorbic Acid in Powdered Drink Mixes

    ERIC Educational Resources Information Center

    Sigmann, Samuella B.; Wheeler, Dale E.

    2004-01-01

    A procedure by which the reactions are used to quantitatively determine the amount of total acid, the amount of total ascorbic acid and the amount of citric acid in a given sample of powdered drink mix, are described. A safe, reliable and low-cost quantitative method to analyze consumer product for acid content is provided.

  11. Quantitative T2-Mapping and T2⁎-Mapping Evaluation of Changes in Cartilage Matrix after Acute Anterior Cruciate Ligament Rupture and the Correlation between the Results of Both Methods.

    PubMed

    Tao, Hongyue; Qiao, Yang; Hu, Yiwen; Xie, Yuxue; Lu, Rong; Yan, Xu; Chen, Shuang

    2018-01-01

    To quantitatively assess changes in cartilage matrix after acute anterior cruciate ligament (ACL) rupture using T2- and T2 ⁎ -mapping and analyze the correlation between the results of both methods. Twenty-three patients and 23 healthy controls were enrolled and underwent quantitative MRI examination. The knee cartilage was segmented into six compartments, including lateral femur (LF), lateral tibia (LT), medial femur (MF), medial tibia (MT), trochlea (Tr), and patella (Pa). T2 and T2 ⁎ values were measured in full-thickness as well as superficial and deep layers of each cartilage compartment. Differences of T2 and T2 ⁎ values between patients and controls were compared using unpaired Student's t -test, and the correlation between their reciprocals was analyzed using Pearson's correlation coefficient. ACL-ruptured patients showed higher T2 and T2 ⁎ values in full-thickness and superficial layers of medial and lateral tibiofemoral joint. Meanwhile, patients exhibited higher T2 ⁎ values in deep layers of lateral tibiofemoral joint. The elevated percentages of T2 and T2 ⁎ value in superficial LT were most significant (20.738%, 17.525%). The reciprocal of T2 ⁎ value was correlated with that of T2 value ( r = 0.886, P < 0.001). The early degeneration could occur in various knee cartilage compartments after acute ACL rupture, especially in the superficial layer of LT. T2 ⁎ -mapping might be more sensitive in detecting deep layer of cartilage than T2-mapping.

  12. Qualitative, semi-quantitative, and quantitative simulation of the osmoregulation system in yeast

    PubMed Central

    Pang, Wei; Coghill, George M.

    2015-01-01

    In this paper we demonstrate how Morven, a computational framework which can perform qualitative, semi-quantitative, and quantitative simulation of dynamical systems using the same model formalism, is applied to study the osmotic stress response pathway in yeast. First the Morven framework itself is briefly introduced in terms of the model formalism employed and output format. We then built a qualitative model for the biophysical process of the osmoregulation in yeast, and a global qualitative-level picture was obtained through qualitative simulation of this model. Furthermore, we constructed a Morven model based on existing quantitative model of the osmoregulation system. This model was then simulated qualitatively, semi-quantitatively, and quantitatively. The obtained simulation results are presented with an analysis. Finally the future development of the Morven framework for modelling the dynamic biological systems is discussed. PMID:25864377

  13. Quantitative Detection of Cracks in Steel Using Eddy Current Pulsed Thermography.

    PubMed

    Shi, Zhanqun; Xu, Xiaoyu; Ma, Jiaojiao; Zhen, Dong; Zhang, Hao

    2018-04-02

    Small cracks are common defects in steel and often lead to catastrophic accidents in industrial applications. Various nondestructive testing methods have been investigated for crack detection; however, most current methods focus on qualitative crack identification and image processing. In this study, eddy current pulsed thermography (ECPT) was applied for quantitative crack detection based on derivative analysis of temperature variation. The effects of the incentive parameters on the temperature variation were analyzed in the simulation study. The crack profile and position are identified in the thermal image based on the Canny edge detection algorithm. Then, one or more trajectories are determined through the crack profile in order to determine the crack boundary through its temperature distribution. The slope curve along the trajectory is obtained. Finally, quantitative analysis of the crack sizes was performed by analyzing the features of the slope curves. The experimental verification showed that the crack sizes could be quantitatively detected with errors of less than 1%. Therefore, the proposed ECPT method was demonstrated to be a feasible and effective nondestructive approach for quantitative crack detection.

  14. Analyzing and improving surface texture by dual-rotation magnetorheological finishing

    NASA Astrophysics Data System (ADS)

    Wang, Yuyue; Zhang, Yun; Feng, Zhijing

    2016-01-01

    The main advantages of magnetorheological finishing (MRF) are its high convergence rate of surface error, the ability of polishing aspheric surfaces and nearly no subsurface damage. However, common MRF produces directional surface texture due to the constant flow direction of the magnetorheological (MR) polishing fluid. This paper studies the mechanism of surface texture formation by texture modeling. Dual-rotation magnetorheological finishing (DRMRF) is presented to suppress directional surface texture after analyzing the results of the texture model for common MRF. The results of the surface texture model for DRMRF and the proposed quantitative method based on mathematical statistics indicate the effective suppression of directional surface texture. An experimental setup is developed and experiments show directional surface texture and no directional surface texture in common MRF and DRMRF, respectively. As a result, the surface roughness of DRMRF is 0.578 nm (root-mean-square value) which is lower than 1.109 nm in common MRF.

  15. Quantitative performance of a polarization diffraction grating polarimeter encoded onto two liquid-crystal-on-silicon displays

    NASA Astrophysics Data System (ADS)

    Cofré, Aarón; Vargas, Asticio; Torres-Ruiz, Fabián A.; Campos, Juan; Lizana, Angel; del Mar Sánchez-López, María; Moreno, Ignacio

    2017-11-01

    We present a quantitative analysis of the performance of a complete snapshot polarimeter based on a polarization diffraction grating (PDGr). The PDGr is generated in a common path polarization interferometer with a Z optical architecture that uses two liquid-crystal on silicon (LCoS) displays to imprint two different phase-only diffraction gratings onto two orthogonal linear states of polarization. As a result, we obtain a programmable PDGr capable to act as a simultaneous polarization state generator (PSG), yielding diffraction orders with different states of polarization. The same system is also shown to operate as a polarization state analyzer (PSA), therefore useful for the realization of a snapshot polarimeter. We analyze its performance using quantitative metrics such as the conditional number, and verify its reliability for the detection of states of polarization.

  16. DIFFERENTIAL ANALYZER

    DOEpatents

    Sorensen, E.G.; Gordon, C.M.

    1959-02-10

    Improvements in analog eomputing machines of the class capable of evaluating differential equations, commonly termed differential analyzers, are described. In general form, the analyzer embodies a plurality of basic computer mechanisms for performing integration, multiplication, and addition, and means for directing the result of any one operation to another computer mechanism performing a further operation. In the device, numerical quantities are represented by the rotation of shafts, or the electrical equivalent of shafts.

  17. Temporal Lobe Epilepsy: Quantitative MR Volumetry in Detection of Hippocampal Atrophy

    PubMed Central

    Farid, Nikdokht; Girard, Holly M.; Kemmotsu, Nobuko; Smith, Michael E.; Magda, Sebastian W.; Lim, Wei Y.; Lee, Roland R.

    2012-01-01

    Purpose: To determine the ability of fully automated volumetric magnetic resonance (MR) imaging to depict hippocampal atrophy (HA) and to help correctly lateralize the seizure focus in patients with temporal lobe epilepsy (TLE). Materials and Methods: This study was conducted with institutional review board approval and in compliance with HIPAA regulations. Volumetric MR imaging data were analyzed for 34 patients with TLE and 116 control subjects. Structural volumes were calculated by using U.S. Food and Drug Administration–cleared software for automated quantitative MR imaging analysis (NeuroQuant). Results of quantitative MR imaging were compared with visual detection of atrophy, and, when available, with histologic specimens. Receiver operating characteristic analyses were performed to determine the optimal sensitivity and specificity of quantitative MR imaging for detecting HA and asymmetry. A linear classifier with cross validation was used to estimate the ability of quantitative MR imaging to help lateralize the seizure focus. Results: Quantitative MR imaging–derived hippocampal asymmetries discriminated patients with TLE from control subjects with high sensitivity (86.7%–89.5%) and specificity (92.2%–94.1%). When a linear classifier was used to discriminate left versus right TLE, hippocampal asymmetry achieved 94% classification accuracy. Volumetric asymmetries of other subcortical structures did not improve classification. Compared with invasive video electroencephalographic recordings, lateralization accuracy was 88% with quantitative MR imaging and 85% with visual inspection of volumetric MR imaging studies but only 76% with visual inspection of clinical MR imaging studies. Conclusion: Quantitative MR imaging can depict the presence and laterality of HA in TLE with accuracy rates that may exceed those achieved with visual inspection of clinical MR imaging studies. Thus, quantitative MR imaging may enhance standard visual analysis, providing a

  18. Results of Studying Astronomy Students’ Science Literacy, Quantitative Literacy, and Information Literacy

    NASA Astrophysics Data System (ADS)

    Buxner, Sanlyn; Impey, Chris David; Follette, Katherine B.; Dokter, Erin F.; McCarthy, Don; Vezino, Beau; Formanek, Martin; Romine, James M.; Brock, Laci; Neiberding, Megan; Prather, Edward E.

    2017-01-01

    Introductory astronomy courses often serve as terminal science courses for non-science majors and present an opportunity to assess non future scientists’ attitudes towards science as well as basic scientific knowledge and scientific analysis skills that may remain unchanged after college. Through a series of studies, we have been able to evaluate students’ basic science knowledge, attitudes towards science, quantitative literacy, and informational literacy. In the Fall of 2015, we conducted a case study of a single class administering all relevant surveys to an undergraduate class of 20 students. We will present our analysis of trends of each of these studies as well as the comparison case study. In general we have found that students basic scientific knowledge has remained stable over the past quarter century. In all of our studies, there is a strong relationship between student attitudes and their science and quantitative knowledge and skills. Additionally, students’ information literacy is strongly connected to their attitudes and basic scientific knowledge. We are currently expanding these studies to include new audiences and will discuss the implications of our findings for instructors.

  19. [Quantitative classification in catering trade and countermeasures of supervision and management in Hunan Province].

    PubMed

    Liu, Xiulan; Chen, Lizhang; He, Xiang

    2012-02-01

    To analyze the status quo of quantitative classification in Hunan Province catering industry, and to discuss the countermeasures in-depth. According to relevant laws and regulations, and after referring to Daily supervision and quantitative scoring sheet and consulting experts, a checklist of key supervision indicators was made. The implementation of quantitative classification in 10 cities in Hunan Province was studied, and the status quo was analyzed. All the 390 catering units implemented quantitative classified management. The larger the catering enterprise, the higher level of quantitative classification. In addition to cafeterias, the smaller the catering units, the higher point of deduction, and snack bars and beverage stores were the highest. For those quantified and classified as C and D, the point of deduction was higher in the procurement and storage of raw materials, operation processing and other aspects. The quantitative classification of Hunan Province has relatively wide coverage. There are hidden risks in food security in small catering units, snack bars, and beverage stores. The food hygienic condition of Hunan Province needs to be improved.

  20. SOME QUANTITATIVE ASPECTS OF THE INSTRUCTIONAL PROCESS.

    ERIC Educational Resources Information Center

    GAVIN, WILLIAM J.; SPITZER, MURRAY

    THE DATA FROM THE SEVERAL STUDIES ANALYZED IN THIS REPORT HAVE BEEN COLLECTED AS PART OF AN ON-GOING EFFORT TO IMPLEMENT THE ABT ASSOCIATES' EDUCATION COST EFFECTIVENESS INSTRUCTIONAL PROCESS SUBMODEL, WHICH IS DEVELOPING TECHNIQUES TO EVALUATE THE QUANTITATIVE, CAUSE-AND-EFFECT RELATIONSHIP BETWEEN THE INSTRUCTIONAL PROCESS AND SCHOLASTIC…

  1. The influence of the design matrix on treatment effect estimates in the quantitative analyses of single-subject experimental design research.

    PubMed

    Moeyaert, Mariola; Ugille, Maaike; Ferron, John M; Beretvas, S Natasha; Van den Noortgate, Wim

    2014-09-01

    The quantitative methods for analyzing single-subject experimental data have expanded during the last decade, including the use of regression models to statistically analyze the data, but still a lot of questions remain. One question is how to specify predictors in a regression model to account for the specifics of the design and estimate the effect size of interest. These quantitative effect sizes are used in retrospective analyses and allow synthesis of single-subject experimental study results which is informative for evidence-based decision making, research and theory building, and policy discussions. We discuss different design matrices that can be used for the most common single-subject experimental designs (SSEDs), namely, the multiple-baseline designs, reversal designs, and alternating treatment designs, and provide empirical illustrations. The purpose of this article is to guide single-subject experimental data analysts interested in analyzing and meta-analyzing SSED data. © The Author(s) 2014.

  2. Quantitatively characterizing the microstructural features of breast ductal carcinoma tissues in different progression stages by Mueller matrix microscope.

    PubMed

    Dong, Yang; Qi, Ji; He, Honghui; He, Chao; Liu, Shaoxiong; Wu, Jian; Elson, Daniel S; Ma, Hui

    2017-08-01

    Polarization imaging has been recognized as a potentially powerful technique for probing the microstructural information and optical properties of complex biological specimens. Recently, we have reported a Mueller matrix microscope by adding the polarization state generator and analyzer (PSG and PSA) to a commercial transmission-light microscope, and applied it to differentiate human liver and cervical cancerous tissues with fibrosis. In this paper, we apply the Mueller matrix microscope for quantitative detection of human breast ductal carcinoma samples at different stages. The Mueller matrix polar decomposition and transformation parameters of the breast ductal tissues in different regions and at different stages are calculated and analyzed. For more quantitative comparisons, several widely-used image texture feature parameters are also calculated to characterize the difference in the polarimetric images. The experimental results indicate that the Mueller matrix microscope and the polarization parameters can facilitate the quantitative detection of breast ductal carcinoma tissues at different stages.

  3. Field test results with the targeted search MCSA. [multi-channel spectrum analyzer for SETI

    NASA Technical Reports Server (NTRS)

    Tarter, J. C.

    1988-01-01

    In April 1985, a 74,000 channel prototype of the multichannel spectrum analyzer (MCSA) that NASA plans to use in a systematic search for extraterrestrial intelligence (SETI) was installed at DSS13, a 26 meter R&D antenna facility at the Goldstone Deep Space Network (DSN) site. Since that time the instrumentation has been used to validate the performance of signal detection algorithms using locally injected signals and the weak carriers from distant spacecraft. This paper describes results from the Goldstone Field Tests and plans to move the prototype equipment to other sites where SETI may be conducted in the future.

  4. Three-dimensional quantitative analysis of adhesive remnants and enamel loss resulting from debonding orthodontic molar tubes.

    PubMed

    Janiszewska-Olszowska, Joanna; Tandecka, Katarzyna; Szatkiewicz, Tomasz; Sporniak-Tutak, Katarzyna; Grocholewicz, Katarzyna

    2014-09-10

    Presenting a new method for direct, quantitative analysis of enamel surface. Measurement of adhesive remnants and enamel loss resulting from debonding molar tubes. Buccal surfaces of fifteen extracted human molars were directly scanned with an optic blue-light 3D scanner to the nearest 2 μm. After 20 s etching molar tubes were bonded and after 24 h storing in 0.9% saline - debonded. Then 3D scanning was repeated. Superimposition and comparison were proceeded and shape alterations of the entire objects were analyzed using specialized computer software. Residual adhesive heights as well as enamel loss depths have been obtained for the entire buccal surfaces. Residual adhesive volume and enamel loss volume have been calculated for every tooth. The maximum height of adhesive remaining on enamel surface was 0.76 mm and the volume on particular teeth ranged from 0.047 mm3 to 4.16 mm3. The median adhesive remnant volume was 0.988 mm3. Mean depths of enamel loss for particular teeth ranged from 0.0076 mm to 0.0416 mm. Highest maximum depth of enamel loss was 0.207 mm. Median volume of enamel loss was 0.104 mm3 and maximum volume was 1.484 mm3. Blue-light 3D scanning is able to provide direct precise scans of the enamel surface, which can be superimposed in order to calculate shape alterations. Debonding molar tubes leaves a certain amount of adhesive remnants on the enamel, however the interface fracture pattern varies for particular teeth and areas of enamel loss are present as well.

  5. Hydrophobic ionic liquids for quantitative bacterial cell lysis with subsequent DNA quantification.

    PubMed

    Fuchs-Telka, Sabine; Fister, Susanne; Mester, Patrick-Julian; Wagner, Martin; Rossmanith, Peter

    2017-02-01

    DNA is one of the most frequently analyzed molecules in the life sciences. In this article we describe a simple and fast protocol for quantitative DNA isolation from bacteria based on hydrophobic ionic liquid supported cell lysis at elevated temperatures (120-150 °C) for subsequent PCR-based analysis. From a set of five hydrophobic ionic liquids, 1-butyl-1-methylpyrrolidinium bis(trifluoromethylsulfonyl)imide was identified as the most suitable for quantitative cell lysis and DNA extraction because of limited quantitative PCR inhibition by the aqueous eluate as well as no detectable DNA uptake. The newly developed method was able to efficiently lyse Gram-negative bacterial cells, whereas Gram-positive cells were protected by their thick cell wall. The performance of the final protocol resulted in quantitative DNA extraction efficiencies for Gram-negative bacteria similar to those obtained with a commercial kit, whereas the number of handling steps, and especially the time required, was dramatically reduced. Graphical Abstract After careful evaluation of five hydrophobic ionic liquids, 1-butyl-1-methylpyrrolidinium bis(trifluoromethylsulfonyl)imide ([BMPyr + ][Ntf 2 - ]) was identified as the most suitable ionic liquid for quantitative cell lysis and DNA extraction. When used for Gram-negative bacteria, the protocol presented is simple and very fast and achieves DNA extraction efficiencies similar to those obtained with a commercial kit. ddH 2 O double-distilled water, qPCR quantitative PCR.

  6. Qualitative, semi-quantitative, and quantitative simulation of the osmoregulation system in yeast.

    PubMed

    Pang, Wei; Coghill, George M

    2015-05-01

    In this paper we demonstrate how Morven, a computational framework which can perform qualitative, semi-quantitative, and quantitative simulation of dynamical systems using the same model formalism, is applied to study the osmotic stress response pathway in yeast. First the Morven framework itself is briefly introduced in terms of the model formalism employed and output format. We then built a qualitative model for the biophysical process of the osmoregulation in yeast, and a global qualitative-level picture was obtained through qualitative simulation of this model. Furthermore, we constructed a Morven model based on existing quantitative model of the osmoregulation system. This model was then simulated qualitatively, semi-quantitatively, and quantitatively. The obtained simulation results are presented with an analysis. Finally the future development of the Morven framework for modelling the dynamic biological systems is discussed. Copyright © 2015 The Authors. Published by Elsevier Ireland Ltd.. All rights reserved.

  7. Guiding center model to interpret neutral particle analyzer results

    NASA Technical Reports Server (NTRS)

    Englert, G. W.; Reinmann, J. J.; Lauver, M. R.

    1974-01-01

    The theoretical model is discussed, which accounts for drift and cyclotron components of ion motion in a partially ionized plasma. Density and velocity distributions are systematically precribed. The flux into the neutral particle analyzer (NPA) from this plasma is determined by summing over all charge exchange neutrals in phase space which are directed into apertures. Especially detailed data, obtained by sweeping the line of sight of the apertures across the plasma of the NASA Lewis HIP-1 burnout device, are presented. Selection of randomized cyclotron velocity distributions about mean azimuthal drift yield energy distributions which compared well with experiment. Use of data obtained with a bending magnet on the NPA showed that separation between energy distribution curves of various mass species correlate well with a drift divided by mean cyclotron energy parameter of the theory. Use of the guiding center model in conjunction with NPA scans across the plasma aid in estimates of ion density and E field variation with plasma radius.

  8. Towards quantitative assessment of calciphylaxis

    NASA Astrophysics Data System (ADS)

    Deserno, Thomas M.; Sárándi, István.; Jose, Abin; Haak, Daniel; Jonas, Stephan; Specht, Paula; Brandenburg, Vincent

    2014-03-01

    Calciphylaxis is a rare disease that has devastating conditions associated with high morbidity and mortality. Calciphylaxis is characterized by systemic medial calcification of the arteries yielding necrotic skin ulcerations. In this paper, we aim at supporting the installation of multi-center registries for calciphylaxis, which includes a photographic documentation of skin necrosis. However, photographs acquired in different centers under different conditions using different equipment and photographers cannot be compared quantitatively. For normalization, we use a simple color pad that is placed into the field of view, segmented from the image, and its color fields are analyzed. In total, 24 colors are printed on that scale. A least-squares approach is used to determine the affine color transform. Furthermore, the card allows scale normalization. We provide a case study for qualitative assessment. In addition, the method is evaluated quantitatively using 10 images of two sets of different captures of the same necrosis. The variability of quantitative measurements based on free hand photography is assessed regarding geometric and color distortions before and after our simple calibration procedure. Using automated image processing, the standard deviation of measurements is significantly reduced. The coefficients of variations yield 5-20% and 2-10% for geometry and color, respectively. Hence, quantitative assessment of calciphylaxis becomes practicable and will impact a better understanding of this rare but fatal disease.

  9. A quantitative study of nanoparticle skin penetration with interactive segmentation.

    PubMed

    Lee, Onseok; Lee, See Hyun; Jeong, Sang Hoon; Kim, Jaeyoung; Ryu, Hwa Jung; Oh, Chilhwan; Son, Sang Wook

    2016-10-01

    In the last decade, the application of nanotechnology techniques has expanded within diverse areas such as pharmacology, medicine, and optical science. Despite such wide-ranging possibilities for implementation into practice, the mechanisms behind nanoparticle skin absorption remain unknown. Moreover, the main mode of investigation has been qualitative analysis. Using interactive segmentation, this study suggests a method of objectively and quantitatively analyzing the mechanisms underlying the skin absorption of nanoparticles. Silica nanoparticles (SNPs) were assessed using transmission electron microscopy and applied to the human skin equivalent model. Captured fluorescence images of this model were used to evaluate degrees of skin penetration. These images underwent interactive segmentation and image processing in addition to statistical quantitative analyses of calculated image parameters including the mean, integrated density, skewness, kurtosis, and area fraction. In images from both groups, the distribution area and intensity of fluorescent silica gradually increased in proportion to time. Since statistical significance was achieved after 2 days in the negative charge group and after 4 days in the positive charge group, there is a periodic difference. Furthermore, the quantity of silica per unit area showed a dramatic change after 6 days in the negative charge group. Although this quantitative result is identical to results obtained by qualitative assessment, it is meaningful in that it was proven by statistical analysis with quantitation by using image processing. The present study suggests that the surface charge of SNPs could play an important role in the percutaneous absorption of NPs. These findings can help achieve a better understanding of the percutaneous transport of NPs. In addition, these results provide important guidance for the design of NPs for biomedical applications.

  10. A Visual Basic program for analyzing oedometer test results and evaluating intergranular void ratio

    NASA Astrophysics Data System (ADS)

    Monkul, M. Murat; Önal, Okan

    2006-06-01

    A visual basic program (POCI) is proposed and explained in order to analyze oedometer test results. Oedometer test results have vital importance from geotechnical point of view, since settlement requirements usually control the design of foundations. The software POCI is developed in order perform the necessary calculations for convential oedometer test. The change of global void ratio and stress-strain characteristics can be observed both numerically and graphically. It enables the users to calculate some parameters such as coefficient of consolidation, compression index, recompression index, and preconsolidation pressure depending on the type and stress history of the soil. Moreover, it adopts the concept of intergranular void ratio which may be important especially in the compression behavior of sandy soils. POCI shows the variation of intergranular void ratio and also enables the users to calculate granular compression index.

  11. Ideal versus School Learning: Analyzing Israeli Secondary School Students' Conceptions of Learning

    ERIC Educational Resources Information Center

    Hadar, Linor

    2009-01-01

    This study explored 130 secondary school students' conceptions of learning using an open-ended task, analyzed both qualitatively and quantitatively. Students' reality of learning comprised two separate spheres, ideal learning and school learning, which rarely interacted. Generally, students commented more about school than ideal learning. Factor…

  12. Analyzing wildfire exposure on Sardinia, Italy

    NASA Astrophysics Data System (ADS)

    Salis, Michele; Ager, Alan A.; Arca, Bachisio; Finney, Mark A.; Alcasena, Fermin; Bacciu, Valentina; Duce, Pierpaolo; Munoz Lozano, Olga; Spano, Donatella

    2014-05-01

    We used simulation modeling based on the minimum travel time algorithm (MTT) to analyze wildfire exposure of key ecological, social and economic features on Sardinia, Italy. Sardinia is the second largest island of the Mediterranean Basin, and in the last fifty years experienced large and dramatic wildfires, which caused losses and threatened urban interfaces, forests and natural areas, and agricultural productions. Historical fires and environmental data for the period 1995-2009 were used as input to estimate fine scale burn probability, conditional flame length, and potential fire size in the study area. With this purpose, we simulated 100,000 wildfire events within the study area, randomly drawing from the observed frequency distribution of burn periods and wind directions for each fire. Estimates of burn probability, excluding non-burnable fuels, ranged from 0 to 1.92x10-3, with a mean value of 6.48x10-5. Overall, the outputs provided a quantitative assessment of wildfire exposure at the landscape scale and captured landscape properties of wildfire exposure. We then examined how the exposure profiles varied among and within selected features and assets located on the island. Spatial variation in modeled outputs resulted in a strong effect of fuel models, coupled with slope and weather. In particular, the combined effect of Mediterranean maquis, woodland areas and complex topography on flame length was relevant, mainly in north-east Sardinia, whereas areas with herbaceous fuels and flat areas were in general characterized by lower fire intensity but higher burn probability. The simulation modeling proposed in this work provides a quantitative approach to inform wildfire risk management activities, and represents one of the first applications of burn probability modeling to capture fire risk and exposure profiles in the Mediterranean basin.

  13. On social inequality: Analyzing the rich-poor disparity

    NASA Astrophysics Data System (ADS)

    Eliazar, Iddo; Cohen, Morrel H.

    2014-05-01

    From the Old Testament to the Communist Manifesto, and from the French Revolution to the Occupy Wall Street protests, social inequality has always been at the focal point of public debate, as well as a major driver of political change. Although being of prime interest since Biblical times, the scientific investigation of the distributions of wealth and income in human societies began only at the close of the nineteenth century, and was pioneered by Pareto, Lorenz, Gini, and Pietra. The methodologies introduced by these trailblazing scholars form the bedrock of the contemporary science of social inequality. Based on this bedrock we present a new quantitative approach to the analysis of wealth and income distributions, which sets its spotlight on the most heated facet of the current global debate on social inequality-the rich-poor disparity. Our approach offers researchers highly applicable quantitative tools to empirically track and statistically analyze the growing gap between the rich and the poor.

  14. Development, application, and results of routine monitoring of Marek's disease virus in broiler house dust using real-time quantitative PCR.

    PubMed

    Walkden-Brown, Stephen W; Islam, A F Aminul; Groves, Peter J; Rubite, Ambrosio; Sharpe, Sue M; Burgess, Susan K

    2013-06-01

    Results are presented from four studies between 2002 and 2011 into the feasibility of routinely monitoring Marek's disease virus serotype 1 (MDV-1) in broiler house dust using real-time quantitative PCR (qPCR) measurement. Study 1 on two farms showed that detection of MDV-1 occurred earlier on average in dust samples tested using qPCR than standard PCR and in spleen samples from five birds per shed assayed for MDV-1 by qPCR or standard PCR. DNA quality following extraction from dust had no effect on detection of MDV-1. Study 2 demonstrated that herpesvirus of turkeys (HVT) and MDV serotype 2 (MDV-2) in addition to MDV-1 could be readily amplified from commercial farm dust samples, often in mixtures. MDV-2 was detected in 11 of 20 samples despite the absence of vaccination with this serotype. Study 3 investigated the reproducibility and sensitivity of the qPCR test and the presence of inhibitors in the samples. Samples extracted and amplified in triplicate showed a high level of reproducibility except at very low levels of virus near the limit of detection. Mixing of samples prior to extraction provided results consistent with the proportions in the mixture. Tests for inhibition showed that if the template contained DNA in the range 0.5-20 ng/microl no inhibition of the reaction was detectable. The sensitivity of the tests in terms of viral copy number (VCN) per milligram of dust was calculated to be in the range 24-600 VCN/mg for MDV-1, 48-1200 VCN/mg for MDV-2, and 182-4560 VCN/mg for HVT. In study 4 the results of 1976 commercial tests carried out for one company were analyzed. Overall 23.1% of samples were positive for MDV-1, 26.1% in unvaccinated and 16.4% in vaccinated chickens. There was marked regional and temporal variation in the proportion of positive samples and the MDV-1 load. The tests were useful in formulating Marek's disease vaccination strategies. The number of samples submitted has increased recently, as has the incidence of positive samples

  15. The Quantitative Methods Boot Camp: Teaching Quantitative Thinking and Computing Skills to Graduate Students in the Life Sciences

    PubMed Central

    Stefan, Melanie I.; Gutlerner, Johanna L.; Born, Richard T.; Springer, Michael

    2015-01-01

    The past decade has seen a rapid increase in the ability of biologists to collect large amounts of data. It is therefore vital that research biologists acquire the necessary skills during their training to visualize, analyze, and interpret such data. To begin to meet this need, we have developed a “boot camp” in quantitative methods for biology graduate students at Harvard Medical School. The goal of this short, intensive course is to enable students to use computational tools to visualize and analyze data, to strengthen their computational thinking skills, and to simulate and thus extend their intuition about the behavior of complex biological systems. The boot camp teaches basic programming using biological examples from statistics, image processing, and data analysis. This integrative approach to teaching programming and quantitative reasoning motivates students’ engagement by demonstrating the relevance of these skills to their work in life science laboratories. Students also have the opportunity to analyze their own data or explore a topic of interest in more detail. The class is taught with a mixture of short lectures, Socratic discussion, and in-class exercises. Students spend approximately 40% of their class time working through both short and long problems. A high instructor-to-student ratio allows students to get assistance or additional challenges when needed, thus enhancing the experience for students at all levels of mastery. Data collected from end-of-course surveys from the last five offerings of the course (between 2012 and 2014) show that students report high learning gains and feel that the course prepares them for solving quantitative and computational problems they will encounter in their research. We outline our course here which, together with the course materials freely available online under a Creative Commons License, should help to facilitate similar efforts by others. PMID:25880064

  16. The quantitative methods boot camp: teaching quantitative thinking and computing skills to graduate students in the life sciences.

    PubMed

    Stefan, Melanie I; Gutlerner, Johanna L; Born, Richard T; Springer, Michael

    2015-04-01

    The past decade has seen a rapid increase in the ability of biologists to collect large amounts of data. It is therefore vital that research biologists acquire the necessary skills during their training to visualize, analyze, and interpret such data. To begin to meet this need, we have developed a "boot camp" in quantitative methods for biology graduate students at Harvard Medical School. The goal of this short, intensive course is to enable students to use computational tools to visualize and analyze data, to strengthen their computational thinking skills, and to simulate and thus extend their intuition about the behavior of complex biological systems. The boot camp teaches basic programming using biological examples from statistics, image processing, and data analysis. This integrative approach to teaching programming and quantitative reasoning motivates students' engagement by demonstrating the relevance of these skills to their work in life science laboratories. Students also have the opportunity to analyze their own data or explore a topic of interest in more detail. The class is taught with a mixture of short lectures, Socratic discussion, and in-class exercises. Students spend approximately 40% of their class time working through both short and long problems. A high instructor-to-student ratio allows students to get assistance or additional challenges when needed, thus enhancing the experience for students at all levels of mastery. Data collected from end-of-course surveys from the last five offerings of the course (between 2012 and 2014) show that students report high learning gains and feel that the course prepares them for solving quantitative and computational problems they will encounter in their research. We outline our course here which, together with the course materials freely available online under a Creative Commons License, should help to facilitate similar efforts by others.

  17. QUANTITATIVE PCR ANALYSIS OF MOLDS IN THE DUST FROM HOMES OF ASTHMATIC CHILDREN IN NORTH CAROLINA

    EPA Science Inventory

    The vacuum bag (VB) dust was analyzed by mold specific quantitative PCR. These results were compared to the analysis survey calculated for each of the homes. The mean and standard deviation (SD) of the ERMI values in the homes of the NC asthmatic children was 16.4 (6.77), compa...

  18. An Quantitative Analysis Method Of Trabecular Pattern In A Bone

    NASA Astrophysics Data System (ADS)

    Idesawa, Masanor; Yatagai, Toyohiko

    1982-11-01

    Orientation and density of trabecular pattern observed in a bone is closely related to its mechanical properties and deseases of a bone are appeared as changes of orientation and/or density distrbution of its trabecular patterns. They have been treated from a qualitative point of view so far because quantitative analysis method has not be established. In this paper, the authors proposed and investigated some quantitative analysis methods of density and orientation of trabecular patterns observed in a bone. These methods can give an index for evaluating orientation of trabecular pattern quantitatively and have been applied to analyze trabecular pattern observed in a head of femur and their availabilities are confirmed. Key Words: Index of pattern orientation, Trabecular pattern, Pattern density, Quantitative analysis

  19. Linearization improves the repeatability of quantitative dynamic contrast-enhanced MRI.

    PubMed

    Jones, Kyle M; Pagel, Mark D; Cárdenas-Rodríguez, Julio

    2018-04-01

    The purpose of this study was to compare the repeatabilities of the linear and nonlinear Tofts and reference region models (RRM) for dynamic contrast-enhanced MRI (DCE-MRI). Simulated and experimental DCE-MRI data from 12 rats with a flank tumor of C6 glioma acquired over three consecutive days were analyzed using four quantitative and semi-quantitative DCE-MRI metrics. The quantitative methods used were: 1) linear Tofts model (LTM), 2) non-linear Tofts model (NTM), 3) linear RRM (LRRM), and 4) non-linear RRM (NRRM). The following semi-quantitative metrics were used: 1) maximum enhancement ratio (MER), 2) time to peak (TTP), 3) initial area under the curve (iauc64), and 4) slope. LTM and NTM were used to estimate K trans , while LRRM and NRRM were used to estimate K trans relative to muscle (R Ktrans ). Repeatability was assessed by calculating the within-subject coefficient of variation (wSCV) and the percent intra-subject variation (iSV) determined with the Gage R&R analysis. The iSV for R Ktrans using LRRM was two-fold lower compared to NRRM at all simulated and experimental conditions. A similar trend was observed for the Tofts model, where LTM was at least 50% more repeatable than the NTM under all experimental and simulated conditions. The semi-quantitative metrics iauc64 and MER were as equally repeatable as K trans and R Ktrans estimated by LTM and LRRM respectively. The iSV for iauc64 and MER were significantly lower than the iSV for slope and TTP. In simulations and experimental results, linearization improves the repeatability of quantitative DCE-MRI by at least 30%, making it as repeatable as semi-quantitative metrics. Copyright © 2017 Elsevier Inc. All rights reserved.

  20. A Quantitative Study Identifying Political Strategies Used by Principals of Dual Language Programs

    ERIC Educational Resources Information Center

    Girard, Guadalupe

    2017-01-01

    Purpose. The purpose of this quantitative study was to identify the external and internal political strategies used by principals that allow them to successfully navigate the political environment surrounding dual language programs. Methodology. This quantitative study used descriptive research to collect, analyze, and report data that identified…

  1. Quantitative analysis of wet-heat inactivation in bovine spongiform encephalopathy

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Matsuura, Yuichi; Ishikawa, Yukiko; Bo, Xiao

    2013-03-01

    Highlights: ► We quantitatively analyzed wet-heat inactivation of the BSE agent. ► Infectivity of the BSE macerate did not survive 155 °C wet-heat treatment. ► Once the sample was dehydrated, infectivity was observed even at 170 °C. ► A quantitative PMCA assay was used to evaluate the degree of BSE inactivation. - Abstract: The bovine spongiform encephalopathy (BSE) agent is resistant to conventional microbial inactivation procedures and thus threatens the safety of cattle products and by-products. To obtain information necessary to assess BSE inactivation, we performed quantitative analysis of wet-heat inactivation of infectivity in BSE-infected cattle spinal cords. Using amore » highly sensitive bioassay, we found that infectivity in BSE cattle macerates fell with increase in temperatures from 133 °C to 150 °C and was not detected in the samples subjected to temperatures above 155 °C. In dry cattle tissues, infectivity was detected even at 170 °C. Thus, BSE infectivity reduces with increase in wet-heat temperatures but is less affected when tissues are dehydrated prior to the wet-heat treatment. The results of the quantitative protein misfolding cyclic amplification assay also demonstrated that the level of the protease-resistant prion protein fell below the bioassay detection limit by wet-heat at 155 °C and higher and could help assess BSE inactivation. Our results show that BSE infectivity is strongly resistant to wet-heat inactivation and that it is necessary to pay attention to BSE decontamination in recycled cattle by-products.« less

  2. Electroencephalography and quantitative electroencephalography in mild traumatic brain injury.

    PubMed

    Haneef, Zulfi; Levin, Harvey S; Frost, James D; Mizrahi, Eli M

    2013-04-15

    Mild traumatic brain injury (mTBI) causes brain injury resulting in electrophysiologic abnormalities visible in electroencephalography (EEG) recordings. Quantitative EEG (qEEG) makes use of quantitative techniques to analyze EEG characteristics such as frequency, amplitude, coherence, power, phase, and symmetry over time independently or in combination. QEEG has been evaluated for its use in making a diagnosis of mTBI and assessing prognosis, including the likelihood of progressing to the postconcussive syndrome (PCS) phase. We review the EEG and qEEG changes of mTBI described in the literature. An attempt is made to separate the findings seen during the acute, subacute, and chronic phases after mTBI. Brief mention is also made of the neurobiological correlates of qEEG using neuroimaging techniques or in histopathology. Although the literature indicates the promise of qEEG in making a diagnosis and indicating prognosis of mTBI, further study is needed to corroborate and refine these methods.

  3. Electroencephalography and Quantitative Electroencephalography in Mild Traumatic Brain Injury

    PubMed Central

    Levin, Harvey S.; Frost, James D.; Mizrahi, Eli M.

    2013-01-01

    Abstract Mild traumatic brain injury (mTBI) causes brain injury resulting in electrophysiologic abnormalities visible in electroencephalography (EEG) recordings. Quantitative EEG (qEEG) makes use of quantitative techniques to analyze EEG characteristics such as frequency, amplitude, coherence, power, phase, and symmetry over time independently or in combination. QEEG has been evaluated for its use in making a diagnosis of mTBI and assessing prognosis, including the likelihood of progressing to the postconcussive syndrome (PCS) phase. We review the EEG and qEEG changes of mTBI described in the literature. An attempt is made to separate the findings seen during the acute, subacute, and chronic phases after mTBI. Brief mention is also made of the neurobiological correlates of qEEG using neuroimaging techniques or in histopathology. Although the literature indicates the promise of qEEG in making a diagnosis and indicating prognosis of mTBI, further study is needed to corroborate and refine these methods. PMID:23249295

  4. On the Reproducibility of Label-Free Quantitative Cross-Linking/Mass Spectrometry

    NASA Astrophysics Data System (ADS)

    Müller, Fränze; Fischer, Lutz; Chen, Zhuo Angel; Auchynnikava, Tania; Rappsilber, Juri

    2018-02-01

    Quantitative cross-linking/mass spectrometry (QCLMS) is an emerging approach to study conformational changes of proteins and multi-subunit complexes. Distinguishing protein conformations requires reproducibly identifying and quantifying cross-linked peptides. Here we analyzed the variation between multiple cross-linking reactions using bis[sulfosuccinimidyl] suberate (BS3)-cross-linked human serum albumin (HSA) and evaluated how reproducible cross-linked peptides can be identified and quantified by LC-MS analysis. To make QCLMS accessible to a broader research community, we developed a workflow that integrates the established software tools MaxQuant for spectra preprocessing, Xi for cross-linked peptide identification, and finally Skyline for quantification (MS1 filtering). Out of the 221 unique residue pairs identified in our sample, 124 were subsequently quantified across 10 analyses with coefficient of variation (CV) values of 14% (injection replica) and 32% (reaction replica). Thus our results demonstrate that the reproducibility of QCLMS is in line with the reproducibility of general quantitative proteomics and we establish a robust workflow for MS1-based quantitation of cross-linked peptides.

  5. PyNeb: a new tool for analyzing emission lines. I. Code description and validation of results

    NASA Astrophysics Data System (ADS)

    Luridiana, V.; Morisset, C.; Shaw, R. A.

    2015-01-01

    Analysis of emission lines in gaseous nebulae yields direct measures of physical conditions and chemical abundances and is the cornerstone of nebular astrophysics. Although the physical problem is conceptually simple, its practical complexity can be overwhelming since the amount of data to be analyzed steadily increases; furthermore, results depend crucially on the input atomic data, whose determination also improves each year. To address these challenges we created PyNeb, an innovative code for analyzing emission lines. PyNeb computes physical conditions and ionic and elemental abundances and produces both theoretical and observational diagnostic plots. It is designed to be portable, modular, and largely customizable in aspects such as the atomic data used, the format of the observational data to be analyzed, and the graphical output. It gives full access to the intermediate quantities of the calculation, making it possible to write scripts tailored to the specific type of analysis one wants to carry out. In the case of collisionally excited lines, PyNeb works by solving the equilibrium equations for an n-level atom; in the case of recombination lines, it works by interpolation in emissivity tables. The code offers a choice of extinction laws and ionization correction factors, which can be complemented by user-provided recipes. It is entirely written in the python programming language and uses standard python libraries. It is fully vectorized, making it apt for analyzing huge amounts of data. The code is stable and has been benchmarked against IRAF/NEBULAR. It is public, fully documented, and has already been satisfactorily used in a number of published papers.

  6. Temporal lobe epilepsy: quantitative MR volumetry in detection of hippocampal atrophy.

    PubMed

    Farid, Nikdokht; Girard, Holly M; Kemmotsu, Nobuko; Smith, Michael E; Magda, Sebastian W; Lim, Wei Y; Lee, Roland R; McDonald, Carrie R

    2012-08-01

    To determine the ability of fully automated volumetric magnetic resonance (MR) imaging to depict hippocampal atrophy (HA) and to help correctly lateralize the seizure focus in patients with temporal lobe epilepsy (TLE). This study was conducted with institutional review board approval and in compliance with HIPAA regulations. Volumetric MR imaging data were analyzed for 34 patients with TLE and 116 control subjects. Structural volumes were calculated by using U.S. Food and Drug Administration-cleared software for automated quantitative MR imaging analysis (NeuroQuant). Results of quantitative MR imaging were compared with visual detection of atrophy, and, when available, with histologic specimens. Receiver operating characteristic analyses were performed to determine the optimal sensitivity and specificity of quantitative MR imaging for detecting HA and asymmetry. A linear classifier with cross validation was used to estimate the ability of quantitative MR imaging to help lateralize the seizure focus. Quantitative MR imaging-derived hippocampal asymmetries discriminated patients with TLE from control subjects with high sensitivity (86.7%-89.5%) and specificity (92.2%-94.1%). When a linear classifier was used to discriminate left versus right TLE, hippocampal asymmetry achieved 94% classification accuracy. Volumetric asymmetries of other subcortical structures did not improve classification. Compared with invasive video electroencephalographic recordings, lateralization accuracy was 88% with quantitative MR imaging and 85% with visual inspection of volumetric MR imaging studies but only 76% with visual inspection of clinical MR imaging studies. Quantitative MR imaging can depict the presence and laterality of HA in TLE with accuracy rates that may exceed those achieved with visual inspection of clinical MR imaging studies. Thus, quantitative MR imaging may enhance standard visual analysis, providing a useful and viable means for translating volumetric analysis into

  7. Validating internal controls for quantitative plant gene expression studies

    PubMed Central

    Brunner, Amy M; Yakovlev, Igor A; Strauss, Steven H

    2004-01-01

    Background Real-time reverse transcription PCR (RT-PCR) has greatly improved the ease and sensitivity of quantitative gene expression studies. However, accurate measurement of gene expression with this method relies on the choice of a valid reference for data normalization. Studies rarely verify that gene expression levels for reference genes are adequately consistent among the samples used, nor compare alternative genes to assess which are most reliable for the experimental conditions analyzed. Results Using real-time RT-PCR to study the expression of 10 poplar (genus Populus) housekeeping genes, we demonstrate a simple method for determining the degree of stability of gene expression over a set of experimental conditions. Based on a traditional method for analyzing the stability of varieties in plant breeding, it defines measures of gene expression stability from analysis of variance (ANOVA) and linear regression. We found that the potential internal control genes differed widely in their expression stability over the different tissues, developmental stages and environmental conditions studied. Conclusion Our results support that quantitative comparisons of candidate reference genes are an important part of real-time RT-PCR studies that seek to precisely evaluate variation in gene expression. The method we demonstrated facilitates statistical and graphical evaluation of gene expression stability. Selection of the best reference gene for a given set of experimental conditions should enable detection of biologically significant changes in gene expression that are too small to be revealed by less precise methods, or when highly variable reference genes are unknowingly used in real-time RT-PCR experiments. PMID:15317655

  8. LabKey Server NAb: A tool for analyzing, visualizing and sharing results from neutralizing antibody assays

    PubMed Central

    2011-01-01

    Background Multiple types of assays allow sensitive detection of virus-specific neutralizing antibodies. For example, the extent of antibody neutralization of HIV-1, SIV and SHIV can be measured in the TZM-bl cell line through the degree of luciferase reporter gene expression after infection. In the past, neutralization curves and titers for this standard assay have been calculated using an Excel macro. Updating all instances of such a macro with new techniques can be unwieldy and introduce non-uniformity across multi-lab teams. Using Excel also poses challenges in centrally storing, sharing and associating raw data files and results. Results We present LabKey Server's NAb tool for organizing, analyzing and securely sharing data, files and results for neutralizing antibody (NAb) assays, including the luciferase-based TZM-bl NAb assay. The customizable tool supports high-throughput experiments and includes a graphical plate template designer, allowing researchers to quickly adapt calculations to new plate layouts. The tool calculates the percent neutralization for each serum dilution based on luminescence measurements, fits a range of neutralization curves to titration results and uses these curves to estimate the neutralizing antibody titers for benchmark dilutions. Results, curve visualizations and raw data files are stored in a database and shared through a secure, web-based interface. NAb results can be integrated with other data sources based on sample identifiers. It is simple to make results public after publication by updating folder security settings. Conclusions Standardized tools for analyzing, archiving and sharing assay results can improve the reproducibility, comparability and reliability of results obtained across many labs. LabKey Server and its NAb tool are freely available as open source software at http://www.labkey.com under the Apache 2.0 license. Many members of the HIV research community can also access the LabKey Server NAb tool without

  9. Quantitative analysis of multiple sclerosis: a feasibility study

    NASA Astrophysics Data System (ADS)

    Li, Lihong; Li, Xiang; Wei, Xinzhou; Sturm, Deborah; Lu, Hongbing; Liang, Zhengrong

    2006-03-01

    Multiple Sclerosis (MS) is an inflammatory and demyelinating disorder of the central nervous system with a presumed immune-mediated etiology. For treatment of MS, the measurements of white matter (WM), gray matter (GM), and cerebral spinal fluid (CSF) are often used in conjunction with clinical evaluation to provide a more objective measure of MS burden. In this paper, we apply a new unifying automatic mixture-based algorithm for segmentation of brain tissues to quantitatively analyze MS. The method takes into account the following effects that commonly appear in MR imaging: 1) The MR data is modeled as a stochastic process with an inherent inhomogeneity effect of smoothly varying intensity; 2) A new partial volume (PV) model is built in establishing the maximum a posterior (MAP) segmentation scheme; 3) Noise artifacts are minimized by a priori Markov random field (MRF) penalty indicating neighborhood correlation from tissue mixture. The volumes of brain tissues (WM, GM) and CSF are extracted from the mixture-based segmentation. Experimental results of feasibility studies on quantitative analysis of MS are presented.

  10. High-throughput quantitative analysis by desorption electrospray ionization mass spectrometry.

    PubMed

    Manicke, Nicholas E; Kistler, Thomas; Ifa, Demian R; Cooks, R Graham; Ouyang, Zheng

    2009-02-01

    A newly developed high-throughput desorption electrospray ionization (DESI) source was characterized in terms of its performance in quantitative analysis. A 96-sample array, containing pharmaceuticals in various matrices, was analyzed in a single run with a total analysis time of 3 min. These solution-phase samples were examined from a hydrophobic PTFE ink printed on glass. The quantitative accuracy, precision, and limit of detection (LOD) were characterized. Chemical background-free samples of propranolol (PRN) with PRN-d(7) as internal standard (IS) and carbamazepine (CBZ) with CBZ-d(10) as IS were examined. So were two other sample sets consisting of PRN/PRN-d(7) at varying concentration in a biological milieu of 10% urine or porcine brain total lipid extract, total lipid concentration 250 ng/microL. The background-free samples, examined in a total analysis time of 1.5 s/sample, showed good quantitative accuracy and precision, with a relative error (RE) and relative standard deviation (RSD) generally less than 3% and 5%, respectively. The samples in urine and the lipid extract required a longer analysis time (2.5 s/sample) and showed RSD values of around 10% for the samples in urine and 4% for the lipid extract samples and RE values of less than 3% for both sets. The LOD for PRN and CBZ when analyzed without chemical background was 10 and 30 fmol, respectively. The LOD of PRN increased to 400 fmol analyzed in 10% urine, and 200 fmol when analyzed in the brain lipid extract.

  11. Quantitative fetal fibronectin and cervical length in symptomatic women: results from a prospective blinded cohort study.

    PubMed

    Levine, Lisa D; Downes, Katheryne L; Romero, Julie A; Pappas, Hope; Elovitz, Michal A

    2018-05-15

    Our objectives were to determine whether quantitative fetal fibronectin (fFN) and cervical length (CL) screening can be used alone or in combination as prognostic tests to identify symptomatic women at the highest or lowest risk for spontaneous preterm birth (sPTB). A prospective, blinded cohort study of women presenting with a singleton gestation to our triage unit between 22-33w6d with preterm labor symptoms was performed. Women with ruptured membranes, moderate/severe bleeding, and dilation >2 cm were excluded. The primary outcome was sPTB <37 weeks. We evaluated test characteristics of quantitative fFN and CL assessment, both separately and in combination, considering traditionally reported cut-points (fFN ≥50 and CL <25), as well as cut-points above and below these measures. We found interactions between fFN >50 and CL <25 and sPTB by parity and obstetric history (p < .05) and therefore stratified results. Test characteristics are presented with positive predictive value (PPV) and negative predictive value (NPV). Five hundred eighty women were enrolled and 537 women were available for analysis. Overall sPTB rate was 11.1%. Among nulliparous women, increasing levels of fFN were associated with increasing risk of sPTB, with PPV going from 26.5% at ≥20 ng/mL to 44.4% at ≥200 ng/mL. A cut-point of 20 ng/mL had higher sensitivity (69.2%) and higher NPV (96.8%) and therefore identified a "low-risk" group. fFN was not informative for multiparous women regardless of prior obstetrical history or quantitative level chosen. For all women, a shorter CL was associated with an increased sPTB risk. Among nulliparas and multiparas without a prior sPTB, a CL <20 mm optimized test characteristics (PPV 25 and 20%, NPV 95.5, and 92.7%, respectively). For multiparas with a prior sPTB, CL <25 mm was more useful. Using fFN and CL in combination for nulliparas did not improve test characteristics over using the individual fFN (p = .74) and CL (p = .31

  12. A quantitative approach to evolution of music and philosophy

    NASA Astrophysics Data System (ADS)

    Vieira, Vilson; Fabbri, Renato; Travieso, Gonzalo; Oliveira, Osvaldo N., Jr.; da Fontoura Costa, Luciano

    2012-08-01

    The development of new statistical and computational methods is increasingly making it possible to bridge the gap between hard sciences and humanities. In this study, we propose an approach based on a quantitative evaluation of attributes of objects in fields of humanities, from which concepts such as dialectics and opposition are formally defined mathematically. As case studies, we analyzed the temporal evolution of classical music and philosophy by obtaining data for 8 features characterizing the corresponding fields for 7 well-known composers and philosophers, which were treated with multivariate statistics and pattern recognition methods. A bootstrap method was applied to avoid statistical bias caused by the small sample data set, with which hundreds of artificial composers and philosophers were generated, influenced by the 7 names originally chosen. Upon defining indices for opposition, skewness and counter-dialectics, we confirmed the intuitive analysis of historians in that classical music evolved according to a master-apprentice tradition, while in philosophy changes were driven by opposition. Though these case studies were meant only to show the possibility of treating phenomena in humanities quantitatively, including a quantitative measure of concepts such as dialectics and opposition, the results are encouraging for further application of the approach presented here to many other areas, since it is entirely generic.

  13. The Quantitative Evaluation of the Clinical and Translational Science Awards (CTSA) Program Based on Science Mapping and Scientometric Analysis

    PubMed Central

    Zhang, Yin; Wang, Lei

    2013-01-01

    Abstract The Clinical and Translational Science Awards (CTSA) program is one of the most important initiatives in translational medical funding. The quantitative evaluation of the efficiency and performance of the CTSA program has a significant referential meaning for the decision making of global translational medical funding. Using science mapping and scientometric analytic tools, this study quantitatively analyzed the scientific articles funded by the CTSA program. The results of the study showed that the quantitative productivities of the CTSA program had a stable increase since 2008. In addition, the emerging trends of the research funded by the CTSA program covered clinical and basic medical research fields. The academic benefits from the CTSA program were assisting its members to build a robust academic home for the Clinical and Translational Science and to attract other financial support. This study provided a quantitative evaluation of the CTSA program based on science mapping and scientometric analysis. Further research is required to compare and optimize other quantitative methods and to integrate various research results. PMID:24330689

  14. The quantitative evaluation of the Clinical and Translational Science Awards (CTSA) program based on science mapping and scientometric analysis.

    PubMed

    Zhang, Yin; Wang, Lei; Diao, Tianxi

    2013-12-01

    The Clinical and Translational Science Awards (CTSA) program is one of the most important initiatives in translational medical funding. The quantitative evaluation of the efficiency and performance of the CTSA program has a significant referential meaning for the decision making of global translational medical funding. Using science mapping and scientometric analytic tools, this study quantitatively analyzed the scientific articles funded by the CTSA program. The results of the study showed that the quantitative productivities of the CTSA program had a stable increase since 2008. In addition, the emerging trends of the research funded by the CTSA program covered clinical and basic medical research fields. The academic benefits from the CTSA program were assisting its members to build a robust academic home for the Clinical and Translational Science and to attract other financial support. This study provided a quantitative evaluation of the CTSA program based on science mapping and scientometric analysis. Further research is required to compare and optimize other quantitative methods and to integrate various research results. © 2013 Wiley Periodicals, Inc.

  15. Computerized image analysis for quantitative neuronal phenotyping in zebrafish.

    PubMed

    Liu, Tianming; Lu, Jianfeng; Wang, Ye; Campbell, William A; Huang, Ling; Zhu, Jinmin; Xia, Weiming; Wong, Stephen T C

    2006-06-15

    An integrated microscope image analysis pipeline is developed for automatic analysis and quantification of phenotypes in zebrafish with altered expression of Alzheimer's disease (AD)-linked genes. We hypothesize that a slight impairment of neuronal integrity in a large number of zebrafish carrying the mutant genotype can be detected through the computerized image analysis method. Key functionalities of our zebrafish image processing pipeline include quantification of neuron loss in zebrafish embryos due to knockdown of AD-linked genes, automatic detection of defective somites, and quantitative measurement of gene expression levels in zebrafish with altered expression of AD-linked genes or treatment with a chemical compound. These quantitative measurements enable the archival of analyzed results and relevant meta-data. The structured database is organized for statistical analysis and data modeling to better understand neuronal integrity and phenotypic changes of zebrafish under different perturbations. Our results show that the computerized analysis is comparable to manual counting with equivalent accuracy and improved efficacy and consistency. Development of such an automated data analysis pipeline represents a significant step forward to achieve accurate and reproducible quantification of neuronal phenotypes in large scale or high-throughput zebrafish imaging studies.

  16. Cloning of DOG1, a quantitative trait locus controlling seed dormancy in Arabidopsis.

    PubMed

    Bentsink, Leónie; Jowett, Jemma; Hanhart, Corrie J; Koornneef, Maarten

    2006-11-07

    Genetic variation for seed dormancy in nature is a typical quantitative trait controlled by multiple loci on which environmental factors have a strong effect. Finding the genes underlying dormancy quantitative trait loci is a major scientific challenge, which also has relevance for agriculture and ecology. In this study we describe the identification of the DELAY OF GERMINATION 1 (DOG1) gene previously identified as a quantitative trait locus involved in the control of seed dormancy. This gene was isolated by a combination of positional cloning and mutant analysis and is absolutely required for the induction of seed dormancy. DOG1 is a member of a small gene family of unknown molecular function, with five members in Arabidopsis. The functional natural allelic variation present in Arabidopsis is caused by polymorphisms in the cis-regulatory region of the DOG1 gene and results in considerable expression differences between the DOG1 alleles of the accessions analyzed.

  17. Quantitative color measurement for black walnut wood.

    Treesearch

    Ali A. Moslemi

    1967-01-01

    Black walnut (Juglans nigra L.) veneer specimens with wide variations in color were evaluated by a quantitative method of color measurement. The internationally adopted CIE system of colorimetry was used to analyze the data. These data were converted to also show them in the Munsell system. Color differences among the walnut veneer specimens were also numerically...

  18. A gold nanoparticle-based semi-quantitative and quantitative ultrasensitive paper sensor for the detection of twenty mycotoxins

    NASA Astrophysics Data System (ADS)

    Kong, Dezhao; Liu, Liqiang; Song, Shanshan; Suryoprabowo, Steven; Li, Aike; Kuang, Hua; Wang, Libing; Xu, Chuanlai

    2016-02-01

    A semi-quantitative and quantitative multi-immunochromatographic (ICA) strip detection assay was developed for the simultaneous detection of twenty types of mycotoxins from five classes, including zearalenones (ZEAs), deoxynivalenols (DONs), T-2 toxins (T-2s), aflatoxins (AFs), and fumonisins (FBs), in cereal food samples. Sensitive and specific monoclonal antibodies were selected for this assay. The semi-quantitative results were obtained within 20 min by the naked eye, with visual limits of detection for ZEAs, DONs, T-2s, AFs and FBs of 0.1-0.5, 2.5-250, 0.5-1, 0.25-1 and 2.5-10 μg kg-1, and cut-off values of 0.25-1, 5-500, 1-10, 0.5-2.5 and 5-25 μg kg-1, respectively. The quantitative results were obtained using a hand-held strip scan reader, with the calculated limits of detection for ZEAs, DONs, T-2s, AFs and FBs of 0.04-0.17, 0.06-49, 0.15-0.22, 0.056-0.49 and 0.53-1.05 μg kg-1, respectively. The analytical results of spiked samples were in accordance with the accurate content in the simultaneous detection analysis. This newly developed ICA strip assay is suitable for the on-site detection and rapid initial screening of mycotoxins in cereal samples, facilitating both semi-quantitative and quantitative determination.A semi-quantitative and quantitative multi-immunochromatographic (ICA) strip detection assay was developed for the simultaneous detection of twenty types of mycotoxins from five classes, including zearalenones (ZEAs), deoxynivalenols (DONs), T-2 toxins (T-2s), aflatoxins (AFs), and fumonisins (FBs), in cereal food samples. Sensitive and specific monoclonal antibodies were selected for this assay. The semi-quantitative results were obtained within 20 min by the naked eye, with visual limits of detection for ZEAs, DONs, T-2s, AFs and FBs of 0.1-0.5, 2.5-250, 0.5-1, 0.25-1 and 2.5-10 μg kg-1, and cut-off values of 0.25-1, 5-500, 1-10, 0.5-2.5 and 5-25 μg kg-1, respectively. The quantitative results were obtained using a hand-held strip scan

  19. Current trends in quantitative proteomics - an update.

    PubMed

    Li, H; Han, J; Pan, J; Liu, T; Parker, C E; Borchers, C H

    2017-05-01

    Proteins can provide insights into biological processes at the functional level, so they are very promising biomarker candidates. The quantification of proteins in biological samples has been routinely used for the diagnosis of diseases and monitoring the treatment. Although large-scale protein quantification in complex samples is still a challenging task, a great amount of effort has been made to advance the technologies that enable quantitative proteomics. Seven years ago, in 2009, we wrote an article about the current trends in quantitative proteomics. In writing this current paper, we realized that, today, we have an even wider selection of potential tools for quantitative proteomics. These tools include new derivatization reagents, novel sampling formats, new types of analyzers and scanning techniques, and recently developed software to assist in assay development and data analysis. In this review article, we will discuss these innovative methods, and their current and potential applications in proteomics. Copyright © 2017 John Wiley & Sons, Ltd. Copyright © 2017 John Wiley & Sons, Ltd.

  20. Quantitative Method of Measuring Metastatic Activity

    NASA Technical Reports Server (NTRS)

    Morrison, Dennis R. (Inventor)

    1999-01-01

    The metastatic potential of tumors can be evaluated by the quantitative detection of urokinase and DNA. The cell sample selected for examination is analyzed for the presence of high levels of urokinase and abnormal DNA using analytical flow cytometry and digital image analysis. Other factors such as membrane associated uroldnase, increased DNA synthesis rates and certain receptors can be used in the method for detection of potentially invasive tumors.

  1. Counter tube window and X-ray fluorescence analyzer study

    NASA Technical Reports Server (NTRS)

    Hertel, R.; Holm, M.

    1973-01-01

    A study was performed to determine the best design tube window and X-ray fluorescence analyzer for quantitative analysis of Venusian dust and condensates. The principal objective of the project was to develop the best counter tube window geometry for the sensing element of the instrument. This included formulation of a mathematical model of the window and optimization of its parameters. The proposed detector and instrument has several important features. The instrument will perform a near real-time analysis of dust in the Venusian atmosphere, and is capable of measuring dust layers less than 1 micron thick. In addition, wide dynamic measurement range will be provided to compensate for extreme variations in count rates. An integral pulse-height analyzer and memory accumulate data and read out spectra for detail computer analysis on the ground.

  2. Quantitative assessment of RNA-protein interactions with high-throughput sequencing-RNA affinity profiling.

    PubMed

    Ozer, Abdullah; Tome, Jacob M; Friedman, Robin C; Gheba, Dan; Schroth, Gary P; Lis, John T

    2015-08-01

    Because RNA-protein interactions have a central role in a wide array of biological processes, methods that enable a quantitative assessment of these interactions in a high-throughput manner are in great demand. Recently, we developed the high-throughput sequencing-RNA affinity profiling (HiTS-RAP) assay that couples sequencing on an Illumina GAIIx genome analyzer with the quantitative assessment of protein-RNA interactions. This assay is able to analyze interactions between one or possibly several proteins with millions of different RNAs in a single experiment. We have successfully used HiTS-RAP to analyze interactions of the EGFP and negative elongation factor subunit E (NELF-E) proteins with their corresponding canonical and mutant RNA aptamers. Here we provide a detailed protocol for HiTS-RAP that can be completed in about a month (8 d hands-on time). This includes the preparation and testing of recombinant proteins and DNA templates, clustering DNA templates on a flowcell, HiTS and protein binding with a GAIIx instrument, and finally data analysis. We also highlight aspects of HiTS-RAP that can be further improved and points of comparison between HiTS-RAP and two other recently developed methods, quantitative analysis of RNA on a massively parallel array (RNA-MaP) and RNA Bind-n-Seq (RBNS), for quantitative analysis of RNA-protein interactions.

  3. Toward Quantitative Small Animal Pinhole SPECT: Assessment of Quantitation Accuracy Prior to Image Compensations

    PubMed Central

    Chen, Chia-Lin; Wang, Yuchuan; Lee, Jason J. S.; Tsui, Benjamin M. W.

    2011-01-01

    Purpose We assessed the quantitation accuracy of small animal pinhole single photon emission computed tomography (SPECT) under the current preclinical settings, where image compensations are not routinely applied. Procedures The effects of several common image-degrading factors and imaging parameters on quantitation accuracy were evaluated using Monte-Carlo simulation methods. Typical preclinical imaging configurations were modeled, and quantitative analyses were performed based on image reconstructions without compensating for attenuation, scatter, and limited system resolution. Results Using mouse-sized phantom studies as examples, attenuation effects alone degraded quantitation accuracy by up to −18% (Tc-99m or In-111) or −41% (I-125). The inclusion of scatter effects changed the above numbers to −12% (Tc-99m or In-111) and −21% (I-125), respectively, indicating the significance of scatter in quantitative I-125 imaging. Region-of-interest (ROI) definitions have greater impacts on regional quantitation accuracy for small sphere sources as compared to attenuation and scatter effects. For the same ROI, SPECT acquisitions using pinhole apertures of different sizes could significantly affect the outcome, whereas the use of different radii-of-rotation yielded negligible differences in quantitation accuracy for the imaging configurations simulated. Conclusions We have systematically quantified the influence of several factors affecting the quantitation accuracy of small animal pinhole SPECT. In order to consistently achieve accurate quantitation within 5% of the truth, comprehensive image compensation methods are needed. PMID:19048346

  4. A quantitative approach to painting styles

    NASA Astrophysics Data System (ADS)

    Vieira, Vilson; Fabbri, Renato; Sbrissa, David; da Fontoura Costa, Luciano; Travieso, Gonzalo

    2015-01-01

    This research extends a method previously applied to music and philosophy (Vilson Vieira et al., 2012), representing the evolution of art as a time-series where relations like dialectics are measured quantitatively. For that, a corpus of paintings of 12 well-known artists from baroque and modern art is analyzed. A set of 99 features is extracted and the features which most contributed to the classification of painters are selected. The projection space obtained provides the basis to the analysis of measurements. These quantitative measures underlie revealing observations about the evolution of painting styles, specially when compared with other humanity fields already analyzed: while music evolved along a master-apprentice tradition (high dialectics) and philosophy by opposition, painting presents another pattern: constant increasing skewness, low opposition between members of the same movement and opposition peaks in the transition between movements. Differences between baroque and modern movements are also observed in the projected "painting space": while baroque paintings are presented as an overlapped cluster, the modern paintings present minor overlapping and are disposed more widely in the projection than the baroque counterparts. This finding suggests that baroque painters shared aesthetics while modern painters tend to "break rules" and develop their own style.

  5. An open-source method to analyze optokinetic reflex responses in larval zebrafish.

    PubMed

    Scheetz, Seth D; Shao, Enhua; Zhou, Yangzhong; Cario, Clinton L; Bai, Qing; Burton, Edward A

    2018-01-01

    Optokinetic reflex (OKR) responses provide a convenient means to evaluate oculomotor, integrative and afferent visual function in larval zebrafish models, which are commonly used to elucidate molecular mechanisms underlying development, disease and repair of the vertebrate nervous system. We developed an open-source MATLAB-based solution for automated quantitative analysis of OKR responses in larval zebrafish. The package includes applications to: (i) generate sinusoidally-transformed animated grating patterns suitable for projection onto a cylindrical screen to elicit the OKR; (ii) determine and record the angular orientations of the eyes in each frame of a video recording showing the OKR response; and (iii) analyze angular orientation data from the tracking program to yield a set of parameters that quantify essential elements of the OKR. The method can be employed without modification using the operating manual provided. In addition, annotated source code is included, allowing users to modify or adapt the software for other applications. We validated the algorithms and measured OKR responses in normal larval zebrafish, showing good agreement with published quantitative data, where available. We provide the first open-source method to elicit and analyze the OKR in larval zebrafish. The wide range of parameters that are automatically quantified by our algorithms significantly expands the scope of quantitative analysis previously reported. Our method for quantifying OKR responses will be useful for numerous applications in neuroscience using the genetically- and chemically-tractable zebrafish model. Published by Elsevier B.V.

  6. Lorentz force particle analyzer

    NASA Astrophysics Data System (ADS)

    Wang, Xiaodong; Thess, André; Moreau, René; Tan, Yanqing; Dai, Shangjun; Tao, Zhen; Yang, Wenzhi; Wang, Bo

    2016-07-01

    A new contactless technique is presented for the detection of micron-sized insulating particles in the flow of an electrically conducting fluid. A transverse magnetic field brakes this flow and tends to become entrained in the flow direction by a Lorentz force, whose reaction force on the magnetic-field-generating system can be measured. The presence of insulating particles suspended in the fluid produce changes in this Lorentz force, generating pulses in it; these pulses enable the particles to be counted and sized. A two-dimensional numerical model that employs a moving mesh method demonstrates the measurement principle when such a particle is present. Two prototypes and a three-dimensional numerical model are used to demonstrate the feasibility of a Lorentz force particle analyzer (LFPA). The findings of this study conclude that such an LFPA, which offers contactless and on-line quantitative measurements, can be applied to an extensive range of applications. These applications include measurements of the cleanliness of high-temperature and aggressive molten metal, such as aluminum and steel alloys, and the clean manufacturing of semiconductors.

  7. Quantitative and Qualitative Analysis of Biomarkers in Fusarium verticillioides

    USDA-ARS?s Scientific Manuscript database

    In this study, a combination HPLC-DART-TOF-MS system was utilized to identify and quantitatively analyze carbohydrates in wild type and mutant strains of Fusarium verticillioides. Carbohydrate fractions were isolated from F. verticillioides cellular extracts by HPLC using a cation-exchange size-excl...

  8. In situ quantitation of ring-conjugated ethylenic lignin-units in spruce thermomechanical pulps by FT-Raman spectroscopy

    Treesearch

    Umesh Agarwal; Sally A. Ralph

    2003-01-01

    With the objective of using FT-Raman to quantitatively analyze ethylenic units in lignin in thermomechanical pulps (TMPs), coniferyl alcohol, coniferin, coniferaldehyde, and G-DHP lignin models were used to first demonstrate that the technique was fully capable of quantifying ring conjugated ethylenic units. Based on this result, the amount of ethylenic units in TMP...

  9. A Microfluidic Platform for High-Throughput Multiplexed Protein Quantitation

    PubMed Central

    Volpetti, Francesca; Garcia-Cordero, Jose; Maerkl, Sebastian J.

    2015-01-01

    We present a high-throughput microfluidic platform capable of quantitating up to 384 biomarkers in 4 distinct samples by immunoassay. The microfluidic device contains 384 unit cells, which can be individually programmed with pairs of capture and detection antibody. Samples are quantitated in each unit cell by four independent MITOMI detection areas, allowing four samples to be analyzed in parallel for a total of 1,536 assays per device. We show that the device can be pre-assembled and stored for weeks at elevated temperature and we performed proof-of-concept experiments simultaneously quantitating IL-6, IL-1β, TNF-α, PSA, and GFP. Finally, we show that the platform can be used to identify functional antibody combinations by screening 64 antibody combinations requiring up to 384 unique assays per device. PMID:25680117

  10. Evaluation of quantitative PCR measurement of bacterial colonization of epithelial cells.

    PubMed

    Schmidt, Marcin T; Olejnik-Schmidt, Agnieszka K; Myszka, Kamila; Borkowska, Monika; Grajek, Włodzimierz

    2010-01-01

    Microbial colonization is an important step in establishing pathogenic or probiotic relations to host cells and in biofilm formation on industrial or medical devices. The aim of this work was to verify the applicability of quantitative PCR (Real-Time PCR) to measure bacterial colonization of epithelial cells. Salmonella enterica and Caco-2 intestinal epithelial cell line was used as a model. To verify sensitivity of the assay a competition of the pathogen cells to probiotic microorganism was tested. The qPCR method was compared to plate count and radiolabel approach, which are well established techniques in this area of research. The three methods returned similar results. The best quantification accuracy had radiolabel method, followed by qPCR. The plate count results showed coefficient of variation two-times higher than this of qPCR. The quantitative PCR proved to be a reliable method for enumeration of microbes in colonization assay. It has several advantages that make it very useful in case of analyzing mixed populations, where several different species or even strains can be monitored at the same time.

  11. Quantitative interaction proteomics using mass spectrometry.

    PubMed

    Wepf, Alexander; Glatter, Timo; Schmidt, Alexander; Aebersold, Ruedi; Gstaiger, Matthias

    2009-03-01

    We present a mass spectrometry-based strategy for the absolute quantification of protein complex components isolated through affinity purification. We quantified bait proteins via isotope-labeled reference peptides corresponding to an affinity tag sequence and prey proteins by label-free correlational quantification using the precursor ion signal intensities of proteotypic peptides generated in reciprocal purifications. We used this method to quantitatively analyze interaction stoichiometries in the human protein phosphatase 2A network.

  12. Quantitative proteomic analysis for high-throughput screening of differential glycoproteins in hepatocellular carcinoma serum

    PubMed Central

    Gao, Hua-Jun; Chen, Ya-Jing; Zuo, Duo; Xiao, Ming-Ming; Li, Ying; Guo, Hua; Zhang, Ning; Chen, Rui-Bing

    2015-01-01

    Objective Hepatocellular carcinoma (HCC) is a leading cause of cancer-related deaths. Novel serum biomarkers are required to increase the sensitivity and specificity of serum screening for early HCC diagnosis. This study employed a quantitative proteomic strategy to analyze the differential expression of serum glycoproteins between HCC and normal control serum samples. Methods Lectin affinity chromatography (LAC) was used to enrich glycoproteins from the serum samples. Quantitative mass spectrometric analysis combined with stable isotope dimethyl labeling and 2D liquid chromatography (LC) separations were performed to examine the differential levels of the detected proteins between HCC and control serum samples. Western blot was used to analyze the differential expression levels of the three serum proteins. Results A total of 2,280 protein groups were identified in the serum samples from HCC patients by using the 2D LC-MS/MS method. Up to 36 proteins were up-regulated in the HCC serum, whereas 19 proteins were down-regulated. Three differential glycoproteins, namely, fibrinogen gamma chain (FGG), FOS-like antigen 2 (FOSL2), and α-1,6-mannosylglycoprotein 6-β-N-acetylglucosaminyltransferase B (MGAT5B) were validated by Western blot. All these three proteins were up-regulated in the HCC serum samples. Conclusion A quantitative glycoproteomic method was established and proven useful to determine potential novel biomarkers for HCC. PMID:26487969

  13. An integrative strategy for quantitative analysis of the N-glycoproteome in complex biological samples

    PubMed Central

    2014-01-01

    Background The complexity of protein glycosylation makes it difficult to characterize glycosylation patterns on a proteomic scale. In this study, we developed an integrated strategy for comparatively analyzing N-glycosylation/glycoproteins quantitatively from complex biological samples in a high-throughput manner. This strategy entailed separating and enriching glycopeptides/glycoproteins using lectin affinity chromatography, and then tandem labeling them with 18O/16O to generate a mass shift of 6 Da between the paired glycopeptides, and finally analyzing them with liquid chromatography-mass spectrometry (LC-MS) and the automatic quantitative method we developed based on Mascot Distiller. Results The accuracy and repeatability of this strategy were first verified using standard glycoproteins; linearity was maintained within a range of 1:10–10:1. The peptide concentration ratios obtained by the self-build quantitative method were similar to both the manually calculated and theoretical values, with a standard deviation (SD) of 0.023–0.186 for glycopeptides. The feasibility of the strategy was further confirmed with serum from hepatocellular carcinoma (HCC) patients and healthy individuals; the expression of 44 glycopeptides and 30 glycoproteins were significantly different between HCC patient and control serum. Conclusions This strategy is accurate, repeatable, and efficient, and may be a useful tool for identification of disease-related N-glycosylation/glycoprotein changes. PMID:24428921

  14. iTRAQ Quantitative Proteomic Comparison of Metastatic and Non-Metastatic Uveal Melanoma Tumors

    PubMed Central

    Crabb, John W.; Hu, Bo; Crabb, John S.; Triozzi, Pierre; Saunthararajah, Yogen; Singh, Arun D.

    2015-01-01

    Background Uveal melanoma is the most common malignancy of the adult eye. The overall mortality rate is high because this aggressive cancer often metastasizes before ophthalmic diagnosis. Quantitative proteomic analysis of primary metastasizing and non-metastasizing tumors was pursued for insights into mechanisms and biomarkers of uveal melanoma metastasis. Methods Eight metastatic and 7 non-metastatic human primary uveal melanoma tumors were analyzed by LC MS/MS iTRAQ technology with Bruch’s membrane/choroid complex from normal postmortem eyes as control tissue. Tryptic peptides from tumor and control proteins were labeled with iTRAQ tags, fractionated by cation exchange chromatography, and analyzed by LC MS/MS. Protein identification utilized the Mascot search engine and the human Uni-Prot/Swiss-Protein database with false discovery ≤ 1%; protein quantitation utilized the Mascot weighted average method. Proteins designated differentially expressed exhibited quantitative differences (p ≤ 0.05, t-test) in a training set of five metastatic and five non-metastatic tumors. Logistic regression models developed from the training set were used to classify the metastatic status of five independent tumors. Results Of 1644 proteins identified and quantified in 5 metastatic and 5 non-metastatic tumors, 12 proteins were found uniquely in ≥ 3 metastatic tumors, 28 were found significantly elevated and 30 significantly decreased only in metastatic tumors, and 31 were designated differentially expressed between metastatic and non-metastatic tumors. Logistic regression modeling of differentially expressed collagen alpha-3(VI) and heat shock protein beta-1 allowed correct prediction of metastasis status for each of five independent tumor specimens. Conclusions The present data provide new clues to molecular differences in metastatic and non-metastatic uveal melanoma tumors. While sample size is limited and validation required, the results support collagen alpha-3(VI) and

  15. Exploring discrepancies between quantitative validation results and the geomorphic plausibility of statistical landslide susceptibility maps

    NASA Astrophysics Data System (ADS)

    Steger, Stefan; Brenning, Alexander; Bell, Rainer; Petschko, Helene; Glade, Thomas

    2016-06-01

    Empirical models are frequently applied to produce landslide susceptibility maps for large areas. Subsequent quantitative validation results are routinely used as the primary criteria to infer the validity and applicability of the final maps or to select one of several models. This study hypothesizes that such direct deductions can be misleading. The main objective was to explore discrepancies between the predictive performance of a landslide susceptibility model and the geomorphic plausibility of subsequent landslide susceptibility maps while a particular emphasis was placed on the influence of incomplete landslide inventories on modelling and validation results. The study was conducted within the Flysch Zone of Lower Austria (1,354 km2) which is known to be highly susceptible to landslides of the slide-type movement. Sixteen susceptibility models were generated by applying two statistical classifiers (logistic regression and generalized additive model) and two machine learning techniques (random forest and support vector machine) separately for two landslide inventories of differing completeness and two predictor sets. The results were validated quantitatively by estimating the area under the receiver operating characteristic curve (AUROC) with single holdout and spatial cross-validation technique. The heuristic evaluation of the geomorphic plausibility of the final results was supported by findings of an exploratory data analysis, an estimation of odds ratios and an evaluation of the spatial structure of the final maps. The results showed that maps generated by different inventories, classifiers and predictors appeared differently while holdout validation revealed similar high predictive performances. Spatial cross-validation proved useful to expose spatially varying inconsistencies of the modelling results while additionally providing evidence for slightly overfitted machine learning-based models. However, the highest predictive performances were obtained for

  16. WE-FG-207B-12: Quantitative Evaluation of a Spectral CT Scanner in a Phantom Study: Results of Spectral Reconstructions

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Duan, X; Arbique, G; Guild, J

    Purpose: To evaluate the quantitative image quality of spectral reconstructions of phantom data from a spectral CT scanner. Methods: The spectral CT scanner (IQon Spectral CT, Philips Healthcare) is equipped with a dual-layer detector and generates conventional 80-140 kVp images and variety of spectral reconstructions, e.g., virtual monochromatic (VM) images, virtual non-contrast (VNC) images, iodine maps, and effective atomic number (Z) images. A cylindrical solid water phantom (Gammex 472, 33 cm diameter and 5 cm thick) with iodine (2.0-20.0 mg I/ml) and calcium (50-600 mg/ml) rod inserts was scanned at 120 kVp and 27 mGy CTDIvol. Spectral reconstructions were evaluatedmore » by comparing image measurements with theoretical values calculated from nominal rod compositions provided by the phantom manufacturer. The theoretical VNC was calculated using water and iodine basis material decomposition, and the theoretical Z was calculated using two common methods, the chemical formula method (Z1) and the dual-energy ratio method (Z2). Results: Beam-hardening-like artifacts between high-attenuation calcium rods (≥300 mg/ml, >800 HU) influenced quantitative measurements, so the quantitative analysis was only performed on iodine rods using the images from the scan with all the calcium rods removed. The CT numbers of the iodine rods in the VM images (50∼150 keV) were close to theoretical values with average difference of 2.4±6.9 HU. Compared with theoretical values, the average difference for iodine concentration, VNC CT number and effective Z of iodine rods were −0.10±0.38 mg/ml, −0.1±8.2 HU, 0.25±0.06 (Z1) and −0.23±0.07 (Z2). Conclusion: The results indicate that the spectral CT scanner generates quantitatively accurate spectral reconstructions at clinically relevant iodine concentrations. Beam-hardening-like artifacts still exist when high-attenuation objects are present and their impact on patient images needs further investigation. YY is an employee of

  17. A fully integrated standalone portable cavity ringdown breath acetone analyzer.

    PubMed

    Sun, Meixiu; Jiang, Chenyu; Gong, Zhiyong; Zhao, Xiaomeng; Chen, Zhuying; Wang, Zhennan; Kang, Meiling; Li, Yingxin; Wang, Chuji

    2015-09-01

    Breath analysis is a promising new technique for nonintrusive disease diagnosis and metabolic status monitoring. One challenging issue in using a breath biomarker for potential particular disease screening is to find a quantitative relationship between the concentration of the breath biomarker and clinical diagnostic parameters of the specific disease. In order to address this issue, we need a new instrument that is capable of conducting real-time, online breath analysis with high data throughput, so that a large scale of clinical test (more subjects) can be achieved in a short period of time. In this work, we report a fully integrated, standalone, portable analyzer based on the cavity ringdown spectroscopy technique for near-real time, online breath acetone measurements. The performance of the portable analyzer in measurements of breath acetone was interrogated and validated by using the certificated gas chromatography-mass spectrometry. The results show that this new analyzer is useful for reliable online (online introduction of a breath sample without pre-treatment) breath acetone analysis with high sensitivity (57 ppb) and high data throughput (one data per second). Subsequently, the validated breath analyzer was employed for acetone measurements in 119 human subjects under various situations. The instrument design, packaging, specifications, and future improvements were also described. From an optical ringdown cavity operated by the lab-set electronics reported previously to this fully integrated standalone new instrument, we have enabled a new scientific tool suited for large scales of breath acetone analysis and created an instrument platform that can even be adopted for study of other breath biomarkers by using different lasers and ringdown mirrors covering corresponding spectral fingerprints.

  18. A fully integrated standalone portable cavity ringdown breath acetone analyzer

    NASA Astrophysics Data System (ADS)

    Sun, Meixiu; Jiang, Chenyu; Gong, Zhiyong; Zhao, Xiaomeng; Chen, Zhuying; Wang, Zhennan; Kang, Meiling; Li, Yingxin; Wang, Chuji

    2015-09-01

    Breath analysis is a promising new technique for nonintrusive disease diagnosis and metabolic status monitoring. One challenging issue in using a breath biomarker for potential particular disease screening is to find a quantitative relationship between the concentration of the breath biomarker and clinical diagnostic parameters of the specific disease. In order to address this issue, we need a new instrument that is capable of conducting real-time, online breath analysis with high data throughput, so that a large scale of clinical test (more subjects) can be achieved in a short period of time. In this work, we report a fully integrated, standalone, portable analyzer based on the cavity ringdown spectroscopy technique for near-real time, online breath acetone measurements. The performance of the portable analyzer in measurements of breath acetone was interrogated and validated by using the certificated gas chromatography-mass spectrometry. The results show that this new analyzer is useful for reliable online (online introduction of a breath sample without pre-treatment) breath acetone analysis with high sensitivity (57 ppb) and high data throughput (one data per second). Subsequently, the validated breath analyzer was employed for acetone measurements in 119 human subjects under various situations. The instrument design, packaging, specifications, and future improvements were also described. From an optical ringdown cavity operated by the lab-set electronics reported previously to this fully integrated standalone new instrument, we have enabled a new scientific tool suited for large scales of breath acetone analysis and created an instrument platform that can even be adopted for study of other breath biomarkers by using different lasers and ringdown mirrors covering corresponding spectral fingerprints.

  19. Second-harmonic patterned polarization-analyzed reflection confocal microscope

    NASA Astrophysics Data System (ADS)

    Okoro, Chukwuemeka; Toussaint, Kimani C.

    2017-08-01

    We introduce the second-harmonic patterned polarization-analyzed reflection confocal (SPPARC) microscope-a multimodal imaging platform that integrates Mueller matrix polarimetry with reflection confocal and second-harmonic generation (SHG) microscopy. SPPARC microscopy provides label-free three-dimensional (3-D), SHG-patterned confocal images that lend themselves to spatially dependent, linear polarimetric analysis for extraction of rich polarization information based on the Mueller calculus. To demonstrate its capabilities, we use SPPARC microscopy to analyze both porcine tendon and ligament samples and find differences in both circular degree-of-polarization and depolarization parameters. Moreover, using the collagen-generated SHG signal as an endogenous counterstain, we show that the technique can be used to provide 3-D polarimetric information of the surrounding extrafibrillar matrix plus cells or EFMC region. The unique characteristics of SPPARC microscopy holds strong potential for it to more accurately and quantitatively describe microstructural changes in collagen-rich samples in three spatial dimensions.

  20. Development of iPad application "Postima" for quantitative analysis of the effects of manual therapy

    NASA Astrophysics Data System (ADS)

    Sugiyama, Naruhisa; Shirakawa, Tomohiro

    2017-07-01

    The technical difficulty of diagnosing joint misalignment and/or dysfunction by quantitative evaluation is commonly acknowledged among manual therapists. Usually, manual therapists make a diagnosis based on a combination of observing patient symptoms and performing physical examinations, both of which rely on subjective criteria and thus contain some uncertainty. We thus sought to investigate the correlations among posture, skeletal misalignment, and pain severity over the course of manual therapy treatment, and to explore the possibility of establishing objective criteria for diagnosis. For this purpose, we developed an iPad application that realizes the measurement of patients' postures and analyzes them quantitatively. We also discuss the results and effectiveness of the measurement and analysis.

  1. SERS quantitative urine creatinine measurement of human subject

    NASA Astrophysics Data System (ADS)

    Wang, Tsuei Lian; Chiang, Hui-hua K.; Lu, Hui-hsin; Hung, Yung-da

    2005-03-01

    SERS method for biomolecular analysis has several potentials and advantages over traditional biochemical approaches, including less specimen contact, non-destructive to specimen, and multiple components analysis. Urine is an easily available body fluid for monitoring the metabolites and renal function of human body. We developed surface-enhanced Raman scattering (SERS) technique using 50nm size gold colloidal particles for quantitative human urine creatinine measurements. This paper shows that SERS shifts of creatinine (104mg/dl) in artificial urine is from 1400cm-1 to 1500cm-1 which was analyzed for quantitative creatinine measurement. Ten human urine samples were obtained from ten healthy persons and analyzed by the SERS technique. Partial least square cross-validation (PLSCV) method was utilized to obtain the estimated creatinine concentration in clinically relevant (55.9mg/dl to 208mg/dl) concentration range. The root-mean square error of cross validation (RMSECV) is 26.1mg/dl. This research demonstrates the feasibility of using SERS for human subject urine creatinine detection, and establishes the SERS platform technique for bodily fluids measurement.

  2. Analyzing online sentiment to predict telephone poll results.

    PubMed

    Fu, King-wa; Chan, Chee-hon

    2013-09-01

    The telephone survey is a common social science research method for capturing public opinion, for example, an individual's values or attitudes, or the government's approval rating. However, reducing domestic landline usage, increasing nonresponse rate, and suffering from response bias of the interviewee's self-reported data pose methodological challenges to such an approach. Because of the labor cost of administration, a phone survey is often conducted on a biweekly or monthly basis, and therefore a daily reflection of public opinion is usually not available. Recently, online sentiment analysis of user-generated content has been deployed to predict public opinion and human behavior. However, its overall effectiveness remains uncertain. This study seeks to examine the temporal association between online sentiment reflected in social media content and phone survey poll results in Hong Kong. Specifically, it aims to find the extent to which online sentiment can predict phone survey results. Using autoregressive integrated moving average time-series analysis, this study suggested that online sentiment scores can lead phone survey results by about 8-15 days, and their correlation coefficients were about 0.16. The finding is significant to the study of social media in social science research, because it supports the conclusion that daily sentiment observed in social media content can serve as a leading predictor for phone survey results, keeping as much as 2 weeks ahead of the monthly announcement of opinion polls. We also discuss the practical and theoretical implications of this study.

  3. MetaFluxNet: the management of metabolic reaction information and quantitative metabolic flux analysis.

    PubMed

    Lee, Dong-Yup; Yun, Hongsoek; Park, Sunwon; Lee, Sang Yup

    2003-11-01

    MetaFluxNet is a program package for managing information on the metabolic reaction network and for quantitatively analyzing metabolic fluxes in an interactive and customized way. It allows users to interpret and examine metabolic behavior in response to genetic and/or environmental modifications. As a result, quantitative in silico simulations of metabolic pathways can be carried out to understand the metabolic status and to design the metabolic engineering strategies. The main features of the program include a well-developed model construction environment, user-friendly interface for metabolic flux analysis (MFA), comparative MFA of strains having different genotypes under various environmental conditions, and automated pathway layout creation. http://mbel.kaist.ac.kr/ A manual for MetaFluxNet is available as PDF file.

  4. Variable selection based near infrared spectroscopy quantitative and qualitative analysis on wheat wet gluten

    NASA Astrophysics Data System (ADS)

    Lü, Chengxu; Jiang, Xunpeng; Zhou, Xingfan; Zhang, Yinqiao; Zhang, Naiqian; Wei, Chongfeng; Mao, Wenhua

    2017-10-01

    Wet gluten is a useful quality indicator for wheat, and short wave near infrared spectroscopy (NIRS) is a high performance technique with the advantage of economic rapid and nondestructive test. To study the feasibility of short wave NIRS analyzing wet gluten directly from wheat seed, 54 representative wheat seed samples were collected and scanned by spectrometer. 8 spectral pretreatment method and genetic algorithm (GA) variable selection method were used to optimize analysis. Both quantitative and qualitative model of wet gluten were built by partial least squares regression and discriminate analysis. For quantitative analysis, normalization is the optimized pretreatment method, 17 wet gluten sensitive variables are selected by GA, and GA model performs a better result than that of all variable model, with R2V=0.88, and RMSEV=1.47. For qualitative analysis, automatic weighted least squares baseline is the optimized pretreatment method, all variable models perform better results than those of GA models. The correct classification rates of 3 class of <24%, 24-30%, >30% wet gluten content are 95.45, 84.52, and 90.00%, respectively. The short wave NIRS technique shows potential for both quantitative and qualitative analysis of wet gluten for wheat seed.

  5. Identification and quantitation of semi-crystalline microplastics using image analysis and differential scanning calorimetry.

    PubMed

    Rodríguez Chialanza, Mauricio; Sierra, Ignacio; Pérez Parada, Andrés; Fornaro, Laura

    2018-06-01

    There are several techniques used to analyze microplastics. These are often based on a combination of visual and spectroscopic techniques. Here we introduce an alternative workflow for identification and mass quantitation through a combination of optical microscopy with image analysis (IA) and differential scanning calorimetry (DSC). We studied four synthetic polymers with environmental concern: low and high density polyethylene (LDPE and HDPE, respectively), polypropylene (PP), and polyethylene terephthalate (PET). Selected experiments were conducted to investigate (i) particle characterization and counting procedures based on image analysis with open-source software, (ii) chemical identification of microplastics based on DSC signal processing, (iii) dependence of particle size on DSC signal, and (iv) quantitation of microplastics mass based on DSC signal. We describe the potential and limitations of these techniques to increase reliability for microplastic analysis. Particle size demonstrated to have particular incidence in the qualitative and quantitative performance of DSC signals. Both, identification (based on characteristic onset temperature) and mass quantitation (based on heat flow) showed to be affected by particle size. As a result, a proper sample treatment which includes sieving of suspended particles is particularly required for this analytical approach.

  6. Quantitative photogrammetric analysis of the Klapp method for treating idiopathic scoliosis.

    PubMed

    Iunes, Denise H; Cecílio, Maria B B; Dozza, Marina A; Almeida, Polyanna R

    2010-01-01

    Few studies have proved that physical therapy techniques are efficient in the treatment of scoliosis. To analyze the efficiency of the Klapp method for the treatment of scoliosis, through a quantitative analysis using computerized biophotogrammetry. Sixteen participants of a mean age of 15+/-2.61 yrs. with idiopathic scoliosis were treated using the Klapp method. To analyze the results from the treatment, they were all of photographed before and after the treatments, following a standardized photographic method. All of the photographs were analyzed quantitatively by the same examiner using the ALCimagem 2000 software. The statistical analyses were performed using the paired t-test with a significance level of 5%. The treatments showed improvements in the angles which evaluated the symmetry of the shoulders, i.e. the acromioclavicular joint angle (AJ; p=0.00) and sternoclavicular joint angle (SJ; p=0.01). There were also improvements in the angle that evaluated the left Thales triangle (DeltaT; p=0.02). Regarding flexibility, there were improvements in the tibiotarsal angle (TTA; p=0.01) and in the hip joint angles (HJA; p=0.00). There were no changes in the vertebral curvatures and nor improvements in head positioning. Only the lumbar curvature, evaluated by the lumbar lordosis angle (LL; p=0.00), changed after the treatments. The Klapp method was an efficient therapeutic technique for treating asymmetries of the trunk and improving its flexibility. However, it was not efficient for pelvic asymmetry modifications in head positioning, cervical lordosis or thoracic kyphosis.

  7. Quantitative description of yttrium aluminate ceramic composition by means of Er+3 microluminescence spectrum

    NASA Astrophysics Data System (ADS)

    Videla, F. A.; Tejerina, M. R.; Moreira-Osorio, L.; Conconi, M. S.; Orzi, D. J. O.; Flores, T.; Ponce, L. V.; Bilmes, G. M.; Torchia, G. A.

    2018-05-01

    The composition of erbium-doped yttrium aluminate ceramics was analyzed by means of confocal luminescence spectroscopy, EDX, and X-ray diffraction. A well-defined linear correlation was found between a proposed estimator computed from the luminescence spectrum and the proportion of ceramic phases coexisting in different samples. This result shows the feasibility of using erbium luminescence spectroscopy to perform a quantitative determination of different phases of yttrium aluminates within a micrometric region in nanograined ceramics.

  8. Tuition Reductions: A Quantitative Analysis of the Prevalence, Circumstances and Outcomes of an Emerging Pricing Strategy in Higher Education

    ERIC Educational Resources Information Center

    Kottich, Sarah

    2017-01-01

    This study analyzed tuition reductions in the private not-for-profit sector of higher education, utilizing a quantitative descriptive and correlational approach with secondary data analysis. It resulted in a listing of 45 institutions with verified tuition reductions from 2007 to 2017, more than previously thought. It found that the…

  9. Thromboelastography platelet mapping in healthy dogs using 1 analyzer versus 2 analyzers.

    PubMed

    Blois, Shauna L; Banerjee, Amrita; Wood, R Darren; Park, Fiona M

    2013-07-01

    The objective of this study was to describe the results of thromboelastography platelet mapping (TEG-PM) carried out using 2 techniques in 20 healthy dogs. Maximum amplitudes (MA) generated by thrombin (MAthrombin), fibrin (MAfibrin), adenosine diphosphate (ADP) receptor activity (MAADP), and thromboxane A2 (TxA2) receptor activity (stimulated by arachidonic acid, MAAA) were recorded. Thromboelastography platelet mapping was carried out according to the manufacturer's guidelines (2-analyzer technique) and using a variation of this method employing only 1 analyzer (1-analyzer technique) on 2 separate blood samples obtained from each dog. Mean [± standard deviation (SD)] MA values for the 1-analyzer/2-analyzer techniques were: MAthrombin = 51.9 mm (± 7.1)/52.5 mm (± 8.0); MAfibrin = 20.7 mm (± 21.8)/23.0 mm (± 26.1); MAADP = 44.5 mm (± 15.6)/45.6 mm (± 17.0); and MAAA = 45.7 mm (± 11.6)/45.0 mm (± 15.4). Mean (± SD) percentage aggregation due to ADP receptor activity was 70.4% (± 32.8)/67.6% (± 33.7). Mean percentage aggregation due to TxA2 receptor activity was 77.3% (± 31.6)/78.1% (± 50.2). Results of TEG-PM were not significantly different for the 1-analyzer and 2-analyzer methods. High correlation was found between the 2 methods for MAfibrin [concordance correlation coefficient (r) = 0.930]; moderate correlation was found for MAthrombin (r = 0.70) and MAADP (r = 0.57); correlation between the 2 methods for MAAA was lower (r = 0.32). Thromboelastography platelet mapping (TEG-PM) should be further investigated to determine if it is a suitable method for measuring platelet dysfunction in dogs with thrombopathy.

  10. Thromboelastography platelet mapping in healthy dogs using 1 analyzer versus 2 analyzers

    PubMed Central

    Blois, Shauna L.; Banerjee, Amrita; Wood, R. Darren; Park, Fiona M.

    2013-01-01

    The objective of this study was to describe the results of thromboelastography platelet mapping (TEG-PM) carried out using 2 techniques in 20 healthy dogs. Maximum amplitudes (MA) generated by thrombin (MAthrombin), fibrin (MAfibrin), adenosine diphosphate (ADP) receptor activity (MAADP), and thromboxane A2 (TxA2) receptor activity (stimulated by arachidonic acid, MAAA) were recorded. Thromboelastography platelet mapping was carried out according to the manufacturer’s guidelines (2-analyzer technique) and using a variation of this method employing only 1 analyzer (1-analyzer technique) on 2 separate blood samples obtained from each dog. Mean [± standard deviation (SD)] MA values for the 1-analyzer/2-analyzer techniques were: MAthrombin = 51.9 mm (± 7.1)/52.5 mm (± 8.0); MAfibrin = 20.7 mm (± 21.8)/23.0 mm (± 26.1); MAADP = 44.5 mm (± 15.6)/45.6 mm (± 17.0); and MAAA = 45.7 mm (± 11.6)/45.0 mm (± 15.4). Mean (± SD) percentage aggregation due to ADP receptor activity was 70.4% (± 32.8)/67.6% (± 33.7). Mean percentage aggregation due to TxA2 receptor activity was 77.3% (± 31.6)/78.1% (± 50.2). Results of TEG-PM were not significantly different for the 1-analyzer and 2-analyzer methods. High correlation was found between the 2 methods for MAfibrin [concordance correlation coefficient (r) = 0.930]; moderate correlation was found for MAthrombin (r = 0.70) and MAADP (r = 0.57); correlation between the 2 methods for MAAA was lower (r = 0.32). Thromboelastography platelet mapping (TEG-PM) should be further investigated to determine if it is a suitable method for measuring platelet dysfunction in dogs with thrombopathy. PMID:24101802

  11. Quantitative genetics

    USDA-ARS?s Scientific Manuscript database

    The majority of economically important traits targeted for cotton improvement are quantitatively inherited. In this chapter, the current state of cotton quantitative genetics is described and separated into four components. These components include: 1) traditional quantitative inheritance analysis, ...

  12. BiQ Analyzer HT: locus-specific analysis of DNA methylation by high-throughput bisulfite sequencing

    PubMed Central

    Lutsik, Pavlo; Feuerbach, Lars; Arand, Julia; Lengauer, Thomas; Walter, Jörn; Bock, Christoph

    2011-01-01

    Bisulfite sequencing is a widely used method for measuring DNA methylation in eukaryotic genomes. The assay provides single-base pair resolution and, given sufficient sequencing depth, its quantitative accuracy is excellent. High-throughput sequencing of bisulfite-converted DNA can be applied either genome wide or targeted to a defined set of genomic loci (e.g. using locus-specific PCR primers or DNA capture probes). Here, we describe BiQ Analyzer HT (http://biq-analyzer-ht.bioinf.mpi-inf.mpg.de/), a user-friendly software tool that supports locus-specific analysis and visualization of high-throughput bisulfite sequencing data. The software facilitates the shift from time-consuming clonal bisulfite sequencing to the more quantitative and cost-efficient use of high-throughput sequencing for studying locus-specific DNA methylation patterns. In addition, it is useful for locus-specific visualization of genome-wide bisulfite sequencing data. PMID:21565797

  13. Legionella in water samples: how can you interpret the results obtained by quantitative PCR?

    PubMed

    Ditommaso, Savina; Ricciardi, Elisa; Giacomuzzi, Monica; Arauco Rivera, Susan R; Zotti, Carla M

    2015-02-01

    Evaluation of the potential risk associated with Legionella has traditionally been determined from culture-based methods. Quantitative polymerase chain reaction (qPCR) is an alternative tool that offers rapid, sensitive and specific detection of Legionella in environmental water samples. In this study we compare the results obtained by conventional qPCR (iQ-Check™ Quanti Legionella spp.; Bio-Rad) and by culture method on artificial samples prepared in Page's saline by addiction of Legionella pneumophila serogroup 1 (ATCC 33152) and we analyse the selective quantification of viable Legionella cells by the qPCR-PMA method. The amount of Legionella DNA (GU) determined by qPCR was 28-fold higher than the load detected by culture (CFU). Applying the qPCR combined with PMA treatment we obtained a reduction of 98.5% of the qPCR signal from dead cells. We observed a dissimilarity in the ability of PMA to suppress the PCR signal in samples with different amounts of bacteria: the effective elimination of detection signals by PMA depended on the concentration of GU and increasing amounts of cells resulted in higher values of reduction. Using the results from this study we created an algorithm to facilitate the interpretation of viable cell level estimation with qPCR-PMA. Copyright © 2014 Elsevier Ltd. All rights reserved.

  14. Sieve-based device for MALDI sample preparation. III. Its power for quantitative measurements.

    PubMed

    Molin, Laura; Cristoni, Simone; Seraglia, Roberta; Traldi, Pietro

    2011-02-01

    The solid sample inhomogeneity is a weak point of traditional MALDI deposition techniques that reflects negatively on quantitative analysis. The recently developed sieve-based device (SBD) sample deposition method, based on the electrospraying of matrix/analyte solutions through a grounded sieve, allows the homogeneous deposition of microcrystals with dimensions smaller than that of the laser spot. In each microcrystal the matrix/analyte molar ratio can be considered constant. Then, by irradiating different portions of the microcrystal distribution an identical response is obtained. This result suggests the employment of SBD in the development of quantitative procedures. For this aim, mixtures of different proteins of known molarity were analyzed, showing a good relationship between molarity and intensity ratios. This behaviour was also observed in the case of proteins with quite different ionic yields. The power of the developed method for quantitative evaluation was also tested by the measurement of the abundance of IGPP[Oxi]GPP[Oxi]GLMGPP (m/z 1219) present in the collagen-α-5(IV) chain precursor, differently expressed in urines from healthy subjects and diabetic-nephropathic patients, confirming its overexpression in the presence of nephropathy. The data obtained indicate that SBD is a particularly effective method for quantitative analysis also in biological fluids of interest. Copyright © 2011 John Wiley & Sons, Ltd.

  15. Quantitative versus qualitative cultures of respiratory secretions for clinical outcomes in patients with ventilator-associated pneumonia.

    PubMed

    Berton, Danilo Cortozi; Kalil, Andre C; Cavalcanti, Manuela; Teixeira, Paulo José Zimermann

    2008-10-08

    Ventilator-associated pneumonia (VAP) is a common infectious disease in intensive care units (ICUs). The best diagnostic approach to resolve this condition remains uncertain. To evaluate whether quantitative cultures of respiratory secretions are effective in reducing mortality in immunocompetent patients with VAP, compared with qualitative cultures. We also considered changes in antibiotic use, length of ICU stay and mechanical ventilation. We searched the Cochrane Central Register of Controlled Trials (CENTRAL) (The Cochrane Library 2007, issue 4), which contains the Acute Respiratory Infections Group's Specialized Register; MEDLINE (1966 to December 2007); EMBASE (1974 to December 2007); and LILACS (1982 to December 2007). Randomized controlled trials (RCTs) comparing respiratory samples processed quantitatively or qualitatively, obtained by invasive or non-invasive methods from immunocompetent patients with VAP, and which analyzed the impact of these methods on antibiotic use and mortality rates. Two review authors independently reviewed and selected trials from the search results, and assessed studies for suitability, methodology and quality. We analyzed data using Review Manager software. We pooled the included studies to yield the risk ratio (RR) for mortality and antibiotic change with 95% confidence intervals (CI). Of the 3931 references identified from the electronic databases, five RCTs (1367 patients) met the inclusion criteria. Three studies compared invasive methods using quantitative cultures versus non-invasive methods using qualitative cultures, and were used to answer the main objective of this review. The other two studies compared invasive versus non-invasive methods, both using quantitative cultures. All five studies were combined to compare invasive versus non-invasive interventions for diagnosing VAP. The studies that compared quantitative and qualitative cultures (1240 patients) showed no statistically significant differences in mortality

  16. Development of an SRM method for absolute quantitation of MYDGF/C19orf10 protein.

    PubMed

    Dwivedi, Ravi C; Krokhin, Oleg V; El-Gabalawy, Hani S; Wilkins, John A

    2016-06-01

    To develop a MS-based selected reaction monitoring (SRM) assay for quantitation of myeloid-derived growth factor (MYDGF) formerly chromosome 19 open reading frame (C19orf10). Candidate reporter peptides were identified in digests of recombinant MYDGF. Isotopically labeled forms of these reporter peptides were employed as internal standards for assay development. Two reference peptides were selected SYLYFQTFFK and GAEIEYAMAYSK with respective LOQ of 42 and 380 attomole per injection. Application of the assay to human serum and synovial fluid determined that the assay sensitivity was reduced and quantitation was not achievable. However, the partial depletion of albumin and immunoglobulin from synovial fluids provided estimates of 300-650 femtomoles per injection (0.7-1.6 nanomolar (nM) fluid concentrations) in three of the six samples analyzed. A validated sensitive assay for the quantitation of MYDGF in biological fluids was developed. However, the endogenous levels of MYDGF in such fluids are at or below the current levels of quantitation. The levels of MYDGF are lower than those previously reported using an ELISA. The current results suggest that additional steps may be required to remove high abundance proteins or to enrich MYDGF for SRM-based quantitation. © 2015 WILEY-VCH Verlag GmbH & Co. KGaA, Weinheim.

  17. Results from E ∥B Neutral Particle Analyzer and Calibration Ion Beam System on C-2U

    NASA Astrophysics Data System (ADS)

    Clary, Ryan; Roquemore, A.; Kolmogorov, A.; Ivanov, A.; Korepanov, S.; Magee, R.; Medley, S.; Smirnov, A.; Tiunov, M.; TAE Team

    2015-11-01

    C-2U is a a high-confinement, advanced beam driven FRC which aims to sustain the configuration for > 5 ms, in excess of typical MHD and fast particle instability times, as well as fast particle slowing down times. Fast particle dynamics are critical to C-2U performance and several diagnostics have been deployed to characterize the fast particle population, including neutron and proton detectors, an electrostatic neutral particle analyzer, and neutral particle bolometers. To increase our understanding of fast particle behavior and supplement existing diagnostics an E ∥B NPA was acquired from PPPL which simultaneously measures H0 and D0 flux between 2 and 22 keV with high energy resolution. In addition, a small, high purity, ion beam system has been constructed and tested to calibrate absolutely fast particle detectors. Here we report results of measurements from the E ∥B analyzer on C-2U and inferred fast particle behavior, as well as the status of the calibration ion beam system.

  18. Quantitative generalizations for catchment sediment yield following forest logging

    Treesearch

    James C. Bathurst; Andrés Iroumé

    2014-01-01

    Published data for temperate forests across the world are analyzed to investigate the potential for generalized quantitative expressions of catchment sediment yield impact in the years immediately following logging. Such generalizations would be useful in a variety of forestry and engineering tasks and would aid the spread of knowledge amongst both relevant...

  19. Results of the First Mars Organic Molecule Analyzer (MOMA) GC-MS Coupling

    NASA Astrophysics Data System (ADS)

    Buch, Arnaud; Pinnick, Veronica; Szopa, Cyril; Danell, Ryan; Grand, Noel; Van Amerom, Friso; Glavin, Daniel; Freissinet, Caroline; Humeau, Olivier; Coll, Patrice; Arevalo, Ricardo; Stalport, Fabien; Brinckerhoff, William; Steininger, Harald; Goesmann, Fred; Mahaffy, Paul; Raulin, Francois

    2014-11-01

    The Mars Organic Molecule Analyzer (MOMA) aboard the ExoMars rover will be a key analytical tool in providing chemical (molecular) information from the solid samples collected by the rover, with a particular focus on the char-acterization of the organic content. The core of the MOMA instrument is a gas chromatograph coupled with a mass spectrometer (GC-MS) which provides the unique capability to characterize a broad range of compounds, including both of volatile and non-volatile species. Samples will be crushed and deposited into sample cups seated in a rotating carousel. Soil samples will be analyzed either by UV laser desorption / ionization (LDI) or pyrolysis gas chromatography ion trap mass spectrometry (pyr-GC-ITMS).The French GC brassboard was coupled to the US ion trap mass spectrometer brassboard in a flight-like con-figuration for several coupling campains. The MOMA GC setup is based on the SAM heritage design with a He reservoir and 4 separate analytical modules including traps, columns and Thermal Conductivity Detectors. Solid samples are sealed and heated in this setup using a manual tapping station, designed and built at MPS in Germany, for GC-MS analysis. The gaseous species eluting from the GC are then ionized by an electron impact ionization source in the MS chamber and analyzed by the linear ion trap mass spectrometer. Volatile and non-volatile compounds were injected in the MOMA instrumental suite. Both of these compounds classes were detected by the TCD and by the MS. MS signal (total ion current) and single mass spectra by comparison with the NIST library, gave us an unambiguous confirmation of these identifications. The mass spectra arise from an average of 10 mass spectra averaged around a given time point in the total ion chromatogram.Based on commercial instrument, the MOMA requirement for sensitivity in the GC-MS mode for organic molecules is 1 pmol. In this test, sensitivity was determined for the GC TCD and MS response to a dilution

  20. Laboratory evolution of the migratory polymorphism in the sand cricket: combining physiology with quantitative genetics.

    PubMed

    Roff, Derek A; Fairbairn, Daphne J

    2007-01-01

    Predicting evolutionary change is the central goal of evolutionary biology because it is the primary means by which we can test evolutionary hypotheses. In this article, we analyze the pattern of evolutionary change in a laboratory population of the wing-dimorphic sand cricket Gryllus firmus resulting from relaxation of selection favoring the migratory (long-winged) morph. Based on a well-characterized trade-off between fecundity and flight capability, we predict that evolution in the laboratory environment should result in a reduction in the proportion of long-winged morphs. We also predict increased fecundity and reduced functionality and weight of the major flight muscles in long-winged females but little change in short-winged (flightless) females. Based on quantitative genetic theory, we predict that the regression equation describing the trade-off between ovary weight and weight of the major flight muscles will show a change in its intercept but not in its slope. Comparisons across generations verify all of these predictions. Further, using values of genetic parameters estimated from previous studies, we show that a quantitative genetic simulation model can account for not only the qualitative changes but also the evolutionary trajectory. These results demonstrate the power of combining quantitative genetic and physiological approaches for understanding the evolution of complex traits.

  1. Validating internal controls for quantitative plant gene expression studies.

    PubMed

    Brunner, Amy M; Yakovlev, Igor A; Strauss, Steven H

    2004-08-18

    Real-time reverse transcription PCR (RT-PCR) has greatly improved the ease and sensitivity of quantitative gene expression studies. However, accurate measurement of gene expression with this method relies on the choice of a valid reference for data normalization. Studies rarely verify that gene expression levels for reference genes are adequately consistent among the samples used, nor compare alternative genes to assess which are most reliable for the experimental conditions analyzed. Using real-time RT-PCR to study the expression of 10 poplar (genus Populus) housekeeping genes, we demonstrate a simple method for determining the degree of stability of gene expression over a set of experimental conditions. Based on a traditional method for analyzing the stability of varieties in plant breeding, it defines measures of gene expression stability from analysis of variance (ANOVA) and linear regression. We found that the potential internal control genes differed widely in their expression stability over the different tissues, developmental stages and environmental conditions studied. Our results support that quantitative comparisons of candidate reference genes are an important part of real-time RT-PCR studies that seek to precisely evaluate variation in gene expression. The method we demonstrated facilitates statistical and graphical evaluation of gene expression stability. Selection of the best reference gene for a given set of experimental conditions should enable detection of biologically significant changes in gene expression that are too small to be revealed by less precise methods, or when highly variable reference genes are unknowingly used in real-time RT-PCR experiments.

  2. Quantitative Analysis of the Usage of the COSMOS Science Education Portal

    NASA Astrophysics Data System (ADS)

    Sotiriou, Sofoklis; Bogner, Franz X.; Neofotistos, George

    2011-08-01

    A quantitative method of mapping the web usage of an innovative educational portal is applied to analyze the behaviour of users of the COSMOS Science Education Portal. The COSMOS Portal contains user-generated resources (that are uploaded by its users). It has been designed to support a science teacher's search, retrieval and access to both, scientific and educational resources. It also aims to introduce in and familiarize teachers with an innovative methodology for designing, expressing and representing educational practices in a commonly understandable way through the use of user-friendly authoring tools that are available through the portal. As a new science education portal that includes user-generated content, the COSMOS Portal encounters the well-known "new product/service challenge": to convince the users to use its tools, which facilitate quite fast lesson planning and lesson preparation activities. To respond to this challenge, the COSMOS Portal operators implemented a validation process by analyzing the usage data of the portal in a 10 month time-period. The data analyzed comprised: (a) the temporal evolution of the number of contributors and the amount of content uploaded to the COSMOS Portal; (b) the number of portal visitors (categorized as all-visitors, new-visitors, and returning-visitors) and (c) visitor loyalty parameters (such as page-views; pages/visit; average time on site; depth of visit; length of visit). The data is augmented with data associated with the usage context (e.g. the time of day when most of the activities in the portal take place). The quantitative results indicate that the exponential growth of the contributors to the COSMOS Portal is followed by an exponential growth of the uploaded content. Furthermore, the web usage statistics demonstrate significant changes in users' behaviour during the period under study, with returning visitors using the COSMOS Portal more frequently, mainly for lesson planning and preparation (in the

  3. Results from using a new dyadic-dependence model to analyze sociocentric physician networks.

    PubMed

    Paul, Sudeshna; Keating, Nancy L; Landon, Bruce E; O'Malley, A James

    2014-09-01

    Professional physician networks can potentially influence clinical practices and quality of care. With the current focus on coordinated care, discerning influences of naturally occurring clusters and other forms of dependence among physicians' relationships based on their attributes and care patterns is an important area of research. In this paper, two directed physician networks: a physician influential conversation network (N = 33) and a physician network obtained from patient visit data (N = 135) are analyzed using a new model that accounts for effect modification of the within-dyad effect of reciprocity and inter-dyad effects involving three (or more) actors. The results from this model include more nuanced effects involving reciprocity and triadic dependence than under incumbent models and more flexible control for these effects in the extraction of other network phenomena, including the relationship between similarity of individuals' attributes (e.g., same-gender, same residency location) and tie-status. In both cases we find extensive evidence of clustering and triadic dependence that if not accounted for confounds the effect of reciprocity and attribute homophily. Findings from our analysis suggest alternative conclusions to those from incumbent models. Copyright © 2014 Elsevier Ltd. All rights reserved.

  4. Development of a relational database to capture and merge clinical history with the quantitative results of radionuclide renography.

    PubMed

    Folks, Russell D; Savir-Baruch, Bital; Garcia, Ernest V; Verdes, Liudmila; Taylor, Andrew T

    2012-12-01

    Our objective was to design and implement a clinical history database capable of linking to our database of quantitative results from (99m)Tc-mercaptoacetyltriglycine (MAG3) renal scans and export a data summary for physicians or our software decision support system. For database development, we used a commercial program. Additional software was developed in Interactive Data Language. MAG3 studies were processed using an in-house enhancement of a commercial program. The relational database has 3 parts: a list of all renal scans (the RENAL database), a set of patients with quantitative processing results (the Q2 database), and a subset of patients from Q2 containing clinical data manually transcribed from the hospital information system (the CLINICAL database). To test interobserver variability, a second physician transcriber reviewed 50 randomly selected patients in the hospital information system and tabulated 2 clinical data items: hydronephrosis and presence of a current stent. The CLINICAL database was developed in stages and contains 342 fields comprising demographic information, clinical history, and findings from up to 11 radiologic procedures. A scripted algorithm is used to reliably match records present in both Q2 and CLINICAL. An Interactive Data Language program then combines data from the 2 databases into an XML (extensible markup language) file for use by the decision support system. A text file is constructed and saved for review by physicians. RENAL contains 2,222 records, Q2 contains 456 records, and CLINICAL contains 152 records. The interobserver variability testing found a 95% match between the 2 observers for presence or absence of ureteral stent (κ = 0.52), a 75% match for hydronephrosis based on narrative summaries of hospitalizations and clinical visits (κ = 0.41), and a 92% match for hydronephrosis based on the imaging report (κ = 0.84). We have developed a relational database system to integrate the quantitative results of MAG3 image

  5. Progress in Quantitative Viral Load Testing: Variability and Impact of the WHO Quantitative International Standards

    PubMed Central

    Sun, Y.; Tang, L.; Procop, G. W.; Hillyard, D. R.; Young, S. A.; Caliendo, A. M.

    2016-01-01

    ABSTRACT It has been hoped that the recent availability of WHO quantitative standards would improve interlaboratory agreement for viral load testing; however, insufficient data are available to evaluate whether this has been the case. Results from 554 laboratories participating in proficiency testing surveys for quantitative PCR assays of cytomegalovirus (CMV), Epstein-Barr virus (EBV), BK virus (BKV), adenovirus (ADV), and human herpesvirus 6 (HHV6) were evaluated to determine overall result variability and then were stratified by assay manufacturer. The impact of calibration to international units/ml (CMV and EBV) on variability was also determined. Viral loads showed a high degree of interlaboratory variability for all tested viruses, with interquartile ranges as high as 1.46 log10 copies/ml and the overall range for a given sample up to 5.66 log10 copies/ml. Some improvement in result variability was seen when international units were adopted. This was particularly the case for EBV viral load results. Variability in viral load results remains a challenge across all viruses tested here; introduction of international quantitative standards may help reduce variability and does so more or less markedly for certain viruses. PMID:27852673

  6. The new AP Physics exams: Integrating qualitative and quantitative reasoning

    NASA Astrophysics Data System (ADS)

    Elby, Andrew

    2015-04-01

    When physics instructors and education researchers emphasize the importance of integrating qualitative and quantitative reasoning in problem solving, they usually mean using those types of reasoning serially and separately: first students should analyze the physical situation qualitatively/conceptually to figure out the relevant equations, then they should process those equations quantitatively to generate a solution, and finally they should use qualitative reasoning to check that answer for plausibility (Heller, Keith, & Anderson, 1992). The new AP Physics 1 and 2 exams will, of course, reward this approach to problem solving. But one kind of free response question will demand and reward a further integration of qualitative and quantitative reasoning, namely mathematical modeling and sense-making--inventing new equations to capture a physical situation and focusing on proportionalities, inverse proportionalities, and other functional relations to infer what the equation ``says'' about the physical world. In this talk, I discuss examples of these qualitative-quantitative translation questions, highlighting how they differ from both standard quantitative and standard qualitative questions. I then discuss the kinds of modeling activities that can help AP and college students develop these skills and habits of mind.

  7. Analysis of ribosomal RNA stability in dead cells of wine yeast by quantitative PCR.

    PubMed

    Sunyer-Figueres, Merce; Wang, Chunxiao; Mas, Albert

    2018-04-02

    During wine production, some yeasts enter a Viable But Not Culturable (VBNC) state, which may influence the quality and stability of the final wine through remnant metabolic activity or by resuscitation. Culture-independent techniques are used for obtaining an accurate estimation of the number of live cells, and quantitative PCR could be the most accurate technique. As a marker of cell viability, rRNA was evaluated by analyzing its stability in dead cells. The species-specific stability of rRNA was tested in Saccharomyces cerevisiae, as well as in three species of non-Saccharomyces yeast (Hanseniaspora uvarum, Torulaspora delbrueckii and Starmerella bacillaris). High temperature and antimicrobial dimethyl dicarbonate (DMDC) treatments were efficient in lysing the yeast cells. rRNA gene and rRNA (as cDNA) were analyzed over 48 h after cell lysis by quantitative PCR. The results confirmed the stability of rRNA for 48 h after the cell lysis treatments. To sum up, rRNA may not be a good marker of cell viability in the wine yeasts that were tested. Copyright © 2018 Elsevier B.V. All rights reserved.

  8. A gas-phase chemiluminescence-based analyzer for waterborne arsenic

    USGS Publications Warehouse

    Idowu, A.D.; Dasgupta, P.K.; Genfa, Z.; Toda, K.; Garbarino, J.R.

    2006-01-01

    We show a practical sequential injection/zone fluidics-based analyzer that measures waterborne arsenic. The approach is capable of differentiating between inorganic As(III) and As(V). The principle is based on generating AsH 3 from the sample in a confined chamber by borohydride reduction at controlled pH, sparging the chamber to drive the AsH3 to a small reflective cell located atop a photomultiplier tube, allowing it to react with ozone generated from ambient air, and measuring the intense chemiluminescence that results. Arsine generation and removal from solution results in isolation from the sample matrix, avoiding the pitfalls encountered in some solution-based analysis techniques. The differential determination of As(III) and As(V) is based on the different pH dependence of the reducibility of these species to AsH3. At pH ???1, both As(III) and As(V) are quantitatively converted to arsine in the presence of NaBH4. At a pH of 4-5, only As(III) is converted to arsine. In the present form, the limit of detection (S/N = 3) is 0.05 ??g/L As at pH ???1 and 0.09 ??g/L As(III) at pH ???4-5 for a 3-mL sample. The analyzer is intrinsically automated and requires 4 min per determination. It is also possible to determine As(III) first at pH 4.5 and then determine the remaining As in a sequential manner; this requires 6 min. There are no significant practical interferences. A new borohydride solution formulation permits month-long reagent stability. ?? 2006 American Chemical Society.

  9. Miniaturization of Fresnel lenses for solar concentration: a quantitative investigation.

    PubMed

    Duerr, Fabian; Meuret, Youri; Thienpont, Hugo

    2010-04-20

    Sizing down the dimensions of solar concentrators for photovoltaic applications offers a number of promising advantages. It provides thinner modules and smaller solar cells, which reduces thermal issues. In this work a plane Fresnel lens design is introduced that is first analyzed with geometrical optics. Because of miniaturization, pure ray tracing may no longer be valid to determine the concentration performance. Therefore, a quantitative wave optical analysis of the miniaturization's influence on the obtained concentration performance is presented. This better quantitative understanding of the impact of diffraction in microstructured Fresnel lenses might help to optimize the design of several applications in nonimaging optics.

  10. Reader's Response: Describing and Analyzing Quantitative Data

    ERIC Educational Resources Information Center

    McGrath, April L.

    2013-01-01

    The work of Harris and Martin (2012) on student motivations for choosing to complete online courses provides information on an important area of development within post-secondary education. As noted by the authors, online learning is an expanding field and learning more about why students choose online courses and their experiences in such courses…

  11. The Interaction Affinity between Vascular Cell Adhesion Molecule-1 (VCAM-1) and Very Late Antigen-4 (VLA-4) Analyzed by Quantitative FRET

    PubMed Central

    Wu, Shu-Han; Karmenyan, Artashes; Chiou, Arthur

    2015-01-01

    Very late antigen-4 (VLA-4), a member of integrin superfamily, interacts with its major counter ligand vascular cell adhesion molecule-1 (VCAM-1) and plays an important role in leukocyte adhesion to vascular endothelium and immunological synapse formation. However, irregular expressions of these proteins may also lead to several autoimmune diseases and metastasis cancer. Thus, quantifying the interaction affinity of the VCAM-1/VLA-4 interaction is of fundamental importance in further understanding the nature of this interaction and drug discovery. In this study, we report an ‘in solution’ steady state organic fluorophore based quantitative fluorescence resonance energy transfer (FRET) assay to quantify this interaction in terms of the dissociation constant (Kd). We have used, in our FRET assay, the Alexa Fluor 488-VLA-4 conjugate as the donor, and Alexa Fluor 546-VCAM-1 as the acceptor. From the FRET signal analysis, Kd of this interaction was determined to be 41.82 ± 2.36 nM. To further confirm our estimation, we have employed surface plasmon resonance (SPR) technique to obtain Kd = 39.60 ± 1.78 nM, which is in good agreement with the result obtained by FRET. This is the first reported work which applies organic fluorophore based ‘in solution’ simple quantitative FRET assay to obtain the dissociation constant of the VCAM-1/VLA-4 interaction, and is also the first quantification of this interaction. Moreover, the value of Kd can serve as an indicator of abnormal protein-protein interactions; hence, this assay can potentially be further developed into a drug screening platform of VLA-4/VCAM-1 as well as other protein-ligand interactions. PMID:25793408

  12. Visual and Quantitative Analysis Methods of Respiratory Patterns for Respiratory Gated PET/CT.

    PubMed

    Son, Hye Joo; Jeong, Young Jin; Yoon, Hyun Jin; Park, Jong-Hwan; Kang, Do-Young

    2016-01-01

    We integrated visual and quantitative methods for analyzing the stability of respiration using four methods: phase space diagrams, Fourier spectra, Poincaré maps, and Lyapunov exponents. Respiratory patterns of 139 patients were grouped based on the combination of the regularity of amplitude, period, and baseline positions. Visual grading was done by inspecting the shape of diagram and classified into two states: regular and irregular. Quantitation was done by measuring standard deviation of x and v coordinates of Poincaré map (SD x , SD v ) or the height of the fundamental peak ( A 1 ) in Fourier spectrum or calculating the difference between maximal upward and downward drift. Each group showed characteristic pattern on visual analysis. There was difference of quantitative parameters (SD x , SD v , A 1 , and MUD-MDD) among four groups (one way ANOVA, p = 0.0001 for MUD-MDD, SD x , and SD v , p = 0.0002 for A 1 ). In ROC analysis, the cutoff values were 0.11 for SD x (AUC: 0.982, p < 0.0001), 0.062 for SD v (AUC: 0.847, p < 0.0001), 0.117 for A 1 (AUC: 0.876, p < 0.0001), and 0.349 for MUD-MDD (AUC: 0.948, p < 0.0001). This is the first study to analyze multiple aspects of respiration using various mathematical constructs and provides quantitative indices of respiratory stability and determining quantitative cutoff value for differentiating regular and irregular respiration.

  13. PatternLab for proteomics 4.0: A one-stop shop for analyzing shotgun proteomic data

    PubMed Central

    Carvalho, Paulo C; Lima, Diogo B; Leprevost, Felipe V; Santos, Marlon D M; Fischer, Juliana S G; Aquino, Priscila F; Moresco, James J; Yates, John R; Barbosa, Valmir C

    2017-01-01

    PatternLab for proteomics is an integrated computational environment that unifies several previously published modules for analyzing shotgun proteomic data. PatternLab contains modules for formatting sequence databases, performing peptide spectrum matching, statistically filtering and organizing shotgun proteomic data, extracting quantitative information from label-free and chemically labeled data, performing statistics for differential proteomics, displaying results in a variety of graphical formats, performing similarity-driven studies with de novo sequencing data, analyzing time-course experiments, and helping with the understanding of the biological significance of data in the light of the Gene Ontology. Here we describe PatternLab for proteomics 4.0, which closely knits together all of these modules in a self-contained environment, covering the principal aspects of proteomic data analysis as a freely available and easily installable software package. All updates to PatternLab, as well as all new features added to it, have been tested over the years on millions of mass spectra. PMID:26658470

  14. Statistical Model to Analyze Quantitative Proteomics Data Obtained by 18O/16O Labeling and Linear Ion Trap Mass Spectrometry

    PubMed Central

    Jorge, Inmaculada; Navarro, Pedro; Martínez-Acedo, Pablo; Núñez, Estefanía; Serrano, Horacio; Alfranca, Arántzazu; Redondo, Juan Miguel; Vázquez, Jesús

    2009-01-01

    Statistical models for the analysis of protein expression changes by stable isotope labeling are still poorly developed, particularly for data obtained by 16O/18O labeling. Besides large scale test experiments to validate the null hypothesis are lacking. Although the study of mechanisms underlying biological actions promoted by vascular endothelial growth factor (VEGF) on endothelial cells is of considerable interest, quantitative proteomics studies on this subject are scarce and have been performed after exposing cells to the factor for long periods of time. In this work we present the largest quantitative proteomics study to date on the short term effects of VEGF on human umbilical vein endothelial cells by 18O/16O labeling. Current statistical models based on normality and variance homogeneity were found unsuitable to describe the null hypothesis in a large scale test experiment performed on these cells, producing false expression changes. A random effects model was developed including four different sources of variance at the spectrum-fitting, scan, peptide, and protein levels. With the new model the number of outliers at scan and peptide levels was negligible in three large scale experiments, and only one false protein expression change was observed in the test experiment among more than 1000 proteins. The new model allowed the detection of significant protein expression changes upon VEGF stimulation for 4 and 8 h. The consistency of the changes observed at 4 h was confirmed by a replica at a smaller scale and further validated by Western blot analysis of some proteins. Most of the observed changes have not been described previously and are consistent with a pattern of protein expression that dynamically changes over time following the evolution of the angiogenic response. With this statistical model the 18O labeling approach emerges as a very promising and robust alternative to perform quantitative proteomics studies at a depth of several thousand proteins

  15. Analyzing Big Data in Psychology: A Split/Analyze/Meta-Analyze Approach

    PubMed Central

    Cheung, Mike W.-L.; Jak, Suzanne

    2016-01-01

    Big data is a field that has traditionally been dominated by disciplines such as computer science and business, where mainly data-driven analyses have been performed. Psychology, a discipline in which a strong emphasis is placed on behavioral theories and empirical research, has the potential to contribute greatly to the big data movement. However, one challenge to psychologists—and probably the most crucial one—is that most researchers may not have the necessary programming and computational skills to analyze big data. In this study we argue that psychologists can also conduct big data research and that, rather than trying to acquire new programming and computational skills, they should focus on their strengths, such as performing psychometric analyses and testing theories using multivariate analyses to explain phenomena. We propose a split/analyze/meta-analyze approach that allows psychologists to easily analyze big data. Two real datasets are used to demonstrate the proposed procedures in R. A new research agenda related to the analysis of big data in psychology is outlined at the end of the study. PMID:27242639

  16. Analyzing Big Data in Psychology: A Split/Analyze/Meta-Analyze Approach.

    PubMed

    Cheung, Mike W-L; Jak, Suzanne

    2016-01-01

    Big data is a field that has traditionally been dominated by disciplines such as computer science and business, where mainly data-driven analyses have been performed. Psychology, a discipline in which a strong emphasis is placed on behavioral theories and empirical research, has the potential to contribute greatly to the big data movement. However, one challenge to psychologists-and probably the most crucial one-is that most researchers may not have the necessary programming and computational skills to analyze big data. In this study we argue that psychologists can also conduct big data research and that, rather than trying to acquire new programming and computational skills, they should focus on their strengths, such as performing psychometric analyses and testing theories using multivariate analyses to explain phenomena. We propose a split/analyze/meta-analyze approach that allows psychologists to easily analyze big data. Two real datasets are used to demonstrate the proposed procedures in R. A new research agenda related to the analysis of big data in psychology is outlined at the end of the study.

  17. An ultra-wideband microwave tomography system: preliminary results.

    PubMed

    Gilmore, Colin; Mojabi, Puyan; Zakaria, Amer; Ostadrahimi, Majid; Kaye, Cam; Noghanian, Sima; Shafai, Lotfollah; Pistorius, Stephen; LoVetri, Joe

    2009-01-01

    We describe a 2D wide-band multi-frequency microwave imaging system intended for biomedical imaging. The system is capable of collecting data from 2-10 GHz, with 24 antenna elements connected to a vector network analyzer via a 2 x 24 port matrix switch. Through the use of two different nonlinear reconstruction schemes: the Multiplicative-Regularized Contrast Source Inversion method and an enhanced version of the Distorted Born Iterative Method, we show preliminary imaging results from dielectric phantoms where data were collected from 3-6 GHz. The early inversion results show that the system is capable of quantitatively reconstructing dielectric objects.

  18. Analysis of vaginal microbicide film hydration kinetics by quantitative imaging refractometry.

    PubMed

    Rinehart, Matthew; Grab, Sheila; Rohan, Lisa; Katz, David; Wax, Adam

    2014-01-01

    We have developed a quantitative imaging refractometry technique, based on holographic phase microscopy, as a tool for investigating microscopic structural changes in water-soluble polymeric materials. Here we apply the approach to analyze the structural degradation of vaginal topical microbicide films due to water uptake. We implemented transmission imaging of 1-mm diameter film samples loaded into a flow chamber with a 1.5×2 mm field of view. After water was flooded into the chamber, interference images were captured and analyzed to obtain high resolution maps of the local refractive index and subsequently the volume fraction and mass density of film material at each spatial location. Here, we compare the hydration dynamics of a panel of films with varying thicknesses and polymer compositions, demonstrating that quantitative imaging refractometry can be an effective tool for evaluating and characterizing the performance of candidate microbicide film designs for anti-HIV drug delivery.

  19. Oxygen effects on glucose measurements with a reference analyzer and three handheld meters.

    PubMed

    Tang, Z; Louie, R F; Payes, M; Chang, K C; Kost, G J

    2000-01-01

    Oxygen may affect glucose meter and reference analyzer measurements. We evaluated the effects of changes in blood oxygen tension (Po2) on Accu-Chek Comfort Curve (Roche Diagnostics, Indianapolis, IN), Precision G, (Abbott Laboratories, Bedford, MA) and One Touch II (Lifescan, Milpitas, CA) glucose meter measurements, and on Yellow Springs Instruments (YSI) (Yellow Springs, OH) reference analyzer measurements. Venous blood drawn from healthy volunteers was adjusted to three glucose levels of 80, 200, and 400 mg/dL, each tonometered with six different Po2 levels (40, 80, 160, 240, 320, and 400 torr). To quantitate oxygen effects on reference analyzer measurements, glucose differences between test sample (Po2 changed) and control (Po2 80 torr) were calculated (YSItest-YSIcontrol). The threshold for determination of oxygen effects was +/-2 SD, where 2 SD was fro

  20. Generalization of the van der Pauw Method: Analyzing Longitudinal Magnetoresistance Asymmetry to Quantify Doping Gradients

    NASA Astrophysics Data System (ADS)

    Grayson, M.; Zhou, Wang; Yoo, Heun-Mo; Prabhu-Gaunkar, S.; Tiemann, L.; Reichl, C.; Wegscheider, W.

    A longitudinal magnetoresistance asymmetry (LMA) between a positive and negative magnetic field is known to occur in both the extreme quantum limit and the classical Drude limit in samples with a nonuniform doping density. By analyzing the current stream function in van der Pauw measurement geometry, it is shown that the electron density gradient can be quantitatively deduced from this LMA in the Drude regime. Results agree with gradients interpolated from local densities calibrated across an entire wafer, establishing a generalization of the van der Pauw method to quantify density gradients. Results will be shown of various semoconductor systems where this method is applied, from bulk doped semiconductors, to exfoliated 2D materials. McCormick Catalyst Award from Northwestern University, EECS Bridge Funding, and AFOSR FA9550-15-1-0247.

  1. Experimental design and data-analysis in label-free quantitative LC/MS proteomics: A tutorial with MSqRob.

    PubMed

    Goeminne, Ludger J E; Gevaert, Kris; Clement, Lieven

    2018-01-16

    Label-free shotgun proteomics is routinely used to assess proteomes. However, extracting relevant information from the massive amounts of generated data remains difficult. This tutorial provides a strong foundation on analysis of quantitative proteomics data. We provide key statistical concepts that help researchers to design proteomics experiments and we showcase how to analyze quantitative proteomics data using our recent free and open-source R package MSqRob, which was developed to implement the peptide-level robust ridge regression method for relative protein quantification described by Goeminne et al. MSqRob can handle virtually any experimental proteomics design and outputs proteins ordered by statistical significance. Moreover, its graphical user interface and interactive diagnostic plots provide easy inspection and also detection of anomalies in the data and flaws in the data analysis, allowing deeper assessment of the validity of results and a critical review of the experimental design. Our tutorial discusses interactive preprocessing, data analysis and visualization of label-free MS-based quantitative proteomics experiments with simple and more complex designs. We provide well-documented scripts to run analyses in bash mode on GitHub, enabling the integration of MSqRob in automated pipelines on cluster environments (https://github.com/statOmics/MSqRob). The concepts outlined in this tutorial aid in designing better experiments and analyzing the resulting data more appropriately. The two case studies using the MSqRob graphical user interface will contribute to a wider adaptation of advanced peptide-based models, resulting in higher quality data analysis workflows and more reproducible results in the proteomics community. We also provide well-documented scripts for experienced users that aim at automating MSqRob on cluster environments. Copyright © 2017 Elsevier B.V. All rights reserved.

  2. Go Figure! Using Quantitative Measures to Enhance Program and Student Success

    ERIC Educational Resources Information Center

    Frost, Leanne H.; Braun, Gwendolyn K.

    2006-01-01

    Using quantitative assessment, Montana State University-Billings substantially improved and expanded its developmental education program and learning center during the past five years. Student-centered questions drove the research efforts. By gathering, analyzing and sharing hard data, the department identified unmet student needs, discovered…

  3. Dynamically monitoring the gene expression of dual fluorophore in the cell cycle with quantitative spectrum analysis

    NASA Astrophysics Data System (ADS)

    Lee, Ja-Yun; Wu, Tzong-Yuan; Hsu, I.-Jen

    2008-04-01

    The cloning and transcription techniques on gene cloned fluorescent proteins have been widely used in many applications. They have been used as reporters of some conditions in a series of reactions. However, it is usually difficult to monitor the specific target with the exactly number of proteins during the process in turbid media, especially at micrometer scales. We successfully revealed an alternative way to monitor the cell cycle behavior and quantitatively analyzed the target cells with green and red fluorescent proteins (GFP and RFP) during different phases of the cell cycle by quantitatively analyzing its behavior and also monitoring its spatial distribution.

  4. Quantitative imaging methods in osteoporosis.

    PubMed

    Oei, Ling; Koromani, Fjorda; Rivadeneira, Fernando; Zillikens, M Carola; Oei, Edwin H G

    2016-12-01

    Osteoporosis is characterized by a decreased bone mass and quality resulting in an increased fracture risk. Quantitative imaging methods are critical in the diagnosis and follow-up of treatment effects in osteoporosis. Prior radiographic vertebral fractures and bone mineral density (BMD) as a quantitative parameter derived from dual-energy X-ray absorptiometry (DXA) are among the strongest known predictors of future osteoporotic fractures. Therefore, current clinical decision making relies heavily on accurate assessment of these imaging features. Further, novel quantitative techniques are being developed to appraise additional characteristics of osteoporosis including three-dimensional bone architecture with quantitative computed tomography (QCT). Dedicated high-resolution (HR) CT equipment is available to enhance image quality. At the other end of the spectrum, by utilizing post-processing techniques such as the trabecular bone score (TBS) information on three-dimensional architecture can be derived from DXA images. Further developments in magnetic resonance imaging (MRI) seem promising to not only capture bone micro-architecture but also characterize processes at the molecular level. This review provides an overview of various quantitative imaging techniques based on different radiological modalities utilized in clinical osteoporosis care and research.

  5. Quantitative proteomics in biological research.

    PubMed

    Wilm, Matthias

    2009-10-01

    Proteomics has enabled the direct investigation of biological material, at first through the analysis of individual proteins, then of lysates from cell cultures, and finally of extracts from tissues and biopsies from entire organisms. Its latest manifestation - quantitative proteomics - allows deeper insight into biological systems. This article reviews the different methods used to extract quantitative information from mass spectra. It follows the technical developments aimed toward global proteomics, the attempt to characterize every expressed protein in a cell by at least one peptide. When applications of the technology are discussed, the focus is placed on yeast biology. In particular, differential quantitative proteomics, the comparison between an experiment and its control, is very discriminating for proteins involved in the process being studied. When trying to understand biological processes on a molecular level, differential quantitative proteomics tends to give a clearer picture than global transcription analyses. As a result, MS has become an even more indispensable tool for biochemically motivated biological research.

  6. Quantitative impact characterization of aeronautical CFRP materials with non-destructive testing methods

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Kiefel, Denis, E-mail: Denis.Kiefel@airbus.com, E-mail: Rainer.Stoessel@airbus.com; Stoessel, Rainer, E-mail: Denis.Kiefel@airbus.com, E-mail: Rainer.Stoessel@airbus.com; Grosse, Christian, E-mail: Grosse@tum.de

    2015-03-31

    In recent years, an increasing number of safety-relevant structures are designed and manufactured from carbon fiber reinforced polymers (CFRP) in order to reduce weight of airplanes by taking the advantage of their specific strength into account. Non-destructive testing (NDT) methods for quantitative defect analysis of damages are liquid- or air-coupled ultrasonic testing (UT), phased array ultrasonic techniques, and active thermography (IR). The advantage of these testing methods is the applicability on large areas. However, their quantitative information is often limited on impact localization and size. In addition to these techniques, Airbus Group Innovations operates a micro x-ray computed tomography (μ-XCT)more » system, which was developed for CFRP characterization. It is an open system which allows different kinds of acquisition, reconstruction, and data evaluation. One main advantage of this μ-XCT system is its high resolution with 3-dimensional analysis and visualization opportunities, which enables to gain important quantitative information for composite part design and stress analysis. Within this study, different NDT methods will be compared at CFRP samples with specified artificial impact damages. The results can be used to select the most suitable NDT-method for specific application cases. Furthermore, novel evaluation and visualization methods for impact analyzes are developed and will be presented.« less

  7. A Reproducible Computerized Method for Quantitation of Capillary Density using Nailfold Capillaroscopy.

    PubMed

    Cheng, Cynthia; Lee, Chadd W; Daskalakis, Constantine

    2015-10-27

    Capillaroscopy is a non-invasive, efficient, relatively inexpensive and easy to learn methodology for directly visualizing the microcirculation. The capillaroscopy technique can provide insight into a patient's microvascular health, leading to a variety of potentially valuable dermatologic, ophthalmologic, rheumatologic and cardiovascular clinical applications. In addition, tumor growth may be dependent on angiogenesis, which can be quantitated by measuring microvessel density within the tumor. However, there is currently little to no standardization of techniques, and only one publication to date reports the reliability of a currently available, complex computer based algorithms for quantitating capillaroscopy data.(1) This paper describes a new, simpler, reliable, standardized capillary counting algorithm for quantitating nailfold capillaroscopy data. A simple, reproducible computerized capillaroscopy algorithm such as this would facilitate more widespread use of the technique among researchers and clinicians. Many researchers currently analyze capillaroscopy images by hand, promoting user fatigue and subjectivity of the results. This paper describes a novel, easy-to-use automated image processing algorithm in addition to a reproducible, semi-automated counting algorithm. This algorithm enables analysis of images in minutes while reducing subjectivity; only a minimal amount of training time (in our experience, less than 1 hr) is needed to learn the technique.

  8. A Reproducible Computerized Method for Quantitation of Capillary Density using Nailfold Capillaroscopy

    PubMed Central

    Daskalakis, Constantine

    2015-01-01

    Capillaroscopy is a non-invasive, efficient, relatively inexpensive and easy to learn methodology for directly visualizing the microcirculation. The capillaroscopy technique can provide insight into a patient’s microvascular health, leading to a variety of potentially valuable dermatologic, ophthalmologic, rheumatologic and cardiovascular clinical applications. In addition, tumor growth may be dependent on angiogenesis, which can be quantitated by measuring microvessel density within the tumor. However, there is currently little to no standardization of techniques, and only one publication to date reports the reliability of a currently available, complex computer based algorithms for quantitating capillaroscopy data.1 This paper describes a new, simpler, reliable, standardized capillary counting algorithm for quantitating nailfold capillaroscopy data. A simple, reproducible computerized capillaroscopy algorithm such as this would facilitate more widespread use of the technique among researchers and clinicians. Many researchers currently analyze capillaroscopy images by hand, promoting user fatigue and subjectivity of the results. This paper describes a novel, easy-to-use automated image processing algorithm in addition to a reproducible, semi-automated counting algorithm. This algorithm enables analysis of images in minutes while reducing subjectivity; only a minimal amount of training time (in our experience, less than 1 hr) is needed to learn the technique. PMID:26554744

  9. Quantitative analysis of the Ge concentration in a SiGe quantum well: comparison of low-energy RBS and SIMS measurements.

    PubMed

    Krecar, D; Rosner, M; Draxler, M; Bauer, P; Hutter, H

    2006-01-01

    The germanium concentration and the position and thickness of the quantum well in molecular beam epitaxy (MBE)-grown SiGe were quantitatively analyzed via low-energy Rutherford backscattering (RBS) and secondary ion mass spectrometry (SIMS). In these samples, the concentrations of Si and Ge were assumed to be constant, except for the quantum well, where the germanium concentration was lower. The thickness of the analyzed quantum well was about 12 nm and it was situated at a depth of about 60 nm below the surface. A dip showed up in the RBS spectra due to the lower germanium concentration in the quantum well, and this was evaluated. Good depth resolution was required in order to obtain quantitative results, and this was obtained by choosing a primary energy of 500 keV and a tilt angle of 51 degrees with respect to the surface normal. Quantitative information was deduced from the raw data by comparing it with SIMNRA simulated spectra. The SIMS measurements were performed with oxygen primary ions. Given the response function of the SIMS instrument (the SIMS depth profile of the germanium delta (delta) layer), and using the forward convolution (point-to-point convolution) model, it is possible to determine the germanium concentration and the thickness of the analyzed quantum well from the raw SIMS data. The aim of this work was to compare the results obtained via RBS and SIMS and to show their potential for use in the semiconductor and microelectronics industry. The detection of trace elements (here the doping element antimony) that could not be evaluated with RBS in low-energy mode is also demonstrated using SIMS instead.

  10. Bayes` theorem and quantitative risk assessment

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Kaplan, S.

    1994-12-31

    This paper argues that for a quantitative risk analysis (QRA) to be useful for public and private decision making, and for rallying the support necessary to implement those decisions, it is necessary that the QRA results be ``trustable.`` Trustable means that the results are based solidly and logically on all the relevant evidence available. This, in turn, means that the quantitative results must be derived from the evidence using Bayes` theorem. Thus, it argues that one should strive to make their QRAs more clearly and explicitly Bayesian, and in this way make them more ``evidence dependent`` than ``personality dependent.``

  11. Quantitative perceptual differences among over-the-counter vaginal products using a standardized methodology: implications for microbicide development☆

    PubMed Central

    Mahan, Ellen D.; Morrow, Kathleen M.; Hayes, John E.

    2015-01-01

    Background Increasing prevalence of HIV infection among women worldwide has motivated the development of female-initiated prevention methods, including gel-based microbicides. User acceptability is vital for microbicide success; however, varying cultural vaginal practices indicate multiple formulations must be developed to appeal to different populations. Perceptual attributes of microbicides have been identified as primary drivers of acceptability; however, previous studies do not allow for direct comparison of these qualities between multiple formulations. Study Design Six vaginal products were analyzed ex vivo using descriptive analysis. Perceptual attributes of samples were identified by trained participants (n=10) and rated quantitatively using scales based on a panel-developed lexicon. Data were analyzed using two-way ANOVAs for each attribute; product differences were assessed via Tukey’s honestly significant difference test. Results Significant differences were found between products for multiple attributes. Patterns were also seen for attributes across intended product usage (i.e., contraceptive, moisturizer or lubricant). For example, Options© Gynol II® (Caldwell Consumer Health, LLC) was significantly stickier and grainier than other products. Conclusions Descriptive analysis, a quantitative approach that is based on consensus lexicon usage among participants, successfully quantified perceptual differences among vaginal products. Since perceptual attributes of products can be directly compared quantitatively, this study represents a novel approach that could be used to inform rational design of microbicides. PMID:21757061

  12. Quantitative comparison of cognitive behavioral therapy and music therapy research: a methodological best-practices analysis to guide future investigation for adult psychiatric patients.

    PubMed

    Silverman, Michael J

    2008-01-01

    While the music therapy profession is relatively young and small in size, it can treat a variety of clinical populations and has established a diverse research base. However, although the profession originated working with persons diagnosed with mental illnesses, there is a considerable lack of quantitative research concerning the effects of music therapy with this population. Music therapy clinicians and researchers have reported on this lack of evidence and the difficulty in conducting psychosocial research on their interventions (Choi, 1997; Silverman, 2003a). While published studies have provided suggestions for future research, no studies have provided detailed propositions for the methodology and design of meticulous high quality randomized controlled psychiatric music therapy research. How do other psychotherapies accomplish their databases and could the music therapy field borrow from their rigorous "methodological best practices" to strengthen its own literature base? Therefore, as the National Institutes of Mental Health state the treatment of choice for evidence-based psychotherapy is cognitive behavioral therapy (CBT), aspects of this psychotherapy's literature base were analyzed. The purpose of this literature analysis was to (a) analyze and identify components of high-quality quantitative CBT research for adult psychiatric consumers, (b) analyze and identify the variables and other elements of existing quantitative psychiatric music therapy research for adult consumers, and (c) compare the two data sets to identify the best methodological designs and variables for future quantitative music therapy research with the mental health population. A table analyzing randomized and thoroughly controlled studies involving the use of CBT for persons with severe mental illnesses is included to determine chief components of high-quality experimental research designs and implementation of quantitative clinical research. The table also shows the same analyzed

  13. Localization and quantitative co-localization of enamelin with amelogenin.

    PubMed

    Gallon, Victoria; Chen, Lisha; Yang, Xiudong; Moradian-Oldak, Janet

    2013-08-01

    Enamelin and amelogenin are vital proteins in enamel formation. The cooperative function of these two proteins controls crystal nucleation and morphology in vitro. We quantitatively analyzed the co-localization between enamelin and amelogenin by confocal microscopy and using two antibodies, one raised against a sequence in the porcine 32 kDa enamelin region and the other raised against full-length recombinant mouse amelogenin. We further investigated the interaction of the porcine 32 kDa enamelin and recombinant amelogenin using immuno-gold labeling. This study reports the quantitative co-localization results for postnatal days 1-8 mandibular mouse molars. We show that amelogenin and enamelin are secreted into the extracellular matrix on the cuspal slopes of the molars at day 1 and that secretion continues to at least day 8. Quantitative co-localization analysis (QCA) was performed in several different configurations using large (45 μm height, 33 μm width) and small (7 μm diameter) regions of interest to elucidate any patterns. Co-localization patterns in day 8 samples revealed that enamelin and amelogenin co-localize near the secretory face of the ameloblasts and appear to be secreted approximately in a 1:1 ratio. The degree of co-localization decreases as the enamel matures, both along the secretory face of ameloblasts and throughout the entire thickness of the enamel. Immuno-reactivity against enamelin is concentrated along the secretory face of ameloblasts, supporting the theory that this protein together with amelogenin is intimately involved in mineral induction at the beginning of enamel formation. Copyright © 2013 Elsevier Inc. All rights reserved.

  14. Evidences of local adaptation in quantitative traits in Prosopis alba (Leguminosae).

    PubMed

    Bessega, C; Pometti, C; Ewens, M; Saidman, B O; Vilardi, J C

    2015-02-01

    Signals of selection on quantitative traits can be detected by the comparison between the genetic differentiation of molecular (neutral) markers and quantitative traits, by multivariate extensions of the same model and by the observation of the additive covariance among relatives. We studied, by three different tests, signals of occurrence of selection in Prosopis alba populations over 15 quantitative traits: three economically important life history traits: height, basal diameter and biomass, 11 leaf morphology traits that may be related with heat-tolerance and physiological responses and spine length that is very important from silvicultural purposes. We analyzed 172 G1-generation trees growing in a common garden belonging to 32 open pollinated families from eight sampling sites in Argentina. The multivariate phenotypes differ significantly among origins, and the highest differentiation corresponded to foliar traits. Molecular genetic markers (SSR) exhibited significant differentiation and allowed us to provide convincing evidence that natural selection is responsible for the patterns of morphological differentiation. The heterogeneous selection over phenotypic traits observed suggested different optima in each population and has important implications for gene resource management. The results suggest that the adaptive significance of traits should be considered together with population provenance in breeding program as a crucial point prior to any selecting program, especially in Prosopis where the first steps are under development.

  15. Quantitative analysis of RNA-protein interactions on a massively parallel array for mapping biophysical and evolutionary landscapes

    PubMed Central

    Buenrostro, Jason D.; Chircus, Lauren M.; Araya, Carlos L.; Layton, Curtis J.; Chang, Howard Y.; Snyder, Michael P.; Greenleaf, William J.

    2015-01-01

    RNA-protein interactions drive fundamental biological processes and are targets for molecular engineering, yet quantitative and comprehensive understanding of the sequence determinants of affinity remains limited. Here we repurpose a high-throughput sequencing instrument to quantitatively measure binding and dissociation of MS2 coat protein to >107 RNA targets generated on a flow-cell surface by in situ transcription and inter-molecular tethering of RNA to DNA. We decompose the binding energy contributions from primary and secondary RNA structure, finding that differences in affinity are often driven by sequence-specific changes in association rates. By analyzing the biophysical constraints and modeling mutational paths describing the molecular evolution of MS2 from low- to high-affinity hairpins, we quantify widespread molecular epistasis, and a long-hypothesized structure-dependent preference for G:U base pairs over C:A intermediates in evolutionary trajectories. Our results suggest that quantitative analysis of RNA on a massively parallel array (RNAMaP) relationships across molecular variants. PMID:24727714

  16. Analysis of Vaginal Microbicide Film Hydration Kinetics by Quantitative Imaging Refractometry

    PubMed Central

    Rinehart, Matthew; Grab, Sheila; Rohan, Lisa; Katz, David; Wax, Adam

    2014-01-01

    We have developed a quantitative imaging refractometry technique, based on holographic phase microscopy, as a tool for investigating microscopic structural changes in water-soluble polymeric materials. Here we apply the approach to analyze the structural degradation of vaginal topical microbicide films due to water uptake. We implemented transmission imaging of 1-mm diameter film samples loaded into a flow chamber with a 1.5×2 mm field of view. After water was flooded into the chamber, interference images were captured and analyzed to obtain high resolution maps of the local refractive index and subsequently the volume fraction and mass density of film material at each spatial location. Here, we compare the hydration dynamics of a panel of films with varying thicknesses and polymer compositions, demonstrating that quantitative imaging refractometry can be an effective tool for evaluating and characterizing the performance of candidate microbicide film designs for anti-HIV drug delivery. PMID:24736376

  17. Analyzing How We Do Analysis and Consume Data, Results from the SciDAC-Data Project

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Ding, P.; Aliaga, L.; Mubarak, M.

    One of the main goals of the Dept. of Energy funded SciDAC-Data project is to analyze the more than 410,000 high energy physics datasets that have been collected, generated and defined over the past two decades by experiments using the Fermilab storage facilities. These datasets have been used as the input to over 5.6 million recorded analysis projects, for which detailed analytics have been gathered. The analytics and meta information for these datasets and analysis projects are being combined with knowledge of their part of the HEP analysis chains for major experiments to understand how modern computing and data deliverymore » is being used. We present the first results of this project, which examine in detail how the CDF, D0, NOvA, MINERvA and MicroBooNE experiments have organized, classified and consumed petascale datasets to produce their physics results. The results include analysis of the correlations in dataset/file overlap, data usage patterns, data popularity, dataset dependency and temporary dataset consumption. The results provide critical insight into how workflows and data delivery schemes can be combined with different caching strategies to more efficiently perform the work required to mine these large HEP data volumes and to understand the physics analysis requirements for the next generation of HEP computing facilities. In particular we present a detailed analysis of the NOvA data organization and consumption model corresponding to their first and second oscillation results (2014-2016) and the first look at the analysis of the Tevatron Run II experiments. We present statistical distributions for the characterization of these data and data driven models describing their consumption« less

  18. Analyzing how we do Analysis and Consume Data, Results from the SciDAC-Data Project

    NASA Astrophysics Data System (ADS)

    Ding, P.; Aliaga, L.; Mubarak, M.; Tsaris, A.; Norman, A.; Lyon, A.; Ross, R.

    2017-10-01

    One of the main goals of the Dept. of Energy funded SciDAC-Data project is to analyze the more than 410,000 high energy physics datasets that have been collected, generated and defined over the past two decades by experiments using the Fermilab storage facilities. These datasets have been used as the input to over 5.6 million recorded analysis projects, for which detailed analytics have been gathered. The analytics and meta information for these datasets and analysis projects are being combined with knowledge of their part of the HEP analysis chains for major experiments to understand how modern computing and data delivery is being used. We present the first results of this project, which examine in detail how the CDF, D0, NOvA, MINERvA and MicroBooNE experiments have organized, classified and consumed petascale datasets to produce their physics results. The results include analysis of the correlations in dataset/file overlap, data usage patterns, data popularity, dataset dependency and temporary dataset consumption. The results provide critical insight into how workflows and data delivery schemes can be combined with different caching strategies to more efficiently perform the work required to mine these large HEP data volumes and to understand the physics analysis requirements for the next generation of HEP computing facilities. In particular we present a detailed analysis of the NOvA data organization and consumption model corresponding to their first and second oscillation results (2014-2016) and the first look at the analysis of the Tevatron Run II experiments. We present statistical distributions for the characterization of these data and data driven models describing their consumption.

  19. A quantitative method for optimized placement of continuous air monitors.

    PubMed

    Whicker, Jeffrey J; Rodgers, John C; Moxley, John S

    2003-11-01

    Alarming continuous air monitors (CAMs) are a critical component for worker protection in facilities that handle large amounts of hazardous materials. In nuclear facilities, continuous air monitors alarm when levels of airborne radioactive materials exceed alarm thresholds, thus prompting workers to exit the room to reduce inhalation exposures. To maintain a high level of worker protection, continuous air monitors are required to detect radioactive aerosol clouds quickly and with good sensitivity. This requires that there are sufficient numbers of continuous air monitors in a room and that they are well positioned. Yet there are no published methodologies to quantitatively determine the optimal number and placement of continuous air monitors in a room. The goal of this study was to develop and test an approach to quantitatively determine optimal number and placement of continuous air monitors in a room. The method we have developed uses tracer aerosol releases (to simulate accidental releases) and the measurement of the temporal and spatial aspects of the dispersion of the tracer aerosol through the room. The aerosol dispersion data is then analyzed to optimize continuous air monitor utilization based on simulated worker exposure. This method was tested in a room within a Department of Energy operated plutonium facility at the Savannah River Site in South Carolina, U.S. Results from this study show that the value of quantitative airflow and aerosol dispersion studies is significant and that worker protection can be significantly improved while balancing the costs associated with CAM programs.

  20. Label-free imaging of intracellular motility by low-coherent quantitative phase microscope in reflection geometry

    NASA Astrophysics Data System (ADS)

    Yamauchi, Toyohiko; Iwai, Hidenao; Yamashita, Yutaka

    2011-11-01

    We demonstrate tomographic imaging of intracellular activity of living cells by a low-coherent quantitative phase microscope. The intracellular organelles, such as the nucleus, nucleolus, and mitochondria, are moving around inside living cells, driven by the cellular physiological activity. In order to visualize the intracellular motility in a label-free manner we have developed a reflection-type quantitative phase microscope which employs the phase shifting interferometric technique with a low-coherent light source. The phase shifting interferometry enables us to quantitatively measure the intensity and phase of the optical field, and the low-coherence interferometry makes it possible to selectively probe a specific sectioning plane in the cell volume. The results quantitatively revealed the depth-resolved fluctuations of intracellular surfaces so that the plasma membrane and the membranes of intracellular organelles were independently measured. The transversal and the vertical spatial resolutions were 0.56 μm and 0.93 μm, respectively, and the mechanical sensitivity of the phase measurement was 1.2 nanometers. The mean-squared displacement was applied as a statistical tool to analyze the temporal fluctuation of the intracellular organelles. To the best of our knowledge, our system visualized depth-resolved intracellular organelles motion for the first time in sub-micrometer resolution without contrast agents.

  1. Utility of a scanning densitometer in analyzing remotely sensed imagery

    NASA Technical Reports Server (NTRS)

    Dooley, J. T.

    1976-01-01

    The utility of a scanning densitometer for analyzing imagery in the NASA Lewis Research Center's regional remote sensing program was evaluated. Uses studied include: (1) quick-look screening of imagery by means of density slicing, magnification, color coding, and edge enhancement; (2) preliminary category classification of both low- and high-resolution data bases; and (3) quantitative measurement of the extent of features within selected areas. The densitometer was capable of providing fast, convenient, and relatively inexpensive preliminary analysis of aerial and satellite photography and scanner imagery involving land cover, water quality, strip mining, and energy conservation.

  2. Method and apparatus for thermographically and quantitatively analyzing a structure for disbonds and/or inclusions

    NASA Technical Reports Server (NTRS)

    Heyman, Joseph S. (Inventor); Winfree, William P. (Inventor); Cramer, K. Elliott (Inventor); Zalamedia, Joseph N. (Inventor)

    1996-01-01

    A heat source such as a magnetic induction/eddy current generator remotely heats a region of a surface of a test structure to a desired depth. For example, the frequency of the heating source can be varied to heat to the desired depth. A thermal sensor senses temperature changes in the heated region as a function of time. A computer compares these sensed temperature changes with calibration standards of a similar sample having known disbond and/or inclusion geography(ies) to analyze the test structure. A plurality of sensors can be arranged linearly to sense vector heat flow.

  3. Quantitative nephelometry

    MedlinePlus

    ... this page: //medlineplus.gov/ency/article/003545.htm Quantitative nephelometry test To use the sharing features on this page, please enable JavaScript. Quantitative nephelometry is a lab test to quickly and ...

  4. A quantitative assay measuring the function of lipase maturation factor 1

    PubMed Central

    Yin, Fen; Doolittle, Mark H.; Péterfy, Miklós

    2009-01-01

    Newly synthesized lipoprotein lipase (LPL) and related members of the lipase gene family require an endoplasmic reticulum maturation factor for attainment of enzyme activity. This factor has been identified as lipase maturation factor 1 (Lmf1), and mutations affecting its function and/or expression result in combined lipase deficiency (cld) and hypertriglyceridemia. To assess the functional impact of Lmf1 sequence variations, both naturally occurring and induced, we report the development of a cell-based assay using LPL activity as a quantitative reporter of Lmf1 function. The assay uses a cell line homozygous for the cld mutation, which renders endogenous Lmf1 nonfunctional. LPL transfected into the mutant cld cell line fails to attain activity; however, cotransfection of LPL with wild-type Lmf1 restores its ability to support normal lipase maturation. In this report, we describe optimized conditions that ensure the detection of a complete range of Lmf1 function (full, partial, or complete loss of function) using LPL activity as the quantitative reporter. To illustrate the dynamic range of the assay, we tested several novel mutations in mouse Lmf1. Our results demonstrate the ability of the assay to detect and analyze Lmf1 mutations having a wide range of effects on Lmf1 function and protein expression. PMID:19471043

  5. Risk assessment of false-positive quantitative real-time PCR results in food, due to detection of DNA originating from dead cells.

    PubMed

    Wolffs, Petra; Norling, Börje; Rådström, Peter

    2005-03-01

    Real-time PCR technology is increasingly used for detection and quantification of pathogens in food samples. A main disadvantage of nucleic acid detection is the inability to distinguish between signals originating from viable cells and DNA released from dead cells. In order to gain knowledge concerning risks of false-positive results due to detection of DNA originating from dead cells, quantitative PCR (qPCR) was used to investigate the degradation kinetics of free DNA in four types of meat samples. Results showed that the fastest degradation rate was observed (1 log unit per 0.5 h) in chicken homogenate, whereas the slowest rate was observed in pork rinse (1 log unit per 120.5 h). Overall results indicated that degradation occurred faster in chicken samples than in pork samples and faster at higher temperatures. Based on these results, it was concluded that, especially in pork samples, there is a risk of false-positive PCR results. This was confirmed in a quantitative study on cell death and signal persistence over a period of 28 days, employing three different methods, i.e. viable counts, direct qPCR, and finally floatation, a recently developed discontinuous density centrifugation method, followed by qPCR. Results showed that direct qPCR resulted in an overestimation of up to 10 times of the amount of cells in the samples compared to viable counts, due to detection of DNA from dead cells. However, after using floatation prior to qPCR, results resembled the viable count data. This indicates that by using of floatation as a sample treatment step prior to qPCR, the risk of false-positive PCR results due to detection of dead cells, can be minimized.

  6. A Sensitivity Model (SM) approach to analyze urban development in Taiwan based on sustainability indicators

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Huang Shuli; Yeh Chiatsung; Budd, William W.

    2009-02-15

    Sustainability indicators have been widely developed to monitor and assess sustainable development. They are expected to guide political decision-making based on their capability to represent states and trends of development. However, using indicators to assess the sustainability of urban strategies and policies has limitations - as they neither reflect the systemic interactions among them, nor provide normative indications in what direction they should be developed. This paper uses a semi-quantitative systematic model tool (Sensitivity Model Tools, SM) to analyze the role of urban development in Taiwan's sustainability. The results indicate that the natural environment in urban area is one ofmore » the most critical components and the urban economic production plays a highly active role in affecting Taiwan's sustainable development. The semi-quantitative simulation model integrates sustainability indicators and urban development policy to provide decision-makers with information about the impacts of their decisions on urban development. The system approach incorporated by this paper can be seen as a necessary, but not sufficient, condition for a sustainability assessment. The participatory process of expert participants for providing judgments on the relations between indicator variables is also discussed.« less

  7. Interdependence of PRECIS Role Operators: A Quantitative Analysis of Their Associations.

    ERIC Educational Resources Information Center

    Mahapatra, Manoranjan; Biswas, Subal Chandra

    1986-01-01

    Analyzes associations among different role operators quantitatively by taking input strings from 200 abstracts, each related to subject fields of taxation, genetic psychology, and Shakespearean drama, and subjecting them to the Chi-square test. Significant associations by other differencing operators and connectives are discussed. A schema of role…

  8. Quantitative Analysis of Qualitative Information from Interviews: A Systematic Literature Review

    ERIC Educational Resources Information Center

    Fakis, Apostolos; Hilliam, Rachel; Stoneley, Helen; Townend, Michael

    2014-01-01

    Background: A systematic literature review was conducted on mixed methods area. Objectives: The overall aim was to explore how qualitative information from interviews has been analyzed using quantitative methods. Methods: A contemporary review was undertaken and based on a predefined protocol. The references were identified using inclusion and…

  9. Missing Value Monitoring Enhances the Robustness in Proteomics Quantitation.

    PubMed

    Matafora, Vittoria; Corno, Andrea; Ciliberto, Andrea; Bachi, Angela

    2017-04-07

    In global proteomic analysis, it is estimated that proteins span from millions to less than 100 copies per cell. The challenge of protein quantitation by classic shotgun proteomic techniques relies on the presence of missing values in peptides belonging to low-abundance proteins that lowers intraruns reproducibility affecting postdata statistical analysis. Here, we present a new analytical workflow MvM (missing value monitoring) able to recover quantitation of missing values generated by shotgun analysis. In particular, we used confident data-dependent acquisition (DDA) quantitation only for proteins measured in all the runs, while we filled the missing values with data-independent acquisition analysis using the library previously generated in DDA. We analyzed cell cycle regulated proteins, as they are low abundance proteins with highly dynamic expression levels. Indeed, we found that cell cycle related proteins are the major components of the missing values-rich proteome. Using the MvM workflow, we doubled the number of robustly quantified cell cycle related proteins, and we reduced the number of missing values achieving robust quantitation for proteins over ∼50 molecules per cell. MvM allows lower quantification variance among replicates for low abundance proteins with respect to DDA analysis, which demonstrates the potential of this novel workflow to measure low abundance, dynamically regulated proteins.

  10. Quantitative magnetic resonance imaging phantoms: A review and the need for a system phantom.

    PubMed

    Keenan, Kathryn E; Ainslie, Maureen; Barker, Alex J; Boss, Michael A; Cecil, Kim M; Charles, Cecil; Chenevert, Thomas L; Clarke, Larry; Evelhoch, Jeffrey L; Finn, Paul; Gembris, Daniel; Gunter, Jeffrey L; Hill, Derek L G; Jack, Clifford R; Jackson, Edward F; Liu, Guoying; Russek, Stephen E; Sharma, Samir D; Steckner, Michael; Stupic, Karl F; Trzasko, Joshua D; Yuan, Chun; Zheng, Jie

    2018-01-01

    The MRI community is using quantitative mapping techniques to complement qualitative imaging. For quantitative imaging to reach its full potential, it is necessary to analyze measurements across systems and longitudinally. Clinical use of quantitative imaging can be facilitated through adoption and use of a standard system phantom, a calibration/standard reference object, to assess the performance of an MRI machine. The International Society of Magnetic Resonance in Medicine AdHoc Committee on Standards for Quantitative Magnetic Resonance was established in February 2007 to facilitate the expansion of MRI as a mainstream modality for multi-institutional measurements, including, among other things, multicenter trials. The goal of the Standards for Quantitative Magnetic Resonance committee was to provide a framework to ensure that quantitative measures derived from MR data are comparable over time, between subjects, between sites, and between vendors. This paper, written by members of the Standards for Quantitative Magnetic Resonance committee, reviews standardization attempts and then details the need, requirements, and implementation plan for a standard system phantom for quantitative MRI. In addition, application-specific phantoms and implementation of quantitative MRI are reviewed. Magn Reson Med 79:48-61, 2018. © 2017 International Society for Magnetic Resonance in Medicine. © 2017 International Society for Magnetic Resonance in Medicine.

  11. Rheological properties of glutaraldehyde-crosslinked collagen solutions analyzed quantitatively using mechanical models.

    PubMed

    Tian, Zhenhua; Duan, Lian; Wu, Lei; Shen, Lirui; Li, Guoying

    2016-06-01

    Understanding the rheological behavior of collagen solutions crosslinked by various amounts of glutaraldehyde (GTA) [GTA/collagen (w/w)=0-0.1] is fundamental either to design optimized products or to ensure stable flow. Under steady shear, all the samples exhibited pseudoplasticity with shear-thinning behavior, and the flow curves were well described by Ostwald-de Waele model and Carreau model. With increased amounts of GTA, the viscosity increased from 6.15 to 168.54 Pa·s at 0.1s(-1), and the pseudoplasticity strengthened (the flow index decreased from 0.549 to 0.117). Additionally, hysteresis loops were evaluated to analyze the thixotropy of the native and crosslinked collagen solutions, and indicated that stronger thixotropic behavior was associated with higher amount of GTA. Furthermore, the values of apparent yield stress were negative, and a flow index <1 for all the systems obtained via Herschel-Bulkley model confirmed that the native and crosslinked collagen solutions belonged to pseudoplastic fluid without apparent yield stress. However, the increment of dynamic denaturation temperature determined by dynamic temperature sweep was not obvious. The viscoelastic properties were examined based on creep-recovery measurements and then simulated using Burger model and a semi-empirical model. The increase in the proportion of recoverable compliance (instantaneous and retardant compliance) reflected that the crosslinked collagen solutions were more resistant to the deformation and exhibited more elastic behavior than the native collagen solution, accompanied by the fact that the compliance value decreased from 39.317 to 0.152 Pa(-1) and the recovery percentage increased from 1.128% to 87.604%. These data indicated that adjusting the amount of GTA could be a suitable mean for manipulating mechanical properties of collagen-based biomaterials. Copyright © 2016 Elsevier B.V. All rights reserved.

  12. Portable paper-based device for quantitative colorimetric assays relying on light reflectance principle.

    PubMed

    Li, Bowei; Fu, Longwen; Zhang, Wei; Feng, Weiwei; Chen, Lingxin

    2014-04-01

    This paper presents a novel paper-based analytical device based on the colorimetric paper assays through its light reflectance. The device is portable, low cost (<20 dollars), and lightweight (only 176 g) that is available to assess the cost-effectiveness and appropriateness of the original health care or on-site detection information. Based on the light reflectance principle, the signal can be obtained directly, stably and user-friendly in our device. We demonstrated the utility and broad applicability of this technique with measurements of different biological and pollution target samples (BSA, glucose, Fe, and nitrite). Moreover, the real samples of Fe (II) and nitrite in the local tap water were successfully analyzed, and compared with the standard UV absorption method, the quantitative results showed good performance, reproducibility, and reliability. This device could provide quantitative information very conveniently and show great potential to broad fields of resource-limited analysis, medical diagnostics, and on-site environmental detection. © 2013 WILEY-VCH Verlag GmbH & Co. KGaA, Weinheim.

  13. Biological Dynamics Markup Language (BDML): an open format for representing quantitative biological dynamics data

    PubMed Central

    Kyoda, Koji; Tohsato, Yukako; Ho, Kenneth H. L.; Onami, Shuichi

    2015-01-01

    Motivation: Recent progress in live-cell imaging and modeling techniques has resulted in generation of a large amount of quantitative data (from experimental measurements and computer simulations) on spatiotemporal dynamics of biological objects such as molecules, cells and organisms. Although many research groups have independently dedicated their efforts to developing software tools for visualizing and analyzing these data, these tools are often not compatible with each other because of different data formats. Results: We developed an open unified format, Biological Dynamics Markup Language (BDML; current version: 0.2), which provides a basic framework for representing quantitative biological dynamics data for objects ranging from molecules to cells to organisms. BDML is based on Extensible Markup Language (XML). Its advantages are machine and human readability and extensibility. BDML will improve the efficiency of development and evaluation of software tools for data visualization and analysis. Availability and implementation: A specification and a schema file for BDML are freely available online at http://ssbd.qbic.riken.jp/bdml/. Contact: sonami@riken.jp Supplementary Information: Supplementary data are available at Bioinformatics online. PMID:25414366

  14. FLASH proton density imaging for improved surface coil intensity correction in quantitative and semi-quantitative SSFP perfusion cardiovascular magnetic resonance.

    PubMed

    Nielles-Vallespin, Sonia; Kellman, Peter; Hsu, Li-Yueh; Arai, Andrew E

    2015-02-17

    A low excitation flip angle (α < 10°) steady-state free precession (SSFP) proton-density (PD) reference scan is often used to estimate the B1-field inhomogeneity for surface coil intensity correction (SCIC) of the saturation-recovery (SR) prepared high flip angle (α = 40-50°) SSFP myocardial perfusion images. The different SSFP off-resonance response for these two flip angles might lead to suboptimal SCIC when there is a spatial variation in the background B0-field. The low flip angle SSFP-PD frames are more prone to parallel imaging banding artifacts in the presence of off-resonance. The use of FLASH-PD frames would eliminate both the banding artifacts and the uneven frequency response in the presence of off-resonance in the surface coil inhomogeneity estimate and improve homogeneity of semi-quantitative and quantitative perfusion measurements. B0-field maps, SSFP and FLASH-PD frames were acquired in 10 healthy volunteers to analyze the SSFP off-resonance response. Furthermore, perfusion scans preceded by both FLASH and SSFP-PD frames from 10 patients with no myocardial infarction were analyzed semi-quantitatively and quantitatively (rest n = 10 and stress n = 1). Intra-subject myocardial blood flow (MBF) coefficient of variation (CoV) over the whole left ventricle (LV), as well as intra-subject peak contrast (CE) and upslope (SLP) standard deviation (SD) over 6 LV sectors were investigated. In the 6 out of 10 cases where artifacts were apparent in the LV ROI of the SSFP-PD images, all three variability metrics were statistically significantly lower when using the FLASH-PD frames as input for the SCIC (CoVMBF-FLASH = 0.3 ± 0.1, CoVMBF-SSFP = 0.4 ± 0.1, p = 0.03; SDCE-FLASH = 10 ± 2, SDCE-SSFP = 32 ± 7, p = 0.01; SDSLP-FLASH = 0.02 ± 0.01, SDSLP-SSFP = 0.06 ± 0.02, p = 0.03). Example rest and stress data sets from the patient pool demonstrate that the low flip angle SSFP protocol

  15. Analyzing Preservice Teachers' Technological Pedagogical Content Knowledge Development in the Context of a Multidimensional Teacher Preparation Program

    ERIC Educational Resources Information Center

    Shinas, Valerie Harlow; Karchmer-Klein, Rachel; Mouza, Chrystalla; Yilmaz-Ozden, Sule; Glutting, Joseph J.

    2015-01-01

    In this quantitative study, correlational and multiple regression analyses were conducted to examine the technological pedagogical content knowledge (TPACK) development of 299 preservice teachers in response to the technology preparation they received during their initial teacher licensure program. Survey data were analyzed to determine the…

  16. Rock surface roughness measurement using CSI technique and analysis of surface characterization by qualitative and quantitative results

    NASA Astrophysics Data System (ADS)

    Mukhtar, Husneni; Montgomery, Paul; Gianto; Susanto, K.

    2016-01-01

    In order to develop image processing that is widely used in geo-processing and analysis, we introduce an alternative technique for the characterization of rock samples. The technique that we have used for characterizing inhomogeneous surfaces is based on Coherence Scanning Interferometry (CSI). An optical probe is first used to scan over the depth of the surface roughness of the sample. Then, to analyse the measured fringe data, we use the Five Sample Adaptive method to obtain quantitative results of the surface shape. To analyse the surface roughness parameters, Hmm and Rq, a new window resizing analysis technique is employed. The results of the morphology and surface roughness analysis show micron and nano-scale information which is characteristic of each rock type and its history. These could be used for mineral identification and studies in rock movement on different surfaces. Image processing is thus used to define the physical parameters of the rock surface.

  17. [Determination of five naphthaquinones in Arnebia euchroma by quantitative analysis multi-components with single-marker].

    PubMed

    Zhao, Wen-Wen; Wu, Zhi-Min; Wu, Xia; Zhao, Hai-Yu; Chen, Xiao-Qing

    2016-10-01

    This study is to determine five naphthaquinones (acetylshikonin, β-acetoxyisovalerylalkannin, isobutylshikonin, β,β'-dimethylacrylalkannin,α-methyl-n-butylshikonin) by quantitative analysis of multi-components with a single marker (QAMS). β,β'-Dimethylacrylalkannin was selected as the internal reference substance, and the relative correlation factors (RCFs) of acetylshikonin, β-acetoxyisovalerylalkannin, isobutylshikonin and α-methyl-n-butylshikonin were calculated. Then the ruggedness of relative correction factors was tested on different instruments and columns. Meanwhile, 16 batches of Arnebia euchroma were analyzed by external standard method (ESM) and QAMS, respectively. The peaks were identifited by LC-MS. The ruggedness of relative correction factors was good. And the analytical results calculated by ESM and QAMS showed no difference. The quantitative method established was feasible and suitable for the quality evaluation of A. euchroma. Copyright© by the Chinese Pharmaceutical Association.

  18. Clinical study of quantitative diagnosis of early cervical cancer based on the classification of acetowhitening kinetics

    NASA Astrophysics Data System (ADS)

    Wu, Tao; Cheung, Tak-Hong; Yim, So-Fan; Qu, Jianan Y.

    2010-03-01

    A quantitative colposcopic imaging system for the diagnosis of early cervical cancer is evaluated in a clinical study. This imaging technology based on 3-D active stereo vision and motion tracking extracts diagnostic information from the kinetics of acetowhitening process measured from the cervix of human subjects in vivo. Acetowhitening kinetics measured from 137 cervical sites of 57 subjects are analyzed and classified using multivariate statistical algorithms. Cross-validation methods are used to evaluate the performance of the diagnostic algorithms. The results show that an algorithm for screening precancer produced 95% sensitivity (SE) and 96% specificity (SP) for discriminating normal and human papillomavirus (HPV)-infected tissues from cervical intraepithelial neoplasia (CIN) lesions. For a diagnostic algorithm, 91% SE and 90% SP are achieved for discriminating normal tissue, HPV infected tissue, and low-grade CIN lesions from high-grade CIN lesions. The results demonstrate that the quantitative colposcopic imaging system could provide objective screening and diagnostic information for early detection of cervical cancer.

  19. [THE COMPARATIVE ANALYSIS OF RESULTS OF DETECTION OF CARCINOGENIC TYPES OF HUMAN PAPILLOMA VIRUS BY QUALITATIVE AND QUANTITATIVE TESTS].

    PubMed

    Kuzmenko, E T; Labigina, A V; Leshenko, O Ya; Rusanov, D N; Kuzmenko, V V; Fedko, L P; Pak, I P

    2015-05-01

    The analysis of results of screening (n = 3208; sexually active citizen aged from 18 to 59 years) was carried out to detect oncogene types of human papilloma virus in using qualitative (1150 females and 720 males) and quantitative (polymerase chain reaction in real-time (843 females and 115 males) techniques. The human papilloma virus of high oncogene type was detected in 65% and 68.4% of females and in 48.6% and 53% of males correspondingly. Among 12 types of human papilloma virus the most frequently diagnosed was human papilloma virus 16 independently of gender of examined and technique of analysis. In females, under application of qualitative tests rate of human papilloma virus 16 made up to 18.3% (n = 280) and under application of quantitative tests Rte of human papilloma virus made up to 14.9% (n = 126; p ≤ 0.05). Under examination of males using qualitative tests rate of human papilloma virus 16 made up to 8.3% (n = 60) and under application of qualitative tests made up to 12.2% (n = 14; p ≥ 0.05). Under application of qualitative tests rate of detection on the rest ofoncogene types of human papilloma virus varied in females from 3.4% to 8.4% and in males from 1.8% to 5.9%. Under application of qualitative tests to females rate of human papilloma virus with high viral load made up to 68.4%, with medium viral load - 2.85% (n = 24) and with low viral load -0.24% (n = 2). Under application of quantitative tests in males rate of detection of types of human papilloma virus made up to 53% and at that in all high viral load was established. In females, the most of oncogene types of human papilloma virus (except for 31, 39, 59) are detected significantly more often than in males.

  20. Issues in Quantitative Analysis of Ultraviolet Imager (UV) Data: Airglow

    NASA Technical Reports Server (NTRS)

    Germany, G. A.; Richards, P. G.; Spann, J. F.; Brittnacher, M. J.; Parks, G. K.

    1999-01-01

    The GGS Ultraviolet Imager (UVI) has proven to be especially valuable in correlative substorm, auroral morphology, and extended statistical studies of the auroral regions. Such studies are based on knowledge of the location, spatial, and temporal behavior of auroral emissions. More quantitative studies, based on absolute radiometric intensities from UVI images, require a more intimate knowledge of the instrument behavior and data processing requirements and are inherently more difficult than studies based on relative knowledge of the oval location. In this study, UVI airglow observations are analyzed and compared with model predictions to illustrate issues that arise in quantitative analysis of UVI images. These issues include instrument calibration, long term changes in sensitivity, and imager flat field response as well as proper background correction. Airglow emissions are chosen for this study because of their relatively straightforward modeling requirements and because of their implications for thermospheric compositional studies. The analysis issues discussed here, however, are identical to those faced in quantitative auroral studies.

  1. Summary of Results from the Mars Phoenix Lander's Thermal Evolved Gas Analyzer

    NASA Technical Reports Server (NTRS)

    Sutter, B.; Ming, D. W.; Boynton, W. V.; Niles, P. B.; Hoffman, J.; Lauer, H. V.; Golden, D. C.

    2009-01-01

    The Mars Phoenix Scout Mission with its diverse instrument suite successfully examined several soils on the Northern plains of Mars. The Thermal and Evolved Gas Analyzer (TEGA) was employed to detect evolved volatiles and organic and inorganic materials by coupling a differential scanning calorimeter (DSC) with a magnetic-sector mass spectrometer (MS) that can detect masses in the 2 to 140 dalton range [1]. Five Martian soils were individually heated to 1000 C in the DSC ovens where evolved gases from mineral decompostion products were examined with the MS. TEGA s DSC has the capability to detect endothermic and exothermic reactions during heating that are characteristic of minerals present in the Martian soil.

  2. Quantitative subsurface analysis using frequency modulated thermal wave imaging

    NASA Astrophysics Data System (ADS)

    Subhani, S. K.; Suresh, B.; Ghali, V. S.

    2018-01-01

    Quantitative depth analysis of the anomaly with an enhanced depth resolution is a challenging task towards the estimation of depth of the subsurface anomaly using thermography. Frequency modulated thermal wave imaging introduced earlier provides a complete depth scanning of the object by stimulating it with a suitable band of frequencies and further analyzing the subsequent thermal response using a suitable post processing approach to resolve subsurface details. But conventional Fourier transform based methods used for post processing unscramble the frequencies with a limited frequency resolution and contribute for a finite depth resolution. Spectral zooming provided by chirp z transform facilitates enhanced frequency resolution which can further improves the depth resolution to axially explore finest subsurface features. Quantitative depth analysis with this augmented depth resolution is proposed to provide a closest estimate to the actual depth of subsurface anomaly. This manuscript experimentally validates this enhanced depth resolution using non stationary thermal wave imaging and offers an ever first and unique solution for quantitative depth estimation in frequency modulated thermal wave imaging.

  3. Multiplexed, quantitative, and targeted metabolite profiling by LC-MS/MRM.

    PubMed

    Wei, Ru; Li, Guodong; Seymour, Albert B

    2014-01-01

    Targeted metabolomics, which focuses on a subset of known metabolites representative of biologically relevant metabolic pathways, is a valuable tool to discover biomarkers and link disease phenotypes to underlying mechanisms or therapeutic modes of action. A key advantage of targeted metabolomics, compared to discovery metabolomics, is its immediate readiness for extracting biological information derived from known metabolites and quantitative measurements. However, simultaneously analyzing hundreds of endogenous metabolites presents a challenge due to their diverse chemical structures and properties. Here we report a method which combines different chromatographic separation conditions, optimal ionization polarities, and the most sensitive triple-quadrupole MS-based data acquisition mode, multiple reaction monitoring (MRM), to quantitatively profile 205 endogenous metabolites in 10 min.

  4. Analyzing refractive index profiles of confined fluids by interferometry.

    PubMed

    Kienle, Daniel F; Kuhl, Tonya L

    2014-12-02

    This work describes an interferometry data analysis method for determining the optical thickness of thin films or any variation in the refractive index of a fluid or film near a surface. In particular, the method described is applied to the analysis of interferometry data taken with a surface force apparatus (SFA). The technique does not require contacting or confining the fluid or film. By analyzing interferometry data taken at many intersurface separation distances out to at least 300 nm, the properties of a film can be quantitatively determined. The film can consist of material deposited on the surface, like a polymer brush, or variation in a fluid's refractive index near a surface resulting from, for example, a concentration gradient, depletion in density, or surface roughness. The method is demonstrated with aqueous polyethylenimine (PEI) adsorbed onto mica substrates, which has a large concentration and therefore refractive index gradient near the mica surface. The PEI layer thickness determined by the proposed method is consistent with the thickness measured by conventional SFA methods. Additionally, a thorough investigation of the effects of random and systematic error in SFA data analysis and modeling via simulations of interferometry is described in detail.

  5. Thermal Analyzer for Planetary Soil (TAPS): an in Situ Instrument for Mineral and Volatile-element Measurements

    NASA Technical Reports Server (NTRS)

    Gooding, J. L.; Ming, D. W.; Gruener, J. E.; Gibbons, F. L.; Allton, J. H.

    1993-01-01

    Thermal Analyzer for Planetary Soil (TAPS) offers a specific implementation for the generic thermal analyzer/evolved-gas analyzer (TA/EGA) function included in the Mars Environmental Survey (MESUR) strawman payload; applications to asteroids and comets are also possible. The baseline TAPS is a single-sample differential scanning calorimeter (DSC), backed by a capacitive-polymer humidity sensor, with an integrated sampling mechanism. After placement on a planetary surface, TAPS acquires 10-50 mg of soil or sediment and heats the sample from ambient temperature to 1000-1300 K. During heating, DSC data are taken for the solid and evolved gases are swept past the water sensor. Through ground based data analysis, multicomponent DSC data are deconvolved and correlated with the water release profile to quantitatively determine the types and relative proportions of volatile-bearing minerals such as clays and other hydrates, carbonates, and nitrates. The rapid-response humidity sensors also achieve quantitative analysis of total water. After conclusion of soil-analysis operations, the humidity sensors become available for meteorology. The baseline design fits within a circular-cylindrical volume less than 1000 cm(sup 3), occupies 1.2 kg mass, and consumes about 2 Whr of power per analysis. Enhanced designs would acquire and analyze multiple samples and employ additional microchemical sensors for analysis of CO2, SO2, NO(x), and other gaseous species. Atmospheric pumps are also being considered as alternatives to pressurized purge gas.

  6. Thermal Analyzer for Planetary Soil (TAPS): an in situ instrument for mineral and volatile-element measurements

    NASA Astrophysics Data System (ADS)

    Gooding, J. L.; Ming, D. W.; Gruener, J. E.; Gibbons, F. L.; Allton, J. H.

    Thermal Analyzer for Planetary Soil (TAPS) offers a specific implementation for the generic thermal analyzer/evolved-gas analyzer (TA/EGA) function included in the Mars Environmental Survey (MESUR) strawman payload; applications to asteroids and comets are also possible. The baseline TAPS is a single-sample differential scanning calorimeter (DSC), backed by a capacitive-polymer humidity sensor, with an integrated sampling mechanism. After placement on a planetary surface, TAPS acquires 10-50 mg of soil or sediment and heats the sample from ambient temperature to 1000-1300 K. During heating, DSC data are taken for the solid and evolved gases are swept past the water sensor. Through ground based data analysis, multicomponent DSC data are deconvolved and correlated with the water release profile to quantitatively determine the types and relative proportions of volatile-bearing minerals such as clays and other hydrates, carbonates, and nitrates. The rapid-response humidity sensors also achieve quantitative analysis of total water. After conclusion of soil-analysis operations, the humidity sensors become available for meteorology. The baseline design fits within a circular-cylindrical volume less than 1000 cm3, occupies 1.2 kg mass, and consumes about 2 Whr of power per analysis. Enhanced designs would acquire and analyze multiple samples and employ additional microchemical sensors for analysis of CO2, SO2, NO(x), and other gaseous species. Atmospheric pumps are also being considered as alternatives to pressurized purge gas.

  7. Ethnic Disparities in Graduate Education: A Selective Review of Quantitative Research, Social Theory, and Quality Initiatives

    ERIC Educational Resources Information Center

    Franklin, Somer L.; Slate, John R.; Joyner, Sheila A.

    2014-01-01

    In this article, we analyzed research studies in the field of graduate education. In particular, we explored the issue of inequity in graduate education through three key lenses of social science analyses. Furthermore, we analyzed selected quantitative research studies that undertook a comparative examination of aggregate trends in enrollment and…

  8. Image Processing for Bioluminescence Resonance Energy Transfer Measurement-BRET-Analyzer.

    PubMed

    Chastagnier, Yan; Moutin, Enora; Hemonnot, Anne-Laure; Perroy, Julie

    2017-01-01

    A growing number of tools now allow live recordings of various signaling pathways and protein-protein interaction dynamics in time and space by ratiometric measurements, such as Bioluminescence Resonance Energy Transfer (BRET) Imaging. Accurate and reproducible analysis of ratiometric measurements has thus become mandatory to interpret quantitative imaging. In order to fulfill this necessity, we have developed an open source toolset for Fiji- BRET-Analyzer -allowing a systematic analysis, from image processing to ratio quantification. We share this open source solution and a step-by-step tutorial at https://github.com/ychastagnier/BRET-Analyzer. This toolset proposes (1) image background subtraction, (2) image alignment over time, (3) a composite thresholding method of the image used as the denominator of the ratio to refine the precise limits of the sample, (4) pixel by pixel division of the images and efficient distribution of the ratio intensity on a pseudocolor scale, and (5) quantification of the ratio mean intensity and standard variation among pixels in chosen areas. In addition to systematize the analysis process, we show that the BRET-Analyzer allows proper reconstitution and quantification of the ratiometric image in time and space, even from heterogeneous subcellular volumes. Indeed, analyzing twice the same images, we demonstrate that compared to standard analysis BRET-Analyzer precisely define the luminescent specimen limits, enlightening proficient strengths from small and big ensembles over time. For example, we followed and quantified, in live, scaffold proteins interaction dynamics in neuronal sub-cellular compartments including dendritic spines, for half an hour. In conclusion, BRET-Analyzer provides a complete, versatile and efficient toolset for automated reproducible and meaningful image ratio analysis.

  9. Quantitative effects of cyanogenesis on an adapted herbivore.

    PubMed

    Ballhorn, D J; Heil, M; Pietrowski, A; Lieberei, R

    2007-12-01

    Plant cyanogenesis means the release of gaseous hydrogen cyanide (HCN) in response to cell damage and is considered as an effective defense against generalist herbivores. In contrast, specialists are generally believed not to be affected negatively by this trait. However, quantitative data on long-term effects of cyanogenesis on specialists are rare. In this study, we used lima bean accessions (Fabaceae: Phaseolus lunatus L.) with high quantitative variability of cyanogenic features comprising cyanogenic potential (HCNp; concentration of cyanogenic precursors) and cyanogenic capacities (HCNc; release of gaseous HCN per unit time). In feeding trials, we analyzed performance of herbivorous Mexican bean beetle (Coleoptera: Coccinellidae: Epilachna varivestis Mulsant) on selected lines characterized by high (HC-plants) and low HCNp (LC-plants). Larval and adult stages of this herbivore feed on a narrow range of legumes and prefer cyanogenic lima bean as host plant. Nevertheless, we found that performance of beetles (larval weight gain per time and body mass of adult beetles) was significantly affected by lima bean HCNp: Body weight decreased and developmental period of larvae and pupae increased on HC-plants during the first generation of beetles and then remained constant for four consecutive generations. In addition, we found continuously decreasing numbers of eggs and larval hatching as inter-generational effects on HC-plants. In contrast to HC-plants, constantly high performance was observed among four generations on LC-plants. Our results demonstrate that Mexican bean beetle, although preferentially feeding on lima bean, is quantitatively affected by the HCNp of its host plant. Effects can only be detected when considering more than one generation. Thus, cyanide-containing precursors can have negative effects even on herbivores adapted to feed on cyanogenic plants.

  10. Evaluation of chemotherapy response in ovarian cancer treatment using quantitative CT image biomarkers: a preliminary study

    NASA Astrophysics Data System (ADS)

    Qiu, Yuchen; Tan, Maxine; McMeekin, Scott; Thai, Theresa; Moore, Kathleen; Ding, Kai; Liu, Hong; Zheng, Bin

    2015-03-01

    The purpose of this study is to identify and apply quantitative image biomarkers for early prediction of the tumor response to the chemotherapy among the ovarian cancer patients participated in the clinical trials of testing new drugs. In the experiment, we retrospectively selected 30 cases from the patients who participated in Phase I clinical trials of new drug or drug agents for ovarian cancer treatment. Each case is composed of two sets of CT images acquired pre- and post-treatment (4-6 weeks after starting treatment). A computer-aided detection (CAD) scheme was developed to extract and analyze the quantitative image features of the metastatic tumors previously tracked by the radiologists using the standard Response Evaluation Criteria in Solid Tumors (RECIST) guideline. The CAD scheme first segmented 3-D tumor volumes from the background using a hybrid tumor segmentation scheme. Then, for each segmented tumor, CAD computed three quantitative image features including the change of tumor volume, tumor CT number (density) and density variance. The feature changes were calculated between the matched tumors tracked on the CT images acquired pre- and post-treatments. Finally, CAD predicted patient's 6-month progression-free survival (PFS) using a decision-tree based classifier. The performance of the CAD scheme was compared with the RECIST category. The result shows that the CAD scheme achieved a prediction accuracy of 76.7% (23/30 cases) with a Kappa coefficient of 0.493, which is significantly higher than the performance of RECIST prediction with a prediction accuracy and Kappa coefficient of 60% (17/30) and 0.062, respectively. This study demonstrated the feasibility of analyzing quantitative image features to improve the early predicting accuracy of the tumor response to the new testing drugs or therapeutic methods for the ovarian cancer patients.

  11. Quantitative prediction of drug side effects based on drug-related features.

    PubMed

    Niu, Yanqing; Zhang, Wen

    2017-09-01

    Unexpected side effects of drugs are great concern in the drug development, and the identification of side effects is an important task. Recently, machine learning methods are proposed to predict the presence or absence of interested side effects for drugs, but it is difficult to make the accurate prediction for all of them. In this paper, we transform side effect profiles of drugs as their quantitative scores, by summing up their side effects with weights. The quantitative scores may measure the dangers of drugs, and thus help to compare the risk of different drugs. Here, we attempt to predict quantitative scores of drugs, namely the quantitative prediction. Specifically, we explore a variety of drug-related features and evaluate their discriminative powers for the quantitative prediction. Then, we consider several feature combination strategies (direct combination, average scoring ensemble combination) to integrate three informative features: chemical substructures, targets, and treatment indications. Finally, the average scoring ensemble model which produces the better performances is used as the final quantitative prediction model. Since weights for side effects are empirical values, we randomly generate different weights in the simulation experiments. The experimental results show that the quantitative method is robust to different weights, and produces satisfying results. Although other state-of-the-art methods cannot make the quantitative prediction directly, the prediction results can be transformed as the quantitative scores. By indirect comparison, the proposed method produces much better results than benchmark methods in the quantitative prediction. In conclusion, the proposed method is promising for the quantitative prediction of side effects, which may work cooperatively with existing state-of-the-art methods to reveal dangers of drugs.

  12. Relating interesting quantitative time series patterns with text events and text features

    NASA Astrophysics Data System (ADS)

    Wanner, Franz; Schreck, Tobias; Jentner, Wolfgang; Sharalieva, Lyubka; Keim, Daniel A.

    2013-12-01

    In many application areas, the key to successful data analysis is the integrated analysis of heterogeneous data. One example is the financial domain, where time-dependent and highly frequent quantitative data (e.g., trading volume and price information) and textual data (e.g., economic and political news reports) need to be considered jointly. Data analysis tools need to support an integrated analysis, which allows studying the relationships between textual news documents and quantitative properties of the stock market price series. In this paper, we describe a workflow and tool that allows a flexible formation of hypotheses about text features and their combinations, which reflect quantitative phenomena observed in stock data. To support such an analysis, we combine the analysis steps of frequent quantitative and text-oriented data using an existing a-priori method. First, based on heuristics we extract interesting intervals and patterns in large time series data. The visual analysis supports the analyst in exploring parameter combinations and their results. The identified time series patterns are then input for the second analysis step, in which all identified intervals of interest are analyzed for frequent patterns co-occurring with financial news. An a-priori method supports the discovery of such sequential temporal patterns. Then, various text features like the degree of sentence nesting, noun phrase complexity, the vocabulary richness, etc. are extracted from the news to obtain meta patterns. Meta patterns are defined by a specific combination of text features which significantly differ from the text features of the remaining news data. Our approach combines a portfolio of visualization and analysis techniques, including time-, cluster- and sequence visualization and analysis functionality. We provide two case studies, showing the effectiveness of our combined quantitative and textual analysis work flow. The workflow can also be generalized to other

  13. Quantitative High-Resolution Genomic Analysis of Single Cancer Cells

    PubMed Central

    Hannemann, Juliane; Meyer-Staeckling, Sönke; Kemming, Dirk; Alpers, Iris; Joosse, Simon A.; Pospisil, Heike; Kurtz, Stefan; Görndt, Jennifer; Püschel, Klaus; Riethdorf, Sabine; Pantel, Klaus; Brandt, Burkhard

    2011-01-01

    During cancer progression, specific genomic aberrations arise that can determine the scope of the disease and can be used as predictive or prognostic markers. The detection of specific gene amplifications or deletions in single blood-borne or disseminated tumour cells that may give rise to the development of metastases is of great clinical interest but technically challenging. In this study, we present a method for quantitative high-resolution genomic analysis of single cells. Cells were isolated under permanent microscopic control followed by high-fidelity whole genome amplification and subsequent analyses by fine tiling array-CGH and qPCR. The assay was applied to single breast cancer cells to analyze the chromosomal region centred by the therapeutical relevant EGFR gene. This method allows precise quantitative analysis of copy number variations in single cell diagnostics. PMID:22140428

  14. Quantitative high-resolution genomic analysis of single cancer cells.

    PubMed

    Hannemann, Juliane; Meyer-Staeckling, Sönke; Kemming, Dirk; Alpers, Iris; Joosse, Simon A; Pospisil, Heike; Kurtz, Stefan; Görndt, Jennifer; Püschel, Klaus; Riethdorf, Sabine; Pantel, Klaus; Brandt, Burkhard

    2011-01-01

    During cancer progression, specific genomic aberrations arise that can determine the scope of the disease and can be used as predictive or prognostic markers. The detection of specific gene amplifications or deletions in single blood-borne or disseminated tumour cells that may give rise to the development of metastases is of great clinical interest but technically challenging. In this study, we present a method for quantitative high-resolution genomic analysis of single cells. Cells were isolated under permanent microscopic control followed by high-fidelity whole genome amplification and subsequent analyses by fine tiling array-CGH and qPCR. The assay was applied to single breast cancer cells to analyze the chromosomal region centred by the therapeutical relevant EGFR gene. This method allows precise quantitative analysis of copy number variations in single cell diagnostics.

  15. RipleyGUI: software for analyzing spatial patterns in 3D cell distributions

    PubMed Central

    Hansson, Kristin; Jafari-Mamaghani, Mehrdad; Krieger, Patrik

    2013-01-01

    The true revolution in the age of digital neuroanatomy is the ability to extensively quantify anatomical structures and thus investigate structure-function relationships in great detail. To facilitate the quantification of neuronal cell patterns we have developed RipleyGUI, a MATLAB-based software that can be used to detect patterns in the 3D distribution of cells. RipleyGUI uses Ripley's K-function to analyze spatial distributions. In addition the software contains statistical tools to determine quantitative statistical differences, and tools for spatial transformations that are useful for analyzing non-stationary point patterns. The software has a graphical user interface making it easy to use without programming experience, and an extensive user manual explaining the basic concepts underlying the different statistical tools used to analyze spatial point patterns. The described analysis tool can be used for determining the spatial organization of neurons that is important for a detailed study of structure-function relationships. For example, neocortex that can be subdivided into six layers based on cell density and cell types can also be analyzed in terms of organizational principles distinguishing the layers. PMID:23658544

  16. [A new method of processing quantitative PCR data].

    PubMed

    Ke, Bing-Shen; Li, Guang-Yun; Chen, Shi-Min; Huang, Xiang-Yan; Chen, Ying-Jian; Xu, Jun

    2003-05-01

    Today standard PCR can't satisfy the need of biotechnique development and clinical research any more. After numerous dynamic research, PE company found there is a linear relation between initial template number and cycling time when the accumulating fluorescent product is detectable.Therefore,they developed a quantitative PCR technique to be used in PE7700 and PE5700. But the error of this technique is too great to satisfy the need of biotechnique development and clinical research. A better quantitative PCR technique is needed. The mathematical model submitted here is combined with the achievement of relative science,and based on the PCR principle and careful analysis of molecular relationship of main members in PCR reaction system. This model describes the function relation between product quantity or fluorescence intensity and initial template number and other reaction conditions, and can reflect the accumulating rule of PCR product molecule accurately. Accurate quantitative PCR analysis can be made use this function relation. Accumulated PCR product quantity can be obtained from initial template number. Using this model to do quantitative PCR analysis,result error is only related to the accuracy of fluorescence intensity or the instrument used. For an example, when the fluorescence intensity is accurate to 6 digits and the template size is between 100 to 1,000,000, the quantitative result accuracy will be more than 99%. The difference of result error is distinct using same condition,same instrument but different analysis method. Moreover,if the PCR quantitative analysis system is used to process data, it will get result 80 times of accuracy than using CT method.

  17. Quantitative Study of Emotional Intelligence and Communication Levels in Information Technology Professionals

    ERIC Educational Resources Information Center

    Hendon, Michalina

    2016-01-01

    This quantitative non-experimental correlational research analyzes the relationship between emotional intelligence and communication due to the lack of this research on information technology professionals in the U.S. One hundred and eleven (111) participants completed a survey that measures both the emotional intelligence and communication…

  18. The Implementation of a Cost Effectiveness Analyzer for Web-Supported Academic Instruction: An Example from Life Science

    ERIC Educational Resources Information Center

    Cohen, Anat; Nachmias, Rafi

    2012-01-01

    This paper describes implementation of a quantitative cost effectiveness analyzer for Web-supported academic instruction that was developed in our University. The paper presents the cost effectiveness analysis of one academic exemplary course in Life Science department and its introducing to the course lecturer for evaluation. The benefits and…

  19. Quantitative and Qualitative Study of Intestinal Flora in Neonates

    PubMed Central

    Sharma, Nidhi; Chaudhry, Rama; Panigrahi, Pinaki

    2012-01-01

    Background: In the neonatal period the developing intestinal barrier function provides a sub-optimal mucosal defense against infection. Establishment of the normal commensal micro-flora plays a vital role in this process. Aims: To determine aerobic and anaerobic bacteria by quantitative and qualitative methods from faecal samples of neonates. Settings and Design: A prospective study was carried out in two groups in a tertiary care hospital, Group A-comprised preterm infant and in group B-full term infants. Materials and Methods: Sixty two preterm infants with the weight < 1500 gm and gestation age < 34 weeks and twenty nine full term infants with 4 weeks of age were included. Quantitation of bacterial load was done by ten-fold serial dilutions on respective media. Statistical Analysis: The data were analyzed by using EPIINFO-Ver 6.04. Results and Conclusions: The predominant aerobic bacterium was Klebsiella pneumoniae. In pre term infants aerobic bacteria were colonized with an average of 2.1 and anaerobic bacteria 0.1. Quantitation showed faecal bacterial colony count ranging from 104-1013 CFU/gms. Gram negative and gram positive bacteria increased gradually over an interval of 2 to 3 weeks. Mean log CFU of gram negative bacteria and gram positive bacteria were statistically insignificant from day 3 to day 14 (P > 0.05). On day 21 there was a significant change in colonization of both bacterial sp (P < 0.05). Potential pathogenic aerobic bacteria dominate the intestinal flora of premature babies nursed in neonatal unit. There is a need to investigate interventions to offset this imbalance in gut micro-ecology of premature babies. PMID:23326075

  20. Analyzing the Curricula of Doctor of Philosophy in Educational Technology-Related Programs in the United States

    ERIC Educational Resources Information Center

    Almaden, Abdullah; Ku, Heng-Yu

    2017-01-01

    The purpose of this study was to analyze on-campus and online PhD programs in educational technology-related fields in the United States. In particular, it sought to evaluate the most common program titles; core, elective, and research courses based on program curricula. The research design was quantitative content analysis and data were collected…

  1. Using multiple PCR and CE with chemiluminescence detection for simultaneous qualitative and quantitative analysis of genetically modified organism.

    PubMed

    Guo, Longhua; Qiu, Bin; Chi, Yuwu; Chen, Guonan

    2008-09-01

    In this paper, an ultrasensitive CE-CL detection system coupled with a novel double-on-column coaxial flow detection interface was developed for the detection of PCR products. A reliable procedure based on this system had been demonstrated for qualitative and quantitative analysis of genetically modified organism-the detection of Roundup Ready Soy (RRS) samples was presented as an example. The promoter, terminator, function and two reference genes of RRS were amplified with multiplex PCR simultaneously. After that, the multiplex PCR products were labeled with acridinium ester at the 5'-terminal through an amino modification and then analyzed by the proposed CE-CL system. Reproducibility of analysis times and peak heights for the CE-CL analysis were determined to be better than 0.91 and 3.07% (RSD, n=15), respectively, for three consecutive days. It was shown that this method could accurately and qualitatively detect RRS standards and the simulative samples. The evaluation in terms of quantitative analysis of RRS provided by this new method was confirmed by comparing our assay results with those of the standard real-time quantitative PCR (RT-QPCR) using SYBR Green I dyes. The results showed a good coherence between the two methods. This approach demonstrated the possibility for accurate qualitative and quantitative detection of GM plants in a single run.

  2. Quantitative analysis of drug distribution by ambient mass spectrometry imaging method with signal extinction normalization strategy and inkjet-printing technology.

    PubMed

    Luo, Zhigang; He, Jingjing; He, Jiuming; Huang, Lan; Song, Xiaowei; Li, Xin; Abliz, Zeper

    2018-03-01

    Quantitative mass spectrometry imaging (MSI) is a robust approach that provides both quantitative and spatial information for drug candidates' research. However, because of complicated signal suppression and interference, acquiring accurate quantitative information from MSI data remains a challenge, especially for whole-body tissue sample. Ambient MSI techniques using spray-based ionization appear to be ideal for pharmaceutical quantitative MSI analysis. However, it is more challenging, as it involves almost no sample preparation and is more susceptible to ion suppression/enhancement. Herein, based on our developed air flow-assisted desorption electrospray ionization (AFADESI)-MSI technology, an ambient quantitative MSI method was introduced by integrating inkjet-printing technology with normalization of the signal extinction coefficient (SEC) using the target compound itself. The method utilized a single calibration curve to quantify multiple tissue types. Basic blue 7 and an antitumor drug candidate (S-(+)-deoxytylophorinidine, CAT) were chosen to initially validate the feasibility and reliability of the quantitative MSI method. Rat tissue sections (heart, kidney, and brain) administered with CAT was then analyzed. The quantitative MSI analysis results were cross-validated by LC-MS/MS analysis data of the same tissues. The consistency suggests that the approach is able to fast obtain the quantitative MSI data without introducing interference into the in-situ environment of the tissue sample, and is potential to provide a high-throughput, economical and reliable approach for drug discovery and development. Copyright © 2017 Elsevier B.V. All rights reserved.

  3. Quantitative, equal carbon response HSQC experiment, QEC-HSQC

    NASA Astrophysics Data System (ADS)

    Mäkelä, Valtteri; Helminen, Jussi; Kilpeläinen, Ilkka; Heikkinen, Sami

    2016-10-01

    Quantitative NMR has become increasingly useful and popular in recent years, with many new and emerging applications in metabolomics, quality control, reaction monitoring and other types of mixture analysis. While sensitive and simple to acquire, the low resolving power of 1D 1H NMR spectra can be a limiting factor when analyzing complex mixtures. This drawback can be solved by observing a different type of nuclei offering improved resolution or with multidimensional experiments, such as HSQC. In this paper, we present a novel Quantitative, Equal Carbon HSQC (QEC-HSQC) experiment providing an equal response across different type of carbons regardless of the number of attached protons, in addition to an uniform response over a wide range of 1JCH couplings. This enables rapid quantification and integration over multiple signals without the need for complete resonance assignments and simplifies the integration of overlapping signals.

  4. Quantitative analysis of lentiviral transgene expression in mice over seven generations.

    PubMed

    Wang, Yong; Song, Yong-tao; Liu, Qin; Liu, Cang'e; Wang, Lu-lu; Liu, Yu; Zhou, Xiao-yang; Wu, Jun; Wei, Hong

    2010-10-01

    Lentiviral transgenesis is now recognized as an extremely efficient and cost-effective method to produce transgenic animals. Transgenes delivered by lentiviral vectors exhibited inheritable expression in many species including those which are refractory to genetic modification such as non-human primates. However, epigenetic modification was frequently observed in lentiviral integrants, and transgene expression found to be inversely correlated with methylation density. Recent data showed that about one-third lentiviral integrants exhibited hypermethylation and low expression, but did not demonstrate whether those integrants with high expression could remain constant expression and hypomethylated during long term germline transmission. In this study, using lentiviral eGFP transgenic mice as the experimental animals, lentiviral eGFP expression levels and its integrant numbers in genome were quantitatively analyzed by fluorescent quantitative polymerase-chain reaction (FQ-PCR), using the house-keeping gene ribosomal protein S18 (Rps18) and the single copy gene fatty acid binding protein of the intestine (Fabpi) as the internal controls respectively. The methylation densities of the integrants were quantitatively analyzed by bisulfite sequencing. We found that the lentiviral integrants with high expression exhibited a relative constant expression level per integrant over at least seven generations. Besides, the individuals containing these integrants exhibited eGFP expression levels which were positively and almost linearly correlated with the integrant numbers in their genomes, suggesting that no remarkable position effect on transgene expression of the integrants analyzed was observed. In addition, over seven generations the methylation density of these integrants did not increase, but rather decreased remarkably, indicating that these high expressing integrants were not subjected to de novo methylation during at least seven generations of germline transmission. Taken

  5. Modern quantitative schlieren techniques

    NASA Astrophysics Data System (ADS)

    Hargather, Michael; Settles, Gary

    2010-11-01

    Schlieren optical techniques have traditionally been used to qualitatively visualize refractive flowfields in transparent media. Modern schlieren optics, however, are increasingly focused on obtaining quantitative information such as temperature and density fields in a flow -- once the sole purview of interferometry -- without the need for coherent illumination. Quantitative data are obtained from schlieren images by integrating the measured refractive index gradient to obtain the refractive index field in an image. Ultimately this is converted to a density or temperature field using the Gladstone-Dale relationship, an equation of state, and geometry assumptions for the flowfield of interest. Several quantitative schlieren methods are reviewed here, including background-oriented schlieren (BOS), schlieren using a weak lens as a "standard," and "rainbow schlieren." Results are presented for the application of these techniques to measure density and temperature fields across a supersonic turbulent boundary layer and a low-speed free-convection boundary layer in air. Modern equipment, including digital cameras, LED light sources, and computer software that make this possible are also discussed.

  6. Mixing Qualitative and Quantitative Methods: Insights into Design and Analysis Issues

    ERIC Educational Resources Information Center

    Lieber, Eli

    2009-01-01

    This article describes and discusses issues related to research design and data analysis in the mixing of qualitative and quantitative methods. It is increasingly desirable to use multiple methods in research, but questions arise as to how best to design and analyze the data generated by mixed methods projects. I offer a conceptualization for such…

  7. Agency Problems and Airport Security: Quantitative and Qualitative Evidence on the Impact of Security Training.

    PubMed

    de Gramatica, Martina; Massacci, Fabio; Shim, Woohyun; Turhan, Uğur; Williams, Julian

    2017-02-01

    We analyze the issue of agency costs in aviation security by combining results from a quantitative economic model with a qualitative study based on semi-structured interviews. Our model extends previous principal-agent models by combining the traditional fixed and varying monetary responses to physical and cognitive effort with nonmonetary welfare and potentially transferable value of employees' own human capital. To provide empirical evidence for the tradeoffs identified in the quantitative model, we have undertaken an extensive interview process with regulators, airport managers, security personnel, and those tasked with training security personnel from an airport operating in a relatively high-risk state, Turkey. Our results indicate that the effectiveness of additional training depends on the mix of "transferable skills" and "emotional" buy-in of the security agents. Principals need to identify on which side of a critical tipping point their agents are to ensure that additional training, with attached expectations of the burden of work, aligns the incentives of employees with the principals' own objectives. © 2016 Society for Risk Analysis.

  8. 10 CFR 26.169 - Reporting Results.

    Code of Federal Regulations, 2012 CFR

    2012-01-01

    ... request. The laboratory shall routinely provide quantitative values for confirmatory opiate test results... requested quantitative values for the test result. (3) For a specimen that has an adulterated or substituted... of the standard curve, the laboratory may report to the MRO that the quantitative value “exceeds the...

  9. 10 CFR 26.169 - Reporting Results.

    Code of Federal Regulations, 2013 CFR

    2013-01-01

    ... request. The laboratory shall routinely provide quantitative values for confirmatory opiate test results... requested quantitative values for the test result. (3) For a specimen that has an adulterated or substituted... of the standard curve, the laboratory may report to the MRO that the quantitative value “exceeds the...

  10. 10 CFR 26.169 - Reporting Results.

    Code of Federal Regulations, 2010 CFR

    2010-01-01

    ... request. The laboratory shall routinely provide quantitative values for confirmatory opiate test results... requested quantitative values for the test result. (3) For a specimen that has an adulterated or substituted... of the standard curve, the laboratory may report to the MRO that the quantitative value “exceeds the...

  11. 10 CFR 26.169 - Reporting Results.

    Code of Federal Regulations, 2011 CFR

    2011-01-01

    ... request. The laboratory shall routinely provide quantitative values for confirmatory opiate test results... requested quantitative values for the test result. (3) For a specimen that has an adulterated or substituted... of the standard curve, the laboratory may report to the MRO that the quantitative value “exceeds the...

  12. 10 CFR 26.169 - Reporting Results.

    Code of Federal Regulations, 2014 CFR

    2014-01-01

    ... request. The laboratory shall routinely provide quantitative values for confirmatory opiate test results... requested quantitative values for the test result. (3) For a specimen that has an adulterated or substituted... of the standard curve, the laboratory may report to the MRO that the quantitative value “exceeds the...

  13. Structured Qualitative Research: Organizing “Mountains of Words” for Data Analysis, both Qualitative and Quantitative

    PubMed Central

    Johnson, Bruce D.; Dunlap, Eloise; Benoit, Ellen

    2008-01-01

    Qualitative research creates mountains of words. U.S. federal funding supports mostly structured qualitative research, which is designed to test hypotheses using semi-quantitative coding and analysis. The authors have 30 years of experience in designing and completing major qualitative research projects, mainly funded by the US National Institute on Drug Abuse [NIDA]. This article reports on strategies for planning, organizing, collecting, managing, storing, retrieving, analyzing, and writing about qualitative data so as to most efficiently manage the mountains of words collected in large-scale ethnographic projects. Multiple benefits accrue from this approach. Several different staff members can contribute to the data collection, even when working from remote locations. Field expenditures are linked to units of work so productivity is measured, many staff in various locations have access to use and analyze the data, quantitative data can be derived from data that is primarily qualitative, and improved efficiencies of resources are developed. The major difficulties involve a need for staff who can program and manage large databases, and who can be skillful analysts of both qualitative and quantitative data. PMID:20222777

  14. Quantitative evaluation research of glare from automotive headlamps

    NASA Astrophysics Data System (ADS)

    Wang, Tiecheng; Qian, Rui; Cao, Ye; Gao, Mingqiu

    2018-01-01

    This study concerns the quantized evaluation research of glare from automotive headlamps. In the actual regulations, only one point in the test screen is set for judging whether driver can bear the light caused by headlamps of opposing vehicle. To evaluating practical effect of glare, we accept a glare zone with the probability distribution information of the oncoming driver's eye position. In this focus area, glare level of headlamp is represented by weighted luminous flux. To confirm the most comfortable illuminance value to human eyes at 50 m, we used test point B50L as observation position, and collected 1,000 subjective evaluation data from 20 test personnel in different ages during two months. Basing on the assessment results, we calculated 0.60 lx as recommended value for standardized testing procedure at 25 m. Then we figured out 0.38 lm as optimum value, and 0.25 / 1.20 lm as limiting values depending on regulations. We tested 40 sample vehicles with different levels to verify the sectional nonlinear quantitative evaluation method we designed, and analyzed the typical test results.

  15. Proficiency testing as a basis for estimating uncertainty of measurement: application to forensic alcohol and toxicology quantitations.

    PubMed

    Wallace, Jack

    2010-05-01

    While forensic laboratories will soon be required to estimate uncertainties of measurement for those quantitations reported to the end users of the information, the procedures for estimating this have been little discussed in the forensic literature. This article illustrates how proficiency test results provide the basis for estimating uncertainties in three instances: (i) For breath alcohol analyzers the interlaboratory precision is taken as a direct measure of uncertainty. This approach applies when the number of proficiency tests is small. (ii) For blood alcohol, the uncertainty is calculated from the differences between the laboratory's proficiency testing results and the mean quantitations determined by the participants; this approach applies when the laboratory has participated in a large number of tests. (iii) For toxicology, either of these approaches is useful for estimating comparability between laboratories, but not for estimating absolute accuracy. It is seen that data from proficiency tests enable estimates of uncertainty that are empirical, simple, thorough, and applicable to a wide range of concentrations.

  16. Measurements in quantitative research: how to select and report on research instruments.

    PubMed

    Hagan, Teresa L

    2014-07-01

    Measures exist to numerically represent degrees of attributes. Quantitative research is based on measurement and is conducted in a systematic, controlled manner. These measures enable researchers to perform statistical tests, analyze differences between groups, and determine the effectiveness of treatments. If something is not measurable, it cannot be tested.

  17. Pleiotropy Analysis of Quantitative Traits at Gene Level by Multivariate Functional Linear Models

    PubMed Central

    Wang, Yifan; Liu, Aiyi; Mills, James L.; Boehnke, Michael; Wilson, Alexander F.; Bailey-Wilson, Joan E.; Xiong, Momiao; Wu, Colin O.; Fan, Ruzong

    2015-01-01

    In genetics, pleiotropy describes the genetic effect of a single gene on multiple phenotypic traits. A common approach is to analyze the phenotypic traits separately using univariate analyses and combine the test results through multiple comparisons. This approach may lead to low power. Multivariate functional linear models are developed to connect genetic variant data to multiple quantitative traits adjusting for covariates for a unified analysis. Three types of approximate F-distribution tests based on Pillai–Bartlett trace, Hotelling–Lawley trace, and Wilks’s Lambda are introduced to test for association between multiple quantitative traits and multiple genetic variants in one genetic region. The approximate F-distribution tests provide much more significant results than those of F-tests of univariate analysis and optimal sequence kernel association test (SKAT-O). Extensive simulations were performed to evaluate the false positive rates and power performance of the proposed models and tests. We show that the approximate F-distribution tests control the type I error rates very well. Overall, simultaneous analysis of multiple traits can increase power performance compared to an individual test of each trait. The proposed methods were applied to analyze (1) four lipid traits in eight European cohorts, and (2) three biochemical traits in the Trinity Students Study. The approximate F-distribution tests provide much more significant results than those of F-tests of univariate analysis and SKAT-O for the three biochemical traits. The approximate F-distribution tests of the proposed functional linear models are more sensitive than those of the traditional multivariate linear models that in turn are more sensitive than SKAT-O in the univariate case. The analysis of the four lipid traits and the three biochemical traits detects more association than SKAT-O in the univariate case. PMID:25809955

  18. Pleiotropy analysis of quantitative traits at gene level by multivariate functional linear models.

    PubMed

    Wang, Yifan; Liu, Aiyi; Mills, James L; Boehnke, Michael; Wilson, Alexander F; Bailey-Wilson, Joan E; Xiong, Momiao; Wu, Colin O; Fan, Ruzong

    2015-05-01

    In genetics, pleiotropy describes the genetic effect of a single gene on multiple phenotypic traits. A common approach is to analyze the phenotypic traits separately using univariate analyses and combine the test results through multiple comparisons. This approach may lead to low power. Multivariate functional linear models are developed to connect genetic variant data to multiple quantitative traits adjusting for covariates for a unified analysis. Three types of approximate F-distribution tests based on Pillai-Bartlett trace, Hotelling-Lawley trace, and Wilks's Lambda are introduced to test for association between multiple quantitative traits and multiple genetic variants in one genetic region. The approximate F-distribution tests provide much more significant results than those of F-tests of univariate analysis and optimal sequence kernel association test (SKAT-O). Extensive simulations were performed to evaluate the false positive rates and power performance of the proposed models and tests. We show that the approximate F-distribution tests control the type I error rates very well. Overall, simultaneous analysis of multiple traits can increase power performance compared to an individual test of each trait. The proposed methods were applied to analyze (1) four lipid traits in eight European cohorts, and (2) three biochemical traits in the Trinity Students Study. The approximate F-distribution tests provide much more significant results than those of F-tests of univariate analysis and SKAT-O for the three biochemical traits. The approximate F-distribution tests of the proposed functional linear models are more sensitive than those of the traditional multivariate linear models that in turn are more sensitive than SKAT-O in the univariate case. The analysis of the four lipid traits and the three biochemical traits detects more association than SKAT-O in the univariate case. © 2015 WILEY PERIODICALS, INC.

  19. Quantitative analysis of professionally trained versus untrained voices.

    PubMed

    Siupsinskiene, Nora

    2003-01-01

    The aim of this study was to compare healthy trained and untrained voices as well as healthy and dysphonic trained voices in adults using combined voice range profile and aerodynamic tests, to define the normal range limiting values of quantitative voice parameters and to select the most informative quantitative voice parameters for separation between healthy and dysphonic trained voices. Three groups of persons were evaluated. One hundred eighty six healthy volunteers were divided into two groups according to voice training: non-professional speakers group consisted of 106 untrained voices persons (36 males and 70 females) and professional speakers group--of 80 trained voices persons (21 males and 59 females). Clinical group consisted of 103 dysphonic professional speakers (23 males and 80 females) with various voice disorders. Eighteen quantitative voice parameters from combined voice range profile (VRP) test were analyzed: 8 of voice range profile, 8 of speaking voice, overall vocal dysfunction degree and coefficient of sound, and aerodynamic maximum phonation time. Analysis showed that healthy professional speakers demonstrated expanded vocal abilities in comparison to healthy non-professional speakers. Quantitative voice range profile parameters- pitch range, high frequency limit, area of high frequencies and coefficient of sound differed significantly between healthy professional and non-professional voices, and were more informative than speaking voice or aerodynamic parameters in showing the voice training. Logistic stepwise regression revealed that VRP area in high frequencies was sufficient to discriminate between healthy and dysphonic professional speakers for male subjects (overall discrimination accuracy--81.8%) and combination of three quantitative parameters (VRP high frequency limit, maximum voice intensity and slope of speaking curve) for female subjects (overall model discrimination accuracy--75.4%). We concluded that quantitative voice assessment

  20. Spatiotemporal Characterization of a Fibrin Clot Using Quantitative Phase Imaging

    PubMed Central

    Gannavarpu, Rajshekhar; Bhaduri, Basanta; Tangella, Krishnarao; Popescu, Gabriel

    2014-01-01

    Studying the dynamics of fibrin clot formation and its morphology is an important problem in biology and has significant impact for several scientific and clinical applications. We present a label-free technique based on quantitative phase imaging to address this problem. Using quantitative phase information, we characterized fibrin polymerization in real-time and present a mathematical model describing the transition from liquid to gel state. By exploiting the inherent optical sectioning capability of our instrument, we measured the three-dimensional structure of the fibrin clot. From this data, we evaluated the fractal nature of the fibrin network and extracted the fractal dimension. Our non-invasive and speckle-free approach analyzes the clotting process without the need for external contrast agents. PMID:25386701

  1. Analyzing the substitution effect on the CoMFA results within the framework of density functional theory (DFT).

    PubMed

    Morales-Bayuelo, Alejandro

    2016-07-01

    Though QSAR was originally developed in the context of physical organic chemistry, it has been applied very extensively to chemicals (drugs) which act on biological systems, in this idea one of the most important QSAR methods is the 3D QSAR model. However, due to the complexity of understanding the results it is necessary to postulate new methodologies to highlight their physical-chemical meaning. In this sense, this work postulates new insights to understand the CoMFA results using molecular quantum similarity and chemical reactivity descriptors within the framework of density functional theory. To obtain these insights a simple theoretical scheme involving quantum similarity (overlap, coulomb operators, their euclidean distances) and chemical reactivity descriptors such as chemical potential (μ), hardness (ɳ), softness (S), electrophilicity (ω), and the Fukui functions, was used to understand the substitution effect. In this sense, this methodology can be applied to analyze the biological activity and the stabilization process in the non-covalent interactions on a particular molecular set taking a reference compound.

  2. Comparison of chemistry analytes between 2 portable, commercially available analyzers and a conventional laboratory analyzer in reptiles.

    PubMed

    McCain, Stephanie L; Flatland, Bente; Schumacher, Juergen P; Clarke Iii, Elsburgh O; Fry, Michael M

    2010-12-01

    Advantages of handheld and small bench-top biochemical analyzers include requirements for smaller sample volume and practicality for use in the field or in practices, but little has been published on the performance of these instruments compared with standard reference methods in analysis of reptilian blood. The aim of this study was to compare reptilian blood biochemical values obtained using the Abaxis VetScan Classic bench-top analyzer and a Heska i-STAT handheld analyzer with values obtained using a Roche Hitachi 911 chemical analyzer. Reptiles, including 14 bearded dragons (Pogona vitticeps), 4 blue-tongued skinks (Tiliqua gigas), 8 Burmese star tortoises (Geochelone platynota), 10 Indian star tortoises (Geochelone elegans), 5 red-tailed boas (Boa constrictor), and 5 Northern pine snakes (Pituophis melanoleucus melanoleucus), were manually restrained, and a single blood sample was obtained and divided for analysis. Results for concentrations of albumin, bile acids, calcium, glucose, phosphates, potassium, sodium, total protein, and uric acid and activities of aspartate aminotransferase and creatine kinase obtained from the VetScan Classic and Hitachi 911 were compared. Results for concentrations of chloride, glucose, potassium, and sodium obtained from the i-STAT and Hitachi 911 were compared. Compared with results from the Hitachi 911, those from the VetScan Classic and i-STAT had variable correlations, and constant or proportional bias was found for many analytes. Bile acid data could not be evaluated because results for 44 of 45 samples fell below the lower linearity limit of the VetScan Classic. Although the 2 portable instruments might provide measurements with clinical utility, there were significant differences compared with the reference analyzer, and development of analyzer-specific reference intervals is recommended. ©2010 American Society for Veterinary Clinical Pathology.

  3. Time-Resolved Fluorescent Immunochromatography of Aflatoxin B1 in Soybean Sauce: A Rapid and Sensitive Quantitative Analysis.

    PubMed

    Wang, Du; Zhang, Zhaowei; Li, Peiwu; Zhang, Qi; Zhang, Wen

    2016-07-14

    Rapid and quantitative sensing of aflatoxin B1 with high sensitivity and specificity has drawn increased attention of studies investigating soybean sauce. A sensitive and rapid quantitative immunochromatographic sensing method was developed for the detection of aflatoxin B1 based on time-resolved fluorescence. It combines the advantages of time-resolved fluorescent sensing and immunochromatography. The dynamic range of a competitive and portable immunoassay was 0.3-10.0 µg·kg(-1), with a limit of detection (LOD) of 0.1 µg·kg(-1) and recoveries of 87.2%-114.3%, within 10 min. The results showed good correlation (R² > 0.99) between time-resolved fluorescent immunochromatographic strip test and high performance liquid chromatography (HPLC). Soybean sauce samples analyzed using time-resolved fluorescent immunochromatographic strip test revealed that 64.2% of samples contained aflatoxin B1 at levels ranging from 0.31 to 12.5 µg·kg(-1). The strip test is a rapid, sensitive, quantitative, and cost-effective on-site screening technique in food safety analysis.

  4. On-Demand Urine Analyzer

    NASA Technical Reports Server (NTRS)

    Farquharson, Stuart; Inscore, Frank; Shende, Chetan

    2010-01-01

    A lab-on-a-chip was developed that is capable of extracting biochemical indicators from urine samples and generating their surface-enhanced Raman spectra (SERS) so that the indicators can be quantified and identified. The development was motivated by the need to monitor and assess the effects of extended weightlessness, which include space motion sickness and loss of bone and muscle mass. The results may lead to developments of effective exercise programs and drug regimes that would maintain astronaut health. The analyzer containing the lab-on-a- chip includes materials to extract 3- methylhistidine (a muscle-loss indicator) and Risedronate (a bone-loss indicator) from the urine sample and detect them at the required concentrations using a Raman analyzer. The lab-on- a-chip has both an extractive material and a SERS-active material. The analyzer could be used to monitor the onset of diseases, such as osteoporosis.

  5. Quantitative Study on Corrosion of Steel Strands Based on Self-Magnetic Flux Leakage.

    PubMed

    Xia, Runchuan; Zhou, Jianting; Zhang, Hong; Liao, Leng; Zhao, Ruiqiang; Zhang, Zeyu

    2018-05-02

    This paper proposed a new computing method to quantitatively and non-destructively determine the corrosion of steel strands by analyzing the self-magnetic flux leakage (SMFL) signals from them. The magnetic dipole model and three growth models (Logistic model, Exponential model, and Linear model) were proposed to theoretically analyze the characteristic value of SMFL. Then, the experimental study on the corrosion detection by the magnetic sensor was carried out. The setup of the magnetic scanning device and signal collection method were also introduced. The results show that the Logistic Growth model is verified as the optimal model for calculating the magnetic field with good fitting effects. Combined with the experimental data analysis, the amplitudes of the calculated values ( B xL ( x,z ) curves) agree with the measured values in general. This method provides significant application prospects for the evaluation of the corrosion and the residual bearing capacity of steel strand.

  6. IWGT report on quantitative approaches to genotoxicity risk ...

    EPA Pesticide Factsheets

    This is the second of two reports from the International Workshops on Genotoxicity Testing (IWGT) Working Group on Quantitative Approaches to Genetic Toxicology Risk Assessment (the QWG). The first report summarized the discussions and recommendations of the QWG related to the need for quantitative dose–response analysis of genetic toxicology data, the existence and appropriate evaluation of threshold responses, and methods to analyze exposure-response relationships and derive points of departure (PoDs) from which acceptable exposure levels could be determined. This report summarizes the QWG discussions and recommendations regarding appropriate approaches to evaluate exposure-related risks of genotoxic damage, including extrapolation below identified PoDs and across test systems and species. Recommendations include the selection of appropriate genetic endpoints and target tissues, uncertainty factors and extrapolation methods to be considered, the importance and use of information on mode of action, toxicokinetics, metabolism, and exposure biomarkers when using quantitative exposure-response data to determine acceptable exposure levels in human populations or to assess the risk associated with known or anticipated exposures. The empirical relationship between genetic damage (mutation and chromosomal aberration) and cancer in animal models was also examined. It was concluded that there is a general correlation between cancer induction and mutagenic and/or clast

  7. Quantitative Analysis of the Usage of the COSMOS Science Education Portal

    ERIC Educational Resources Information Center

    Sotiriou, Sofoklis; Bogner, Franz X.; Neofotistos, George

    2011-01-01

    A quantitative method of mapping the web usage of an innovative educational portal is applied to analyze the behaviour of users of the COSMOS Science Education Portal. The COSMOS Portal contains user-generated resources (that are uploaded by its users). It has been designed to support a science teacher's search, retrieval and access to both,…

  8. A Quantitative and Combinatorial Approach to Non-Linear Meanings of Multiplication

    ERIC Educational Resources Information Center

    Tillema, Erik; Gatza, Andrew

    2016-01-01

    We provide a conceptual analysis of how combinatorics problems have the potential to support students to establish non-linear meanings of multiplication (NLMM). The problems we analyze we have used in a series of studies with 6th, 8th, and 10th grade students. We situate the analysis in prior work on students' quantitative and multiplicative…

  9. Dispersal of Invasive Forest Insects via Recreational Firewood: A Quantitative Analysis

    Treesearch

    Frank H. Koch; Denys Yemshanov; Roger D. Magarey; William D. Smith

    2012-01-01

    Recreational travel is a recognized vector for the spread of invasive species in North America. However, there has been little quantitative analysis of the risks posed by such travel and the associated transport of firewood. In this study, we analyzed the risk of forest insect spread with firewood and estimated related dispersal parameters for application in...

  10. Quantitative MR assessment of structural changes in white matter of children treated for ALL

    NASA Astrophysics Data System (ADS)

    Reddick, Wilburn E.; Glass, John O.; Mulhern, Raymond K.

    2001-07-01

    Our research builds on the hypothesis that white matter damage resulting from therapy spans a continuum of severity that can be reliably probed using non-invasive MR technology. This project focuses on children treated for ALL with a regimen containing seven courses of high-dose methotrexate (HDMTX) which is known to cause leukoencephalopathy. Axial FLAIR, T1-, T2-, and PD-weighted images were acquired, registered and then analyzed with a hybrid neural network segmentation algorithm to identify normal brain parenchyma and leukoencephalopathy. Quantitative T1 and T2 maps were also analyzed at the level of the basal ganglia and the centrum semiovale. The segmented images were used as mask to identify regions of normal appearing white matter (NAWM) and leukoencephalopathy in the quantitative T1 and T2 maps. We assessed the longitudinal changes in volume, T1 and T2 in NAWM and leukoencephalopathy for 42 patients. The segmentation analysis revealed that 69% of patients had leukoencephalopathy after receiving seven courses of HDMTX. The leukoencephalopathy affected approximately 17% of the patients' white matter volume on average (range 2% - 38%). Relaxation rates in the NAWM were not significantly changed between the 1st and 7th courses. Regions of leukoencephalopathy exhibited a 13% elevation in T1 and a 37% elevation in T2 relaxation rates.

  11. GProX, a user-friendly platform for bioinformatics analysis and visualization of quantitative proteomics data.

    PubMed

    Rigbolt, Kristoffer T G; Vanselow, Jens T; Blagoev, Blagoy

    2011-08-01

    Recent technological advances have made it possible to identify and quantify thousands of proteins in a single proteomics experiment. As a result of these developments, the analysis of data has become the bottleneck of proteomics experiment. To provide the proteomics community with a user-friendly platform for comprehensive analysis, inspection and visualization of quantitative proteomics data we developed the Graphical Proteomics Data Explorer (GProX)(1). The program requires no special bioinformatics training, as all functions of GProX are accessible within its graphical user-friendly interface which will be intuitive to most users. Basic features facilitate the uncomplicated management and organization of large data sets and complex experimental setups as well as the inspection and graphical plotting of quantitative data. These are complemented by readily available high-level analysis options such as database querying, clustering based on abundance ratios, feature enrichment tests for e.g. GO terms and pathway analysis tools. A number of plotting options for visualization of quantitative proteomics data is available and most analysis functions in GProX create customizable high quality graphical displays in both vector and bitmap formats. The generic import requirements allow data originating from essentially all mass spectrometry platforms, quantitation strategies and software to be analyzed in the program. GProX represents a powerful approach to proteomics data analysis providing proteomics experimenters with a toolbox for bioinformatics analysis of quantitative proteomics data. The program is released as open-source and can be freely downloaded from the project webpage at http://gprox.sourceforge.net.

  12. GProX, a User-Friendly Platform for Bioinformatics Analysis and Visualization of Quantitative Proteomics Data*

    PubMed Central

    Rigbolt, Kristoffer T. G.; Vanselow, Jens T.; Blagoev, Blagoy

    2011-01-01

    Recent technological advances have made it possible to identify and quantify thousands of proteins in a single proteomics experiment. As a result of these developments, the analysis of data has become the bottleneck of proteomics experiment. To provide the proteomics community with a user-friendly platform for comprehensive analysis, inspection and visualization of quantitative proteomics data we developed the Graphical Proteomics Data Explorer (GProX)1. The program requires no special bioinformatics training, as all functions of GProX are accessible within its graphical user-friendly interface which will be intuitive to most users. Basic features facilitate the uncomplicated management and organization of large data sets and complex experimental setups as well as the inspection and graphical plotting of quantitative data. These are complemented by readily available high-level analysis options such as database querying, clustering based on abundance ratios, feature enrichment tests for e.g. GO terms and pathway analysis tools. A number of plotting options for visualization of quantitative proteomics data is available and most analysis functions in GProX create customizable high quality graphical displays in both vector and bitmap formats. The generic import requirements allow data originating from essentially all mass spectrometry platforms, quantitation strategies and software to be analyzed in the program. GProX represents a powerful approach to proteomics data analysis providing proteomics experimenters with a toolbox for bioinformatics analysis of quantitative proteomics data. The program is released as open-source and can be freely downloaded from the project webpage at http://gprox.sourceforge.net. PMID:21602510

  13. Quantitative analysis on PUVA-induced skin photodamages using optical coherence tomography

    NASA Astrophysics Data System (ADS)

    Zhai, Juan; Guo, Zhouyi; Liu, Zhiming; Xiong, Honglian; Zeng, Changchun; Jin, Ying

    2009-08-01

    Psoralen plus ultraviolet A radiation (PUVA) therapy is a very important clinical treatment of skin diseases such as vitiligo and psoriasis, but associated with an increased risk of skin photodamages especially photoaging. Since skin biopsy alters the original skin morphology and always requires an iatrogenic trauma, optical coherence tomography (OCT) appears to be a promising technique to study skin damage in vivo. In this study, the Balb/c mice had 8-methoxypsralen (8-MOP) treatment prior to UVA radiation was used as PUVA-induced photo-damaged modal. The OCT imaging of photo-damaged group (modal) and normal group (control) in vivo was obtained of mice dorsal skin at 0, 24, 48, 72 hours after irradiation respectively. And then the results were quantitatively analyzed combined with histological information. The experimental results showed that, PUVA-induced photo-damaged skin had an increase in epidermal thickness (ET), a reduction of attenuation coefficient in OCT images signal, and an increase in brightness of the epidermis layer compared with the control group. In conclusion, noninvasive high-resolution imaging techniques such as OCT may be a promising tool for photobiological studies aimed at assessing photo-damage and repair processes in vivo. It can be used to quantitative analysis of changes in photo-damaged skin, such as the ET and collagen in dermis, provides a theoretical basis for treatment and prevention of skin photodamages.

  14. Role of metal ions in catalysis by HIV integrase analyzed using a quantitative PCR disintegration assay.

    PubMed

    Diamond, Tracy L; Bushman, Frederic D

    2006-01-01

    Paired metal ions have been proposed to be central to the catalytic mechanisms of RNase H nucleases, bacterial transposases, Holliday junction resolvases, retroviral integrases and many other enzymes. Here we present a sensitive assay for DNA transesterification in which catalysis by human immunodeficiency virus-type 1 (HIV-1) integrase (IN) connects two DNA strands (disintegration reaction), allowing detection using quantitative PCR (qPCR). We present evidence suggesting that the three acidic residues of the IN active site function through metal binding using metal rescue. In this method, the catalytic acidic residues were each substituted with cysteines. Mn2+ binds tightly to the sulfur atoms of the cysteine residues, but Mg2+ does not. We found that Mn2+, but not Mg2+, could rescue catalysis of each cysteine-substituted enzyme, providing evidence for functionally important metal binding by all three residues. We also used the PCR-boosted assay to show that HIV-1 IN could carry out transesterification reactions involving DNA 5' hydroxyl groups as well as 3' hydroxyls as nucleophiles. Lastly, we show that Mn2+ by itself (i.e. without enzyme) can catalyze formation of a low level of PCR-amplifiable product under extreme conditions, allowing us to estimate the rate enhancement due to the IN-protein scaffold as at least 60 million-fold.

  15. Role of metal ions in catalysis by HIV integrase analyzed using a quantitative PCR disintegration assay

    PubMed Central

    Diamond, Tracy L.; Bushman, Frederic D.

    2006-01-01

    Paired metal ions have been proposed to be central to the catalytic mechanisms of RNase H nucleases, bacterial transposases, Holliday junction resolvases, retroviral integrases and many other enzymes. Here we present a sensitive assay for DNA transesterification in which catalysis by human immunodeficiency virus-type 1 (HIV-1) integrase (IN) connects two DNA strands (disintegration reaction), allowing detection using quantitative PCR (qPCR). We present evidence suggesting that the three acidic residues of the IN active site function through metal binding using metal rescue. In this method, the catalytic acidic residues were each substituted with cysteines. Mn2+ binds tightly to the sulfur atoms of the cysteine residues, but Mg2+ does not. We found that Mn2+, but not Mg2+, could rescue catalysis of each cysteine-substituted enzyme, providing evidence for functionally important metal binding by all three residues. We also used the PCR-boosted assay to show that HIV-1 IN could carry out transesterification reactions involving DNA 5′ hydroxyl groups as well as 3′ hydroxyls as nucleophiles. Lastly, we show that Mn2+ by itself (i.e. without enzyme) can catalyze formation of a low level of PCR-amplifiable product under extreme conditions, allowing us to estimate the rate enhancement due to the IN-protein scaffold as at least 60 million-fold. PMID:17085478

  16. Improving quantitative gas chromatography-electron ionization mass spectrometry results using a modified ion source: demonstration for a pharmaceutical application.

    PubMed

    D'Autry, Ward; Wolfs, Kris; Hoogmartens, Jos; Adams, Erwin; Van Schepdael, Ann

    2011-07-01

    Gas chromatography-mass spectrometry is a well established analytical technique. However, mass spectrometers with electron ionization sources may suffer from signal drifts, hereby negatively influencing quantitative performance. To demonstrate this phenomenon for a real application, a static headspace-gas chromatography method in combination with electron ionization-quadrupole mass spectrometry was optimized for the determination of residual dichloromethane in coronary stent coatings. Validating the method, the quantitative performance of an original stainless steel ion source was compared to that of a modified ion source. Ion source modification included the application of a gold coating on the repeller and exit plate. Several validation aspects such as limit of detection, limit of quantification, linearity and precision were evaluated using both ion sources. It was found that, as expected, the stainless steel ion source suffered from signal drift. As a consequence, non-linearity and high RSD values for repeated analyses were obtained. An additional experiment was performed to check whether an internal standard compound would lead to better results. It was found that the signal drift patterns of the analyte and internal standard were different, consequently leading to high RSD values for the response factor. With the modified ion source however, a more stable signal was observed resulting in acceptable linearity and precision. Moreover, it was also found that sensitivity improved compared to the stainless steel ion source. Finally, the optimized method with the modified ion source was applied to determine residual dichloromethane in the coating of coronary stents. The solvent was detected but found to be below the limit of quantification. Copyright © 2011 Elsevier B.V. All rights reserved.

  17. Shallow Investigations of the Deep Seafloor: Quantitative Morphology in the Levant Basin, Eastern Mediterranean

    NASA Astrophysics Data System (ADS)

    Kanari, M.; Ketter, T.; Tibor, G.; Schattner, U.

    2017-12-01

    We aim to characterize the seafloor morphology and its shallow sub-surface structures and deformations in the deep part of the Levant basin (eastern Mediterranean) using recently acquired high-resolution shallow seismic reflection data and multibeam bathymetry, which allow quantitative analysis of morphology and structure. The Levant basin at the eastern Mediterranean is considered a passive continental margin, where most of the recent geological processes were related in literature to salt tectonics rooted at the Messinian deposits from 6Ma. We analyzed two sets of recently acquired high-resolution data from multibeam bathymetry and 3.5 kHz Chirp sub-bottom seismic reflection in the deep basin of the continental shelf offshore Israel (water depths up to 2100 m). Semi-automatic mapping of seafloor features and seismic data interpretation resulted in quantitative morphological analysis of the seafloor and its underlying sediment with penetration depth up to 60 m. The quantitative analysis and its interpretation are still in progress. Preliminary results reveal distinct morphologies of four major elements: channels, faults, folds and sediment waves, validated by seismic data. From the spatial distribution and orientation analyses of these phenomena, we identify two primary process types which dominate the formation of the seafloor in the Levant basin: structural and sedimentary. Characterization of the geological and geomorphological processes forming the seafloor helps to better understand the transport mechanisms and the relations between sediment transport and deposition in deep water and the shallower parts of the shelf and slope.

  18. Quantitative Methodology: A Guide for Emerging Physical Education and Adapted Physical Education Researchers

    ERIC Educational Resources Information Center

    Haegele, Justin A.; Hodge, Samuel R.

    2015-01-01

    Emerging professionals, particularly senior-level undergraduate and graduate students in kinesiology who have an interest in physical education for individuals with and without disabilities, should understand the basic assumptions of the quantitative research paradigm. Knowledge of basic assumptions is critical for conducting, analyzing, and…

  19. Quantitative single-molecule imaging by confocal laser scanning microscopy.

    PubMed

    Vukojevic, Vladana; Heidkamp, Marcus; Ming, Yu; Johansson, Björn; Terenius, Lars; Rigler, Rudolf

    2008-11-25

    A new approach to quantitative single-molecule imaging by confocal laser scanning microscopy (CLSM) is presented. It relies on fluorescence intensity distribution to analyze the molecular occurrence statistics captured by digital imaging and enables direct determination of the number of fluorescent molecules and their diffusion rates without resorting to temporal or spatial autocorrelation analyses. Digital images of fluorescent molecules were recorded by using fast scanning and avalanche photodiode detectors. In this way the signal-to-background ratio was significantly improved, enabling direct quantitative imaging by CLSM. The potential of the proposed approach is demonstrated by using standard solutions of fluorescent dyes, fluorescently labeled DNA molecules, quantum dots, and the Enhanced Green Fluorescent Protein in solution and in live cells. The method was verified by using fluorescence correlation spectroscopy. The relevance for biological applications, in particular, for live cell imaging, is discussed.

  20. Eifel maars: Quantitative shape characterization of juvenile ash particles (Eifel Volcanic Field, Germany)

    NASA Astrophysics Data System (ADS)

    Rausch, Juanita; Grobéty, Bernard; Vonlanthen, Pierre

    2015-01-01

    The Eifel region in western central Germany is the type locality for maar volcanism, which is classically interpreted to be the result of explosive eruptions due to shallow interaction between magma and external water (i.e. phreatomagmatic eruptions). Sedimentary structures, deposit features and particle morphology found in many maar deposits of the West Eifel Volcanic Field (WEVF), in contrast to deposits in the East Eifel Volcanic Field (EEVF), lack the diagnostic criteria of typical phreatomagmatic deposits. The aim of this study was to determine quantitatively the shape of WEVF and EEVF maar ash particles in order to infer the governing eruption style in Eifel maar volcanoes. The quantitative shape characterization was done by analyzing fractal dimensions of particle contours (125-250 μm sieve fraction) obtained from Scanning electron microscopy (SEM) and SEM micro-computed tomography (SEM micro-CT) images. The fractal analysis (dilation method) and the fractal spectrum technique confirmed that the WEVF and EEVF maar particles have contrasting multifractal shapes. Whereas the low small-scale dimensions of EEVF particles (Eppelsberg Green Unit) coincide with previously published values for phreatomagmatic particles, the WEVF particles (Meerfelder Maar, Pulvermaar and Ulmener Maar) have larger values indicating more complex small-scale features, which are characteristic for magmatic particles. These quantitative results are strengthening the qualitative microscopic observations, that the studied WEVF maar eruptions are rather dominated by magmatic processes. The different eruption styles in the two volcanic fields can be explained by the different geological and hydrological settings found in both regions and the different chemical compositions of the magmas.

  1. Application of Quantitative Microbial Risk Assessment to analyze the public health risk from poor drinking water quality in a low income area in Accra, Ghana.

    PubMed

    Machdar, E; van der Steen, N P; Raschid-Sally, L; Lens, P N L

    2013-04-01

    In Accra, Ghana, a majority of inhabitants lives in over-crowded areas with limited access to piped water supply, which is often also intermittent. This study assessed in a densely populated area the risk from microbial contamination of various sources of drinking water, by conducting a Quantitative Microbiological Risk Assessment (QMRA) to estimate the risk to human health from microorganism exposure and dose-response relationships. Furthermore the cost-effectiveness in reducing the disease burden through targeted interventions was evaluated. Five risk pathways for drinking water were identified through a survey (110 families), namely household storage, private yard taps, communal taps, communal wells and water sachets. Samples from each source were analyzed for Escherichia coli and Ascaris contamination. Published ratios between E. coli and other pathogens were used for the QMRA and disease burden calculations. The major part of the burden of disease originated from E. coli O157:H7 (78%) and the least important contributor was Cryptosporidium (0.01%). Other pathogens contributed 16% (Campylobacter), 5% (Rotavirus) and 0.3% (Ascaris). The sum of the disease burden of these pathogens was 0.5 DALYs per person per year, which is much higher than the WHO reference level. The major contamination pathway was found to be household storage. Disinfection of water at household level was the most cost-effective intervention (<5 USD/DALY-averted) together with hygiene education. Water supply network improvements were significantly less cost-effective. Copyright © 2013 Elsevier B.V. All rights reserved.

  2. Evaluation of Mindray BC-3600 hematology analyzer in a university hospital.

    PubMed

    Shu, G; Lu, H; Du, H; Shi, J; Wu, G

    2013-02-01

    The BC-3600 Auto Hematology Analyzer (hereinafter call BC-3600) is a quantitative, automated hematology analyzer and leukocyte differential counter for In Vitro Diagnostic Use in clinical laboratories. The analyzer was evaluated and compared with the Mindray BC-3200 3-part differential (BC-3200) and Sysmex XE-2100 5-part differential (XE-2100) Hematology Analyzer in the hematology laboratory of a university hospital. The BC-3600 was evaluated according to guidelines published by Clinical and Laboratory Standards Institute (CLSI), the International Committee for Standardization in Hematology (ICSH), and Department of Food and Drug Administration (FDA). There were no background, minimal carryover (<0.5%), and excellent linearity for white blood cell (WBC), hemoglobin (Hb) level, red blood cell (RBC), and platelet (PLT) counts (r > 0.999). Precision was good at all levels for the routine cell blood count (CBC) parameters: CV% being ≤2.0, except for platelet count (PLT) at the low level with CV% of ≤5.0% and WBC at the low level with CV% of <3.0%. Correlation between the BC-3600 and BC-3200, XE-2100 were excellent (r > 0.99) for all major CBC parameters. It is concluded that the overall performance of the BC-3600 is excellent and compares well with that of BC-3200 and XE-2100. © 2012 Blackwell Publishing Ltd.

  3. Quantitative research.

    PubMed

    Watson, Roger

    2015-04-01

    This article describes the basic tenets of quantitative research. The concepts of dependent and independent variables are addressed and the concept of measurement and its associated issues, such as error, reliability and validity, are explored. Experiments and surveys – the principal research designs in quantitative research – are described and key features explained. The importance of the double-blind randomised controlled trial is emphasised, alongside the importance of longitudinal surveys, as opposed to cross-sectional surveys. Essential features of data storage are covered, with an emphasis on safe, anonymous storage. Finally, the article explores the analysis of quantitative data, considering what may be analysed and the main uses of statistics in analysis.

  4. Electric Field Quantitative Measurement System and Method

    NASA Technical Reports Server (NTRS)

    Generazio, Edward R. (Inventor)

    2016-01-01

    A method and system are provided for making a quantitative measurement of an electric field. A plurality of antennas separated from one another by known distances are arrayed in a region that extends in at least one dimension. A voltage difference between at least one selected pair of antennas is measured. Each voltage difference is divided by the known distance associated with the selected pair of antennas corresponding thereto to generate a resulting quantity. The plurality of resulting quantities defined over the region quantitatively describe an electric field therein.

  5. Diagnostic performance of semi-quantitative and quantitative stress CMR perfusion analysis: a meta-analysis.

    PubMed

    van Dijk, R; van Assen, M; Vliegenthart, R; de Bock, G H; van der Harst, P; Oudkerk, M

    2017-11-27

    analysis our results show similar diagnostic accuracy comparing anatomical (AUC 0.86(0.83-0.89)) and functional reference standards (AUC 0.88(0.84-0.90)). Only the per territory analysis sensitivity did not show significant heterogeneity. None of the groups showed signs of publication bias. The clinical value of semi-quantitative and quantitative CMR perfusion analysis remains uncertain due to extensive inter-study heterogeneity and large differences in CMR perfusion acquisition protocols, reference standards, and methods of assessment of myocardial perfusion parameters. For wide spread implementation, standardization of CMR perfusion techniques is essential. CRD42016040176 .

  6. Total organic carbon analyzer

    NASA Technical Reports Server (NTRS)

    Godec, Richard G.; Kosenka, Paul P.; Smith, Brian D.; Hutte, Richard S.; Webb, Johanna V.; Sauer, Richard L.

    1991-01-01

    The development and testing of a breadboard version of a highly sensitive total-organic-carbon (TOC) analyzer are reported. Attention is given to the system components including the CO2 sensor, oxidation reactor, acidification module, and the sample-inlet system. Research is reported for an experimental reagentless oxidation reactor, and good results are reported for linearity, sensitivity, and selectivity in the CO2 sensor. The TOC analyzer is developed with gravity-independent components and is designed for minimal additions of chemical reagents. The reagentless oxidation reactor is based on electrolysis and UV photolysis and is shown to be potentially useful. The stability of the breadboard instrument is shown to be good on a day-to-day basis, and the analyzer is capable of 5 sample analyses per day for a period of about 80 days. The instrument can provide accurate TOC and TIC measurements over a concentration range of 20 ppb to 50 ppm C.

  7. Analyzing Tropical Waves Using the Parallel Ensemble Empirical Model Decomposition Method: Preliminary Results from Hurricane Sandy

    NASA Technical Reports Server (NTRS)

    Shen, Bo-Wen; Cheung, Samson; Li, Jui-Lin F.; Wu, Yu-ling

    2013-01-01

    In this study, we discuss the performance of the parallel ensemble empirical mode decomposition (EMD) in the analysis of tropical waves that are associated with tropical cyclone (TC) formation. To efficiently analyze high-resolution, global, multiple-dimensional data sets, we first implement multilevel parallelism into the ensemble EMD (EEMD) and obtain a parallel speedup of 720 using 200 eight-core processors. We then apply the parallel EEMD (PEEMD) to extract the intrinsic mode functions (IMFs) from preselected data sets that represent (1) idealized tropical waves and (2) large-scale environmental flows associated with Hurricane Sandy (2012). Results indicate that the PEEMD is efficient and effective in revealing the major wave characteristics of the data, such as wavelengths and periods, by sifting out the dominant (wave) components. This approach has a potential for hurricane climate study by examining the statistical relationship between tropical waves and TC formation.

  8. The Quantitative Preparation of Future Geoscience Graduate Students

    NASA Astrophysics Data System (ADS)

    Manduca, C. A.; Hancock, G. S.

    2006-12-01

    Modern geoscience is a highly quantitative science. In February, a small group of faculty and graduate students from across the country met to discuss the quantitative preparation of geoscience majors for graduate school. The group included ten faculty supervising graduate students in quantitative areas spanning the earth, atmosphere, and ocean sciences; five current graduate students in these areas; and five faculty teaching undergraduate students in the spectrum of institutions preparing students for graduate work. Discussion focused in four key ares: Are incoming graduate students adequately prepared for the quantitative aspects of graduate geoscience programs? What are the essential quantitative skills are that are required for success in graduate school? What are perceived as the important courses to prepare students for the quantitative aspects of graduate school? What programs/resources would be valuable in helping faculty/departments improve the quantitative preparation of students? The participants concluded that strengthening the quantitative preparation of undergraduate geoscience majors would increase their opportunities in graduate school. While specifics differed amongst disciplines, a special importance was placed on developing the ability to use quantitative skills to solve geoscience problems. This requires the ability to pose problems so they can be addressed quantitatively, understand the relationship between quantitative concepts and physical representations, visualize mathematics, test the reasonableness of quantitative results, creatively move forward from existing models/techniques/approaches, and move between quantitative and verbal descriptions. A list of important quantitative competencies desirable in incoming graduate students includes mechanical skills in basic mathematics, functions, multi-variate analysis, statistics and calculus, as well as skills in logical analysis and the ability to learn independently in quantitative ways

  9. Quantitative analysis of in vivo mucosal bacterial biofilms.

    PubMed

    Singhal, Deepti; Boase, Sam; Field, John; Jardeleza, Camille; Foreman, Andrew; Wormald, Peter-John

    2012-01-01

    Quantitative assays of mucosal biofilms on ex vivo samples are challenging using the currently applied specialized microscopic techniques to identify them. The COMSTAT2 computer program has been applied to in vitro biofilm models for quantifying biofilm structures seen on confocal scanning laser microscopy (CSLM). The aim of this study was to quantify Staphylococcus aureus (S. aureus) biofilms seen via CSLM on ex situ samples of sinonasal mucosa, using the COMSTAT2 program. S. aureus biofilms were grown in frontal sinuses of 4 merino sheep as per a previously standardized sheep sinusitis model for biofilms. Two sinonasal mucosal samples, 10 mm × 10 mm in size, from each of the 2 sinuses of the 4 sheep were analyzed for biofilm presence with Baclight stain and CSLM. Two random image stacks of mucosa with S. aureus biofilm were recorded from each sample, and analyzed using COMSTAT2 software that translates image stacks into a simplified 3-dimensional matrix of biofilm mass by eliminating surrounding host tissue. Three independent observers analyzed images using COMSTAT2 and 3 repeated rounds of analyses were done to calculate biofilm biomass. The COMSTAT2 application uses an observer-dependent threshold setting to translate CSLM biofilm images into a simplified 3-dimensional output for quantitative analysis. Intraclass correlation coefficient (ICC) between thresholds set by the 3 observers for each image stacks was 0.59 (p = 0.0003). Threshold values set at different points of time by a single observer also showed significant correlation as seen by ICC of 0.80 (p < 0.001). COMSTAT2 can be applied to quantify and study the complex 3-dimensional biofilm structures that are recorded via CSLM on mucosal tissue like the sinonasal mucosa. Copyright © 2011 American Rhinologic Society-American Academy of Otolaryngic Allergy, LLC.

  10. The efficacy of semi-quantitative urine protein-to-creatinine (P/C) ratio for the detection of significant proteinuria in urine specimens in health screening settings.

    PubMed

    Chang, Chih-Chun; Su, Ming-Jang; Ho, Jung-Li; Tsai, Yu-Hui; Tsai, Wei-Ting; Lee, Shu-Jene; Yen, Tzung-Hai; Chu, Fang-Yeh

    2016-01-01

    Urine protein detection could be underestimated using the conventional dipstick method because of variations in urine aliquots. This study aimed to assess the efficacy of the semi-quantitative urine protein-to-creatinine (P/C) ratio compared with other laboratory methods. Random urine samples were requested from patients undergoing chronic kidney disease screening. Significant proteinuria was determined by the quantitative P/C ratio of at least 150 mg protein/g creatinine. The semi-quantitative P/C ratio, dipstick protein and quantitative protein concentrations were compared and analyzed. In the 2932 urine aliquots, 156 (5.3 %) urine samples were considered as diluted and 60 (39.2 %) were found as significant proteinuria. The semi-quantitative P/C ratio testing had the best sensitivity (70.0 %) and specificity (95.9 %) as well as the lowest underestimation rate (0.37 %) when compared to other laboratory methods in the study. In the semi-quantitative P/C ratio test, 19 (12.2 %) had positive, 52 (33.3 %) had diluted, and 85 (54.5 %) had negative results. Of those with positive results, 7 (36.8 %) were positive detected by traditional dipstick urine protein test, and 9 (47.4 %) were positive detected by quantitative urine protein test. Additionally, of those with diluted results, 25 (48.1 %) had significant proteinuria, and all were assigned as no significant proteinuria by both tests. The semi-quantitative urine P/C ratio is clinically applicable based on its better sensitivity and screening ability for significant proteinuria than other laboratory methods, particularly in diluted urine samples. To establish an effective strategy for CKD prevention, urine protein screening with semi-quantitative P/C ratio could be considered.

  11. A Quantitative Gas Chromatographic Ethanol Determination.

    ERIC Educational Resources Information Center

    Leary, James J.

    1983-01-01

    Describes a gas chromatographic experiment for the quantitative determination of volume percent ethanol in water ethanol solutions. Background information, procedures, and typical results are included. Accuracy and precision of results are both on the order of two percent. (JN)

  12. Principles, performance, and applications of spectral reconstitution (SR) in quantitative analysis of oils by Fourier transform infrared spectroscopy (FT-IR).

    PubMed

    García-González, Diego L; Sedman, Jacqueline; van de Voort, Frederik R

    2013-04-01

    Spectral reconstitution (SR) is a dilution technique developed to facilitate the rapid, automated, and quantitative analysis of viscous oil samples by Fourier transform infrared spectroscopy (FT-IR). This technique involves determining the dilution factor through measurement of an absorption band of a suitable spectral marker added to the diluent, and then spectrally removing the diluent from the sample and multiplying the resulting spectrum to compensate for the effect of dilution on the band intensities. The facsimile spectrum of the neat oil thus obtained can then be qualitatively or quantitatively analyzed for the parameter(s) of interest. The quantitative performance of the SR technique was examined with two transition-metal carbonyl complexes as spectral markers, chromium hexacarbonyl and methylcyclopentadienyl manganese tricarbonyl. The estimation of the volume fraction (VF) of the diluent in a model system, consisting of canola oil diluted to various extents with odorless mineral spirits, served as the basis for assessment of these markers. The relationship between the VF estimates and the true volume fraction (VF(t)) was found to be strongly dependent on the dilution ratio and also depended, to a lesser extent, on the spectral resolution. These dependences are attributable to the effect of changes in matrix polarity on the bandwidth of the ν(CO) marker bands. Excellent VF(t) estimates were obtained by making a polarity correction devised with a variance-spectrum-delineated correction equation. In the absence of such a correction, SR was shown to introduce only a minor and constant bias, provided that polarity differences among all the diluted samples analyzed were minimal. This bias can be built into the calibration of a quantitative FT-IR analytical method by subjecting appropriate calibration standards to the same SR procedure as the samples to be analyzed. The primary purpose of the SR technique is to simplify preparation of diluted samples such that

  13. Spectral Feature Analysis for Quantitative Estimation of Cyanobacteria Chlorophyll-A

    NASA Astrophysics Data System (ADS)

    Lin, Yi; Ye, Zhanglin; Zhang, Yugan; Yu, Jie

    2016-06-01

    In recent years, lake eutrophication caused a large of Cyanobacteria bloom which not only brought serious ecological disaster but also restricted the sustainable development of regional economy in our country. Chlorophyll-a is a very important environmental factor to monitor water quality, especially for lake eutrophication. Remote sensed technique has been widely utilized in estimating the concentration of chlorophyll-a by different kind of vegetation indices and monitoring its distribution in lakes, rivers or along coastline. For each vegetation index, its quantitative estimation accuracy for different satellite data might change since there might be a discrepancy of spectral resolution and channel center between different satellites. The purpose this paper is to analyze the spectral feature of chlorophyll-a with hyperspectral data (totally 651 bands) and use the result to choose the optimal band combination for different satellites. The analysis method developed here in this study could be useful to recognize and monitor cyanobacteria bloom automatically and accrately. In our experiment, the reflectance (from 350nm to 1000nm) of wild cyanobacteria in different consistency (from 0 to 1362.11ug/L) and the corresponding chlorophyll-a concentration were measured simultaneously. Two kinds of hyperspectral vegetation indices were applied in this study: simple ratio (SR) and narrow band normalized difference vegetation index (NDVI), both of which consists of any two bands in the entire 651 narrow bands. Then multivariate statistical analysis was used to construct the linear, power and exponential models. After analyzing the correlation between chlorophyll-a and single band reflectance, SR, NDVI respetively, the optimal spectral index for quantitative estimation of cyanobacteria chlorophyll-a, as well corresponding central wavelength and band width were extracted. Results show that: Under the condition of water disturbance, SR and NDVI are both suitable for quantitative

  14. Quantitative detection of astaxanthin and cantaxanthin in Atlantic salmon by resonance Raman spectroscopy

    NASA Astrophysics Data System (ADS)

    Ermakov, Igor V.; Ermakova, Maia R.; Gellermann, Werner

    2006-02-01

    Two major carotenoids species found in salmonids muscle tissues are astaxanthin and cantaxanthin. They are taken up from fish food and are responsible for the attractive red-orange color of salmon filet. Since carotenoids are powerful antioxidants and biomarkers of nutrient consumption, they are thought to indicate fish health and resistance to diseases in fish farm environments. Therefore, a rapid, accurate, quantitative optical technique for measuring carotenoid content in salmon tissues is of economic interest. We demonstrate the possibility of using fast, selective, quantitative detection of astaxanthin and cantaxanthin in salmon muscle tissues, employing resonance Raman spectroscopy. Analyzing strong Raman signals originating from the carbon-carbon double bond stretch vibrations of the carotenoid molecules under blue laser excitation, we are able to characterize quantitatively the concentrations of carotenoids in salmon muscle tissue. To validate the technique, we compared Raman data with absorption measurements of carotenoid extracts in acetone. A close correspondence was observed in absorption spectra for tissue extract in acetone and a pure astaxanthin solution. Raman results show a linear dependence between Raman and absorption data. The proposed technique holds promise as a method of rapid screening of carotenoid levels in fish muscle tissues and may be attractive for the fish farm industry to assess the dietary status of salmon, risk for infective diseases, and product quality control.

  15. Quantitative Graphics in Newspapers.

    ERIC Educational Resources Information Center

    Tankard, James W., Jr.

    The use of quantitative graphics in newspapers requires achieving a balance between being accurate and getting the attention of the reader. The statistical representations in newspapers are drawn by graphic designers whose key technique is fusion--the striking combination of two visual images. This technique often results in visual puns,…

  16. Identifying, Analyzing, and Communicating Rural: A Quantitative Perspective

    ERIC Educational Resources Information Center

    Koziol, Natalie A.; Arthur, Ann M.; Hawley, Leslie R.; Bovaird, James A.; Bash, Kirstie L.; McCormick, Carina; Welch, Greg W.

    2015-01-01

    Defining rural is a critical task for rural education researchers, as it has implications for all phases of a study. However, it is also a difficult task due to the many ways in which rural can be theoretically, conceptually, and empirically operationalized. This article provides researchers with specific guidance on important theoretical and…

  17. The quantitative control and matching of an optical false color composite imaging system

    NASA Astrophysics Data System (ADS)

    Zhou, Chengxian; Dai, Zixin; Pan, Xizhe; Li, Yinxi

    1993-10-01

    Design of an imaging system for optical false color composite (OFCC) capable of high-precision density-exposure time control and color balance is presented. The system provides high quality FCC image data that can be analyzed using a quantitative calculation method. The quality requirement to each part of the image generation system is defined, and the distribution of satellite remote sensing image information is analyzed. The proposed technology makes it possible to present the remote sensing image data more effectively and accurately.

  18. Quantitative and Qualitative Analyzes of the Explosive Cyclones that Reached the Antarctic Coast in the First Half of 2017

    NASA Astrophysics Data System (ADS)

    Pires, L. B. M.; Romao, M.; Freitas, A. C. V.

    2017-12-01

    An explosive cyclone is a kind of extratropical cyclone which shows a drop in pressure of at least 24 hPa in 24 hours. These are usually intense and they have rapid displacement which hinders their predictability. It is likely that climate change is causing an increase in this type of event in the Antarctic coast and, if this increase is confirmed, the regime of winds and temperatures may be changing. If there are more incidences of explosive cyclones, probably the Antarctic winds are becoming more intense and the temperatures in some places are becoming lower and in others are becoming higher. In the northern portion of the Antarctic Peninsula a decrease in temperature already has been recorded over the last 15 years, while a higher incidence of explosive cyclones over the region also has been found during this period. Studies also have suggested that the drop in temperatures in the Antarctic may be associated with the changes in wind direction, but the cause of these wind direction changes is unknown. Explosive cyclones, which change the wind patterns when they reach certain areas therefore may be contributing to this change in the Antarctic climate. This study is part of the "Explosive Cyclones on the Antarctic Coast" (EXCANC) Project conducted by the World Environmental Conservancy organization. This project analyzes data from meteorological stations strategically scattered throughout the coast and operated by various international Antarctic programs, and also utilizes satellite images. Results show that during the first half of 2017 the highest number of events were recorded at the Australian Casey station with 10 cases, followed by the French station of Dumont D'Urville with 7 cases. Halley's English station recorded its first explosive cyclone this year. Intensity analyzes also are shown.

  19. Validation of PCR methods for quantitation of genetically modified plants in food.

    PubMed

    Hübner, P; Waiblinger, H U; Pietsch, K; Brodmann, P

    2001-01-01

    For enforcement of the recently introduced labeling threshold for genetically modified organisms (GMOs) in food ingredients, quantitative detection methods such as quantitative competitive (QC-PCR) and real-time PCR are applied by official food control laboratories. The experiences of 3 European food control laboratories in validating such methods were compared to describe realistic performance characteristics of quantitative PCR detection methods. The limit of quantitation (LOQ) of GMO-specific, real-time PCR was experimentally determined to reach 30-50 target molecules, which is close to theoretical prediction. Starting PCR with 200 ng genomic plant DNA, the LOQ depends primarily on the genome size of the target plant and ranges from 0.02% for rice to 0.7% for wheat. The precision of quantitative PCR detection methods, expressed as relative standard deviation (RSD), varied from 10 to 30%. Using Bt176 corn containing test samples and applying Bt176 specific QC-PCR, mean values deviated from true values by -7to 18%, with an average of 2+/-10%. Ruggedness of real-time PCR detection methods was assessed in an interlaboratory study analyzing commercial, homogeneous food samples. Roundup Ready soybean DNA contents were determined in the range of 0.3 to 36%, relative to soybean DNA, with RSDs of about 25%. Taking the precision of quantitative PCR detection methods into account, suitable sample plans and sample sizes for GMO analysis are suggested. Because quantitative GMO detection methods measure GMO contents of samples in relation to reference material (calibrants), high priority must be given to international agreements and standardization on certified reference materials.

  20. Discrepancies between qualitative and quantitative evaluation of randomised controlled trial results: achieving clarity through mixed methods triangulation.

    PubMed

    Tonkin-Crine, Sarah; Anthierens, Sibyl; Hood, Kerenza; Yardley, Lucy; Cals, Jochen W L; Francis, Nick A; Coenen, Samuel; van der Velden, Alike W; Godycki-Cwirko, Maciek; Llor, Carl; Butler, Chris C; Verheij, Theo J M; Goossens, Herman; Little, Paul

    2016-05-12

    Mixed methods are commonly used in health services research; however, data are not often integrated to explore complementarity of findings. A triangulation protocol is one approach to integrating such data. A retrospective triangulation protocol was carried out on mixed methods data collected as part of a process evaluation of a trial. The multi-country randomised controlled trial found that a web-based training in communication skills (including use of a patient booklet) and the use of a C-reactive protein (CRP) point-of-care test decreased antibiotic prescribing by general practitioners (GPs) for acute cough. The process evaluation investigated GPs' and patients' experiences of taking part in the trial. Three analysts independently compared findings across four data sets: qualitative data collected view semi-structured interviews with (1) 62 patients and (2) 66 GPs and quantitative data collected via questionnaires with (3) 2886 patients and (4) 346 GPs. Pairwise comparisons were made between data sets and were categorised as agreement, partial agreement, dissonance or silence. Three instances of dissonance occurred in 39 independent findings. GPs and patients reported different views on the use of a CRP test. GPs felt that the test was useful in convincing patients to accept a no-antibiotic decision, but patient data suggested that this was unnecessary if a full explanation was given. Whilst qualitative data indicated all patients were generally satisfied with their consultation, quantitative data indicated highest levels of satisfaction for those receiving a detailed explanation from their GP with a booklet giving advice on self-care. Both qualitative and quantitative data sets indicated higher patient enablement for those in the communication groups who had received a booklet. Use of CRP tests does not appear to engage patients or influence illness perceptions and its effect is more centred on changing clinician behaviour. Communication skills and the patient

  1. 3D quantitative comparative analysis of long bone diaphysis variations in microanatomy and cross-sectional geometry.

    PubMed

    Houssaye, Alexandra; Taverne, Maxime; Cornette, Raphaël

    2018-05-01

    Long bone inner structure and cross-sectional geometry display a strong functional signal, leading to convergences, and are widely analyzed in comparative anatomy at small and large taxonomic scales. Long bone microanatomical studies have essentially been conducted on transverse sections but also on a few longitudinal ones. Recent studies highlighted the interest in analyzing variations of the inner structure along the diaphysis using a qualitative as well as a quantitative approach. With the development of microtomography, it has become possible to study three-dimensional (3D) bone microanatomy and, in more detail, the form-function relationships of these features. This study focused on the selection of quantitative parameters to describe in detail the cross-sectional shape changes and distribution of the osseous tissue along the diaphysis. Two-dimensional (2D) virtual transverse sections were also performed in the two usual reference planes and results were compared with those obtained based on the whole diaphysis analysis. The sample consisted in 14 humeri and 14 femora of various mammalian taxa that are essentially terrestrial. Comparative quantitative analyses between different datasets made it possible to highlight the parameters that are strongly impacted by size and phylogeny and the redundant ones, and thus to estimate their relevance for use in form-function analyses. The analysis illustrated that results based on 2D transverse sections are similar for both sectional planes; thus if a strong bias exists when mixing sections from the two reference planes in the same analysis, it would not problematic to use either one plane or the other in comparative studies. However, this may no longer hold for taxa showing a much stronger variation in bone microstructure along the diaphysis. Finally, the analysis demonstrated the significant contribution of the parameters describing variations along the diaphysis, and thus the interest in performing 3D analyses; this

  2. GC-FID coupled with chemometrics for quantitative and chemical fingerprinting analysis of Alpinia oxyphylla oil.

    PubMed

    Miao, Qing; Kong, Weijun; Zhao, Xiangsheng; Yang, Shihai; Yang, Meihua

    2015-01-01

    Analytical methods for quantitative analysis and chemical fingerprinting of volatile oils from Alpinia oxyphylla were established. The volatile oils were prepared by hydrodistillation, and the yields were between 0.82% and 1.33%. The developed gas chromatography-flame ionization detection (GC-FID) method showed good specificity, linearity, reproducibility, stability and recovery, and could be used satisfactorily for quantitative analysis. The results showed that the volatile oils contained 2.31-77.30 μL/mL p-cymene and 12.38-99.34 mg/mL nootkatone. A GC-FID fingerprinting method was established, and the profiles were analyzed using chemometrics. GC-MS was used to identify the principal compounds in the GC-FID profiles. The profiles of almost all the samples were consistent and stable. The harvesting time and source were major factors that affected the profile, while the volatile oil yield and the nootkatone content had minor secondary effects. Copyright © 2014 Elsevier B.V. All rights reserved.

  3. Quantitative method of measuring cancer cell urokinase and metastatic potential

    NASA Technical Reports Server (NTRS)

    Morrison, Dennis R. (Inventor)

    1993-01-01

    The metastatic potential of tumors can be evaluated by the quantitative detection of urokinase and DNA. The cell sample selected for examination is analyzed for the presence of high levels of urokinase and abnormal DNA using analytical flow cytometry and digital image analysis. Other factors such as membrane associated urokinase, increased DNA synthesis rates and certain receptors can be used in the method for detection of potentially invasive tumors.

  4. Quantitative Insights into the Fast Pyrolysis of Extracted Cellulose, Hemicelluloses, and Lignin

    PubMed Central

    Windt, Michael; Ziegler, Bernhard; Appelt, Jörn; Saake, Bodo; Meier, Dietrich; Bridgwater, Anthony

    2017-01-01

    Abstract The transformation of lignocellulosic biomass into bio‐based commodity chemicals is technically possible. Among thermochemical processes, fast pyrolysis, a relatively mature technology that has now reached a commercial level, produces a high yield of an organic‐rich liquid stream. Despite recent efforts to elucidate the degradation paths of biomass during pyrolysis, the selectivity and recovery rates of bio‐compounds remain low. In an attempt to clarify the general degradation scheme of biomass fast pyrolysis and provide a quantitative insight, the use of fast pyrolysis microreactors is combined with spectroscopic techniques (i.e., mass spectrometry and NMR spectroscopy) and mixtures of unlabeled and 13C‐enriched materials. The first stage of the work aimed to select the type of reactor to use to ensure control of the pyrolysis regime. A comparison of the chemical fragmentation patterns of “primary” fast pyrolysis volatiles detected by using GC‐MS between two small‐scale microreactors showed the inevitable occurrence of secondary reactions. In the second stage, liquid fractions that are also made of primary fast pyrolysis condensates were analyzed by using quantitative liquid‐state 13C NMR spectroscopy to provide a quantitative distribution of functional groups. The compilation of these results into a map that displays the distribution of functional groups according to the individual and main constituents of biomass (i.e., hemicelluloses, cellulose and lignin) confirmed the origin of individual chemicals within the fast pyrolysis liquids. PMID:28644517

  5. Quantitative Insights into the Fast Pyrolysis of Extracted Cellulose, Hemicelluloses, and Lignin.

    PubMed

    Carrier, Marion; Windt, Michael; Ziegler, Bernhard; Appelt, Jörn; Saake, Bodo; Meier, Dietrich; Bridgwater, Anthony

    2017-08-24

    The transformation of lignocellulosic biomass into bio-based commodity chemicals is technically possible. Among thermochemical processes, fast pyrolysis, a relatively mature technology that has now reached a commercial level, produces a high yield of an organic-rich liquid stream. Despite recent efforts to elucidate the degradation paths of biomass during pyrolysis, the selectivity and recovery rates of bio-compounds remain low. In an attempt to clarify the general degradation scheme of biomass fast pyrolysis and provide a quantitative insight, the use of fast pyrolysis microreactors is combined with spectroscopic techniques (i.e., mass spectrometry and NMR spectroscopy) and mixtures of unlabeled and 13 C-enriched materials. The first stage of the work aimed to select the type of reactor to use to ensure control of the pyrolysis regime. A comparison of the chemical fragmentation patterns of "primary" fast pyrolysis volatiles detected by using GC-MS between two small-scale microreactors showed the inevitable occurrence of secondary reactions. In the second stage, liquid fractions that are also made of primary fast pyrolysis condensates were analyzed by using quantitative liquid-state 13 C NMR spectroscopy to provide a quantitative distribution of functional groups. The compilation of these results into a map that displays the distribution of functional groups according to the individual and main constituents of biomass (i.e., hemicelluloses, cellulose and lignin) confirmed the origin of individual chemicals within the fast pyrolysis liquids. © 2017 The Authors. Published by Wiley-VCH Verlag GmbH & Co. KGaA.

  6. QUANTITATIVE ELISA OF POLYCHLORINATED BIPHENYLS IN AN OILY SOIL MATRIX USING SUPERCRITICAL FLUID EXTRACTION

    EPA Science Inventory

    Soil samples from the GenCorp Lawrence Brownfields site were analyzed with a commercial semi-quantitative enzyme-linked immunosorbent assay (ELISA) using a methanol shake extraction. Many of the soil samples were extremely oily, with total petroleum hydrocarbon levels up to 240...

  7. Primary enzyme quantitation

    DOEpatents

    Saunders, G.C.

    1982-03-04

    The disclosure relates to the quantitation of a primary enzyme concentration by utilizing a substrate for the primary enzyme labeled with a second enzyme which is an indicator enzyme. Enzyme catalysis of the substrate occurs and results in release of the indicator enzyme in an amount directly proportional to the amount of primary enzyme present. By quantifying the free indicator enzyme one determines the amount of primary enzyme present.

  8. A WebGIS-based system for analyzing and visualizing air quality data for Shanghai Municipality

    NASA Astrophysics Data System (ADS)

    Wang, Manyi; Liu, Chaoshun; Gao, Wei

    2014-10-01

    An online visual analytical system based on Java Web and WebGIS for air quality data for Shanghai Municipality was designed and implemented to quantitatively analyze and qualitatively visualize air quality data. By analyzing the architecture of WebGIS and Java Web, we firstly designed the overall scheme for system architecture, then put forward the software and hardware environment and also determined the main function modules for the system. The visual system was ultimately established with the DIV + CSS layout method combined with JSP, JavaScript, and some other computer programming languages based on the Java programming environment. Moreover, Struts, Spring, and Hibernate frameworks (SSH) were integrated in the system for the purpose of easy maintenance and expansion. To provide mapping service and spatial analysis functions, we selected ArcGIS for Server as the GIS server. We also used Oracle database and ESRI file geodatabase to store spatial data and non-spatial data in order to ensure the data security. In addition, the response data from the Web server are resampled to implement rapid visualization through the browser. The experimental successes indicate that this system can quickly respond to user's requests, and efficiently return the accurate processing results.

  9. From themes to hypotheses: following up with quantitative methods.

    PubMed

    Morgan, David L

    2015-06-01

    One important category of mixed-methods research designs consists of quantitative studies that follow up on qualitative research. In this case, the themes that serve as the results from the qualitative methods generate hypotheses for testing through the quantitative methods. That process requires operationalization to translate the concepts from the qualitative themes into quantitative variables. This article illustrates these procedures with examples that range from simple operationalization to the evaluation of complex models. It concludes with an argument for not only following up qualitative work with quantitative studies but also the reverse, and doing so by going beyond integrating methods within single projects to include broader mutual attention from qualitative and quantitative researchers who work in the same field. © The Author(s) 2015.

  10. A quantitative method to analyze the quality of EIA information in wind energy development and avian/bat assessments

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Chang, Tony, E-mail: tc282@nau.edu; Nielsen, Erik, E-mail: erik.nielsen@nau.edu; Auberle, William, E-mail: william.auberle@nau.edu

    2013-01-15

    The environmental impact assessment (EIA) has been a tool for decision makers since the enactment of the National Environmental Policy Act (NEPA). Since that time, few analyses have been performed to verify the quality of information and content within EIAs. High quality information within assessments is vital in order for decision makers, stake holders, and the public to understand the potential impact of proposed actions on the ecosystem and wildlife species. Low quality information has been a major cause for litigation and economic loss. Since 1999, wind energy development has seen an exponential growth with unknown levels of impact onmore » wildlife species, in particular bird and bat species. The purpose of this article is to: (1) develop, validate, and apply a quantitative index to review avian/bat assessment quality for wind energy EIAs; and (2) assess the trends and status of avian/bat assessment quality in a sample of wind energy EIAs. This research presents the development and testing of the Avian and Bat Assessment Quality Index (ABAQI), a new approach to quantify information quality of ecological assessments within wind energy development EIAs in relation to avian and bat species based on review areas and factors derived from 23 state wind/wildlife siting guidance documents. The ABAQI was tested through a review of 49 publicly available EIA documents and validated by identifying high variation in avian and bat assessments quality for wind energy developments. Of all the reviewed EIAs, 66% failed to provide high levels of preconstruction avian and bat survey information, compared to recommended factors from state guidelines. This suggests the need for greater consistency from recommended guidelines by state, and mandatory compliance by EIA preparers to avoid possible habitat and species loss, wind energy development shut down, and future lawsuits. - Highlights: Black-Right-Pointing-Pointer We developed, validated, and applied a quantitative index to

  11. Second Generation International Space Station (ISS) Total Organic Carbon Analyzer (TOCA) Verification Testing and On-Orbit Performance Results

    NASA Technical Reports Server (NTRS)

    Bentley, Nicole L.; Thomas, Evan A.; VanWie, Michael; Morrison, Chad; Stinson, Richard G.

    2010-01-01

    The Total Organic Carbon Analyzer (TOGA) is designed to autonomously determine recovered water quality as a function of TOC. The current TOGA has been on the International Space Station since November 2008. Functional checkout and operations revealed complex operating considerations. Specifically, failure of the hydrogen catalyst resulted in the development of an innovative oxidation analysis method. This method reduces the activation time and limits the hydrogen produced during analysis, while retaining the ability to indicate TOC concentrations within 25% accuracy. Subsequent testing and comparison to archived samples returned from the Station and tested on the ground yield high confidence in this method, and in the quality of the recovered water.

  12. Building quantitative, three-dimensional atlases of gene expression and morphology at cellular resolution.

    PubMed

    Knowles, David W; Biggin, Mark D

    2013-01-01

    Animals comprise dynamic three-dimensional arrays of cells that express gene products in intricate spatial and temporal patterns that determine cellular differentiation and morphogenesis. A rigorous understanding of these developmental processes requires automated methods that quantitatively record and analyze complex morphologies and their associated patterns of gene expression at cellular resolution. Here we summarize light microscopy-based approaches to establish permanent, quantitative datasets-atlases-that record this information. We focus on experiments that capture data for whole embryos or large areas of tissue in three dimensions, often at multiple time points. We compare and contrast the advantages and limitations of different methods and highlight some of the discoveries made. We emphasize the need for interdisciplinary collaborations and integrated experimental pipelines that link sample preparation, image acquisition, image analysis, database design, visualization, and quantitative analysis. Copyright © 2013 Wiley Periodicals, Inc.

  13. Quantitative breast tissue characterization using grating-based x-ray phase-contrast imaging

    NASA Astrophysics Data System (ADS)

    Willner, M.; Herzen, J.; Grandl, S.; Auweter, S.; Mayr, D.; Hipp, A.; Chabior, M.; Sarapata, A.; Achterhold, K.; Zanette, I.; Weitkamp, T.; Sztrókay, A.; Hellerhoff, K.; Reiser, M.; Pfeiffer, F.

    2014-04-01

    X-ray phase-contrast imaging has received growing interest in recent years due to its high capability in visualizing soft tissue. Breast imaging became the focus of particular attention as it is considered the most promising candidate for a first clinical application of this contrast modality. In this study, we investigate quantitative breast tissue characterization using grating-based phase-contrast computed tomography (CT) at conventional polychromatic x-ray sources. Different breast specimens have been scanned at a laboratory phase-contrast imaging setup and were correlated to histopathology. Ascertained tumor types include phylloides tumor, fibroadenoma and infiltrating lobular carcinoma. Identified tissue types comprising adipose, fibroglandular and tumor tissue have been analyzed in terms of phase-contrast Hounsfield units and are compared to high-quality, high-resolution data obtained with monochromatic synchrotron radiation, as well as calculated values based on tabulated tissue properties. The results give a good impression of the method’s prospects and limitations for potential tumor detection and the associated demands on such a phase-contrast breast CT system. Furthermore, the evaluated quantitative tissue values serve as a reference for simulations and the design of dedicated phantoms for phase-contrast mammography.

  14. Diagnostic accuracy of semi-quantitative and quantitative culture techniques for the diagnosis of catheter-related infections in newborns and molecular typing of isolated microorganisms.

    PubMed

    Riboli, Danilo Flávio Moraes; Lyra, João César; Silva, Eliane Pessoa; Valadão, Luisa Leite; Bentlin, Maria Regina; Corrente, José Eduardo; Rugolo, Ligia Maria Suppo de Souza; da Cunha, Maria de Lourdes Ribeiro de Souza

    2014-05-22

    Catheter-related bloodstream infections (CR-BSIs) have become the most common cause of healthcare-associated bloodstream infections in neonatal intensive care units (ICUs). Microbiological evidence implicating catheters as the source of bloodstream infection is necessary to establish the diagnosis of CR-BSIs. Semi-quantitative culture is used to determine the presence of microorganisms on the external catheter surface, whereas quantitative culture also isolates microorganisms present inside the catheter. The main objective of this study was to determine the sensitivity and specificity of these two techniques for the diagnosis of CR-BSIs in newborns from a neonatal ICU. In addition, PFGE was used for similarity analysis of the microorganisms isolated from catheters and blood cultures. Semi-quantitative and quantitative methods were used for the culture of catheter tips obtained from newborns. Strains isolated from catheter tips and blood cultures which exhibited the same antimicrobial susceptibility profile were included in the study as positive cases of CR-BSI. PFGE of the microorganisms isolated from catheters and blood cultures was performed for similarity analysis and detection of clones in the ICU. A total of 584 catheter tips from 399 patients seen between November 2005 and June 2012 were analyzed. Twenty-nine cases of CR-BSI were confirmed. Coagulase-negative staphylococci (CoNS) were the most frequently isolated microorganisms, including S. epidermidis as the most prevalent species (65.5%), followed by S. haemolyticus (10.3%), yeasts (10.3%), K. pneumoniae (6.9%), S. aureus (3.4%), and E. coli (3.4%). The sensitivity of the semi-quantitative and quantitative techniques was 72.7% and 59.3%, respectively, and specificity was 95.7% and 94.4%. The diagnosis of CR-BSIs based on PFGE analysis of similarity between strains isolated from catheter tips and blood cultures showed 82.6% sensitivity and 100% specificity. The semi-quantitative culture method showed higher

  15. Nontargeted quantitation of lipid classes using hydrophilic interaction liquid chromatography-electrospray ionization mass spectrometry with single internal standard and response factor approach.

    PubMed

    Cífková, Eva; Holčapek, Michal; Lísa, Miroslav; Ovčačíková, Magdaléna; Lyčka, Antonín; Lynen, Frédéric; Sandra, Pat

    2012-11-20

    The identification and quantitation of a wide range of lipids in complex biological samples is an essential requirement for the lipidomic studies. High-performance liquid chromatography-mass spectrometry (HPLC/MS) has the highest potential to obtain detailed information on the whole lipidome, but the reliable quantitation of multiple lipid classes is still a challenging task. In this work, we describe a new method for the nontargeted quantitation of polar lipid classes separated by hydrophilic interaction liquid chromatography (HILIC) followed by positive-ion electrospray ionization mass spectrometry (ESI-MS) using a single internal lipid standard to which all class specific response factors (RFs) are related to. The developed method enables the nontargeted quantitation of lipid classes and molecules inside these classes in contrast to the conventional targeted quantitation, which is based on predefined selected reaction monitoring (SRM) transitions for selected lipids only. In the nontargeted quantitation method described here, concentrations of lipid classes are obtained by the peak integration in HILIC chromatograms multiplied by their RFs related to the single internal standard (i.e., sphingosyl PE, d17:1/12:0) used as common reference for all polar lipid classes. The accuracy, reproducibility and robustness of the method have been checked by various means: (1) the comparison with conventional lipidomic quantitation using SRM scans on a triple quadrupole (QqQ) mass analyzer, (2) (31)P nuclear magnetic resonance (NMR) quantitation of the total lipid extract, (3) method robustness test using subsequent measurements by three different persons, (4) method transfer to different HPLC/MS systems using different chromatographic conditions, and (5) comparison with previously published results for identical samples, especially human reference plasma from the National Institute of Standards and Technology (NIST human plasma). Results on human plasma, egg yolk and porcine

  16. iMet-Q: A User-Friendly Tool for Label-Free Metabolomics Quantitation Using Dynamic Peak-Width Determination

    PubMed Central

    Chang, Hui-Yin; Chen, Ching-Tai; Lih, T. Mamie; Lynn, Ke-Shiuan; Juo, Chiun-Gung; Hsu, Wen-Lian; Sung, Ting-Yi

    2016-01-01

    Efficient and accurate quantitation of metabolites from LC-MS data has become an important topic. Here we present an automated tool, called iMet-Q (intelligent Metabolomic Quantitation), for label-free metabolomics quantitation from high-throughput MS1 data. By performing peak detection and peak alignment, iMet-Q provides a summary of quantitation results and reports ion abundance at both replicate level and sample level. Furthermore, it gives the charge states and isotope ratios of detected metabolite peaks to facilitate metabolite identification. An in-house standard mixture and a public Arabidopsis metabolome data set were analyzed by iMet-Q. Three public quantitation tools, including XCMS, MetAlign, and MZmine 2, were used for performance comparison. From the mixture data set, seven standard metabolites were detected by the four quantitation tools, for which iMet-Q had a smaller quantitation error of 12% in both profile and centroid data sets. Our tool also correctly determined the charge states of seven standard metabolites. By searching the mass values for those standard metabolites against Human Metabolome Database, we obtained a total of 183 metabolite candidates. With the isotope ratios calculated by iMet-Q, 49% (89 out of 183) metabolite candidates were filtered out. From the public Arabidopsis data set reported with two internal standards and 167 elucidated metabolites, iMet-Q detected all of the peaks corresponding to the internal standards and 167 metabolites. Meanwhile, our tool had small abundance variation (≤0.19) when quantifying the two internal standards and had higher abundance correlation (≥0.92) when quantifying the 167 metabolites. iMet-Q provides user-friendly interfaces and is publicly available for download at http://ms.iis.sinica.edu.tw/comics/Software_iMet-Q.html. PMID:26784691

  17. High Throughput Protein Quantitation using MRM Viewer Software and Dynamic MRM on a Triple Quadruple Mass Spectrometer

    PubMed Central

    Miller, C.; Waddell, K.; Tang, N.

    2010-01-01

    RP-122 Peptide quantitation using Multiple Reaction Monitoring (MRM) has been established as an important methodology for biomarker verification andvalidation.This requires high throughput combined with high sensitivity to analyze potentially thousands of target peptides in each sample.Dynamic MRM allows the system to only acquire the required MRMs of the peptide during a retention window corresponding to when each peptide is eluting. This reduces the number of concurrent MRM and therefore improves quantitation and sensitivity. MRM Selector allows the user to generate an MRM transition list with retention time information from discovery data obtained on a QTOF MS system.This list can be directly imported into the triple quadrupole acquisition software.However, situations can exist where a) the list of MRMs contain an excess of MRM transitions allowable under the ideal acquisition conditions chosen ( allowing for cycle time and chromatography conditions), or b) too many transitions in a certain retention time region which would result in an unacceptably low dwell time and cycle time.A new tool - MRM viewer has been developed to help users automatically generate multiple dynamic MRM methods from a single MRM list.In this study, a list of 3293 MRM transitions from a human plasma sample was compiled.A single dynamic MRM method with 3293 transitions results in a minimum dwell time of 2.18ms.Using MRM viewer we can generate three dynamic MRM methods with a minimum dwell time of 20ms which can give a better quality MRM quantitation.This tool facilitates both high throughput and high sensitivity for MRM quantitation.

  18. Quantitative Oxygenation Venography from MRI Phase

    PubMed Central

    Fan, Audrey P.; Bilgic, Berkin; Gagnon, Louis; Witzel, Thomas; Bhat, Himanshu; Rosen, Bruce R.; Adalsteinsson, Elfar

    2014-01-01

    Purpose To demonstrate acquisition and processing methods for quantitative oxygenation venograms that map in vivo oxygen saturation (SvO2) along cerebral venous vasculature. Methods Regularized quantitative susceptibility mapping (QSM) is used to reconstruct susceptibility values and estimate SvO2 in veins. QSM with ℓ1 and ℓ2 regularization are compared in numerical simulations of vessel structures with known magnetic susceptibility. Dual-echo, flow-compensated phase images are collected in three healthy volunteers to create QSM images. Bright veins in the susceptibility maps are vectorized and used to form a three-dimensional vascular mesh, or venogram, along which to display SvO2 values from QSM. Results Quantitative oxygenation venograms that map SvO2 along brain vessels of arbitrary orientation and geometry are shown in vivo. SvO2 values in major cerebral veins lie within the normal physiological range reported by 15O positron emission tomography. SvO2 from QSM is consistent with previous MR susceptometry methods for vessel segments oriented parallel to the main magnetic field. In vessel simulations, ℓ1 regularization results in less than 10% SvO2 absolute error across all vessel tilt orientations and provides more accurate SvO2 estimation than ℓ2 regularization. Conclusion The proposed analysis of susceptibility images enables reliable mapping of quantitative SvO2 along venograms and may facilitate clinical use of venous oxygenation imaging. PMID:24006229

  19. A Critical Appraisal of Techniques, Software Packages, and Standards for Quantitative Proteomic Analysis

    PubMed Central

    Lawless, Craig; Hubbard, Simon J.; Fan, Jun; Bessant, Conrad; Hermjakob, Henning; Jones, Andrew R.

    2012-01-01

    Abstract New methods for performing quantitative proteome analyses based on differential labeling protocols or label-free techniques are reported in the literature on an almost monthly basis. In parallel, a correspondingly vast number of software tools for the analysis of quantitative proteomics data has also been described in the literature and produced by private companies. In this article we focus on the review of some of the most popular techniques in the field and present a critical appraisal of several software packages available to process and analyze the data produced. We also describe the importance of community standards to support the wide range of software, which may assist researchers in the analysis of data using different platforms and protocols. It is intended that this review will serve bench scientists both as a useful reference and a guide to the selection and use of different pipelines to perform quantitative proteomics data analysis. We have produced a web-based tool (http://www.proteosuite.org/?q=other_resources) to help researchers find appropriate software for their local instrumentation, available file formats, and quantitative methodology. PMID:22804616

  20. Quantitative ROESY analysis of computational models: structural studies of citalopram and β-cyclodextrin complexes by (1) H-NMR and computational methods.

    PubMed

    Ali, Syed Mashhood; Shamim, Shazia

    2015-07-01

    Complexation of racemic citalopram with β-cyclodextrin (β-CD) in aqueous medium was investigated to determine atom-accurate structure of the inclusion complexes. (1) H-NMR chemical shift change data of β-CD cavity protons in the presence of citalopram confirmed the formation of 1 : 1 inclusion complexes. ROESY spectrum confirmed the presence of aromatic ring in the β-CD cavity but whether one of the two or both rings was not clear. Molecular mechanics and molecular dynamic calculations showed the entry of fluoro-ring from wider side of β-CD cavity as the most favored mode of inclusion. Minimum energy computational models were analyzed for their accuracy in atomic coordinates by comparison of calculated and experimental intermolecular ROESY peak intensities, which were not found in agreement. Several least energy computational models were refined and analyzed till calculated and experimental intensities were compatible. The results demonstrate that computational models of CD complexes need to be analyzed for atom-accuracy and quantitative ROESY analysis is a promising method. Moreover, the study also validates that the quantitative use of ROESY is feasible even with longer mixing times if peak intensity ratios instead of absolute intensities are used. Copyright © 2015 John Wiley & Sons, Ltd.

  1. Quantitative Percussion Diagnostics For Evaluating Bond Integrity Between Composite Laminates

    NASA Astrophysics Data System (ADS)

    Poveromo, Scott Leonard

    Conventional nondestructive testing (NDT) techniques used to detect defects in composites are not able to determine intact bond integrity within a composite structure and are costly to use on large and complex shaped surfaces. To overcome current NDT limitations, a new technology was utilized based on quantitative percussion diagnostics (QPD) to better quantify bond quality in fiber reinforced composite materials. Experimental results indicate that this technology is capable of detecting 'kiss' bonds (very low adhesive shear strength), caused by the application of release agents on the bonding surfaces, between flat composite laminates bonded together with epoxy adhesive. Specifically, the local value of the loss coefficient determined from quantitative percussion testing was found to be significantly greater for a release coated panel compared to that for a well bonded sample. Also, the local value of the probe force or force returned to the probe after impact was observed to be lower for the release coated panels. The increase in loss coefficient and decrease in probe force are thought to be due to greater internal friction during the percussion event for poorly bonded specimens. NDT standards were also fabricated by varying the cure parameters of an epoxy film adhesive. Results from QPD for the variable cure NDT standards and lap shear strength measurements taken of mechanical test specimens were compared and analyzed. Finally, experimental results have been compared to a finite element analysis to understand the visco-elastic behavior of the laminates during percussion testing. This comparison shows how a lower quality bond leads to a reduction in the percussion force by biasing strain in the percussion tested side of the panel.

  2. Tau-U: A Quantitative Approach for Analysis of Single-Case Experimental Data in Aphasia.

    PubMed

    Lee, Jaime B; Cherney, Leora R

    2018-03-01

    Tau-U is a quantitative approach for analyzing single-case experimental design (SCED) data. It combines nonoverlap between phases with intervention phase trend and can correct for a baseline trend (Parker, Vannest, & Davis, 2011). We demonstrate the utility of Tau-U by comparing it with the standardized mean difference approach (Busk & Serlin, 1992) that is widely reported within the aphasia SCED literature. Repeated writing measures from 3 participants with chronic aphasia who received computer-based writing treatment are analyzed visually and quantitatively using both Tau-U and the standardized mean difference approach. Visual analysis alone was insufficient for determining an effect between the intervention and writing improvement. The standardized mean difference yielded effect sizes ranging from 4.18 to 26.72 for trained items and 1.25 to 3.20 for untrained items. Tau-U yielded significant (p < .05) effect sizes for 2 of 3 participants for trained probes and 1 of 3 participants for untrained probes. A baseline trend correction was applied to data from 2 of 3 participants. Tau-U has the unique advantage of allowing for the correction of an undesirable baseline trend. Although further study is needed, Tau-U shows promise as a quantitative approach to augment visual analysis of SCED data in aphasia.

  3. An integrative strategy for quantitative analysis of the N-glycoproteome in complex biological samples.

    PubMed

    Wang, Ji; Zhou, Chuang; Zhang, Wei; Yao, Jun; Lu, Haojie; Dong, Qiongzhu; Zhou, Haijun; Qin, Lunxiu

    2014-01-15

    The complexity of protein glycosylation makes it difficult to characterize glycosylation patterns on a proteomic scale. In this study, we developed an integrated strategy for comparatively analyzing N-glycosylation/glycoproteins quantitatively from complex biological samples in a high-throughput manner. This strategy entailed separating and enriching glycopeptides/glycoproteins using lectin affinity chromatography, and then tandem labeling them with 18O/16O to generate a mass shift of 6 Da between the paired glycopeptides, and finally analyzing them with liquid chromatography-mass spectrometry (LC-MS) and the automatic quantitative method we developed based on Mascot Distiller. The accuracy and repeatability of this strategy were first verified using standard glycoproteins; linearity was maintained within a range of 1:10-10:1. The peptide concentration ratios obtained by the self-build quantitative method were similar to both the manually calculated and theoretical values, with a standard deviation (SD) of 0.023-0.186 for glycopeptides. The feasibility of the strategy was further confirmed with serum from hepatocellular carcinoma (HCC) patients and healthy individuals; the expression of 44 glycopeptides and 30 glycoproteins were significantly different between HCC patient and control serum. This strategy is accurate, repeatable, and efficient, and may be a useful tool for identification of disease-related N-glycosylation/glycoprotein changes.

  4. Receiver-operating-characteristic analysis of an automated program for analyzing striatal uptake of 123I-ioflupane SPECT images: calibration using visual reads.

    PubMed

    Kuo, Phillip Hsin; Avery, Ryan; Krupinski, Elizabeth; Lei, Hong; Bauer, Adam; Sherman, Scott; McMillan, Natalie; Seibyl, John; Zubal, George

    2013-03-01

    A fully automated objective striatal analysis (OSA) program that quantitates dopamine transporter uptake in subjects with suspected Parkinson's disease was applied to images from clinical (123)I-ioflupane studies. The striatal binding ratios or alternatively the specific binding ratio (SBR) of the lowest putamen uptake was computed, and receiver-operating-characteristic (ROC) analysis was applied to 94 subjects to determine the best discriminator using this quantitative method. Ninety-four (123)I-ioflupane SPECT scans were analyzed from patients referred to our clinical imaging department and were reconstructed using the manufacturer-supplied reconstruction and filtering parameters for the radiotracer. Three trained readers conducted independent visual interpretations and reported each case as either normal or showing dopaminergic deficit (abnormal). The same images were analyzed using the OSA software, which locates the striatal and occipital structures and places regions of interest on the caudate and putamen. Additionally, the OSA places a region of interest on the occipital region that is used to calculate the background-subtracted SBR. The lower SBR of the 2 putamen regions was taken as the quantitative report. The 33 normal (bilateral comma-shaped striata) and 61 abnormal (unilateral or bilateral dopaminergic deficit) studies were analyzed to generate ROC curves. Twenty-nine of the scans were interpreted as normal and 59 as abnormal by all 3 readers. For 12 scans, the 3 readers did not unanimously agree in their interpretations (discordant). The ROC analysis, which used the visual-majority-consensus interpretation from the readers as the gold standard, yielded an area under the curve of 0.958 when using 1.08 as the threshold SBR for the lowest putamen. The sensitivity and specificity of the automated quantitative analysis were 95% and 89%, respectively. The OSA program delivers SBR quantitative values that have a high sensitivity and specificity, compared

  5. Multigrid-based reconstruction algorithm for quantitative photoacoustic tomography

    PubMed Central

    Li, Shengfu; Montcel, Bruno; Yuan, Zhen; Liu, Wanyu; Vray, Didier

    2015-01-01

    This paper proposes a multigrid inversion framework for quantitative photoacoustic tomography reconstruction. The forward model of optical fluence distribution and the inverse problem are solved at multiple resolutions. A fixed-point iteration scheme is formulated for each resolution and used as a cost function. The simulated and experimental results for quantitative photoacoustic tomography reconstruction show that the proposed multigrid inversion can dramatically reduce the required number of iterations for the optimization process without loss of reliability in the results. PMID:26203371

  6. Quantitative Phase Imaging in a Volume Holographic Microscope

    NASA Astrophysics Data System (ADS)

    Waller, Laura; Luo, Yuan; Barbastathis, George

    2010-04-01

    We demonstrate a method for quantitative phase imaging in a Volume Holographic Microscope (VHM) from a single exposure, describe the properties of the system and show experimental results. The VHM system uses a multiplexed volume hologram (VH) to laterally separate images from different focal planes. This 3D intensity information is then used to solve the transport of intensity (TIE) equation and recover phase quantitatively. We discuss the modifications to the technique that were made in order to give accurate results.

  7. Quantitative cell biology: the essential role of theory.

    PubMed

    Howard, Jonathon

    2014-11-05

    Quantitative biology is a hot area, as evidenced by the recent establishment of institutes, graduate programs, and conferences with that name. But what is quantitative biology? What should it be? And how can it contribute to solving the big questions in biology? The past decade has seen very rapid development of quantitative experimental techniques, especially at the single-molecule and single-cell levels. In this essay, I argue that quantitative biology is much more than just the quantitation of these experimental results. Instead, it should be the application of the scientific method by which measurement is directed toward testing theories. In this view, quantitative biology is the recognition that theory and models play critical roles in biology, as they do in physics and engineering. By tying together experiment and theory, quantitative biology promises a deeper understanding of underlying mechanisms, when the theory works, or to new discoveries, when it does not. © 2014 Howard. This article is distributed by The American Society for Cell Biology under license from the author(s). Two months after publication it is available to the public under an Attribution–Noncommercial–Share Alike 3.0 Unported Creative Commons License (http://creativecommons.org/licenses/by-nc-sa/3.0).

  8. Quantitative instruments used to assess children's sense of smell: a review article.

    PubMed

    Moura, Raissa Gomes Fonseca; Cunha, Daniele Andrade; Gomes, Ana Carolina de Lima Gusmão; Silva, Hilton Justino da

    2014-01-01

    To systematically gather from the literature available the quantitative instruments used to assess the sense of smell in studies carried out with children. The present study included a survey in the Pubmed and Bireme platforms and in the databases of MedLine, Lilacs, regional SciELO and Web of Science, followed by selection and critical analysis of the articles found and chosen. We selected original articles related to the topic in question, conducted only with children in Portuguese, English, and Spanish. We excluded studies addressing other phases of human development, exclusively or concurrently with the pediatric population; studies on animals; literature review articles; dissertations; book chapters; case study articles; and editorials. A book report protocol was created for this study, including the following information: author, department, year, location, population/sample, age, purpose of the study, methods, and main results. We found 8,451 articles by typing keywords and identifiers. Out of this total, 5,928 were excluded by the title, 2,366 by the abstract, and 123 after we read the full text. Thus, 34 articles were selected, of which 28 were repeated in the databases, totalizing 6 articles analyzed in this review. We observed a lack of standardization of the quantitative instruments used to assess children's sense of smell, with great variability in the methodology of the tests, which reduces the effectiveness and reliability of the results.

  9. Comparative usefulness of inflammatory markers to indicate bacterial infection-analyzed according to blood culture results and related clinical factors.

    PubMed

    Nishikawa, Hirokazu; Shirano, Michinori; Kasamatsu, Yu; Morimura, Ayumi; Iida, Ko; Kishi, Tomomi; Goto, Tetsushi; Okamoto, Saki; Ehara, Eiji

    2016-01-01

    To assess relationships of inflammatory markers and 2 related clinical factors with blood culture results, we retrospectively investigated inpatients' blood culture and blood chemistry findings that were recorded from January to December 2014 using electronic medical records and analyzed the data of 852 subjects (426 culture-positive and 426 culture-negative). Results suggested that the risk of positive blood culture statistically increased as inflammatory marker levels and the number of related factors increased. Concerning the effectiveness of inflammatory markers, when the outcome definition was also changed for C-reactive protein (CRP), the odds ratio had a similar value, whereas when the outcome definition of blood culture positivity was used for procalcitonin (PCT), the greatest effectiveness of that was detected. Therefore, the current results suggest that PCT is more useful than CRP as an auxiliary indication of bacterial infection. Copyright © 2016 Elsevier Inc. All rights reserved.

  10. Analysis of genetic effects of nuclear-cytoplasmic interaction on quantitative traits: genetic model for diploid plants.

    PubMed

    Han, Lide; Yang, Jian; Zhu, Jun

    2007-06-01

    A genetic model was proposed for simultaneously analyzing genetic effects of nuclear, cytoplasm, and nuclear-cytoplasmic interaction (NCI) as well as their genotype by environment (GE) interaction for quantitative traits of diploid plants. In the model, the NCI effects were further partitioned into additive and dominance nuclear-cytoplasmic interaction components. Mixed linear model approaches were used for statistical analysis. On the basis of diallel cross designs, Monte Carlo simulations showed that the genetic model was robust for estimating variance components under several situations without specific effects. Random genetic effects were predicted by an adjusted unbiased prediction (AUP) method. Data on four quantitative traits (boll number, lint percentage, fiber length, and micronaire) in Upland cotton (Gossypium hirsutum L.) were analyzed as a worked example to show the effectiveness of the model.

  11. Quantitative study of flavonoids in leaves of citrus plants.

    PubMed

    Kawaii, S; Tomono, Y; Katase, E; Ogawa, K; Yano, M; Koizumi, M; Ito, C; Furukawa, H

    2000-09-01

    Leaf flavonoids were quantitatively determined in 68 representative or economically important Citrus species, cultivars, and near-Citrus relatives. Contents of 23 flavonoids including 6 polymethoxylated flavones were analyzed by means of reversed phase HPLC analysis. Principal component analysis revealed that the 7 associations according to Tanaka's classification were observed, but some do overlap each other. Group VII species could be divided into two different subgroups, namely, the first-10-species class and the last-19-species class according to Tanaka's classification numbers.

  12. Multiparametric Quantitative Ultrasound Imaging in Assessment of Chronic Kidney Disease.

    PubMed

    Gao, Jing; Perlman, Alan; Kalache, Safa; Berman, Nathaniel; Seshan, Surya; Salvatore, Steven; Smith, Lindsey; Wehrli, Natasha; Waldron, Levi; Kodali, Hanish; Chevalier, James

    2017-11-01

    To evaluate the value of multiparametric quantitative ultrasound imaging in assessing chronic kidney disease (CKD) using kidney biopsy pathologic findings as reference standards. We prospectively measured multiparametric quantitative ultrasound markers with grayscale, spectral Doppler, and acoustic radiation force impulse imaging in 25 patients with CKD before kidney biopsy and 10 healthy volunteers. Based on all pathologic (glomerulosclerosis, interstitial fibrosis/tubular atrophy, arteriosclerosis, and edema) scores, the patients with CKD were classified into mild (no grade 3 and <2 of grade 2) and moderate to severe (at least 2 of grade 2 or 1 of grade 3) CKD groups. Multiparametric quantitative ultrasound parameters included kidney length, cortical thickness, pixel intensity, parenchymal shear wave velocity, intrarenal artery peak systolic velocity (PSV), end-diastolic velocity (EDV), and resistive index. We tested the difference in quantitative ultrasound parameters among mild CKD, moderate to severe CKD, and healthy controls using analysis of variance, analyzed correlations of quantitative ultrasound parameters with pathologic scores and the estimated glomerular filtration rate (GFR) using Pearson correlation coefficients, and examined the diagnostic performance of quantitative ultrasound parameters in determining moderate CKD and an estimated GFR of less than 60 mL/min/1.73 m 2 using receiver operating characteristic curve analysis. There were significant differences in cortical thickness, pixel intensity, PSV, and EDV among the 3 groups (all P < .01). Among quantitative ultrasound parameters, the top areas under the receiver operating characteristic curves for PSV and EDV were 0.88 and 0.97, respectively, for determining pathologic moderate to severe CKD, and 0.76 and 0.86 for estimated GFR of less than 60 mL/min/1.73 m 2 . Moderate to good correlations were found for PSV, EDV, and pixel intensity with pathologic scores and estimated GFR. The

  13. Matrix effect and correction by standard addition in quantitative liquid chromatographic-mass spectrometric analysis of diarrhetic shellfish poisoning toxins.

    PubMed

    Ito, Shinya; Tsukada, Katsuo

    2002-01-11

    An evaluation of the feasibility of liquid chromatography-mass spectrometry (LC-MS) with atmospheric pressure ionization was made for quantitation of four diarrhetic shellfish poisoning toxins, okadaic acid, dinophysistoxin-1, pectenotoxin-6 and yessotoxin in scallops. When LC-MS was applied to the analysis of scallop extracts, large signal suppressions were observed due to coeluting substances from the column. To compensate for these matrix signal suppressions, the standard addition method was applied. First, the sample was analyzed and then the sample involving the addition of calibration standards is analyzed. Although this method requires two LC-MS runs per analysis, effective correction of quantitative errors was found.

  14. Critical Quantitative Inquiry in Context

    ERIC Educational Resources Information Center

    Stage, Frances K.; Wells, Ryan S.

    2014-01-01

    This chapter briefly traces the development of the concept of critical quantitative inquiry, provides an expanded conceptualization of the tasks of critical quantitative research, offers theoretical explanation and justification for critical research using quantitative methods, and previews the work of quantitative criticalists presented in this…

  15. Quantitative dispersion microscopy

    PubMed Central

    Fu, Dan; Choi, Wonshik; Sung, Yongjin; Yaqoob, Zahid; Dasari, Ramachandra R.; Feld, Michael

    2010-01-01

    Refractive index dispersion is an intrinsic optical property and a useful source of contrast in biological imaging studies. In this report, we present the first dispersion phase imaging of living eukaryotic cells. We have developed quantitative dispersion microscopy based on the principle of quantitative phase microscopy. The dual-wavelength quantitative phase microscope makes phase measurements at 310 nm and 400 nm wavelengths to quantify dispersion (refractive index increment ratio) of live cells. The measured dispersion of living HeLa cells is found to be around 1.088, which agrees well with that measured directly for protein solutions using total internal reflection. This technique, together with the dry mass and morphology measurements provided by quantitative phase microscopy, could prove to be a useful tool for distinguishing different types of biomaterials and studying spatial inhomogeneities of biological samples. PMID:21113234

  16. Development of a real-time and quantitative thrombus sensor for an extracorporeal centrifugal blood pump by near-infrared light

    PubMed Central

    Sakota, Daisuke; Fujiwara, Tatsuki; Ohuchi, Katsuhiro; Kuwana, Katsuyuki; Yamazaki, Hiroyuki; Kosaka, Ryo; Nishida, Masahiro; Mizuno, Tomohiro; Arai, Hirokuni; Maruyama, Osamu

    2017-01-01

    We developed an optical thrombus sensor for a monopivot extracorporeal centrifugal blood pump. In this study, we investigated its quantitative performance for thrombus detection in acute animal experiments of left ventricular assist using the pump on pathogen-free pigs. Optical fibers were set in the driver unit of the pump. The incident light at the near-infrared wavelength of 810 nm was aimed at the pivot bearing, and the resulting scattered light was guided to the optical fibers. The detected signal was analyzed to obtain the thrombus formation level. As a result, real-time and quantitative monitoring of the thrombus surface area on the pivot bearing was achieved with an accuracy of 3.6 ± 2.3 mm2. In addition, the sensing method using the near-infrared light was not influenced by changes in the oxygen saturation and the hematocrit. It is expected that the developed sensor will be useful for optimal anticoagulation management for long-term extracorporeal circulation therapies. PMID:29359096

  17. Development of a real-time and quantitative thrombus sensor for an extracorporeal centrifugal blood pump by near-infrared light.

    PubMed

    Sakota, Daisuke; Fujiwara, Tatsuki; Ohuchi, Katsuhiro; Kuwana, Katsuyuki; Yamazaki, Hiroyuki; Kosaka, Ryo; Nishida, Masahiro; Mizuno, Tomohiro; Arai, Hirokuni; Maruyama, Osamu

    2018-01-01

    We developed an optical thrombus sensor for a monopivot extracorporeal centrifugal blood pump. In this study, we investigated its quantitative performance for thrombus detection in acute animal experiments of left ventricular assist using the pump on pathogen-free pigs. Optical fibers were set in the driver unit of the pump. The incident light at the near-infrared wavelength of 810 nm was aimed at the pivot bearing, and the resulting scattered light was guided to the optical fibers. The detected signal was analyzed to obtain the thrombus formation level. As a result, real-time and quantitative monitoring of the thrombus surface area on the pivot bearing was achieved with an accuracy of 3.6 ± 2.3 mm 2 . In addition, the sensing method using the near-infrared light was not influenced by changes in the oxygen saturation and the hematocrit. It is expected that the developed sensor will be useful for optimal anticoagulation management for long-term extracorporeal circulation therapies.

  18. Determination of mercury in ayurvedic dietary supplements that are not rasa shastra using the hydra-C direct mercury analyzer.

    PubMed

    Abdalla, Amir A; Smith, Robert E

    2013-01-01

    Mercury has been determined in Ayurvedic dietary supplements (Trifala, Trifala Guggulu, Turmeric, Mahasudarshan, Yograj, Shatawari, Hingwastika, Shatavari, and Shilajit) by inductively coupled plasma-mass spectrometry (ICP-MS) and direct mercury analysis using the Hydra-C direct mercury analyzer (Teledyne Leeman Labs Hudson, NH, USA). Similar results were obtained from the two methods, but the direct mercury analysis method was much faster and safer and required no microwave digestion (unlike ICP-MS). Levels of mercury ranged from 0.002 to 56  μ g/g in samples of dietary supplements. Standard reference materials Ephedra 3240 and tomato leaves that were from the National Institute of Standard and Technology (NIST) and dogfish liver (DOLT3) that was from the Canadian Research Council were analyzed using Hydra-C method. Average mercury recoveries were 102% (RSD% 0.0018), 100% (RSD% 0.0009), and 101% (RSD% 0.0729), respectively. Hydra-C method Limit Of Quantitation was 0.5 ng.

  19. Determination of Mercury in Ayurvedic Dietary Supplements That Are Not Rasa Shastra Using the Hydra-C Direct Mercury Analyzer

    PubMed Central

    Abdalla, Amir A.; Smith, Robert E.

    2013-01-01

    Mercury has been determined in Ayurvedic dietary supplements (Trifala, Trifala Guggulu, Turmeric, Mahasudarshan, Yograj, Shatawari, Hingwastika, Shatavari, and Shilajit) by inductively coupled plasma-mass spectrometry (ICP-MS) and direct mercury analysis using the Hydra-C direct mercury analyzer (Teledyne Leeman Labs Hudson, NH, USA). Similar results were obtained from the two methods, but the direct mercury analysis method was much faster and safer and required no microwave digestion (unlike ICP-MS). Levels of mercury ranged from 0.002 to 56 μg/g in samples of dietary supplements. Standard reference materials Ephedra 3240 and tomato leaves that were from the National Institute of Standard and Technology (NIST) and dogfish liver (DOLT3) that was from the Canadian Research Council were analyzed using Hydra-C method. Average mercury recoveries were 102% (RSD% 0.0018), 100% (RSD% 0.0009), and 101% (RSD% 0.0729), respectively. Hydra-C method Limit Of Quantitation was 0.5 ng. PMID:23710181

  20. Accurate ECG diagnosis of atrial tachyarrhythmias using quantitative analysis: a prospective diagnostic and cost-effectiveness study.

    PubMed

    Krummen, David E; Patel, Mitul; Nguyen, Hong; Ho, Gordon; Kazi, Dhruv S; Clopton, Paul; Holland, Marian C; Greenberg, Scott L; Feld, Gregory K; Faddis, Mitchell N; Narayan, Sanjiv M

    2010-11-01

    Quantitative ECG Analysis. Optimal atrial tachyarrhythmia management is facilitated by accurate electrocardiogram interpretation, yet typical atrial flutter (AFl) may present without sawtooth F-waves or RR regularity, and atrial fibrillation (AF) may be difficult to separate from atypical AFl or rapid focal atrial tachycardia (AT). We analyzed whether improved diagnostic accuracy using a validated analysis tool significantly impacts costs and patient care. We performed a prospective, blinded, multicenter study using a novel quantitative computerized algorithm to identify atrial tachyarrhythmia mechanism from the surface ECG in patients referred for electrophysiology study (EPS). In 122 consecutive patients (age 60 ± 12 years) referred for EPS, 91 sustained atrial tachyarrhythmias were studied. ECGs were also interpreted by 9 physicians from 3 specialties for comparison and to allow healthcare system modeling. Diagnostic accuracy was compared to the diagnosis at EPS. A Markov model was used to estimate the impact of improved arrhythmia diagnosis. We found 13% of typical AFl ECGs had neither sawtooth flutter waves nor RR regularity, and were misdiagnosed by the majority of clinicians (0/6 correctly diagnosed by consensus visual interpretation) but correctly by quantitative analysis in 83% (5/6, P = 0.03). AF diagnosis was also improved through use of the algorithm (92%) versus visual interpretation (primary care: 76%, P < 0.01). Economically, we found that these improvements in diagnostic accuracy resulted in an average cost-savings of $1,303 and 0.007 quality-adjusted-life-years per patient. Typical AFl and AF are frequently misdiagnosed using visual criteria. Quantitative analysis improves diagnostic accuracy and results in improved healthcare costs and patient outcomes. © 2010 Wiley Periodicals, Inc.

  1. Errors in retarding potential analyzers caused by nonuniformity of the grid-plane potential.

    NASA Technical Reports Server (NTRS)

    Hanson, W. B.; Frame, D. R.; Midgley, J. E.

    1972-01-01

    One aspect of the degradation in performance of retarding potential analyzers caused by potential depressions in the retarding grid is quantitatively estimated from laboratory measurements and theoretical calculations. A simple expression is obtained that permits the use of laboratory measurements of grid properties to make first-order corrections to flight data. Systematic positive errors in ion temperature of approximately 16% for the Ogo 4 instrument and 3% for the Ogo 6 instrument are deduced. The effects of the transverse electric fields arising from the grid potential depressions are not treated.

  2. Organizing "mountains of words" for data analysis, both qualitative and quantitative.

    PubMed

    Johnson, Bruce D; Dunlap, Eloise; Benoit, Ellen

    2010-04-01

    Qualitative research creates mountains of words. U.S. federal funding supports mostly structured qualitative research, which is designed to test hypotheses using semiquantitative coding and analysis. This article reports on strategies for planning, organizing, collecting, managing, storing, retrieving, analyzing, and writing about qualitative data so as to most efficiently manage the mountains of words collected in large-scale ethnographic projects. Multiple benefits accrue from this approach. Field expenditures are linked to units of work so productivity is measured, many staff in various locations have access to use and analyze the data, quantitative data can be derived from data that is primarily qualitative, and improved efficiencies of resources are developed.

  3. A Comparison of Multivariate and Pre-Processing Methods for Quantitative Laser-Induced Breakdown Spectroscopy of Geologic Samples

    NASA Technical Reports Server (NTRS)

    Anderson, R. B.; Morris, R. V.; Clegg, S. M.; Bell, J. F., III; Humphries, S. D.; Wiens, R. C.

    2011-01-01

    The ChemCam instrument selected for the Curiosity rover is capable of remote laser-induced breakdown spectroscopy (LIBS).[1] We used a remote LIBS instrument similar to ChemCam to analyze 197 geologic slab samples and 32 pressed-powder geostandards. The slab samples are well-characterized and have been used to validate the calibration of previous instruments on Mars missions, including CRISM [2], OMEGA [3], the MER Pancam [4], Mini-TES [5], and Moessbauer [6] instruments and the Phoenix SSI [7]. The resulting dataset was used to compare multivariate methods for quantitative LIBS and to determine the effect of grain size on calculations. Three multivariate methods - partial least squares (PLS), multilayer perceptron artificial neural networks (MLP ANNs) and cascade correlation (CC) ANNs - were used to generate models and extract the quantitative composition of unknown samples. PLS can be used to predict one element (PLS1) or multiple elements (PLS2) at a time, as can the neural network methods. Although MLP and CC ANNs were successful in some cases, PLS generally produced the most accurate and precise results.

  4. Biological Dynamics Markup Language (BDML): an open format for representing quantitative biological dynamics data.

    PubMed

    Kyoda, Koji; Tohsato, Yukako; Ho, Kenneth H L; Onami, Shuichi

    2015-04-01

    Recent progress in live-cell imaging and modeling techniques has resulted in generation of a large amount of quantitative data (from experimental measurements and computer simulations) on spatiotemporal dynamics of biological objects such as molecules, cells and organisms. Although many research groups have independently dedicated their efforts to developing software tools for visualizing and analyzing these data, these tools are often not compatible with each other because of different data formats. We developed an open unified format, Biological Dynamics Markup Language (BDML; current version: 0.2), which provides a basic framework for representing quantitative biological dynamics data for objects ranging from molecules to cells to organisms. BDML is based on Extensible Markup Language (XML). Its advantages are machine and human readability and extensibility. BDML will improve the efficiency of development and evaluation of software tools for data visualization and analysis. A specification and a schema file for BDML are freely available online at http://ssbd.qbic.riken.jp/bdml/. Supplementary data are available at Bioinformatics online. © The Author 2014. Published by Oxford University Press.

  5. A sampling framework for incorporating quantitative mass spectrometry data in protein interaction analysis.

    PubMed

    Tucker, George; Loh, Po-Ru; Berger, Bonnie

    2013-10-04

    Comprehensive protein-protein interaction (PPI) maps are a powerful resource for uncovering the molecular basis of genetic interactions and providing mechanistic insights. Over the past decade, high-throughput experimental techniques have been developed to generate PPI maps at proteome scale, first using yeast two-hybrid approaches and more recently via affinity purification combined with mass spectrometry (AP-MS). Unfortunately, data from both protocols are prone to both high false positive and false negative rates. To address these issues, many methods have been developed to post-process raw PPI data. However, with few exceptions, these methods only analyze binary experimental data (in which each potential interaction tested is deemed either observed or unobserved), neglecting quantitative information available from AP-MS such as spectral counts. We propose a novel method for incorporating quantitative information from AP-MS data into existing PPI inference methods that analyze binary interaction data. Our approach introduces a probabilistic framework that models the statistical noise inherent in observations of co-purifications. Using a sampling-based approach, we model the uncertainty of interactions with low spectral counts by generating an ensemble of possible alternative experimental outcomes. We then apply the existing method of choice to each alternative outcome and aggregate results over the ensemble. We validate our approach on three recent AP-MS data sets and demonstrate performance comparable to or better than state-of-the-art methods. Additionally, we provide an in-depth discussion comparing the theoretical bases of existing approaches and identify common aspects that may be key to their performance. Our sampling framework extends the existing body of work on PPI analysis using binary interaction data to apply to the richer quantitative data now commonly available through AP-MS assays. This framework is quite general, and many enhancements are likely

  6. Quantitation of sugar content in pyrolysis liquids after acid hydrolysis using high-performance liquid chromatography without neutralization.

    PubMed

    Johnston, Patrick A; Brown, Robert C

    2014-08-13

    A rapid method for the quantitation of total sugars in pyrolysis liquids using high-performance liquid chromatography (HPLC) was developed. The method avoids the tedious and time-consuming sample preparation required by current analytical methods. It is possible to directly analyze hydrolyzed pyrolysis liquids, bypassing the neutralization step usually required in determination of total sugars. A comparison with traditional methods was used to determine the validity of the results. The calibration curve coefficient of determination on all standard compounds was >0.999 using a refractive index detector. The relative standard deviation for the new method was 1.13%. The spiked sugar recoveries on the pyrolysis liquid samples were between 104 and 105%. The research demonstrates that it is possible to obtain excellent accuracy and efficiency using HPLC to quantitate glucose after acid hydrolysis of polymeric and oligomeric sugars found in fast pyrolysis bio-oils without neutralization.

  7. Quantitative Tagless Copurification: A Method to Validate and Identify Protein-Protein Interactions

    DOE PAGES

    Shatsky, Maxim; Dong, Ming; Liu, Haichuan; ...

    2016-04-20

    Identifying protein-protein interactions (PPIs) at an acceptable false discovery rate (FDR) is challenging. Previously we identified several hundred PPIs from affinity purification - mass spectrometry (AP-MS) data for the bacteria Escherichia coli and Desulfovibrio vulgaris. These two interactomes have lower FDRs than any of the nine interactomes proposed previously for bacteria and are more enriched in PPIs validated by other data than the nine earlier interactomes. To more thoroughly determine the accuracy of ours or other interactomes and to discover further PPIs de novo, here we present a quantitative tagless method that employs iTRAQ MS to measure the copurification ofmore » endogenous proteins through orthogonal chromatography steps. 5273 fractions from a four-step fractionation of a D. vulgaris protein extract were assayed, resulting in the detection of 1242 proteins. Protein partners from our D. vulgaris and E. coli AP-MS interactomes copurify as frequently as pairs belonging to three benchmark data sets of well-characterized PPIs. In contrast, the protein pairs from the nine other bacterial interactomes copurify two- to 20-fold less often. We also identify 200 high confidence D. vulgaris PPIs based on tagless copurification and colocalization in the genome. These PPIs are as strongly validated by other data as our AP-MS interactomes and overlap with our AP-MS interactome for D.vulgaris within 3% of expectation, once FDRs and false negative rates are taken into account. Finally, we reanalyzed data from two quantitative tagless screens of human cell extracts. We estimate that the novel PPIs reported in these studies have an FDR of at least 85% and find that less than 7% of the novel PPIs identified in each screen overlap. Our results establish that a quantitative tagless method can be used to validate and identify PPIs, but that such data must be analyzed carefully to minimize the FDR.« less

  8. Quantitative Tagless Copurification: A Method to Validate and Identify Protein-Protein Interactions

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Shatsky, Maxim; Dong, Ming; Liu, Haichuan

    Identifying protein-protein interactions (PPIs) at an acceptable false discovery rate (FDR) is challenging. Previously we identified several hundred PPIs from affinity purification - mass spectrometry (AP-MS) data for the bacteria Escherichia coli and Desulfovibrio vulgaris. These two interactomes have lower FDRs than any of the nine interactomes proposed previously for bacteria and are more enriched in PPIs validated by other data than the nine earlier interactomes. To more thoroughly determine the accuracy of ours or other interactomes and to discover further PPIs de novo, here we present a quantitative tagless method that employs iTRAQ MS to measure the copurification ofmore » endogenous proteins through orthogonal chromatography steps. 5273 fractions from a four-step fractionation of a D. vulgaris protein extract were assayed, resulting in the detection of 1242 proteins. Protein partners from our D. vulgaris and E. coli AP-MS interactomes copurify as frequently as pairs belonging to three benchmark data sets of well-characterized PPIs. In contrast, the protein pairs from the nine other bacterial interactomes copurify two- to 20-fold less often. We also identify 200 high confidence D. vulgaris PPIs based on tagless copurification and colocalization in the genome. These PPIs are as strongly validated by other data as our AP-MS interactomes and overlap with our AP-MS interactome for D.vulgaris within 3% of expectation, once FDRs and false negative rates are taken into account. Finally, we reanalyzed data from two quantitative tagless screens of human cell extracts. We estimate that the novel PPIs reported in these studies have an FDR of at least 85% and find that less than 7% of the novel PPIs identified in each screen overlap. Our results establish that a quantitative tagless method can be used to validate and identify PPIs, but that such data must be analyzed carefully to minimize the FDR.« less

  9. Development of a quantitative diagnostic method of estrogen receptor expression levels by immunohistochemistry using organic fluorescent material-assembled nanoparticles

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Gonda, Kohsuke, E-mail: gonda@med.tohoku.ac.jp; Miyashita, Minoru; Watanabe, Mika

    2012-09-28

    sections to quantitatively examine the two methods. The results demonstrated that our nanoparticle staining analyzed a wide range of ER expression levels with higher accuracy and quantitative sensitivity than DAB staining. This enhancement in the diagnostic accuracy and sensitivity for ERs using our immunostaining method will improve the prediction of responses to therapies that target ERs and progesterone receptors that are induced by a downstream ER signal.« less

  10. Using PSEA-Quant for Protein Set Enrichment Analysis of Quantitative Mass Spectrometry-Based Proteomics

    PubMed Central

    Lavallée-Adam, Mathieu

    2017-01-01

    PSEA-Quant analyzes quantitative mass spectrometry-based proteomics datasets to identify enrichments of annotations contained in repositories such as the Gene Ontology and Molecular Signature databases. It allows users to identify the annotations that are significantly enriched for reproducibly quantified high abundance proteins. PSEA-Quant is available on the web and as a command-line tool. It is compatible with all label-free and isotopic labeling-based quantitative proteomics methods. This protocol describes how to use PSEA-Quant and interpret its output. The importance of each parameter as well as troubleshooting approaches are also discussed. PMID:27010334

  11. Quantitative Resistance: More Than Just Perception of a Pathogen.

    PubMed

    Corwin, Jason A; Kliebenstein, Daniel J

    2017-04-01

    Molecular plant pathology has focused on studying large-effect qualitative resistance loci that predominantly function in detecting pathogens and/or transmitting signals resulting from pathogen detection. By contrast, less is known about quantitative resistance loci, particularly the molecular mechanisms controlling variation in quantitative resistance. Recent studies have provided insight into these mechanisms, showing that genetic variation at hundreds of causal genes may underpin quantitative resistance. Loci controlling quantitative resistance contain some of the same causal genes that mediate qualitative resistance, but the predominant mechanisms of quantitative resistance extend beyond pathogen recognition. Indeed, most causal genes for quantitative resistance encode specific defense-related outputs such as strengthening of the cell wall or defense compound biosynthesis. Extending previous work on qualitative resistance to focus on the mechanisms of quantitative resistance, such as the link between perception of microbe-associated molecular patterns and growth, has shown that the mechanisms underlying these defense outputs are also highly polygenic. Studies that include genetic variation in the pathogen have begun to highlight a potential need to rethink how the field considers broad-spectrum resistance and how it is affected by genetic variation within pathogen species and between pathogen species. These studies are broadening our understanding of quantitative resistance and highlighting the potentially vast scale of the genetic basis of quantitative resistance. © 2017 American Society of Plant Biologists. All rights reserved.

  12. Quantitative Resistance: More Than Just Perception of a Pathogen

    PubMed Central

    2017-01-01

    Molecular plant pathology has focused on studying large-effect qualitative resistance loci that predominantly function in detecting pathogens and/or transmitting signals resulting from pathogen detection. By contrast, less is known about quantitative resistance loci, particularly the molecular mechanisms controlling variation in quantitative resistance. Recent studies have provided insight into these mechanisms, showing that genetic variation at hundreds of causal genes may underpin quantitative resistance. Loci controlling quantitative resistance contain some of the same causal genes that mediate qualitative resistance, but the predominant mechanisms of quantitative resistance extend beyond pathogen recognition. Indeed, most causal genes for quantitative resistance encode specific defense-related outputs such as strengthening of the cell wall or defense compound biosynthesis. Extending previous work on qualitative resistance to focus on the mechanisms of quantitative resistance, such as the link between perception of microbe-associated molecular patterns and growth, has shown that the mechanisms underlying these defense outputs are also highly polygenic. Studies that include genetic variation in the pathogen have begun to highlight a potential need to rethink how the field considers broad-spectrum resistance and how it is affected by genetic variation within pathogen species and between pathogen species. These studies are broadening our understanding of quantitative resistance and highlighting the potentially vast scale of the genetic basis of quantitative resistance. PMID:28302676

  13. System Design of One-chip Wave Particle Interaction Analyzer for SCOPE mission.

    NASA Astrophysics Data System (ADS)

    Fukuhara, Hajime; Ueda, Yoshikatsu; Kojima, Hiro; Yamakawa, Hiroshi

    In past science spacecrafts such like GEOTAIL, we usually capture electric and magnetic field waveforms and observe energetic eletron and ion particles as velocity distributions by each sensor. We analyze plasma wave-particle interactions by these respective data and the discussions are sometimes restricted by the difference of time resolution and by the data loss in desired regions. One-chip Wave Particle Interaction Analyzer (OWPIA) conducts direct quantitative observations of wave-particle interaction by direct 'E dot v' calculation on-board. This new instruments have a capability to use all plasma waveform data and electron particle informations. In the OWPIA system, we have to calibrate the digital observation data and transform the same coordinate system. All necessary calculations are processed in Field Programmable Gate Array(FPGA). In our study, we introduce a basic concept of the OWPIA system and a optimization method for each calculation functions installed in FPGA. And we also discuss the process speed, the FPGA utilization efficiency, the total power consumption.

  14. Quantitation of permethylated N-glycans through multiple-reaction monitoring (MRM) LC-MS/MS.

    PubMed

    Zhou, Shiyue; Hu, Yunli; DeSantos-Garcia, Janie L; Mechref, Yehia

    2015-04-01

    The important biological roles of glycans and their implications in disease development and progression have created a demand for the development of sensitive quantitative glycomics methods. Quantitation of glycans existing at low abundance is still analytically challenging. In this study, an N-linked glycans quantitation method using multiple-reaction monitoring (MRM) on a triple quadrupole instrument was developed. Optimum normalized collision energy (CE) for both sialylated and fucosylated N-glycan was determined to be 30%, whereas it was found to be 35% for either fucosylated or sialylated N-glycans. The optimum CE for mannose and complex type N-glycan was determined to be 35%. Additionally, the use of three transitions was shown to facilitate reliable quantitation. A total of 88 N-glycan compositions in human blood serum were quantified using this MRM approach. Reliable detection and quantitation of these glycans was achieved when the equivalence of 0.005 μL of blood serum was analyzed. Accordingly, N-glycans down to the 100th of a μL level can be reliably quantified in pooled human blood serum, spanning a dynamic concentration range of three orders of magnitude. MRM was also effectively utilized to quantitatively compare the expression of N-glycans derived from brain-targeting breast carcinoma cells (MDA-MB-231BR) and metastatic breast cancer cells (MDA-MB-231). Thus, the described MRM method of permethylated N-glycan enables a rapid and reliable identification and quantitation of glycans derived from glycoproteins purified or present in complex biological samples.

  15. Quantitation of Permethylated N-Glycans through Multiple-Reaction Monitoring (MRM) LC-MS/MS

    PubMed Central

    Zhou, Shiyue; Hu, Yunli; DeSantos-Garcia, Janie L.; Mechref, Yehia

    2015-01-01

    The important biological roles of glycans and their implications in disease development and progression have created a demand for the development of sensitive quantitative glycomics methods. Quantitation of glycans existing at low abundance is still analytically challenging. In this study, an N-linked glycans quantitation method using multiple reaction monitoring (MRM) on a triple quadrupole instrument was developed. Optimum normalized collision energy (CE) for both sialylated and fucosylated N-glycan structures was determined to be 30% while it was found to be 35% for either fucosylated or sialylated structures The optimum CE for mannose and complex type N-glycan structures was determined to be 35%. Additionally, the use of three transitions was shown to facilitate reliable quantitation. A total of 88 N-glycan structures in human blood serum were quantified using this MRM approach. Reliable detection and quantitation of these structures was achieved when the equivalence of 0.005 μL of blood serum was analyzed. Accordingly, N-glycans down to the 100th of a μL level can be reliably quantified in pooled human blood serum, spanning a dynamic concentration range of three orders of magnitudes. MRM was also effectively utilized to quantitatively compare the expression of N-glycans derived from brain-targeting breast carcinoma cells (MDA-MB-231BR) and metastatic breast cancer cells (MDA-MB-231). Thus, the described MRM method of permethylated N-glycan structures enables a rapid and reliable identification and quantitation of glycans derived from glycoproteins purified or present in complex biological samples. PMID:25698222

  16. Quantitation of Permethylated N-Glycans through Multiple-Reaction Monitoring (MRM) LC-MS/MS

    NASA Astrophysics Data System (ADS)

    Zhou, Shiyue; Hu, Yunli; DeSantos-Garcia, Janie L.; Mechref, Yehia

    2015-04-01

    The important biological roles of glycans and their implications in disease development and progression have created a demand for the development of sensitive quantitative glycomics methods. Quantitation of glycans existing at low abundance is still analytically challenging. In this study, an N-linked glycans quantitation method using multiple-reaction monitoring (MRM) on a triple quadrupole instrument was developed. Optimum normalized collision energy (CE) for both sialylated and fucosylated N-glycan was determined to be 30%, whereas it was found to be 35% for either fucosylated or sialylated N-glycans. The optimum CE for mannose and complex type N-glycan was determined to be 35%. Additionally, the use of three transitions was shown to facilitate reliable quantitation. A total of 88 N-glycan compositions in human blood serum were quantified using this MRM approach. Reliable detection and quantitation of these glycans was achieved when the equivalence of 0.005 μL of blood serum was analyzed. Accordingly, N-glycans down to the 100th of a μL level can be reliably quantified in pooled human blood serum, spanning a dynamic concentration range of three orders of magnitude. MRM was also effectively utilized to quantitatively compare the expression of N-glycans derived from brain-targeting breast carcinoma cells (MDA-MB-231BR) and metastatic breast cancer cells (MDA-MB-231). Thus, the described MRM method of permethylated N-glycan enables a rapid and reliable identification and quantitation of glycans derived from glycoproteins purified or present in complex biological samples.

  17. Preliminary clinical results: an analyzing tool for 2D optical imaging in detection of active inflammation in rheumatoid arthritis

    NASA Astrophysics Data System (ADS)

    Adi Aizudin Bin Radin Nasirudin, Radin; Meier, Reinhard; Ahari, Carmen; Sievert, Matti; Fiebich, Martin; Rummeny, Ernst J.; No"l, Peter B.

    2011-03-01

    Optical imaging (OI) is a relatively new method in detecting active inflammation of hand joints of patients suffering from rheumatoid arthritis (RA). With the high number of people affected by this disease especially in western countries, the availability of OI as an early diagnostic imaging method is clinically highly relevant. In this paper, we present a newly in-house developed OI analyzing tool and a clinical evaluation study. Our analyzing tool extends the capability of existing OI tools. We include many features in the tool, such as region-based image analysis, hyper perfusion curve analysis, and multi-modality image fusion to aid clinicians in localizing and determining the intensity of inflammation in joints. Additionally, image data management options, such as the full integration of PACS/RIS, are included. In our clinical study we demonstrate how OI facilitates the detection of active inflammation in rheumatoid arthritis. The preliminary clinical results indicate a sensitivity of 43.5%, a specificity of 80.3%, an accuracy of 65.7%, a positive predictive value of 76.6%, and a negative predictive value of 64.9% in relation to clinical results from MRI. The accuracy of inflammation detection serves as evidence to the potential of OI as a useful imaging modality for early detection of active inflammation in patients with rheumatoid arthritis. With our in-house developed tool we extend the usefulness of OI imaging in the clinical arena. Overall, we show that OI is a fast, inexpensive, non-invasive and nonionizing yet highly sensitive and accurate imaging modality.-

  18. A field- and laboratory-based quantitative analysis of alluvium: Relating analytical results to TIMS data

    NASA Technical Reports Server (NTRS)

    Wenrich, Melissa L.; Hamilton, Victoria E.; Christensen, Philip R.

    1995-01-01

    Thermal Infrared Multispectral Scanner (TIMS) data were acquired over the McDowell Mountains northeast of Scottsdale, Arizona during August 1994. The raw data were processed to emphasize lithologic differences using a decorrelation stretch and assigning bands 5, 3, and 1 to red, green, and blue, respectively. Processed data of alluvium flanking the mountains exhibit moderate color variation. The objective of this study was to determine, using a quantitative approach, what environmental variable(s), in the absence of bedrock, is/are responsible for influencing the spectral properties of the desert alluvial surface.

  19. The Role of Introductory Geosciences in Students' Quantitative Literacy

    NASA Astrophysics Data System (ADS)

    Wenner, J. M.; Manduca, C.; Baer, E. M.

    2006-12-01

    Quantitative literacy is more than mathematics; it is about reasoning with data. Colleges and universities have begun to recognize the distinction between mathematics and quantitative literacy, modifying curricula to reflect the need for numerate citizens. Although students may view geology as 'rocks for jocks', the geosciences are truthfully rife with data, making introductory geoscience topics excellent context for developing the quantitative literacy of students with diverse backgrounds. In addition, many news items that deal with quantitative skills, such as the global warming phenomenon, have their basis in the Earth sciences and can serve as timely examples of the importance of quantitative literacy for all students in introductory geology classrooms. Participants at a workshop held in 2006, 'Infusing Quantitative Literacy into Introductory Geoscience Courses,' discussed and explored the challenges and opportunities associated with the inclusion of quantitative material and brainstormed about effective practices for imparting quantitative literacy to students with diverse backgrounds. The tangible results of this workshop add to the growing collection of quantitative materials available through the DLESE- and NSF-supported Teaching Quantitative Skills in the Geosciences website, housed at SERC. There, faculty can find a collection of pages devoted to the successful incorporation of quantitative literacy in introductory geoscience. The resources on the website are designed to help faculty to increase their comfort with presenting quantitative ideas to students with diverse mathematical abilities. A methods section on "Teaching Quantitative Literacy" (http://serc.carleton.edu/quantskills/methods/quantlit/index.html) focuses on connecting quantitative concepts with geoscience context and provides tips, trouble-shooting advice and examples of quantitative activities. The goal in this section is to provide faculty with material that can be readily incorporated

  20. Quantitative impedance measurements for eddy current model validation

    NASA Astrophysics Data System (ADS)

    Khan, T. A.; Nakagawa, N.

    2000-05-01

    This paper reports on a series of laboratory-based impedance measurement data, collected by the use of a quantitatively accurate, mechanically controlled measurement station. The purpose of the measurement is to validate a BEM-based eddy current model against experiment. We have therefore selected two "validation probes," which are both split-D differential probes. Their internal structures and dimensions are extracted from x-ray CT scan data, and thus known within the measurement tolerance. A series of measurements was carried out, using the validation probes and two Ti-6Al-4V block specimens, one containing two 1-mm long fatigue cracks, and the other containing six EDM notches of a range of sizes. Motor-controlled XY scanner performed raster scans over the cracks, with the probe riding on the surface with a spring-loaded mechanism to maintain the lift off. Both an impedance analyzer and a commercial EC instrument were used in the measurement. The probes were driven in both differential and single-coil modes for the specific purpose of model validation. The differential measurements were done exclusively by the eddyscope, while the single-coil data were taken with both the impedance analyzer and the eddyscope. From the single-coil measurements, we obtained the transfer function to translate the voltage output of the eddyscope into impedance values, and then used it to translate the differential measurement data into impedance results. The presentation will highlight the schematics of the measurement procedure, a representative of raw data, explanation of the post data-processing procedure, and then a series of resulting 2D flaw impedance results. A noise estimation will be given also, in order to quantify the accuracy of these measurements, and to be used in probability-of-detection estimation.—This work was supported by the NSF Industry/University Cooperative Research Program.

  1. Quantitative electrophysiological monitoring of anti-histamine drug effects on live cells via reusable sensor platforms.

    PubMed

    Pham Ba, Viet Anh; Cho, Dong-Guk; Kim, Daesan; Yoo, Haneul; Ta, Van-Thao; Hong, Seunghun

    2017-08-15

    We demonstrated the quantitative electrophysiological monitoring of histamine and anti-histamine drug effects on live cells via reusable sensor platforms based on carbon nanotube transistors. This method enabled us to monitor the real-time electrophysiological responses of a single HeLa cell to histamine with different concentrations. The measured electrophysiological responses were attributed to the activity of histamine type 1 receptors on a HeLa cell membrane by histamine. Furthermore, the effects of anti-histamine drugs such as cetirizine or chlorphenamine on the electrophysiological activities of HeLa cells were also evaluated quantitatively. Significantly, we utilized only a single device to monitor the responses of multiple HeLa cells to each drug, which allowed us to quantitatively analyze the antihistamine drug effects on live cells without errors from the device-to-device variation in device characteristics. Such quantitative evaluation capability of our method would promise versatile applications such as drug screening and nanoscale bio sensor researches. Copyright © 2017 Elsevier B.V. All rights reserved.

  2. Quantitative 3D imaging of yeast by hard X-ray tomography.

    PubMed

    Zheng, Ting; Li, Wenjie; Guan, Yong; Song, Xiangxia; Xiong, Ying; Liu, Gang; Tian, Yangchao

    2012-05-01

    Full-field hard X-ray tomography could be used to obtain three-dimensional (3D) nanoscale structures of biological samples. The image of the fission yeast, Schizosaccharomyces pombe, was clearly visualized based on Zernike phase contrast imaging technique and heavy metal staining method at a spatial resolution better than 50 nm at the energy of 8 keV. The distributions and shapes of the organelles during the cell cycle were clearly visualized and two types of organelle were distinguished. The results for cells during various phases were compared and the ratios of organelle volume to cell volume can be analyzed quantitatively. It showed that the ratios remained constant between growth and division phase and increased strongly in stationary phase, following the shape and size of two types of organelles changes. Our results demonstrated that hard X-ray microscopy was a complementary method for imaging and revealing structural information for biological samples. Copyright © 2011 Wiley Periodicals, Inc.

  3. Quantitation of fixative-induced morphologic and antigenic variation in mouse and human breast cancers

    PubMed Central

    Cardiff, Robert D; Hubbard, Neil E; Engelberg, Jesse A; Munn, Robert J; Miller, Claramae H; Walls, Judith E; Chen, Jane Q; Velásquez-García, Héctor A; Galvez, Jose J; Bell, Katie J; Beckett, Laurel A; Li, Yue-Ju; Borowsky, Alexander D

    2013-01-01

    Quantitative Image Analysis (QIA) of digitized whole slide images for morphometric parameters and immunohistochemistry of breast cancer antigens was used to evaluate the technical reproducibility, biological variability, and intratumoral heterogeneity in three transplantable mouse mammary tumor models of human breast cancer. The relative preservation of structure and immunogenicity of the three mouse models and three human breast cancers was also compared when fixed with representatives of four distinct classes of fixatives. The three mouse mammary tumor cell models were an ER + /PR + model (SSM2), a Her2 + model (NDL), and a triple negative model (MET1). The four breast cancer antigens were ER, PR, Her2, and Ki67. The fixatives included examples of (1) strong cross-linkers, (2) weak cross-linkers, (3) coagulants, and (4) combination fixatives. Each parameter was quantitatively analyzed using modified Aperio Technologies ImageScope algorithms. Careful pre-analytical adjustments to the algorithms were required to provide accurate results. The QIA permitted rigorous statistical analysis of results and grading by rank order. The analyses suggested excellent technical reproducibility and confirmed biological heterogeneity within each tumor. The strong cross-linker fixatives, such as formalin, consistently ranked higher than weak cross-linker, coagulant and combination fixatives in both the morphometric and immunohistochemical parameters. PMID:23399853

  4. Pentobarbital quantitation using EMIT serum barbiturate assay reagents: application to monitoring of high-dose pentobarbital therapy.

    PubMed

    Pape, B E; Cary, P L; Clay, L C; Godolphin, W

    1983-01-01

    Pentobarbital serum concentrations associated with a high-dose therapeutic regimen were determined using EMIT immunoassay reagents. Replicate analyses of serum controls resulted in a within-assay coefficient of variation of 5.0% and a between-assay coefficient of variation of 10%. Regression analysis of 44 serum samples analyzed by this technique (y) and a reference procedure (x) were y = 0.98x + 3.6 (r = 0.98; x = ultraviolet spectroscopy) and y = 1.04x + 2.4 (r = 0.96; x = high-performance liquid chromatography). Clinical evaluation of the results indicates the immunoassay is sufficiently sensitive and selective for pentobarbital to allow accurate quantitation within the therapeutic range associated with high-dose therapy.

  5. A Quantitative Real-Time PCR-Based Strategy for Molecular Evaluation of Nicotine Conversion in Burley Tobacco.

    PubMed

    Sun, Bo; Xue, Sheng-Ling; Zhang, Fen; Luo, Zhao-Peng; Wu, Ming-Zhu; Chen, Qing; Tang, Hao-Ru; Lin, Fu-Cheng; Yang, Jun

    2015-11-17

    Nornicotine production in Nicotiana tabacum is undesirable because it is the precursor of the carcinogen N'-nitrosonornicotine. In some individual burley tobacco plants, a large proportion of the nicotine can be converted to nornicotine, and this process of nicotine conversion is mediated primarily by enzymatic N-demethylation of nicotine which is controlled mainly by CYP82E4. Here we report a novel strategy based on quantitative real-time polymerase chain reaction (qPCR) method, which analyzed the ratio of nicotine conversion through examining the transcript level of CYP82E4 in burley leaves and do not need ethylene induction before detected. The assay was linear in a range from 1 × 10¹ to 1 × 10⁵ copies/mL of serially diluted standards, and also showed high specificity and reproducibility (93%-99%). To assess its applicability, 55 plants of burley cultivar Ky8959 at leaf maturing stage were analyzed, and the results were in accordance with those from gas chromatograph-mass spectrometry (GC-MS) method. Moreover, a linear correlation existed between conversion level and CYP82E4 transcript abundance. Taken together, the quantitative real-time PCR assay is standardized, rapid and reproducible for estimation of nicotine conversion level in vivo, which is expected to shed new light on monitoring of burley tobacco converter.

  6. Statistical differences between relative quantitative molecular fingerprints from microbial communities.

    PubMed

    Portillo, M C; Gonzalez, J M

    2008-08-01

    Molecular fingerprints of microbial communities are a common method for the analysis and comparison of environmental samples. The significance of differences between microbial community fingerprints was analyzed considering the presence of different phylotypes and their relative abundance. A method is proposed by simulating coverage of the analyzed communities as a function of sampling size applying a Cramér-von Mises statistic. Comparisons were performed by a Monte Carlo testing procedure. As an example, this procedure was used to compare several sediment samples from freshwater ponds using a relative quantitative PCR-DGGE profiling technique. The method was able to discriminate among different samples based on their molecular fingerprints, and confirmed the lack of differences between aliquots from a single sample.

  7. Small- and Large-Effect Quantitative Trait Locus Interactions Underlie Variation in Yeast Sporulation Efficiency

    PubMed Central

    Lorenz, Kim; Cohen, Barak A.

    2012-01-01

    Quantitative trait loci (QTL) with small effects on phenotypic variation can be difficult to detect and analyze. Because of this a large fraction of the genetic architecture of many complex traits is not well understood. Here we use sporulation efficiency in Saccharomyces cerevisiae as a model complex trait to identify and study small-effect QTL. In crosses where the large-effect quantitative trait nucleotides (QTN) have been genetically fixed we identify small-effect QTL that explain approximately half of the remaining variation not explained by the major effects. We find that small-effect QTL are often physically linked to large-effect QTL and that there are extensive genetic interactions between small- and large-effect QTL. A more complete understanding of quantitative traits will require a better understanding of the numbers, effect sizes, and genetic interactions of small-effect QTL. PMID:22942125

  8. Effect of eight solvents on ethanol analysis by Dräger 7110 Evidential breath analyzer.

    PubMed

    Laakso, Olli; Pennanen, Teemu; Himberg, Kimmo; Kuitunen, Tapio; Himberg, Jaakko-Juhani

    2004-09-01

    The Dräger 7110 MK III FIN Evidential breath analyzer is classified as a quantitative analyzer capable to provide sufficient evidence for establishing legal intoxication. The purpose of this study was to evaluate ethanol specificity of this instrument in the presence of other solvents. Effects of eight possible interfering compounds on ethanol analysis were determined in a procedure simulating a human breathing. Most of the compounds studied had either a negligible effect on ethanol analysis (acetone, methyl ethyl ketone, and methyl isobutyl ketone) or were detected in very low concentrations before influencing ethanol readings (methanol, ethyl acetate, and diethyl ether). However, 1-propanol and 2-propanol increased the ethanol readings significantly. Thus, Dräger ethanol readings should be interpreted carefully in the presence of propanol.

  9. A New Algorithm Using Cross-Assignment for Label-Free Quantitation with LC/LTQ-FT MS

    PubMed Central

    Andreev, Victor P.; Li, Lingyun; Cao, Lei; Gu, Ye; Rejtar, Tomas; Wu, Shiaw-Lin; Karger, Barry L.

    2008-01-01

    A new algorithm is described for label-free quantitation of relative protein abundances across multiple complex proteomic samples. Q-MEND is based on the denoising and peak picking algorithm, MEND, previously developed in our laboratory. Q-MEND takes advantage of the high resolution and mass accuracy of the hybrid LTQFT MS mass spectrometer (or other high resolution mass spectrometers, such as a Q-TOF MS). The strategy, termed “cross-assignment”, is introduced to increase substantially the number of quantitated proteins. In this approach, all MS/MS identifications for the set of analyzed samples are combined into a master ID list, and then each LC/MS run is searched for the features that can be assigned to a specific identification from that master list. The reliability of quantitation is enhanced by quantitating separately all peptide charge states, along with a scoring procedure to filter out less reliable peptide abundance measurements. The effectiveness of Q-MEND is illustrated in the relative quantitative analysis of E.coli samples spiked with known amounts of non-E.coli protein digests. A mean quantitation accuracy of 7% and mean precision of 15% is demonstrated. Q-MEND can perform relative quantitation of a set of LC/MS datasets without manual intervention and can generate files compatible with the Guidelines for Proteomic Data Publication. PMID:17441747

  10. A new algorithm using cross-assignment for label-free quantitation with LC-LTQ-FT MS.

    PubMed

    Andreev, Victor P; Li, Lingyun; Cao, Lei; Gu, Ye; Rejtar, Tomas; Wu, Shiaw-Lin; Karger, Barry L

    2007-06-01

    A new algorithm is described for label-free quantitation of relative protein abundances across multiple complex proteomic samples. Q-MEND is based on the denoising and peak picking algorithm, MEND, previously developed in our laboratory. Q-MEND takes advantage of the high resolution and mass accuracy of the hybrid LTQ-FT MS mass spectrometer (or other high-resolution mass spectrometers, such as a Q-TOF MS). The strategy, termed "cross-assignment", is introduced to increase substantially the number of quantitated proteins. In this approach, all MS/MS identifications for the set of analyzed samples are combined into a master ID list, and then each LC-MS run is searched for the features that can be assigned to a specific identification from that master list. The reliability of quantitation is enhanced by quantitating separately all peptide charge states, along with a scoring procedure to filter out less reliable peptide abundance measurements. The effectiveness of Q-MEND is illustrated in the relative quantitative analysis of Escherichia coli samples spiked with known amounts of non-E. coli protein digests. A mean quantitation accuracy of 7% and mean precision of 15% is demonstrated. Q-MEND can perform relative quantitation of a set of LC-MS data sets without manual intervention and can generate files compatible with the Guidelines for Proteomic Data Publication.

  11. Safety of laboratory analyzers for infection testing - results of the market surveillance by the BfArM until end 2007

    PubMed Central

    2009-01-01

    control (8). Manufacturers issued corrective measures in 66 cases (90.4%) from which 49 and 17 were related to laboratory analyzers and their consumables, respectively. Based on the underlying root causes of product failures these were predominantly customer information (48), recalls (40), software-updates (30) and design changes (9) in the product group of laboratory analyzers as well as customer information (16), recalls (12) and modifications of production and quality management (11) in the group of consumables. The results and experiences obtained since 1999 suggest that the system for post marketing surveillance of IVD is an established tool to ensure product safety, even though the current system can be further enhanced. PMID:20156760

  12. Quantitative CT characterization of pediatric lung development using routine clinical imaging

    PubMed Central

    Stein, Jill M.; Walkup, Laura L.; Brody, Alan S.; Fleck, Robert J.

    2016-01-01

    Background The use of quantitative CT analysis in children is limited by lack of normal values of lung parenchymal attenuation. These characteristics are important because normal lung development yields significant parenchymal attenuation changes as children age. Objective To perform quantitative characterization of normal pediatric lung parenchymal X-ray CT attenuation under routine clinical conditions in order to establish a baseline comparison to that seen in pathological lung conditions. Materials and methods We conducted a retrospective query of normal CT chest examinations in children ages 0–7 years from 2004 to 2014 using standard clinical protocol. During these examinations semi-automated lung parenchymal segmentation was performed to measure lung volume and mean lung attenuation. Results We analyzed 42 CT examinations in 39 children, ages 3 days to 83 months (mean ± standard deviation [SD] = 42±27 months). Lung volume ranged 0.10–1.72 liters (L). Mean lung attenuation was much higher in children younger than 12 months, with values as high as −380 Hounsfield units (HU) in neonates (lung volume 0.10 L). Lung volume decreased to approximately −650 HU by age 2 years (lung volume 0.47 L), with subsequently slower exponential decrease toward a relatively constant value of −860 HU as age and lung volume increased. Conclusion Normal lung parenchymal X-ray CT attenuation decreases with increasing lung volume and age; lung attenuation decreases rapidly in the first 2 years of age and more slowly thereafter. This change in normal lung attenuation should be taken into account as quantitative CT methods are translated to pediatric pulmonary imaging. PMID:27576458

  13. Analyzing cross-college course enrollments via contextual graph mining

    PubMed Central

    Liu, Xiaozhong; Chen, Yan

    2017-01-01

    The ability to predict what courses a student may enroll in the coming semester plays a pivotal role in the allocation of learning resources, which is a hot topic in the domain of educational data mining. In this study, we propose an innovative approach to characterize students’ cross-college course enrollments by leveraging a novel contextual graph. Specifically, different kinds of variables, such as students, courses, colleges and diplomas, as well as various types of variable relations, are utilized to depict the context of each variable, and then a representation learning algorithm node2vec is applied to extracting sophisticated graph-based features for the enrollment analysis. In this manner, the relations between any pair of variables can be measured quantitatively, which enables the variable type to transform from nominal to ratio. These graph-based features are examined by the random forest algorithm, and experiments on 24,663 students, 1,674 courses and 417,590 enrollment records demonstrate that the contextual graph can successfully improve analyzing the cross-college course enrollments, where three of the graph-based features have significantly stronger impacts on prediction accuracy than the others. Besides, the empirical results also indicate that the student’s course preference is the most important factor in predicting future course enrollments, which is consistent to the previous studies that acknowledge the course interest is a key point for course recommendations. PMID:29186171

  14. Analyzing cross-college course enrollments via contextual graph mining.

    PubMed

    Wang, Yongzhen; Liu, Xiaozhong; Chen, Yan

    2017-01-01

    The ability to predict what courses a student may enroll in the coming semester plays a pivotal role in the allocation of learning resources, which is a hot topic in the domain of educational data mining. In this study, we propose an innovative approach to characterize students' cross-college course enrollments by leveraging a novel contextual graph. Specifically, different kinds of variables, such as students, courses, colleges and diplomas, as well as various types of variable relations, are utilized to depict the context of each variable, and then a representation learning algorithm node2vec is applied to extracting sophisticated graph-based features for the enrollment analysis. In this manner, the relations between any pair of variables can be measured quantitatively, which enables the variable type to transform from nominal to ratio. These graph-based features are examined by the random forest algorithm, and experiments on 24,663 students, 1,674 courses and 417,590 enrollment records demonstrate that the contextual graph can successfully improve analyzing the cross-college course enrollments, where three of the graph-based features have significantly stronger impacts on prediction accuracy than the others. Besides, the empirical results also indicate that the student's course preference is the most important factor in predicting future course enrollments, which is consistent to the previous studies that acknowledge the course interest is a key point for course recommendations.

  15. Validated semiquantitative/quantitative screening of 51 drugs in whole blood as silylated derivatives by gas chromatography-selected ion monitoring mass spectrometry and gas chromatography electron capture detection.

    PubMed

    Gunnar, Teemu; Mykkänen, Sirpa; Ariniemi, Kari; Lillsunde, Pirjo

    2004-07-05

    A comprehensively validated procedure is presented for simultaneous semiquantitative/quantitative screening of 51 drugs of abuse or drugs potentially hazardous for traffic safety in serum, plasma or whole blood. Benzodiazepines (12), cannabinoids (3), opioids (8), cocaine, antidepressants (13), antipsychotics (5) and antiepileptics (2) as well as zolpidem, zaleplon, zopiclone, meprobamate, carisoprodol, tizanidine and orphenadrine and internal standard flurazepam, were isolated by high-yield liquid-liquid extraction (LLE). The dried extracts were derivatized by two-step silylation and analyzed by the combination of two different gas chromatographic (GC) separations with both electron capture detection (ECD) and mass spectrometry (MS) operating in a selected ion-monitoring (SIM) mode. Quantitative or semiquantitative results were obtained for each substance based on four-point calibration. In the validation tests, accuracy, reproducibility, linearity, limit of detection (LOD) and limit of quantitation (LOQ), selectivity, as well as extraction efficiency and stability of standard stock solutions were tested, and derivatization was optimized in detail. Intra- and inter-day precisions were within 2.5-21.8 and 6.0-22.5%, and square of correlation coefficients of linearity ranged from 0.9896 to 0.9999. The limit of quantitation (LOQ) varied from 2 to 2000 ng/ml due to a variety of the relevant concentrations of the analyzed substances in blood. The method is feasible for highly sensitive, reliable and possibly routinely performed clinical and forensic toxicological analyses.

  16. Quantitative analysis of gallstones using laser-induced breakdown spectroscopy

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Singh, Vivek K.; Singh, Vinita; Rai, Awadhesh K.

    2008-11-01

    The utility of laser-induced breakdown spectroscopy (LIBS) for categorizing different types of gallbladder stone has been demonstrated by analyzing their major and minor constituents. LIBS spectra of three types of gallstone have been recorded in the 200-900 nm spectral region. Calcium is found to be the major element in all types of gallbladder stone. The spectrophotometric method has been used to classify the stones. A calibration-free LIBS method has been used for the quantitative analysis of metal elements, and the results have been compared with those obtained from inductively coupled plasma atomic emission spectroscopy (ICP-AES) measurements. The single-shot LIBS spectramore » from different points on the cross section (in steps of 0.5 mm from one end to the other) of gallstones have also been recorded to study the variation of constituents from the center to the surface. The presence of different metal elements and their possible role in gallstone formation is discussed.« less

  17. Quantitative Image Analysis Techniques with High-Speed Schlieren Photography

    NASA Technical Reports Server (NTRS)

    Pollard, Victoria J.; Herron, Andrew J.

    2017-01-01

    Optical flow visualization techniques such as schlieren and shadowgraph photography are essential to understanding fluid flow when interpreting acquired wind tunnel test data. Output of the standard implementations of these visualization techniques in test facilities are often limited only to qualitative interpretation of the resulting images. Although various quantitative optical techniques have been developed, these techniques often require special equipment or are focused on obtaining very precise and accurate data about the visualized flow. These systems are not practical in small, production wind tunnel test facilities. However, high-speed photography capability has become a common upgrade to many test facilities in order to better capture images of unsteady flow phenomena such as oscillating shocks and flow separation. This paper describes novel techniques utilized by the authors to analyze captured high-speed schlieren and shadowgraph imagery from wind tunnel testing for quantification of observed unsteady flow frequency content. Such techniques have applications in parametric geometry studies and in small facilities where more specialized equipment may not be available.

  18. A First Laboratory Utilizing NMR for Undergraduate Education: Characterization of Edible Fats and Oils by Quantitative [superscript 13]C NMR

    ERIC Educational Resources Information Center

    Fry, Charles G.; Hofstetter, Heike; Bowman, Matthew D.

    2017-01-01

    Quantitative [superscript 13]C NMR provides a straightforward method of analyzing edible oils in undergraduate chemistry laboratories. [superscript 13]C spectra are relatively easy to understand, and are much simpler to analyze and workup than corresponding [superscript 1]H spectra. Average chain length, degree of saturation, and average…

  19. Determining quantitative immunophenotypes and evaluating their implications

    NASA Astrophysics Data System (ADS)

    Redelman, Douglas; Hudig, Dorothy; Berner, Dave; Castell, Linda M.; Roberts, Don; Ensign, Wayne

    2002-05-01

    Quantitative immunophenotypes varied widely among > 100 healthy young males but were maintained at characteristic levels within individuals. The initial results (SPIE Proceedings 4260:226) that examined cell numbers and the quantitative expression of adhesion and lineage-specific molecules, e.g., CD2 and CD14, have now been confirmed and extended to include the quantitative expression of inducible molecules such as HLA-DR and perforin (Pf). Some properties, such as the ratio of T helper (Th) to T cytotoxic/suppressor (Tc/s) cells, are known to be genetically determined. Other properties, e.g., the T:B cell ratio, the amount of CD19 per B cell, etc., behaved similarly and may also be inherited traits. Since some patterns observed in these healthy individuals resembled those found in pathological situations we tested whether the patterns could be associated with the occurrence of disease. The current studies shows that there were associations between quantitative immunophenotypes and the subsequent incidence and severity of disease. For example, individuals with characteristically low levels of HLA-DR or B cells or reduced numbers of Pf+ Tc/s cells had more frequent and/or more severe upper respiratory infections. Quantitative immunophenotypes will be more widely measured if the necessary standards are available and if appropriate procedures are made more accessible.

  20. Qualitative versus quantitative methods in psychiatric research.

    PubMed

    Razafsha, Mahdi; Behforuzi, Hura; Azari, Hassan; Zhang, Zhiqun; Wang, Kevin K; Kobeissy, Firas H; Gold, Mark S

    2012-01-01

    Qualitative studies are gaining their credibility after a period of being misinterpreted as "not being quantitative." Qualitative method is a broad umbrella term for research methodologies that describe and explain individuals' experiences, behaviors, interactions, and social contexts. In-depth interview, focus groups, and participant observation are among the qualitative methods of inquiry commonly used in psychiatry. Researchers measure the frequency of occurring events using quantitative methods; however, qualitative methods provide a broader understanding and a more thorough reasoning behind the event. Hence, it is considered to be of special importance in psychiatry. Besides hypothesis generation in earlier phases of the research, qualitative methods can be employed in questionnaire design, diagnostic criteria establishment, feasibility studies, as well as studies of attitude and beliefs. Animal models are another area that qualitative methods can be employed, especially when naturalistic observation of animal behavior is important. However, since qualitative results can be researcher's own view, they need to be statistically confirmed, quantitative methods. The tendency to combine both qualitative and quantitative methods as complementary methods has emerged over recent years. By applying both methods of research, scientists can take advantage of interpretative characteristics of qualitative methods as well as experimental dimensions of quantitative methods.

  1. [Morphology and diagnostics of mechanisms of chest bone fractures and their use in analyzing forensic medicine results].

    PubMed

    Sobol, Julia; Kordel, Krzysztof; Kołowski, Janusz; Kis-Wojciechowska, Margit; Przybylski, Zygmunt

    2007-01-01

    The study presents the analysis of 343 available protocols of autopsy results. In the reviewed material, the authors noted that of 343 autopsies performed in 2005, in 92 cases, rib fractures were present. The study reviews the articles on the morphology and determination of the mechanism of rib fracturing. The authors describe the majority of factors that influence the type of fracture, as well as the current views on the possibility of applying the knowledge of morphology and mechanisms of rib fracturing in opinionating in traffic accidents, injuries inflicted to victims of assault and battery, in interpreting autopsy findings in victims of falls from high altitude or crushing by heavy objects, as well as in differentiating between primary and secondary injuries, and also identifying fractures occurring during resuscitation. Fractures in children are presented separately. The authors also analyze the issue of establishing the sequence of fractures.

  2. Quantitative imaging biomarkers: a review of statistical methods for technical performance assessment.

    PubMed

    Raunig, David L; McShane, Lisa M; Pennello, Gene; Gatsonis, Constantine; Carson, Paul L; Voyvodic, James T; Wahl, Richard L; Kurland, Brenda F; Schwarz, Adam J; Gönen, Mithat; Zahlmann, Gudrun; Kondratovich, Marina V; O'Donnell, Kevin; Petrick, Nicholas; Cole, Patricia E; Garra, Brian; Sullivan, Daniel C

    2015-02-01

    Technological developments and greater rigor in the quantitative measurement of biological features in medical images have given rise to an increased interest in using quantitative imaging biomarkers to measure changes in these features. Critical to the performance of a quantitative imaging biomarker in preclinical or clinical settings are three primary metrology areas of interest: measurement linearity and bias, repeatability, and the ability to consistently reproduce equivalent results when conditions change, as would be expected in any clinical trial. Unfortunately, performance studies to date differ greatly in designs, analysis method, and metrics used to assess a quantitative imaging biomarker for clinical use. It is therefore difficult or not possible to integrate results from different studies or to use reported results to design studies. The Radiological Society of North America and the Quantitative Imaging Biomarker Alliance with technical, radiological, and statistical experts developed a set of technical performance analysis methods, metrics, and study designs that provide terminology, metrics, and methods consistent with widely accepted metrological standards. This document provides a consistent framework for the conduct and evaluation of quantitative imaging biomarker performance studies so that results from multiple studies can be compared, contrasted, or combined. © The Author(s) 2014 Reprints and permissions: sagepub.co.uk/journalsPermissions.nav.

  3. Does contraceptive treatment in wildlife result in side effects? A review of quantitative and anecdotal evidence.

    PubMed

    Gray, Meeghan E; Cameron, Elissa Z

    2010-01-01

    The efficacy of contraceptive treatments has been extensively tested, and several formulations are effective at reducing fertility in a range of species. However, these formulations should minimally impact the behavior of individuals and populations before a contraceptive is used for population manipulation, but these effects have received less attention. Potential side effects have been identified theoretically and we reviewed published studies that have investigated side effects on behavior and physiology of individuals or population-level effects, which provided mixed results. Physiological side effects were most prevalent. Most studies reported a lack of secondary effects, but were usually based on qualitative data or anecdotes. A meta-analysis on quantitative studies of side effects showed that secondary effects consistently occur across all categories and all contraceptive types. This contrasts with the qualitative studies, suggesting that anecdotal reports are insufficient to investigate secondary impacts of contraceptive treatment. We conclude that more research is needed to address fundamental questions about secondary effects of contraceptive treatment and experiments are fundamental to conclusions. In addition, researchers are missing a vital opportunity to use contraceptives as an experimental tool to test the influence of reproduction, sex and fertility on the behavior of wildlife species.

  4. Quantitative Literacy: Geosciences and Beyond

    NASA Astrophysics Data System (ADS)

    Richardson, R. M.; McCallum, W. G.

    2002-12-01

    Quantitative literacy seems like such a natural for the geosciences, right? The field has gone from its origin as a largely descriptive discipline to one where it is hard to imagine failing to bring a full range of mathematical tools to the solution of geological problems. Although there are many definitions of quantitative literacy, we have proposed one that is analogous to the UNESCO definition of conventional literacy: "A quantitatively literate person is one who, with understanding, can both read and represent quantitative information arising in his or her everyday life." Central to this definition is the concept that a curriculum for quantitative literacy must go beyond the basic ability to "read and write" mathematics and develop conceptual understanding. It is also critical that a curriculum for quantitative literacy be engaged with a context, be it everyday life, humanities, geoscience or other sciences, business, engineering, or technology. Thus, our definition works both within and outside the sciences. What role do geoscience faculty have in helping students become quantitatively literate? Is it our role, or that of the mathematicians? How does quantitative literacy vary between different scientific and engineering fields? Or between science and nonscience fields? We will argue that successful quantitative literacy curricula must be an across-the-curriculum responsibility. We will share examples of how quantitative literacy can be developed within a geoscience curriculum, beginning with introductory classes for nonmajors (using the Mauna Loa CO2 data set) through graduate courses in inverse theory (using singular value decomposition). We will highlight six approaches to across-the curriculum efforts from national models: collaboration between mathematics and other faculty; gateway testing; intensive instructional support; workshops for nonmathematics faculty; quantitative reasoning requirement; and individual initiative by nonmathematics faculty.

  5. UK audit of quantitative thyroid uptake imaging.

    PubMed

    Taylor, Jonathan C; Murray, Anthony W; Hall, David O; Barnfield, Mark C; O'Shaugnessy, Emma R; Carson, Kathryn J; Cullis, James; Towey, David J; Kenny, Bob

    2017-07-01

    A national audit of quantitative thyroid uptake imaging was conducted by the Nuclear Medicine Software Quality Group of the Institute of Physics and Engineering in Medicine in 2014/2015. The aims of the audit were to measure and assess the variability in thyroid uptake results across the UK and to compare local protocols with British Nuclear Medicine Society (BNMS) guidelines. Participants were invited through a combination of emails on a public mailbase and targeted invitations from regional co-ordinators. All participants were given a set of images from which to calculate quantitative measures and a spreadsheet for capturing results. The image data consisted of two sets of 10 anterior thyroid images, half of which were acquired after administration of Tc-pertechnetate and the other half after administration of I-iodide. Images of the administration syringes or thyroid phantoms were also included. In total, 54 participants responded to the audit. The median number of scans conducted per year was 50. A majority of centres had at least one noncompliance in comparison with BNMS guidelines. Of most concern was the widespread lack of injection-site imaging. Quantitative results showed that both intersite and intrasite variability were low for the Tc dataset. The coefficient of quartile deviation was between 0.03 and 0.13 for measurements of overall percentage uptake. Although the number of returns for the I dataset was smaller, the level of variability between participants was greater (the coefficient of quartile deviation was between 0.17 and 0.25). A UK-wide audit showed that thyroid uptake imaging is still a common test in the UK. It was found that most centres do not adhere to all aspects of the BNMS practice guidelines but that quantitative results are reasonably consistent for Tc-based scans.

  6. A Quantitative Analysis of the Increase in Public School Segregation in Delaware: 1989-2006

    ERIC Educational Resources Information Center

    Glenn, William J.

    2011-01-01

    This study analyzes the increase in school segregation in Delaware from a quantitative perspective. The article tests the hypothesis that the declaration of unitary status that released the Wilmington area school districts from their desegregation order caused the increase in segregation. The research reveals that the declaration of unitary status…

  7. Stress Analyzer

    NASA Technical Reports Server (NTRS)

    1990-01-01

    SPATE 900 Dynamic Stress Analyzer is an acronym for Stress Pattern Analysis by Thermal Emission. It detects stress-induced temperature changes in a structure and indicates the degree of stress. Ometron, Inc.'s SPATE 9000 consists of a scan unit and a data display. The scan unit contains an infrared channel focused on the test structure to collect thermal radiation, and a visual channel used to set up the scan area and interrogate the stress display. Stress data is produced by detecting minute temperature changes, down to one-thousandth of a degree Centigrade, resulting from the application to the structure of dynamic loading. The electronic data processing system correlates the temperature changes with a reference signal to determine stress level.

  8. Optimization of Statistical Methods Impact on Quantitative Proteomics Data.

    PubMed

    Pursiheimo, Anna; Vehmas, Anni P; Afzal, Saira; Suomi, Tomi; Chand, Thaman; Strauss, Leena; Poutanen, Matti; Rokka, Anne; Corthals, Garry L; Elo, Laura L

    2015-10-02

    As tools for quantitative label-free mass spectrometry (MS) rapidly develop, a consensus about the best practices is not apparent. In the work described here we compared popular statistical methods for detecting differential protein expression from quantitative MS data using both controlled experiments with known quantitative differences for specific proteins used as standards as well as "real" experiments where differences in protein abundance are not known a priori. Our results suggest that data-driven reproducibility-optimization can consistently produce reliable differential expression rankings for label-free proteome tools and are straightforward in their application.

  9. Quantitative, spectrally-resolved intraoperative fluorescence imaging

    PubMed Central

    Valdés, Pablo A.; Leblond, Frederic; Jacobs, Valerie L.; Wilson, Brian C.; Paulsen, Keith D.; Roberts, David W.

    2012-01-01

    Intraoperative visual fluorescence imaging (vFI) has emerged as a promising aid to surgical guidance, but does not fully exploit the potential of the fluorescent agents that are currently available. Here, we introduce a quantitative fluorescence imaging (qFI) approach that converts spectrally-resolved data into images of absolute fluorophore concentration pixel-by-pixel across the surgical field of view (FOV). The resulting estimates are linear, accurate, and precise relative to true values, and spectral decomposition of multiple fluorophores is also achieved. Experiments with protoporphyrin IX in a glioma rodent model demonstrate in vivo quantitative and spectrally-resolved fluorescence imaging of infiltrating tumor margins for the first time. Moreover, we present images from human surgery which detect residual tumor not evident with state-of-the-art vFI. The wide-field qFI technique has broad implications for intraoperative surgical guidance because it provides near real-time quantitative assessment of multiple fluorescent biomarkers across the operative field. PMID:23152935

  10. Quantitative analysis for peripheral vascularity assessment based on clinical photoacoustic and ultrasound images

    NASA Astrophysics Data System (ADS)

    Murakoshi, Dai; Hirota, Kazuhiro; Ishii, Hiroyasu; Hashimoto, Atsushi; Ebata, Tetsurou; Irisawa, Kaku; Wada, Takatsugu; Hayakawa, Toshiro; Itoh, Kenji; Ishihara, Miya

    2018-02-01

    Photoacoustic (PA) imaging technology is expected to be applied to clinical assessment for peripheral vascularity. We started a clinical evaluation with the prototype PA imaging system we recently developed. Prototype PA imaging system was composed with in-house Q-switched Alexandrite laser system which emits short-pulsed laser with 750 nm wavelength, handheld ultrasound transducer where illumination optics were integrated and signal processing for PA image reconstruction implemented in the clinical ultrasound (US) system. For the purpose of quantitative assessment of PA images, an image analyzing function has been developed and applied to clinical PA images. In this analyzing function, vascularity derived from PA signal intensity ranged for prescribed threshold was defined as a numerical index of vessel fulfillment and calculated for the prescribed region of interest (ROI). Skin surface was automatically detected by utilizing B-mode image acquired simultaneously with PA image. Skinsurface position is utilized to place the ROI objectively while avoiding unwanted signals such as artifacts which were imposed due to melanin pigment in the epidermal layer which absorbs laser emission and generates strong PA signals. Multiple images were available to support the scanned image set for 3D viewing. PA images for several fingers of patients with systemic sclerosis (SSc) were quantitatively assessed. Since the artifact region is trimmed off in PA images, the visibility of vessels with rather low PA signal intensity on the 3D projection image was enhanced and the reliability of the quantitative analysis was improved.

  11. Collection of quantitative chemical release field data.

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Demirgian, J.; Macha, S.; Loyola Univ.

    1999-01-01

    Detection and quantitation of chemicals in the environment requires Fourier-transform infrared (FTIR) instruments that are properly calibrated and tested. This calibration and testing requires field testing using matrices that are representative of actual instrument use conditions. Three methods commonly used for developing calibration files and training sets in the field are a closed optical cell or chamber, a large-scale chemical release, and a small-scale chemical release. There is no best method. The advantages and limitations of each method should be considered in evaluating field results. Proper calibration characterizes the sensitivity of an instrument, its ability to detect a component inmore » different matrices, and the quantitative accuracy and precision of the results.« less

  12. Quantitative analysis of titanium concentration using calibration-free laser-induced breakdown spectroscopy (LIBS)

    NASA Astrophysics Data System (ADS)

    Zaitun; Prasetyo, S.; Suliyanti, M. M.; Isnaeni; Herbani, Y.

    2018-03-01

    Laser-induced breakdown spectroscopy (LIBS) can be used for quantitative and qualitative analysis. Calibration-free LIBS (CF-LIBS) is a method to quantitatively analyze concentration of elements in a sample in local thermodynamic equilibrium conditions without using available matrix-matched calibration. In this study, we apply CF-LIBS for quantitative analysis of Ti in TiO2 sample. TiO2 powder sample was mixed with polyvinyl alcohol and formed into pellets. An Nd:YAG pulsed laser at a wavelength of 1064 nm was focused onto the sample to generate plasma. The spectrum of plasma was recorded using spectrophotometer then compared to NIST spectral line to determine energy levels and other parameters. The value of plasma temperature obtained using Boltzmann plot is 8127.29 K and electron density from calculation is 2.49×1016 cm-3. Finally, the concentration of Ti in TiO2 sample from this study is 97% that is in proximity with the sample certificate.

  13. Meta-analysis of quantitative pleiotropic traits for next-generation sequencing with multivariate functional linear models

    PubMed Central

    Chiu, Chi-yang; Jung, Jeesun; Chen, Wei; Weeks, Daniel E; Ren, Haobo; Boehnke, Michael; Amos, Christopher I; Liu, Aiyi; Mills, James L; Ting Lee, Mei-ling; Xiong, Momiao; Fan, Ruzong

    2017-01-01

    To analyze next-generation sequencing data, multivariate functional linear models are developed for a meta-analysis of multiple studies to connect genetic variant data to multiple quantitative traits adjusting for covariates. The goal is to take the advantage of both meta-analysis and pleiotropic analysis in order to improve power and to carry out a unified association analysis of multiple studies and multiple traits of complex disorders. Three types of approximate F -distributions based on Pillai–Bartlett trace, Hotelling–Lawley trace, and Wilks's Lambda are introduced to test for association between multiple quantitative traits and multiple genetic variants. Simulation analysis is performed to evaluate false-positive rates and power of the proposed tests. The proposed methods are applied to analyze lipid traits in eight European cohorts. It is shown that it is more advantageous to perform multivariate analysis than univariate analysis in general, and it is more advantageous to perform meta-analysis of multiple studies instead of analyzing the individual studies separately. The proposed models require individual observations. The value of the current paper can be seen at least for two reasons: (a) the proposed methods can be applied to studies that have individual genotype data; (b) the proposed methods can be used as a criterion for future work that uses summary statistics to build test statistics to meta-analyze the data. PMID:28000696

  14. Quantitative analysis of diffusion tensor imaging (DTI) using statistical parametric mapping (SPM) for brain disorders

    NASA Astrophysics Data System (ADS)

    Lee, Jae-Seung; Im, In-Chul; Kang, Su-Man; Goo, Eun-Hoe; Kwak, Byung-Joon

    2013-07-01

    This study aimed to quantitatively analyze data from diffusion tensor imaging (DTI) using statistical parametric mapping (SPM) in patients with brain disorders and to assess its potential utility for analyzing brain function. DTI was obtained by performing 3.0-T magnetic resonance imaging for patients with Alzheimer's disease (AD) and vascular dementia (VD), and the data were analyzed using Matlab-based SPM software. The two-sample t-test was used for error analysis of the location of the activated pixels. We compared regions of white matter where the fractional anisotropy (FA) values were low and the apparent diffusion coefficients (ADCs) were increased. In the AD group, the FA values were low in the right superior temporal gyrus, right inferior temporal gyrus, right sub-lobar insula, and right occipital lingual gyrus whereas the ADCs were significantly increased in the right inferior frontal gyrus and right middle frontal gyrus. In the VD group, the FA values were low in the right superior temporal gyrus, right inferior temporal gyrus, right limbic cingulate gyrus, and right sub-lobar caudate tail whereas the ADCs were significantly increased in the left lateral globus pallidus and left medial globus pallidus. In conclusion by using DTI and SPM analysis, we were able to not only determine the structural state of the regions affected by brain disorders but also quantitatively analyze and assess brain function.

  15. Meta-analysis of quantitative pleiotropic traits for next-generation sequencing with multivariate functional linear models.

    PubMed

    Chiu, Chi-Yang; Jung, Jeesun; Chen, Wei; Weeks, Daniel E; Ren, Haobo; Boehnke, Michael; Amos, Christopher I; Liu, Aiyi; Mills, James L; Ting Lee, Mei-Ling; Xiong, Momiao; Fan, Ruzong

    2017-02-01

    To analyze next-generation sequencing data, multivariate functional linear models are developed for a meta-analysis of multiple studies to connect genetic variant data to multiple quantitative traits adjusting for covariates. The goal is to take the advantage of both meta-analysis and pleiotropic analysis in order to improve power and to carry out a unified association analysis of multiple studies and multiple traits of complex disorders. Three types of approximate F -distributions based on Pillai-Bartlett trace, Hotelling-Lawley trace, and Wilks's Lambda are introduced to test for association between multiple quantitative traits and multiple genetic variants. Simulation analysis is performed to evaluate false-positive rates and power of the proposed tests. The proposed methods are applied to analyze lipid traits in eight European cohorts. It is shown that it is more advantageous to perform multivariate analysis than univariate analysis in general, and it is more advantageous to perform meta-analysis of multiple studies instead of analyzing the individual studies separately. The proposed models require individual observations. The value of the current paper can be seen at least for two reasons: (a) the proposed methods can be applied to studies that have individual genotype data; (b) the proposed methods can be used as a criterion for future work that uses summary statistics to build test statistics to meta-analyze the data.

  16. Gas Analyzer

    NASA Technical Reports Server (NTRS)

    1989-01-01

    The M200 originated in the 1970's under an Ames Research Center/Stanford University contract to develop a small, lightweight gas analyzer for Viking Landers. Although the unit was not used on the spacecraft, it was further developed by The National Institute for Occupational Safety and Health (NIOSH). Three researchers from the project later formed Microsensor Technology, Inc. (MTI) to commercialize the analyzer. The original version (Micromonitor 500) was introduced in 1982, and the M200 in 1988. The M200, a more advanced version, features dual gas chromatograph which separate a gaseous mixture into components and measure concentrations of each gas. It is useful for monitoring gas leaks, chemical spills, etc. Many analyses are completed in less than 30 seconds, and a wide range of mixtures can be analyzed.

  17. Process Analyzer

    NASA Technical Reports Server (NTRS)

    1994-01-01

    The ChemScan UV-6100 is a spectrometry system originally developed by Biotronics Technologies, Inc. under a Small Business Innovation Research (SBIR) contract. It is marketed to the water and wastewater treatment industries, replacing "grab sampling" with on-line data collection. It analyzes the light absorbance characteristics of a water sample, simultaneously detects hundreds of individual wavelengths absorbed by chemical substances in a process solution, and quantifies the information. Spectral data is then processed by ChemScan analyzer and compared with calibration files in the system's memory in order to calculate concentrations of chemical substances that cause UV light absorbance in specific patterns. Monitored substances can be analyzed for quality and quantity. Applications include detection of a variety of substances, and the information provided enables an operator to control a process more efficiently.

  18. Design and analysis issues in quantitative proteomics studies.

    PubMed

    Karp, Natasha A; Lilley, Kathryn S

    2007-09-01

    Quantitative proteomics is the comparison of distinct proteomes which enables the identification of protein species which exhibit changes in expression or post-translational state in response to a given stimulus. Many different quantitative techniques are being utilized and generate large datasets. Independent of the technique used, these large datasets need robust data analysis to ensure valid conclusions are drawn from such studies. Approaches to address the problems that arise with large datasets are discussed to give insight into the types of statistical analyses of data appropriate for the various experimental strategies that can be employed by quantitative proteomic studies. This review also highlights the importance of employing a robust experimental design and highlights various issues surrounding the design of experiments. The concepts and examples discussed within will show how robust design and analysis will lead to confident results that will ensure quantitative proteomics delivers.

  19. A Quantitative Infrared Spectroscopy Experiment.

    ERIC Educational Resources Information Center

    Krahling, Mark D.; Eliason, Robert

    1985-01-01

    Although infrared spectroscopy is used primarily for qualitative identifications, it is possible to use it as a quantitative tool as well. The use of a standard curve to determine percent methanol in a 2,2,2-trifluoroethanol sample is described. Background information, experimental procedures, and results obtained are provided. (JN)

  20. Transient segregation behavior in Cd1-xZnxTe with low Zn content-A qualitative and quantitative analysis

    NASA Astrophysics Data System (ADS)

    Neubert, M.; Jurisch, M.

    2015-06-01

    The paper analyzes experimental compositional profiles in Vertical Bridgman (VB, VGF) grown (Cd,Zn)Te crystals, found in the literature. The origin of the observed axial ZnTe-distribution profiles is attributed to dendritic growth after initial nucleation from supercooled melts. The analysis was done by utilizing a boundary layer model providing a very good approximation of the experimental data. Besides the discussion of the qualitative results also a quantitative analysis of the fitted model parameters is presented as far as it is possible by the utilized model.

  1. A quantitative comparison of corrective and perfective maintenance

    NASA Technical Reports Server (NTRS)

    Henry, Joel; Cain, James

    1994-01-01

    This paper presents a quantitative comparison of corrective and perfective software maintenance activities. The comparison utilizes basic data collected throughout the maintenance process. The data collected are extensive and allow the impact of both types of maintenance to be quantitatively evaluated and compared. Basic statistical techniques test relationships between and among process and product data. The results show interesting similarities and important differences in both process and product characteristics.

  2. Intra-laboratory validation of chronic bee paralysis virus quantitation using an accredited standardised real-time quantitative RT-PCR method.

    PubMed

    Blanchard, Philippe; Regnault, Julie; Schurr, Frank; Dubois, Eric; Ribière, Magali

    2012-03-01

    Chronic bee paralysis virus (CBPV) is responsible for chronic bee paralysis, an infectious and contagious disease in adult honey bees (Apis mellifera L.). A real-time RT-PCR assay to quantitate the CBPV load is now available. To propose this assay as a reference method, it was characterised further in an intra-laboratory study during which the reliability and the repeatability of results and the performance of the assay were confirmed. The qPCR assay alone and the whole quantitation method (from sample RNA extraction to analysis) were both assessed following the ISO/IEC 17025 standard and the recent XP U47-600 standard issued by the French Standards Institute. The performance of the qPCR assay and of the overall CBPV quantitation method were validated over a 6 log range from 10(2) to 10(8) with a detection limit of 50 and 100 CBPV RNA copies, respectively, and the protocol of the real-time RT-qPCR assay for CBPV quantitation was approved by the French Accreditation Committee. Copyright © 2011 Elsevier B.V. All rights reserved.

  3. The comparison of automated urine analyzers with manual microscopic examination for urinalysis automated urine analyzers and manual urinalysis.

    PubMed

    İnce, Fatma Demet; Ellidağ, Hamit Yaşar; Koseoğlu, Mehmet; Şimşek, Neşe; Yalçın, Hülya; Zengin, Mustafa Osman

    2016-08-01

    Urinalysis is one of the most commonly performed tests in the clinical laboratory. However, manual microscopic sediment examination is labor-intensive, time-consuming, and lacks standardization in high-volume laboratories. In this study, the concordance of analyses between manual microscopic examination and two different automatic urine sediment analyzers has been evaluated. 209 urine samples were analyzed by the Iris iQ200 ELITE (İris Diagnostics, USA), Dirui FUS-200 (DIRUI Industrial Co., China) automatic urine sediment analyzers and by manual microscopic examination. The degree of concordance (Kappa coefficient) and the rates within the same grading were evaluated. For erythrocytes, leukocytes, epithelial cells, bacteria, crystals and yeasts, the degree of concordance between the two instruments was better than the degree of concordance between the manual microscopic method and the individual devices. There was no concordance between all methods for casts. The results from the automated analyzers for erythrocytes, leukocytes and epithelial cells were similar to the result of microscopic examination. However, in order to avoid any error or uncertainty, some images (particularly: dysmorphic cells, bacteria, yeasts, casts and crystals) have to be analyzed by manual microscopic examination by trained staff. Therefore, the software programs which are used in automatic urine sediment analysers need further development to recognize urinary shaped elements more accurately. Automated systems are important in terms of time saving and standardization.

  4. Evaluation of a rapid quantitative determination method of PSA concentration with gold immunochromatographic strips.

    PubMed

    Wu, Cheng-Ching; Lin, Hung-Yu; Wang, Chao-Ping; Lu, Li-Fen; Yu, Teng-Hung; Hung, Wei-Chin; Houng, Jer-Yiing; Chung, Fu-Mei; Lee, Yau-Jiunn; Hu, Jin-Jia

    2015-11-03

    Prostate cancer remains the most common cancer in men. Qualitative or semi-quantitative immunochromatographic measurements of prostate specific antigen (PSA) have been shown to be simple, noninvasive and feasible. The aim of this study was to evaluate an optimized gold immunochromatographic strip device for the detection of PSA, in which the results can be analysed using a Chromogenic Rapid Test Reader to quantitatively assess the test results. This reader measures the reflectance of the signal line via a charge-coupled device camera. For quantitative analysis, PSA concentration was computed via a calibration equation. Capillary blood samples from 305 men were evaluated, and two independent observers interpreted the test results after 12 min. Blood samples were also collected and tested with a conventional quantitative assay. Sensitivity, specificity, positive and negative predictive values, and accuracy of the PSA rapid quantitative test system were 100, 96.6, 89.5, 100, and 97.4 %, respectively. Reproducibility of the test was 99.2, and interobserver variation was 8 % with a false positive rate of 3.4 %. The correlation coefficient between the ordinary quantitative assay and the rapid quantitative test was 0.960. The PSA rapid quantitative test system provided results quickly and was easy to use, so that tests using this system can be easily performed at outpatient clinics or elsewhere. This system may also be useful for initial cancer screening and for point-of-care testing, because results can be obtained within 12 min and at a cost lower than that of conventional quantitative assays.

  5. Retinal status analysis method based on feature extraction and quantitative grading in OCT images.

    PubMed

    Fu, Dongmei; Tong, Hejun; Zheng, Shuang; Luo, Ling; Gao, Fulin; Minar, Jiri

    2016-07-22

    Optical coherence tomography (OCT) is widely used in ophthalmology for viewing the morphology of the retina, which is important for disease detection and assessing therapeutic effect. The diagnosis of retinal diseases is based primarily on the subjective analysis of OCT images by trained ophthalmologists. This paper describes an OCT images automatic analysis method for computer-aided disease diagnosis and it is a critical part of the eye fundus diagnosis. This study analyzed 300 OCT images acquired by Optovue Avanti RTVue XR (Optovue Corp., Fremont, CA). Firstly, the normal retinal reference model based on retinal boundaries was presented. Subsequently, two kinds of quantitative methods based on geometric features and morphological features were proposed. This paper put forward a retinal abnormal grading decision-making method which was used in actual analysis and evaluation of multiple OCT images. This paper showed detailed analysis process by four retinal OCT images with different abnormal degrees. The final grading results verified that the analysis method can distinguish abnormal severity and lesion regions. This paper presented the simulation of the 150 test images, where the results of analysis of retinal status showed that the sensitivity was 0.94 and specificity was 0.92.The proposed method can speed up diagnostic process and objectively evaluate the retinal status. This paper aims on studies of retinal status automatic analysis method based on feature extraction and quantitative grading in OCT images. The proposed method can obtain the parameters and the features that are associated with retinal morphology. Quantitative analysis and evaluation of these features are combined with reference model which can realize the target image abnormal judgment and provide a reference for disease diagnosis.

  6. A temperature-controlled photoelectrochemical cell for quantitative product analysis.

    PubMed

    Corson, Elizabeth R; Creel, Erin B; Kim, Youngsang; Urban, Jeffrey J; Kostecki, Robert; McCloskey, Bryan D

    2018-05-01

    In this study, we describe the design and operation of a temperature-controlled photoelectrochemical cell for analysis of gaseous and liquid products formed at an illuminated working electrode. This cell is specifically designed to quantitatively analyze photoelectrochemical processes that yield multiple gas and liquid products at low current densities and exhibit limiting reactant concentrations that prevent these processes from being studied in traditional single chamber electrolytic cells. The geometry of the cell presented in this paper enables front-illumination of the photoelectrode and maximizes the electrode surface area to electrolyte volume ratio to increase liquid product concentration and hence enhances ex situ spectroscopic sensitivity toward them. Gas is bubbled through the electrolyte in the working electrode chamber during operation to maintain a saturated reactant concentration and to continuously mix the electrolyte. Gaseous products are detected by an in-line gas chromatograph, and liquid products are analyzed ex situ by nuclear magnetic resonance. Cell performance was validated by examining carbon dioxide reduction on a silver foil electrode, showing comparable results both to those reported in the literature and identical experiments performed in a standard parallel-electrode electrochemical cell. To demonstrate a photoelectrochemical application of the cell, CO 2 reduction experiments were carried out on a plasmonic nanostructured silver photocathode and showed different product distributions under dark and illuminated conditions.

  7. A temperature-controlled photoelectrochemical cell for quantitative product analysis

    NASA Astrophysics Data System (ADS)

    Corson, Elizabeth R.; Creel, Erin B.; Kim, Youngsang; Urban, Jeffrey J.; Kostecki, Robert; McCloskey, Bryan D.

    2018-05-01

    In this study, we describe the design and operation of a temperature-controlled photoelectrochemical cell for analysis of gaseous and liquid products formed at an illuminated working electrode. This cell is specifically designed to quantitatively analyze photoelectrochemical processes that yield multiple gas and liquid products at low current densities and exhibit limiting reactant concentrations that prevent these processes from being studied in traditional single chamber electrolytic cells. The geometry of the cell presented in this paper enables front-illumination of the photoelectrode and maximizes the electrode surface area to electrolyte volume ratio to increase liquid product concentration and hence enhances ex situ spectroscopic sensitivity toward them. Gas is bubbled through the electrolyte in the working electrode chamber during operation to maintain a saturated reactant concentration and to continuously mix the electrolyte. Gaseous products are detected by an in-line gas chromatograph, and liquid products are analyzed ex situ by nuclear magnetic resonance. Cell performance was validated by examining carbon dioxide reduction on a silver foil electrode, showing comparable results both to those reported in the literature and identical experiments performed in a standard parallel-electrode electrochemical cell. To demonstrate a photoelectrochemical application of the cell, CO2 reduction experiments were carried out on a plasmonic nanostructured silver photocathode and showed different product distributions under dark and illuminated conditions.

  8. Quantitative study of Xanthosoma violaceum leaf surfaces using RIMAPS and variogram techniques.

    PubMed

    Favret, Eduardo A; Fuentes, Néstor O; Molina, Ana M

    2006-08-01

    Two new imaging techniques (rotated image with maximum averaged power spectrum (RIMAPS) and variogram) are presented for the study and description of leaf surfaces. Xanthosoma violaceum was analyzed to illustrate the characteristics of both techniques. Both techniques produce a quantitative description of leaf surface topography. RIMAPS combines digitized images rotation with Fourier transform, and it is used to detect patterns orientation and characteristics of surface topography. Variogram relates the mathematical variance of a surface with the area of the sample window observed. It gives the typical scale lengths of the surface patterns. RIMAPS detects the morphological variations of the surface topography pattern between fresh and dried (herbarium) samples of the leaf. The variogram method finds the characteristic dimensions of the leaf microstructure, i.e., cell length, papillae diameter, etc., showing that there are not significant differences between dry and fresh samples. The results obtained show the robustness of RIMAPS and variogram analyses to detect, distinguish, and characterize leaf surfaces, as well as give scale lengths. Both techniques are tools for the biologist to study variations of the leaf surface when different patterns are present. The use of RIMAPS and variogram opens a wide spectrum of possibilities by providing a systematic, quantitative description of the leaf surface topography.

  9. Quantitation of peptides from non-invasive skin tapings using isotope dilution and tandem mass spectrometry.

    PubMed

    Reisdorph, Nichole; Armstrong, Michael; Powell, Roger; Quinn, Kevin; Legg, Kevin; Leung, Donald; Reisdorph, Rick

    2018-05-01

    Previous work from our laboratories utilized a novel skin taping method and mass spectrometry-based proteomics to discover clinical biomarkers of skin conditions; these included atopic dermatitis, Staphylococcus aureus colonization, and eczema herpeticum. While suitable for discovery purposes, semi-quantitative proteomics is generally time-consuming and expensive. Furthermore, depending on the method used, discovery-based proteomics can result in high variation and inadequate sensitivity to detect low abundant peptides. Therefore, we strove to develop a rapid, sensitive, and reproducible method to quantitate disease-related proteins from skin tapings. We utilized isotopically-labeled peptides and tandem mass spectrometry to obtain absolute quantitation values on 14 peptides from 7 proteins; these proteins had shown previous importance in skin disease. The method demonstrated good reproducibility, dynamic range, and linearity (R 2  > 0.993) when n = 3 standards were analyzed across 0.05-2.5 pmol. The method was used to determine if differences exist between skin proteins in a small group of atopic versus non-atopic individuals (n = 12). While only minimal differences were found, peptides were detected in all samples and exhibited good correlation between peptides for 5 of the 7 proteins (R 2  = 0.71-0.98). This method can be applied to larger cohorts to further establish the relationships of these proteins to skin disease. Copyright © 2017. Published by Elsevier B.V.

  10. Dominant Epistasis Between Two Quantitative Trait Loci Governing Sporulation Efficiency in Yeast Saccharomyces cerevisiae

    PubMed Central

    Bergman, Juraj; Mitrikeski, Petar T.

    2015-01-01

    Summary Sporulation efficiency in the yeast Saccharomyces cerevisiae is a well-established model for studying quantitative traits. A variety of genes and nucleotides causing different sporulation efficiencies in laboratory, as well as in wild strains, has already been extensively characterised (mainly by reciprocal hemizygosity analysis and nucleotide exchange methods). We applied a different strategy in order to analyze the variation in sporulation efficiency of laboratory yeast strains. Coupling classical quantitative genetic analysis with simulations of phenotypic distributions (a method we call phenotype modelling) enabled us to obtain a detailed picture of the quantitative trait loci (QTLs) relationships underlying the phenotypic variation of this trait. Using this approach, we were able to uncover a dominant epistatic inheritance of loci governing the phenotype. Moreover, a molecular analysis of known causative quantitative trait genes and nucleotides allowed for the detection of novel alleles, potentially responsible for the observed phenotypic variation. Based on the molecular data, we hypothesise that the observed dominant epistatic relationship could be caused by the interaction of multiple quantitative trait nucleotides distributed across a 60--kb QTL region located on chromosome XIV and the RME1 locus on chromosome VII. Furthermore, we propose a model of molecular pathways which possibly underlie the phenotypic variation of this trait. PMID:27904371

  11. Using PSEA-Quant for Protein Set Enrichment Analysis of Quantitative Mass Spectrometry-Based Proteomics.

    PubMed

    Lavallée-Adam, Mathieu; Yates, John R

    2016-03-24

    PSEA-Quant analyzes quantitative mass spectrometry-based proteomics datasets to identify enrichments of annotations contained in repositories such as the Gene Ontology and Molecular Signature databases. It allows users to identify the annotations that are significantly enriched for reproducibly quantified high abundance proteins. PSEA-Quant is available on the Web and as a command-line tool. It is compatible with all label-free and isotopic labeling-based quantitative proteomics methods. This protocol describes how to use PSEA-Quant and interpret its output. The importance of each parameter as well as troubleshooting approaches are also discussed. © 2016 by John Wiley & Sons, Inc. Copyright © 2016 John Wiley & Sons, Inc.

  12. Quantitative and Qualitative Evaluations of the Enhanced Logo-autobiography Program for Korean-American Women.

    PubMed

    Sung, Kyung Mi; Bernstein, Kunsook

    2017-12-01

    This study extends Bernstein et al.'s (2016) investigation of the effects of the Enhanced Logo-autobiography Program on Korean-American women's depressive symptoms, coping strategies, purpose in life, and posttraumatic growth by analyzing quantitative and qualitative data. This study's participants significantly improved on quantitative measures of depression, coping strategies, purpose in life, and post-traumatic growth at eight weeks post-intervention and follow-up. The qualitative content analysis revealed 17 themes with five essential themes. The program's activity to promote purpose in life through posttraumatic growth facilitated participants' recovery from traumatic experiences. Standardized guidelines are needed to conduct this program in Korean community centers.

  13. A qualitative and quantitative study on the enkephalinergic innervation of the pig gastrointestinal tract.

    PubMed

    Porcher, C; Julé, Y; Henry, M

    2000-03-01

    Enkephalins are involved in neural control of digestive functions such as motility, secretion, and absorption. To better understand their role in pigs, we analyzed the qualitative and quantitative distribution of enkephalin immunoreactivity (ENK-IR) in components of the intestinal wall from the esophagus to the anal sphincter. Immunohistochemical labelings were analyzed using conventional fluorescence and confocal microscopy. ENK-IR was compared with the synaptophysin immunoreactivity (SYN-IR). The results show that maximal ENK-IR levels in the entire digestive tract are reached in the myenteric plexuses and, to a lesser extent, in the external submucous plexus and the circular muscle layer. In the longitudinal muscle layer, ENK-IR was present in the esophagus, stomach, rectum, and anal sphincter, whereas it was absent from the duodenum to the distal colon. In the ENK-IR plexuses and muscle layers, more than 60% of the nerve fibers identified by SYN-IR expressed ENK-IR. No ENK-IR was observed in the internal submucous plexus and the mucosa; the latter was found to contain ENK-IR endocrine cells. These results strongly suggest that, in pigs, enkephalins play a major role in the regulatory mechanisms that underlie the neural control of digestive motility.

  14. Quantitative Glycomics Strategies*

    PubMed Central

    Mechref, Yehia; Hu, Yunli; Desantos-Garcia, Janie L.; Hussein, Ahmed; Tang, Haixu

    2013-01-01

    The correlations between protein glycosylation and many biological processes and diseases are increasing the demand for quantitative glycomics strategies enabling sensitive monitoring of changes in the abundance and structure of glycans. This is currently attained through multiple strategies employing several analytical techniques such as capillary electrophoresis, liquid chromatography, and mass spectrometry. The detection and quantification of glycans often involve labeling with ionic and/or hydrophobic reagents. This step is needed in order to enhance detection in spectroscopic and mass spectrometric measurements. Recently, labeling with stable isotopic reagents has also been presented as a very viable strategy enabling relative quantitation. The different strategies available for reliable and sensitive quantitative glycomics are herein described and discussed. PMID:23325767

  15. Guidelines for reporting quantitative mass spectrometry based experiments in proteomics.

    PubMed

    Martínez-Bartolomé, Salvador; Deutsch, Eric W; Binz, Pierre-Alain; Jones, Andrew R; Eisenacher, Martin; Mayer, Gerhard; Campos, Alex; Canals, Francesc; Bech-Serra, Joan-Josep; Carrascal, Montserrat; Gay, Marina; Paradela, Alberto; Navajas, Rosana; Marcilla, Miguel; Hernáez, María Luisa; Gutiérrez-Blázquez, María Dolores; Velarde, Luis Felipe Clemente; Aloria, Kerman; Beaskoetxea, Jabier; Medina-Aunon, J Alberto; Albar, Juan P

    2013-12-16

    Mass spectrometry is already a well-established protein identification tool and recent methodological and technological developments have also made possible the extraction of quantitative data of protein abundance in large-scale studies. Several strategies for absolute and relative quantitative proteomics and the statistical assessment of quantifications are possible, each having specific measurements and therefore, different data analysis workflows. The guidelines for Mass Spectrometry Quantification allow the description of a wide range of quantitative approaches, including labeled and label-free techniques and also targeted approaches such as Selected Reaction Monitoring (SRM). The HUPO Proteomics Standards Initiative (HUPO-PSI) has invested considerable efforts to improve the standardization of proteomics data handling, representation and sharing through the development of data standards, reporting guidelines, controlled vocabularies and tooling. In this manuscript, we describe a key output from the HUPO-PSI-namely the MIAPE Quant guidelines, which have developed in parallel with the corresponding data exchange format mzQuantML [1]. The MIAPE Quant guidelines describe the HUPO-PSI proposal concerning the minimum information to be reported when a quantitative data set, derived from mass spectrometry (MS), is submitted to a database or as supplementary information to a journal. The guidelines have been developed with input from a broad spectrum of stakeholders in the proteomics field to represent a true consensus view of the most important data types and metadata, required for a quantitative experiment to be analyzed critically or a data analysis pipeline to be reproduced. It is anticipated that they will influence or be directly adopted as part of journal guidelines for publication and by public proteomics databases and thus may have an impact on proteomics laboratories across the world. This article is part of a Special Issue entitled: Standardization and

  16. Communication Profiles of Psychiatric Residents and Attending Physicians in Medication-Management Appointments: A Quantitative Pilot Study

    ERIC Educational Resources Information Center

    Castillo, Enrico G.; Pincus, Harold A.; Wieland, Melissa; Roter, Debra; Larson, Susan; Houck, Patricia; Reynolds, Charles F.; Cruz, Mario

    2012-01-01

    Objective: The authors quantitatively examined differences in psychiatric residents' and attending physicians' communication profiles and voice tones. Methods: Audiotaped recordings of 49 resident-patient and 35 attending-patient medication-management appointments at four ambulatory sites were analyzed with the Roter Interaction Analysis System…

  17. The pyPHaz software, an interactive tool to analyze and visualize results from probabilistic hazard assessments

    NASA Astrophysics Data System (ADS)

    Tonini, Roberto; Selva, Jacopo; Costa, Antonio; Sandri, Laura

    2014-05-01

    Probabilistic Hazard Assessment (PHA) is becoming an essential tool for risk mitigation policies, since it allows to quantify the hazard due to hazardous phenomena and, differently from the deterministic approach, it accounts for both aleatory and epistemic uncertainties. On the other hand, one of the main disadvantages of PHA methods is that their results are not easy to understand and interpret by people who are not specialist in probabilistic tools. For scientists, this leads to the issue of providing tools that can be easily used and understood by decision makers (i.e., risk managers or local authorities). The work here presented fits into the problem of simplifying the transfer between scientific knowledge and land protection policies, by providing an interface between scientists, who produce PHA's results, and decision makers, who use PHA's results for risk analyses. In this framework we present pyPHaz, an open tool developed and designed to visualize and analyze PHA results due to one or more phenomena affecting a specific area of interest. The software implementation has been fully developed with the free and open-source Python programming language and some featured Python-based libraries and modules. The pyPHaz tool allows to visualize the Hazard Curves (HC) calculated in a selected target area together with different levels of uncertainty (mean and percentiles) on maps that can be interactively created and modified by the user, thanks to a dedicated Graphical User Interface (GUI). Moreover, the tool can be used to compare the results of different PHA models and to merge them, by creating ensemble models. The pyPHaz software has been designed with the features of storing and accessing all the data through a MySQL database and of being able to read as input the XML-based standard file formats defined in the frame of GEM (Global Earthquake Model). This format model is easy to extend also to any other kind of hazard, as it will be shown in the applications

  18. Relationship between DNA damage response, initiated by camptothecin or oxidative stress, and DNA replication, analyzed by quantitative 3D image analysis.

    PubMed

    Berniak, K; Rybak, P; Bernas, T; Zarębski, M; Biela, E; Zhao, H; Darzynkiewicz, Z; Dobrucki, J W

    2013-10-01

    A method of quantitative analysis of spatial (3D) relationship between discrete nuclear events detected by confocal microscopy is described and applied in analysis of a dependence between sites of DNA damage signaling (γH2AX foci) and DNA replication (EdU incorporation) in cells subjected to treatments with camptothecin (Cpt) or hydrogen peroxide (H2O2). Cpt induces γH2AX foci, likely reporting formation of DNA double-strand breaks (DSBs), almost exclusively at sites of DNA replication. This finding is consistent with the known mechanism of induction of DSBs by DNA topoisomerase I (topo1) inhibitors at the sites of collisions of the moving replication forks with topo1-DNA "cleavable complexes" stabilized by Cpt. Whereas an increased level of H2AX histone phosphorylation is seen in S-phase of cells subjected to H2O2, only a minor proportion of γH2AX foci coincide with DNA replication sites. Thus, the increased level of H2AX phosphorylation induced by H2O2 is not a direct consequence of formation of DNA lesions at the sites of moving DNA replication forks. These data suggest that oxidative stress induced by H2O2 and formation of the primary H2O2-induced lesions (8-oxo-7,8-dihydroguanosine) inhibits replication globally and triggers formation of γH2AX at various distances from replication forks. Quantitative analysis of a frequency of DNA replication sites and γH2AX foci suggests also that stalling of replicating forks by Cpt leads to activation of new DNA replication origins. © 2013 International Society for Advancement of Cytometry. Copyright © 2013 International Society for Advancement of Cytometry.

  19. Quantitative aspects of inductively coupled plasma mass spectrometry

    NASA Astrophysics Data System (ADS)

    Bulska, Ewa; Wagner, Barbara

    2016-10-01

    Accurate determination of elements in various kinds of samples is essential for many areas, including environmental science, medicine, as well as industry. Inductively coupled plasma mass spectrometry (ICP-MS) is a powerful tool enabling multi-elemental analysis of numerous matrices with high sensitivity and good precision. Various calibration approaches can be used to perform accurate quantitative measurements by ICP-MS. They include the use of pure standards, matrix-matched standards, or relevant certified reference materials, assuring traceability of the reported results. This review critically evaluates the advantages and limitations of different calibration approaches, which are used in quantitative analyses by ICP-MS. Examples of such analyses are provided. This article is part of the themed issue 'Quantitative mass spectrometry'.

  20. Regenerate Healing Outcomes in Unilateral Mandibular Distraction Osteogenesis Using Quantitative Histomorphometry

    PubMed Central

    Schwarz, Daniel A.; Arman, Krikor G.; Kakwan, Mehreen S.; Jamali, Ameen M.; Elmeligy, Ayman A.; Buchman, Steven R.

    2015-01-01

    Background The authors’ goal was to ascertain regenerate bone-healing metrics using quantitative histomorphometry at a single consolidation period. Methods Rats underwent either mandibular distraction osteogenesis (n=7) or partially reduced fractures (n=7); their contralateral mandibles were used as controls (n=11). External fixators were secured and unilateral osteotomies performed, followed by either mandibular distraction osteogenesis (4 days’ latency, then 0.3 mm every 12 hours for 8 days; 5.1 mm) or partially reduced fractures (fixed immediately postoperatively; 2.1 mm); both groups underwent 4 weeks of consolidation. After tissue processing, bone volume/tissue volume ratio, osteoid volume/tissue volume ratio, and osteocyte count per high-power field were analyzed by means of quantitative histomorphometry. Results Contralateral mandibles had statistically greater bone volume/tissue volume ratio and osteocyte count per high-power field compared with both mandibular distraction osteogenesis and partially reduced fractures by almost 50 percent, whereas osteoid volume/tissue volume ratio was statistically greater in both mandibular distraction osteogenesis specimens and partially reduced fractures compared with contralateral mandibles. No statistical difference in bone volume/tissue volume ratio, osteoid volume/tissue volume ratio, or osteocyte count per high-power field was found between mandibular distraction osteogenesis specimens and partially reduced fractures. Conclusions The authors’ findings demonstrate significantly decreased bone quantity and maturity in mandibular distraction osteogenesis specimens and partially reduced fractures compared with contralateral mandibles using the clinically analogous protocols. If these results are extrapolated clinically, treatment strategies may require modification to ensure reliable, predictable, and improved outcomes. PMID:20463629

  1. Variables affecting the quantitation of CD22 in neoplastic B cells.

    PubMed

    Jasper, Gregory A; Arun, Indu; Venzon, David; Kreitman, Robert J; Wayne, Alan S; Yuan, Constance M; Marti, Gerald E; Stetler-Stevenson, Maryalice

    2011-03-01

    Quantitative flow cytometry (QFCM) is being applied in the clinical flow cytometry laboratory for diagnosis, prognosis, and assessment of patients receiving antibody-based therapy. ABC values and the effect of technical variables on CD22 quantitation in acute lymphoblastic leukemia (ALL), chronic lymphocytic leukemia (CLL), mantle cell lymphoma (MCL), follicular lymphoma (FCL), hairy cell leukemia (HCL) and normal B cells were studied. The QuantiBrite System® was used to determine the level of CD22 expression (mean antibody bound per cell, ABC) by malignant and normal B cells. The intra-assay variability, number of cells required for precision, effect of delayed processing as well as shipment of peripheral blood specimens (delayed processing and exposure to noncontrolled environments), and the effect of paraformaldehyde fixation on assay results were studied. The QuantiBRITE method of measuring CD22 ABC is precise (median CV 1.6%, 95% confidence interval, 1.2-2.3%) but a threshold of 250 malignant cells is required for reliable CD22 ABC values. Delayed processing and overnight shipment of specimens resulted in significantly different ABC values whereas fixation for up to 12 h had no significant effect. ABC measurements determined that CD22 expression is lower than normal in ALL, CLL, FCL, and MCL but higher than normal in HCL. CD22 expression was atypical in the hematolymphoid malignancies studied and may have diagnostic utility. Technical variables such as cell number analyzed and delayed processing or overnight shipment of specimens impact significantly on the measurement of antigen expression by QFCM in the clinical laboratory. Published 2010 Wiley-Liss, Inc.

  2. Affinity for Quantitative Tools: Undergraduate Marketing Students Moving beyond Quantitative Anxiety

    ERIC Educational Resources Information Center

    Tarasi, Crina O.; Wilson, J. Holton; Puri, Cheenu; Divine, Richard L.

    2013-01-01

    Marketing students are known as less likely to have an affinity for the quantitative aspects of the marketing discipline. In this article, we study the reasons why this might be true and develop a parsimonious 20-item scale for measuring quantitative affinity in undergraduate marketing students. The scale was administered to a sample of business…

  3. Comparative study of contrast-enhanced ultrasound qualitative and quantitative analysis for identifying benign and malignant breast tumor lumps.

    PubMed

    Liu, Jian; Gao, Yun-Hua; Li, Ding-Dong; Gao, Yan-Chun; Hou, Ling-Mi; Xie, Ting

    2014-01-01

    To compare the value of contrast-enhanced ultrasound (CEUS) qualitative and quantitative analysis in the identification of breast tumor lumps. Qualitative and quantitative indicators of CEUS for 73 cases of breast tumor lumps were retrospectively analyzed by univariate and multivariate approaches. Logistic regression was applied and ROC curves were drawn for evaluation and comparison. The CEUS qualitative indicator-generated regression equation contained three indicators, namely enhanced homogeneity, diameter line expansion and peak intensity grading, which demonstrated prediction accuracy for benign and malignant breast tumor lumps of 91.8%; the quantitative indicator-generated regression equation only contained one indicator, namely the relative peak intensity, and its prediction accuracy was 61.5%. The corresponding areas under the ROC curve for qualitative and quantitative analyses were 91.3% and 75.7%, respectively, which exhibited a statistically significant difference by the Z test (P<0.05). The ability of CEUS qualitative analysis to identify breast tumor lumps is better than with quantitative analysis.

  4. Quantitative Proteomic Analysis of Duck Ovarian Follicles Infected with Duck Tembusu Virus by Label-Free LC-MS

    PubMed Central

    Han, Kaikai; Zhao, Dongmin; Liu, Yuzhuo; Liu, Qingtao; Huang, Xinmei; Yang, Jing; An, Fengjiao; Li, Yin

    2016-01-01

    Duck Tembusu virus (DTMUV) is a newly emerging pathogenic flavivirus that has caused massive economic losses to the duck industry in China. DTMUV infection mainly results in significant decreases in egg production in egg-laying ducks within 1–2 weeks post infection. However, information on the comparative protein expression of host tissues in response to DTMUV infection is limited. In the present study, the cellular protein response to DTMUV infection in duck ovarian follicles was analyzed using nano-flow high-performance liquid chromatography-electrospray tandem mass spectrometry. Quantitative proteomic analysis revealed 131 differentially expressed proteins, among which 53 were up regulated and 78 were down regulated. The identified proteins were involved in the regulation of essential processes such as cellular structure and integrity, RNA processing, protein biosynthesis and modification, vesicle transport, signal transduction, and mitochondrial pathway. Some selected proteins that were found to be regulated in DTMUV-infected tissues were screened by quantitative real-time PCR to examine their regulation at the transcriptional level, western blot analysis was used to validate the changes of some selected proteins on translational level. To our knowledge, this study is the first to analyze the proteomic changes in duck ovarian follicles following DTMUV infection. The protein-related information obtained in this study may be useful to understand the host response to DTMUV infection and the inherent mechanism of DTMUV replication and pathogenicity. PMID:27066001

  5. Quantitative analysis of task selection for brain-computer interfaces

    NASA Astrophysics Data System (ADS)

    Llera, Alberto; Gómez, Vicenç; Kappen, Hilbert J.

    2014-10-01

    Objective. To assess quantitatively the impact of task selection in the performance of brain-computer interfaces (BCI). Approach. We consider the task-pairs derived from multi-class BCI imagery movement tasks in three different datasets. We analyze for the first time the benefits of task selection on a large-scale basis (109 users) and evaluate the possibility of transferring task-pair information across days for a given subject. Main results. Selecting the subject-dependent optimal task-pair among three different imagery movement tasks results in approximately 20% potential increase in the number of users that can be expected to control a binary BCI. The improvement is observed with respect to the best task-pair fixed across subjects. The best task-pair selected for each subject individually during a first day of recordings is generally a good task-pair in subsequent days. In general, task learning from the user side has a positive influence in the generalization of the optimal task-pair, but special attention should be given to inexperienced subjects. Significance. These results add significant evidence to existing literature that advocates task selection as a necessary step towards usable BCIs. This contribution motivates further research focused on deriving adaptive methods for task selection on larger sets of mental tasks in practical online scenarios.

  6. Quantitative Classification of Rice (Oryza sativa L.) Root Length and Diameter Using Image Analysis.

    PubMed

    Gu, Dongxiang; Zhen, Fengxian; Hannaway, David B; Zhu, Yan; Liu, Leilei; Cao, Weixing; Tang, Liang

    2017-01-01

    Quantitative study of root morphological characteristics of plants is helpful for understanding the relationships between their morphology and function. However, few studies and little detailed and accurate information of root characteristics were reported in fine-rooted plants like rice (Oryza sativa L.). The aims of this study were to quantitatively classify fine lateral roots (FLRs), thick lateral roots (TLRs), and nodal roots (NRs) and analyze their dynamics of mean diameter (MD), lengths and surface area percentage with growth stages in rice plant. Pot experiments were carried out during three years with three rice cultivars, three nitrogen (N) rates and three water regimes. In cultivar experiment, among the three cultivars, root length of 'Yangdao 6' was longest, while the MD of its FLR was the smallest, and the mean diameters for TLR and NR were the largest, the surface area percentage (SAP) of TLRs (SAPT) was the highest, indicating that Yangdao 6 has better nitrogen and water uptake ability. High N rate increased the length of different types of roots and increased the MD of lateral roots, decreased the SAP of FLRs (SAPF) and TLRs, but increased the SAP of NRs (SAPN). Moderate decrease of water supply increased root length and diameter, water stress increased the SAPF and SAPT, but decreased SAPN. The quantitative results indicate that rice plant tends to increase lateral roots to get more surface area for nitrogen and water uptake when available assimilates are limiting under nitrogen and water stress environments.

  7. Accurate Quantitation and Analysis of Nitrofuran Metabolites, Chloramphenicol, and Florfenicol in Seafood by Ultrahigh-Performance Liquid Chromatography-Tandem Mass Spectrometry: Method Validation and Regulatory Samples.

    PubMed

    Aldeek, Fadi; Hsieh, Kevin C; Ugochukwu, Obiadada N; Gerard, Ghislain; Hammack, Walter

    2018-05-23

    We developed and validated a method for the extraction, identification, and quantitation of four nitrofuran metabolites, 3-amino-2-oxazolidinone (AOZ), 3-amino-5-morpholinomethyl-2-oxazolidinone (AMOZ), semicarbazide (SC), and 1-aminohydantoin (AHD), as well as chloramphenicol and florfenicol in a variety of seafood commodities. Samples were extracted by liquid-liquid extraction techniques, analyzed by ultrahigh-performance liquid chromatography-tandem mass spectrometry (UHPLC-MS/MS), and quantitated using commercially sourced, derivatized nitrofuran metabolites, with their isotopically labeled internal standards in-solvent. We obtained recoveries of 90-100% at various fortification levels. The limit of detection (LOD) was set at 0.25 ng/g for AMOZ and AOZ, 1 ng/g for AHD and SC, and 0.1 ng/g for the phenicols. Various extraction methods, standard stability, derivatization efficiency, and improvements to conventional quantitation techniques were also investigated. We successfully applied this method to the identification and quantitation of nitrofuran metabolites and phenicols in 102 imported seafood products. Our results revealed that four of the samples contained residues from banned veterinary drugs.

  8. Slow erosion of a quantitative apple resistance to Venturia inaequalis based on an isolate-specific Quantitative Trait Locus.

    PubMed

    Caffier, Valérie; Le Cam, Bruno; Al Rifaï, Mehdi; Bellanger, Marie-Noëlle; Comby, Morgane; Denancé, Caroline; Didelot, Frédérique; Expert, Pascale; Kerdraon, Tifenn; Lemarquand, Arnaud; Ravon, Elisa; Durel, Charles-Eric

    2016-10-01

    Quantitative plant resistance affects the aggressiveness of pathogens and is usually considered more durable than qualitative resistance. However, the efficiency of a quantitative resistance based on an isolate-specific Quantitative Trait Locus (QTL) is expected to decrease over time due to the selection of isolates with a high level of aggressiveness on resistant plants. To test this hypothesis, we surveyed scab incidence over an eight-year period in an orchard planted with susceptible and quantitatively resistant apple genotypes. We sampled 79 Venturia inaequalis isolates from this orchard at three dates and we tested their level of aggressiveness under controlled conditions. Isolates sampled on resistant genotypes triggered higher lesion density and exhibited a higher sporulation rate on apple carrying the resistance allele of the QTL T1 compared to isolates sampled on susceptible genotypes. Due to this ability to select aggressive isolates, we expected the QTL T1 to be non-durable. However, our results showed that the quantitative resistance based on the QTL T1 remained efficient in orchard over an eight-year period, with only a slow decrease in efficiency and no detectable increase of the aggressiveness of fungal isolates over time. We conclude that knowledge on the specificity of a QTL is not sufficient to evaluate its durability. Deciphering molecular mechanisms associated with resistance QTLs, genetic determinants of aggressiveness and putative trade-offs within pathogen populations is needed to help in understanding the erosion processes. Copyright © 2016 Elsevier B.V. All rights reserved.

  9. Glycan reductive isotope labeling for quantitative glycomics.

    PubMed

    Xia, Baoyun; Feasley, Christa L; Sachdev, Goverdhan P; Smith, David F; Cummings, Richard D

    2009-04-15

    Many diseases and disorders are characterized by quantitative and/or qualitative changes in complex carbohydrates. Mass spectrometry methods show promise in monitoring and detecting these important biological changes. Here we report a new glycomics method, termed glycan reductive isotope labeling (GRIL), where free glycans are derivatized by reductive amination with the differentially coded stable isotope tags [(12)C(6)]aniline and [(13)C(6)]aniline. These dual-labeled aniline-tagged glycans can be recovered by reverse-phase chromatography and can be quantified based on ultraviolet (UV) absorbance and relative ion abundances. Unlike previously reported isotopically coded reagents for glycans, GRIL does not contain deuterium, which can be chromatographically resolved. Our method shows no chromatographic resolution of differentially labeled glycans. Mixtures of differentially tagged glycans can be directly compared and quantified using mass spectrometric techniques. We demonstrate the use of GRIL to determine relative differences in glycan amount and composition. We analyze free glycans and glycans enzymatically or chemically released from a variety of standard glycoproteins, as well as human and mouse serum glycoproteins, using this method. This technique allows linear relative quantitation of glycans over a 10-fold concentration range and can accurately quantify sub-picomole levels of released glycans, providing a needed advancement in the field of glycomics.

  10. Chemotaxis of cancer cells in three-dimensional environment monitored label-free by quantitative phase digital holographic microscopy

    NASA Astrophysics Data System (ADS)

    Kemper, Björn; Schnekenburger, Jürgen; Ketelhut, Steffi

    2017-02-01

    We investigated the capabilities of digital holographic microscopy (DHM) for label-free quantification of the response of living single cells to chemical stimuli in 3D assays. Fibro sarcoma cells were observed in a collagen matrix inside 3D chemotaxis chambers with a Mach-Zehnder interferometer-based DHM setup. From the obtained series of quantitative phase images, the migration trajectories of single cells were retrieved by automated cell tracking and subsequently analyzed for maximum migration distance and motility. Our results demonstrate DHM as a highly reliable and efficient tool for label-free quantification of chemotaxis in 2D and 3D environments.

  11. Traceability Assessment and Performance Evaluation of Results for Measurement of Abbott Clinical Chemistry Assays on 4 Chemistry Analyzers.

    PubMed

    Lim, Jinsook; Song, Kyung Eun; Song, Sang Hoon; Choi, Hyun-Jung; Koo, Sun Hoe; Kwon, Gye Choel

    2016-05-01

    -The traceability of clinical results to internationally recognized and accepted reference materials and reference measurement procedures has become increasingly important. Therefore, the establishment of traceability has become a mandatory requirement for all in vitro diagnostics devices. -To evaluate the traceability of the Abbott Architect c8000 system (Abbott Laboratories, Abbott Park, Illinois), consisting of calibrators and reagents, across 4 different chemistry analyzers, and to evaluate its general performance on the Toshiba 2000FR NEO (Toshiba Medical Systems Corporation, Otawara-shi, Tochigi-ken, Japan). -For assessment of traceability, secondary reference materials were evaluated 5 times, and then bias was calculated. Precision, linearity, and carryover were determined according to the guidelines of the Clinical and Laboratory Standards Institute (Wayne, Pennsylvania). -The biases from 4 different analyzers ranged from -2.33% to 2.70% on the Toshiba 2000FR NEO, -2.33% to 5.12% on the Roche Hitachi 7600 (Roche Diagnostics International, Basel, Switzerland), -0.93% to 2.87% on the Roche Modular, and -2.16% to 2.86% on the Abbott Architect c16000. The total coefficients of variance of all analytes were less than 5%. The coefficients of determination (R(2)) were more than 0.9900. The carryover rate ranged from -0.54% to 0.17%. -Abbott clinical chemistry assays met the performance criteria based on desirable biological variation for precision, bias, and total error. They also showed excellent linearity and carryover. Therefore, these clinical chemistry assays were found to be accurate and reliable and are readily applicable on the various platforms used in this study.

  12. From information theory to quantitative description of steric effects.

    PubMed

    Alipour, Mojtaba; Safari, Zahra

    2016-07-21

    Immense efforts have been made in the literature to apply the information theory descriptors for investigating the electronic structure theory of various systems. In the present study, the information theoretic quantities, such as Fisher information, Shannon entropy, Onicescu information energy, and Ghosh-Berkowitz-Parr entropy, have been used to present a quantitative description for one of the most widely used concepts in chemistry, namely the steric effects. Taking the experimental steric scales for the different compounds as benchmark sets, there are reasonable linear relationships between the experimental scales of the steric effects and theoretical values of steric energies calculated from information theory functionals. Perusing the results obtained from the information theoretic quantities with the two representations of electron density and shape function, the Shannon entropy has the best performance for the purpose. On the one hand, the usefulness of considering the contributions of functional groups steric energies and geometries, and on the other hand, dissecting the effects of both global and local information measures simultaneously have also been explored. Furthermore, the utility of the information functionals for the description of steric effects in several chemical transformations, such as electrophilic and nucleophilic reactions and host-guest chemistry, has been analyzed. The functionals of information theory correlate remarkably with the stability of systems and experimental scales. Overall, these findings show that the information theoretic quantities can be introduced as quantitative measures of steric effects and provide further evidences of the quality of information theory toward helping theoreticians and experimentalists to interpret different problems in real systems.

  13. Quantitative Analysis of TDLUs using Adaptive Morphological Shape Techniques

    PubMed Central

    Rosebrock, Adrian; Caban, Jesus J.; Figueroa, Jonine; Gierach, Gretchen; Linville, Laura; Hewitt, Stephen; Sherman, Mark

    2014-01-01

    Within the complex branching system of the breast, terminal duct lobular units (TDLUs) are the anatomical location where most cancer originates. With aging, TDLUs undergo physiological involution, reflected in a loss of structural components (acini) and a reduction in total number. Data suggest that women undergoing benign breast biopsies that do not show age appropriate involution are at increased risk of developing breast cancer. To date, TDLU assessments have generally been made by qualitative visual assessment, rather than by objective quantitative analysis. This paper introduces a technique to automatically estimate a set of quantitative measurements and use those variables to more objectively describe and classify TDLUs. To validate the accuracy of our system, we compared the computer-based morphological properties of 51 TDLUs in breast tissues donated for research by volunteers in the Susan G. Komen Tissue Bank and compared results to those of a pathologist, demonstrating 70% agreement. Secondly, in order to show that our method is applicable to a wider range of datasets, we analyzed 52 TDLUs from biopsies performed for clinical indications in the National Cancer Institute’s Breast Radiology Evaluation and Study of Tissues (BREAST) Stamp Project and obtained 82% correlation with visual assessment. Lastly, we demonstrate the ability to uncover novel measures when researching the structural properties of the acini by applying machine learning and clustering techniques. Through our study we found that while the number of acini per TDLU increases exponentially with the TDLU diameter, the average elongation and roundness remain constant. PMID:25722829

  14. "What else are you worried about?" - Integrating textual responses into quantitative social science research.

    PubMed

    Rohrer, Julia M; Brümmer, Martin; Schmukle, Stefan C; Goebel, Jan; Wagner, Gert G

    2017-01-01

    Open-ended questions have routinely been included in large-scale survey and panel studies, yet there is some perplexity about how to actually incorporate the answers to such questions into quantitative social science research. Tools developed recently in the domain of natural language processing offer a wide range of options for the automated analysis of such textual data, but their implementation has lagged behind. In this study, we demonstrate straightforward procedures that can be applied to process and analyze textual data for the purposes of quantitative social science research. Using more than 35,000 textual answers to the question "What else are you worried about?" from participants of the German Socio-economic Panel Study (SOEP), we (1) analyzed characteristics of respondents that determined whether they answered the open-ended question, (2) used the textual data to detect relevant topics that were reported by the respondents, and (3) linked the features of the respondents to the worries they reported in their textual data. The potential uses as well as the limitations of the automated analysis of textual data are discussed.

  15. The LARSYS educational package: Instructor's notes. [instructional materials for training people to analyze remotely sensed data

    NASA Technical Reports Server (NTRS)

    Lindenlaub, J. C.; Davis, S. M.

    1974-01-01

    Materials are presented for assisting instructors in teaching the LARSYS Educational Package, which is a set of instructional materials to train people to analyze remotely sensed multispectral data. The seven units of the package are described. These units are: quantitative remote sensing, overview of the LARSYS software system, the 2780 remote terminal, demonstration of LARSYS on the 2780 remote terminal, exercises, guide to multispectral data analysis, and a case study using LARSYS for analysis of LANDSAT data.

  16. Effect of once-yearly zoledronic acid on the spine and hip as measured by quantitative computed tomography: results of the HORIZON Pivotal Fracture Trial

    PubMed Central

    Lang, T.; Boonen, S.; Cummings, S.; Delmas, P. D.; Cauley, J. A.; Horowitz, Z.; Kerzberg, E.; Bianchi, G.; Kendler, D.; Leung, P.; Man, Z.; Mesenbrink, P.; Eriksen, E. F.; Black, D. M.

    2016-01-01

    Summary Changes in bone mineral density and bone strength following treatment with zoledronic acid (ZOL) were measured by quantitative computed analysis (QCT) or dual-energy X-ray absorptiometry (DXA). ZOL treatment increased spine and hip BMD vs placebo, assessed by QCT and DXA. Changes in trabecular bone resulted in increased bone strength. Introduction To investigate bone mineral density (BMD) changes in trabecular and cortical bone, estimated by quantitative computed analysis (QCT) or dual-energy X-ray absorptiometry (DXA), and whether zoledronic acid 5 mg (ZOL) affects bone strength. Methods In 233 women from a randomized, controlled trial of once-yearly ZOL, lumbar spine, total hip, femoral neck, and trochanter were assessed by DXA and QCT (baseline, Month 36). Mean percentage changes from baseline and between-treatment differences (ZOL vs placebo, t-test) were evaluated. Results Mean between-treatment differences for lumbar spine BMD were significant by DXA (7.0%, p<0.01) and QCT (5.7%, p<0.0001). Between-treatment differences were significant for trabecular spine (p=0.0017) [non-parametric test], trabecular trochanter (10.7%, p<0.0001), total hip (10.8%, p<0.0001), and compressive strength indices at femoral neck (8.6%, p=0.0001), and trochanter (14.1%, p<0.0001). Conclusions Once-yearly ZOL increased hip and spine BMD vs placebo, assessed by QCT vs DXA. Changes in trabecular bone resulted in increased indices of compressive strength. PMID:19802508

  17. Quantitation of lysergic acid diethylamide in urine using atmospheric pressure matrix-assisted laser desorption/ionization ion trap mass spectrometry.

    PubMed

    Cui, Meng; McCooeye, Margaret A; Fraser, Catharine; Mester, Zoltán

    2004-12-01

    A quantitative method was developed for analysis of lysergic acid diethylamide (LSD) in urine using atmospheric pressure matrix-assisted laser desorption/ionization ion trap mass spectrometry (AP MALDI-ITMS). Following solid-phase extraction of LSD from urine samples, extracts were analyzed by AP MALDI-ITMS. The identity of LSD was confirmed by fragmentation of the [M + H](+) ion using tandem mass spectrometry. The quantification of LSD was achieved using stable-isotope-labeled LSD (LSD-d(3)) as the internal standard. The [M + H](+) ion fragmented to produce a dominant fragment ion, which was used for a selected reaction monitoring (SRM) method for quantitative analysis of LSD. SRM was compared with selected ion monitoring and produced a wider linear range and lower limit of quantification. For SRM analysis of samples of LSD spiked in urine, the calibration curve was linear in the range of 1-100 ng/mL with a coefficient of determination, r(2), of 0.9917. This assay was used to determine LSD in urine samples and the AP MALDI-MS results were comparable to the HPLC/ ESI-MS results.

  18. Using qualitative and quantitative methods to evaluate small-scale disease management pilot programs.

    PubMed

    Esposito, Dominick; Taylor, Erin Fries; Gold, Marsha

    2009-02-01

    Interest in disease management programs continues to grow as managed care plans, the federal and state governments, and other organizations consider such efforts as a means to improve health care quality and reduce costs. These efforts vary in size, scope, and target population. While large-scale programs provide the means to measure impacts, evaluation of smaller interventions remains valuable as they often represent the early planning stages of larger initiatives. This paper describes a multi-method approach for evaluating small interventions that sought to improve the quality of care for Medicaid beneficiaries with multiple chronic conditions. Our approach relied on quantitative and qualitative methods to develop a complete understanding of each intervention. Quantitative data in the form of both process measures, such as case manager contacts, and outcome measures, such as hospital use, were reported and analyzed. Qualitative information was collected through interviews and the development of logic models to document the flow of intervention activities and how they were intended to affect outcomes. The logic models helped us to understand the underlying reasons for the success or lack thereof of each intervention. The analysis provides useful information on several fronts. First, qualitative data provided valuable information about implementation. Second, process measures helped determine whether implementation occurred as anticipated. Third, outcome measures indicated the potential for favorable results later, possibly suggesting further study. Finally, the evaluation of qualitative and quantitative data in combination helped us assess the potential promise of each intervention and identify common themes and challenges across all interventions.

  19. A Portable Analyzer for Pouch-Actuated, Immunoassay Cassettes

    PubMed Central

    Qiu, Xianbo; Liu, Changchun; Mauk, Michael G.; Hart, Robert W.; Chen, Dafeng; Qiu, Jing; Kientz, Terry; Fiene, Jonathan; Bau, Haim H.

    2011-01-01

    A portable, small footprint, light, general purpose analyzer (processor) to control the flow in immunoassay cassettes and to facilitate the detection of test results is described. The durable analyzer accepts disposable cassettes that contain pouches and reaction chambers for various unit operations such as hydration of dry reagents, stirring, and incubation. The analyzer includes individually controlled, linear actuators to compress the pouches in the cassette, which facilitates the pumping and mixing of sample and reagents, and to close diaphragm-based valves for flow control. The same types of actuators are used to compress pouches and actuate valves. The analyzer also houses a compact OEM scanner/reader to excite fluorescence and detect emission from labels. The analyzer is hydraulically isolated from the cassette, reducing the possibility of cross-contamination. The analyzer facilitates programmable, automated execution of a sequence of operations such as pumping and valving in a timely fashion, reducing the level of expertise required from the operator and the possibility for errors. The analyzer’s design is modular and expandable to accommodate cassettes of various complexities and additional functionalities. In this paper, the utility of the analyzer has been demonstrated with the execution of a simple, consecutive, lateral flow assay of a model biological system and the test results were detected with up converting phosphor labels that are excited at infrared frequencies and emit in the visible spectrum. PMID:22125359

  20. Alchemy: A Web 2.0 Real-time Quality Assurance Platform for Human Immunodeficiency Virus, Hepatitis C Virus, and BK Virus Quantitation Assays

    PubMed Central

    Agosto-Arroyo, Emmanuel; Coshatt, Gina M.; Winokur, Thomas S.; Harada, Shuko; Park, Seung L.

    2017-01-01

    Background: The molecular diagnostics laboratory faces the challenge of improving test turnaround time (TAT). Low and consistent TATs are of great clinical and regulatory importance, especially for molecular virology tests. Laboratory information systems (LISs) contain all the data elements necessary to do accurate quality assurance (QA) reporting of TAT and other measures, but these reports are in most cases still performed manually: a time-consuming and error-prone task. The aim of this study was to develop a web-based real-time QA platform that would automate QA reporting in the molecular diagnostics laboratory at our institution, and minimize the time expended in preparing these reports. Methods: Using a standard Linux, Nginx, MariaDB, PHP stack virtual machine running atop a Dell Precision 5810, we designed and built a web-based QA platform, code-named Alchemy. Data files pulled periodically from the LIS in comma-separated value format were used to autogenerate QA reports for the human immunodeficiency virus (HIV) quantitation, hepatitis C virus (HCV) quantitation, and BK virus (BKV) quantitation. Alchemy allowed the user to select a specific timeframe to be analyzed and calculated key QA statistics in real-time, including the average TAT in days, tests falling outside the expected TAT ranges, and test result ranges. Results: Before implementing Alchemy, reporting QA for the HIV, HCV, and BKV quantitation assays took 45–60 min of personnel time per test every month. With Alchemy, that time has decreased to 15 min total per month. Alchemy allowed the user to select specific periods of time and analyzed the TAT data in-depth without the need of extensive manual calculations. Conclusions: Alchemy has significantly decreased the time and the human error associated with QA report generation in our molecular diagnostics laboratory. Other tests will be added to this web-based platform in future updates. This effort shows the utility of informatician