Science.gov

Sample records for accuracy assessment methods

  1. A fast RCS accuracy assessment method for passive radar calibrators

    NASA Astrophysics Data System (ADS)

    Zhou, Yongsheng; Li, Chuanrong; Tang, Lingli; Ma, Lingling; Liu, QI

    2016-10-01

    In microwave radar radiometric calibration, the corner reflector acts as the standard reference target but its structure is usually deformed during the transportation and installation, or deformed by wind and gravity while permanently installed outdoor, which will decrease the RCS accuracy and therefore the radiometric calibration accuracy. A fast RCS accuracy measurement method based on 3-D measuring instrument and RCS simulation was proposed in this paper for tracking the characteristic variation of the corner reflector. In the first step, RCS simulation algorithm was selected and its simulation accuracy was assessed. In the second step, the 3-D measuring instrument was selected and its measuring accuracy was evaluated. Once the accuracy of the selected RCS simulation algorithm and 3-D measuring instrument was satisfied for the RCS accuracy assessment, the 3-D structure of the corner reflector would be obtained by the 3-D measuring instrument, and then the RCSs of the obtained 3-D structure and corresponding ideal structure would be calculated respectively based on the selected RCS simulation algorithm. The final RCS accuracy was the absolute difference of the two RCS calculation results. The advantage of the proposed method was that it could be applied outdoor easily, avoiding the correlation among the plate edge length error, plate orthogonality error, plate curvature error. The accuracy of this method is higher than the method using distortion equation. In the end of the paper, a measurement example was presented in order to show the performance of the proposed method.

  2. Survey methods for assessing land cover map accuracy

    USGS Publications Warehouse

    Nusser, S.M.; Klaas, E.E.

    2003-01-01

    The increasing availability of digital photographic materials has fueled efforts by agencies and organizations to generate land cover maps for states, regions, and the United States as a whole. Regardless of the information sources and classification methods used, land cover maps are subject to numerous sources of error. In order to understand the quality of the information contained in these maps, it is desirable to generate statistically valid estimates of accuracy rates describing misclassification errors. We explored a full sample survey framework for creating accuracy assessment study designs that balance statistical and operational considerations in relation to study objectives for a regional assessment of GAP land cover maps. We focused not only on appropriate sample designs and estimation approaches, but on aspects of the data collection process, such as gaining cooperation of land owners and using pixel clusters as an observation unit. The approach was tested in a pilot study to assess the accuracy of Iowa GAP land cover maps. A stratified two-stage cluster sampling design addressed sample size requirements for land covers and the need for geographic spread while minimizing operational effort. Recruitment methods used for private land owners yielded high response rates, minimizing a source of nonresponse error. Collecting data for a 9-pixel cluster centered on the sampled pixel was simple to implement, and provided better information on rarer vegetation classes as well as substantial gains in precision relative to observing data at a single-pixel.

  3. Assessing the Accuracy of Ancestral Protein Reconstruction Methods

    PubMed Central

    Williams, Paul D; Pollock, David D; Blackburne, Benjamin P; Goldstein, Richard A

    2006-01-01

    The phylogenetic inference of ancestral protein sequences is a powerful technique for the study of molecular evolution, but any conclusions drawn from such studies are only as good as the accuracy of the reconstruction method. Every inference method leads to errors in the ancestral protein sequence, resulting in potentially misleading estimates of the ancestral protein's properties. To assess the accuracy of ancestral protein reconstruction methods, we performed computational population evolution simulations featuring near-neutral evolution under purifying selection, speciation, and divergence using an off-lattice protein model where fitness depends on the ability to be stable in a specified target structure. We were thus able to compare the thermodynamic properties of the true ancestral sequences with the properties of “ancestral sequences” inferred by maximum parsimony, maximum likelihood, and Bayesian methods. Surprisingly, we found that methods such as maximum parsimony and maximum likelihood that reconstruct a “best guess” amino acid at each position overestimate thermostability, while a Bayesian method that sometimes chooses less-probable residues from the posterior probability distribution does not. Maximum likelihood and maximum parsimony apparently tend to eliminate variants at a position that are slightly detrimental to structural stability simply because such detrimental variants are less frequent. Other properties of ancestral proteins might be similarly overestimated. This suggests that ancestral reconstruction studies require greater care to come to credible conclusions regarding functional evolution. Inferred functional patterns that mimic reconstruction bias should be reevaluated. PMID:16789817

  4. GEOSPATIAL DATA ACCURACY ASSESSMENT

    EPA Science Inventory

    The development of robust accuracy assessment methods for the validation of spatial data represent's a difficult scientific challenge for the geospatial science community. The importance and timeliness of this issue is related directly to the dramatic escalation in the developmen...

  5. Assessing the Accuracy of Classwide Direct Observation Methods: Two Analyses Using Simulated and Naturalistic Data

    ERIC Educational Resources Information Center

    Dart, Evan H.; Radley, Keith C.; Briesch, Amy M.; Furlow, Christopher M.; Cavell, Hannah J.; Briesch, Amy M.

    2016-01-01

    Two studies investigated the accuracy of eight different interval-based group observation methods that are commonly used to assess the effects of classwide interventions. In Study 1, a Microsoft Visual Basic program was created to simulate a large set of observational data. Binary data were randomly generated at the student level to represent…

  6. A novel method for assessing the 3-D orientation accuracy of inertial/magnetic sensors.

    PubMed

    Faber, Gert S; Chang, Chien-Chi; Rizun, Peter; Dennerlein, Jack T

    2013-10-18

    A novel method for assessing the accuracy of inertial/magnetic sensors is presented. The method, referred to as the "residual matrix" method, is advantageous because it decouples the sensor's error with respect to Earth's gravity vector (attitude residual error: pitch and roll) from the sensor's error with respect to magnetic north (heading residual error), while remaining insensitive to singularity problems when the second Euler rotation is close to ±90°. As a demonstration, the accuracy of an inertial/magnetic sensor mounted to a participant's forearm was evaluated during a reaching task in a laboratory. Sensor orientation was measured internally (by the inertial/magnetic sensor) and externally using an optoelectronic measurement system with a marker cluster rigidly attached to the sensor's enclosure. Roll, pitch and heading residuals were calculated using the proposed novel method, as well as using a common orientation assessment method where the residuals are defined as the difference between the Euler angles measured by the inertial sensor and those measured by the optoelectronic system. Using the proposed residual matrix method, the roll and pitch residuals remained less than 1° and, as expected, no statistically significant difference between these two measures of attitude accuracy was found; the heading residuals were significantly larger than the attitude residuals but remained below 2°. Using the direct Euler angle comparison method, the residuals were in general larger due to singularity issues, and the expected significant difference between inertial/magnetic sensor attitude and heading accuracy was not present.

  7. A SVD-based method to assess the uniqueness and accuracy of SPECT geometrical calibration.

    PubMed

    Ma, Tianyu; Yao, Rutao; Shao, Yiping; Zhou, Rong

    2009-12-01

    Geometrical calibration is critical to obtaining high resolution and artifact-free reconstructed image for SPECT and CT systems. Most published calibration methods use analytical approach to determine the uniqueness condition for a specific calibration problem, and the calibration accuracy is often evaluated through empirical studies. In this work, we present a general method to assess the characteristics of both the uniqueness and the quantitative accuracy of the calibration. The method uses a singular value decomposition (SVD) based approach to analyze the Jacobian matrix from a least-square cost function for the calibration. With this method, the uniqueness of the calibration can be identified by assessing the nonsingularity of the Jacobian matrix, and the estimation accuracy of the calibration parameters can be quantified by analyzing the SVD components. A direct application of this method is that the efficacy of a calibration configuration can be quantitatively evaluated by choosing a figure-of-merit, e.g., the minimum required number of projection samplings to achieve desired calibration accuracy. The proposed method was validated with a slit-slat SPECT system through numerical simulation studies and experimental measurements with point sources and an ultra-micro hot-rod phantom. The predicted calibration accuracy from the numerical studies was confirmed by the experimental point source calibrations at approximately 0.1 mm for both the center of rotation (COR) estimation of a rotation stage and the slit aperture position (SAP) estimation of a slit-slat collimator by an optimized system calibration protocol. The reconstructed images of a hot rod phantom showed satisfactory spatial resolution with a proper calibration and showed visible resolution degradation with artificially introduced 0.3 mm COR estimation error. The proposed method can be applied to other SPECT and CT imaging systems to analyze calibration method assessment and calibration protocol

  8. New technology in dietary assessment: a review of digital methods in improving food record accuracy.

    PubMed

    Stumbo, Phyllis J

    2013-02-01

    Methods for conducting dietary assessment in the United States date back to the early twentieth century. Methods of assessment encompassed dietary records, written and spoken dietary recalls, FFQ using pencil and paper and more recently computer and internet applications. Emerging innovations involve camera and mobile telephone technology to capture food and meal images. This paper describes six projects sponsored by the United States National Institutes of Health that use digital methods to improve food records and two mobile phone applications using crowdsourcing. The techniques under development show promise for improving accuracy of food records.

  9. In vivo estimation of the glenohumeral joint centre by functional methods: accuracy and repeatability assessment.

    PubMed

    Lempereur, Mathieu; Leboeuf, Fabien; Brochard, Sylvain; Rousset, Jean; Burdin, Valérie; Rémy-Néris, Olivier

    2010-01-19

    Several algorithms have been proposed for determining the centre of rotation of ball joints. These algorithms are used rather to locate the hip joint centre. Few studies have focused on the determination of the glenohumeral joint centre. However, no studies have assessed the accuracy and repeatability of functional methods for glenohumeral joint centre. This paper aims at evaluating the accuracy and the repeatability with which the glenohumeral joint rotation centre (GHRC) can be estimated in vivo by functional methods. The reference joint centre is the glenohumeral anatomical centre obtained by medical imaging. Five functional methods were tested: the algorithm of Gamage and Lasenby (2002), bias compensated (Halvorsen, 2003), symmetrical centre of rotation estimation (Ehrig et al., 2006), normalization method (Chang and Pollard, 2007), helical axis (Woltring et al., 1985). The glenohumeral anatomical centre (GHAC) was deduced from the fitting of the humeral head. Four subjects performed three cycles of three different movements (flexion/extension, abduction/adduction and circumduction). For each test, the location of the glenohumeral joint centre was estimated by the five methods. Analyses focused on the 3D location, on the repeatability of location and on the accuracy by computing the Euclidian distance between the estimated GHRC and the GHAC. For all the methods, the error repeatability was inferior to 8.25 mm. This study showed that there are significant differences between the five functional methods. The smallest distance between the estimated joint centre and the centre of the humeral head was obtained with the method of Gamage and Lasenby (2002).

  10. Theory and methods for accuracy assessment of thematic maps using fuzzy sets

    SciTech Connect

    Gopal, S.; Woodcock, C. )

    1994-02-01

    The use of fuzzy sets in map accuracy assessment expands the amount of information that can be provided regarding the nature, frequency, magnitude, and source of errors in a thematic map. The need for using fuzzy sets arises from the observation that all map locations do not fit unambiguously in a single map category. Fuzzy sets allow for varying levels of set membership for multiple map categories. A linguistic measurement scale allows the kinds of comments commonly made during map evaluations to be used to quantify map accuracy. Four tables result from the use of fuzzy functions, and when taken together they provide more information than traditional confusion matrices. The use of a hypothetical dataset helps illustrate the benefits of the new methods. It is hoped that the enhanced ability to evaluate maps resulting from the use of fuzzy sets will improve our understanding of uncertainty in maps and facilitate improved error modeling. 40 refs.

  11. How could the replica method improve accuracy of performance assessment of channel coding?

    NASA Astrophysics Data System (ADS)

    Kabashima, Yoshiyuki

    2009-12-01

    We explore the relation between the techniques of statistical mechanics and information theory for assessing the performance of channel coding. We base our study on a framework developed by Gallager in IEEE Trans. Inform. Theory IT-11, 3 (1965), where the minimum decoding error probability is upper-bounded by an average of a generalized Chernoff's bound over a code ensemble. We show that the resulting bound in the framework can be directly assessed by the replica method, which has been developed in statistical mechanics of disordered systems, whereas in Gallager's original methodology further replacement by another bound utilizing Jensen's inequality is necessary. Our approach associates a seemingly ad hoc restriction with respect to an adjustable parameter for optimizing the bound with a phase transition between two replica symmetric solutions, and can improve the accuracy of performance assessments of general code ensembles including low density parity check codes, although its mathematical justification is still open.

  12. Numerical accuracy assessment

    NASA Astrophysics Data System (ADS)

    Boerstoel, J. W.

    1988-12-01

    A framework is provided for numerical accuracy assessment. The purpose of numerical flow simulations is formulated. This formulation concerns the classes of aeronautical configurations (boundaries), the desired flow physics (flow equations and their properties), the classes of flow conditions on flow boundaries (boundary conditions), and the initial flow conditions. Next, accuracy and economical performance requirements are defined; the final numerical flow simulation results of interest should have a guaranteed accuracy, and be produced for an acceptable FLOP-price. Within this context, the validation of numerical processes with respect to the well known topics of consistency, stability, and convergence when the mesh is refined must be done by numerical experimentation because theory gives only partial answers. This requires careful design of text cases for numerical experimentation. Finally, the results of a few recent evaluation exercises of numerical experiments with a large number of codes on a few test cases are summarized.

  13. A Method for Assessing Ground-Truth Accuracy of the 5DCT Technique

    PubMed Central

    Dou, T. H.; Thomas, D. H.; O'Connell, D.; Lamb, J.M.; Lee, P.; Low, D.A.

    2015-01-01

    Purpose To develop a technique that assesses the accuracy of the breathing phase-specific volume image generation process by patient-specific breathing motion model using the original free-breathing CT scans as ground truths. Methods 16 lung cancer patients underwent a previously published protocol in which 25 free-breathing fast helical CT scans were acquired with a simultaneous breathing surrogate. A patient-specific motion model was constructed based on the tissue displacements determined by a state-of-the-art deformable image registration. The first image was arbitrarily selected as the reference image. The motion model was used, along with the free-breathing phase information of the original 25 image datasets, to generate a set of deformation vector fields (DVF) that mapped the reference image to the 24 non-reference images. The high-pitch helically acquired original scans served as ground truths because they captured the instantaneous tissue positions during free breathing. Image similarity between the simulated and the original scans was assessed using deformable registration that evaluated the point-wise discordance throughout the lungs. Results Qualitative comparisons using image overlays showed excellent agreement between the simulated and the original images. Even large 2 cm diaphragm displacements were very well modeled, as was sliding motion across the lung-chest wall boundary. The mean error across the patient cohort was 1.15±0.37 mm, while the mean 95th percentile error was 2.47±0.78 mm. Conclusion The proposed ground truth based technique provided voxel-by-voxel accuracy analysis that could identify organ or tumor-specific motion modeling errors for treatment planning. Despite a large variety of breathing patterns and lung deformations during the free-breathing scanning session, the 5DCT technique was able to accurately reproduce the original helical CT scans, suggesting its applicability to a wide range of patients. PMID:26530763

  14. A Method for Assessing the Accuracy of a Photogrammetry System for Precision Deployable Structures

    NASA Technical Reports Server (NTRS)

    Moore, Ashley

    2005-01-01

    The measurement techniques used to validate analytical models of large deployable structures are an integral Part of the technology development process and must be precise and accurate. Photogrammetry and videogrammetry are viable, accurate, and unobtrusive methods for measuring such large Structures. Photogrammetry uses Software to determine the three-dimensional position of a target using camera images. Videogrammetry is based on the same principle, except a series of timed images are analyzed. This work addresses the accuracy of a digital photogrammetry system used for measurement of large, deployable space structures at JPL. First, photogrammetry tests are performed on a precision space truss test article, and the images are processed using Photomodeler software. The accuracy of the Photomodeler results is determined through, comparison with measurements of the test article taken by an external testing group using the VSTARS photogrammetry system. These two measurements are then compared with Australis photogrammetry software that simulates a measurement test to predict its accuracy. The software is then used to study how particular factors, such as camera resolution and placement, affect the system accuracy to help design the setup for the videogrammetry system that will offer the highest level of accuracy for measurement of deploying structures.

  15. An improved multivariate analytical method to assess the accuracy of acoustic sediment classification maps.

    NASA Astrophysics Data System (ADS)

    Biondo, M.; Bartholomä, A.

    2014-12-01

    High resolution hydro acoustic methods have been successfully employed for the detailed classification of sedimentary habitats. The fine-scale mapping of very heterogeneous, patchy sedimentary facies, and the compound effect of multiple non-linear physical processes on the acoustic signal, cause the classification of backscatter images to be subject to a great level of uncertainty. Standard procedures for assessing the accuracy of acoustic classification maps are not yet established. This study applies different statistical techniques to automated classified acoustic images with the aim of i) quantifying the ability of backscatter to resolve grain size distributions ii) understanding complex patterns influenced by factors other than grain size variations iii) designing innovative repeatable statistical procedures to spatially assess classification uncertainties. A high-frequency (450 kHz) sidescan sonar survey, carried out in the year 2012 in the shallow upper-mesotidal inlet the Jade Bay (German North Sea), allowed to map 100 km2 of surficial sediment with a resolution and coverage never acquired before in the area. The backscatter mosaic was ground-truthed using a large dataset of sediment grab sample information (2009-2011). Multivariate procedures were employed for modelling the relationship between acoustic descriptors and granulometric variables in order to evaluate the correctness of acoustic classes allocation and sediment group separation. Complex patterns in the acoustic signal appeared to be controlled by the combined effect of surface roughness, sorting and mean grain size variations. The area is dominated by silt and fine sand in very mixed compositions; in this fine grained matrix, percentages of gravel resulted to be the prevailing factor affecting backscatter variability. In the absence of coarse material, sorting mostly affected the ability to detect gradual but significant changes in seabed types. Misclassification due to temporal discrepancies

  16. Accuracy Assessment of Crown Delineation Methods for the Individual Trees Using LIDAR Data

    NASA Astrophysics Data System (ADS)

    Chang, K. T.; Lin, C.; Lin, Y. C.; Liu, J. K.

    2016-06-01

    Forest canopy density and height are used as variables in a number of environmental applications, including the estimation of biomass, forest extent and condition, and biodiversity. The airborne Light Detection and Ranging (LiDAR) is very useful to estimate forest canopy parameters according to the generated canopy height models (CHMs). The purpose of this work is to introduce an algorithm to delineate crown parameters, e.g. tree height and crown radii based on the generated rasterized CHMs. And accuracy assessment for the extraction of volumetric parameters of a single tree is also performed via manual measurement using corresponding aerial photo pairs. A LiDAR dataset of a golf course acquired by Leica ALS70-HP is used in this study. Two algorithms, i.e. a traditional one with the subtraction of a digital elevation model (DEM) from a digital surface model (DSM), and a pit-free approach are conducted to generate the CHMs firstly. Then two algorithms, a multilevel morphological active-contour (MMAC) and a variable window filter (VWF), are implemented and used in this study for individual tree delineation. Finally, experimental results of two automatic estimation methods for individual trees can be evaluated with manually measured stand-level parameters, i.e. tree height and crown diameter. The resulting CHM generated by a simple subtraction is full of empty pixels (called "pits") that will give vital impact on subsequent analysis for individual tree delineation. The experimental results indicated that if more individual trees can be extracted, tree crown shape will became more completely in the CHM data after the pit-free process.

  17. Accuracy assessment of the ERP prediction method based on analysis of 100-year ERP series

    NASA Astrophysics Data System (ADS)

    Malkin, Z.; Tissen, V. M.

    2012-12-01

    A new method has been developed at the Siberian Research Institute of Metrology (SNIIM) for highly accurate prediction of UT1 and Pole motion (PM). In this study, a detailed comparison was made of real-time UT1 predictions made in 2006-2011 and PMpredictions made in 2009-2011making use of the SNIIM method with simultaneous predictions computed at the International Earth Rotation and Reference Systems Service (IERS), USNO. Obtained results have shown that proposed method provides better accuracy at different prediction lengths.

  18. [Assessment of overall spatial accuracy in image guided stereotactic body radiotherapy using a spine registration method].

    PubMed

    Nakazawa, Hisato; Uchiyama, Yukio; Komori, Masataka; Hayashi, Naoki

    2014-06-01

    Stereotactic body radiotherapy (SBRT) for lung and liver tumors is always performed under image guidance, a technique used to confirm the accuracy of setup positioning by fusing planning digitally reconstructed radiographs with X-ray, fluoroscopic, or computed tomography (CT) images, using bony structures, tumor shadows, or metallic markers as landmarks. The Japanese SBRT guidelines state that bony spinal structures should be used as the main landmarks for patient setup. In this study, we used the Novalis system as a linear accelerator for SBRT of lung and liver tumors. The current study compared the differences between spine registration and target registration and calculated total spatial accuracy including setup uncertainty derived from our image registration results and the geometric uncertainty of the Novalis system. We were able to evaluate clearly whether overall spatial accuracy is achieved within a setup margin (SM) for planning target volume (PTV) in treatment planning. After being granted approval by the Hospital and University Ethics Committee, we retrospectively analyzed eleven patients with lung tumor and seven patients with liver tumor. The results showed the total spatial accuracy to be within a tolerable range for SM of treatment planning. We therefore regard our method to be suitable for image fusion involving 2-dimensional X-ray images during the treatment planning stage of SBRT for lung and liver tumors.

  19. Accuracy assessment of blind and semi-blind restoration methods for hyperspectral images

    NASA Astrophysics Data System (ADS)

    Zhang, Mo; Vozel, Benoit; Chehdi, Kacem; Uss, Mykhail; Abramov, Sergey; Lukin, Vladimir

    2016-10-01

    Hyperspectral images acquired by remote sensing systems are generally degraded by noise and can be sometimes more severely degraded by blur. When no knowledge is available about the degradations present or the original image, blind restoration methods must be considered. Otherwise, when a partial information is needed, semi-blind restoration methods can be considered. Numerous semi-blind and quite advanced methods are available in the literature. So to get better insights and feedback on the applicability and potential efficiency of a representative set of four semi-blind methods recently proposed, we have performed a comparative study of these methods in objective terms of blur filter and original image error estimation accuracy. In particular, we have paid special attention to the accurate recovering in the spectral dimension of original spectral signatures. We have analyzed peculiarities and factors restricting the applicability of these methods. Our tests are performed on a synthetic hyperspectral image, degraded with various synthetic blurs (out-of-focus, gaussian, motion) and with signal independent noise of typical levels such as those encountered in real hyperspectral images. This synthetic image has been built from various samples from classified areas of a real-life hyperspectral image, in order to benefit from realistic reference spectral signatures to recover after synthetic degradation. Conclusions, practical recommendations and perspectives are drawn from the results experimentally obtained.

  20. Assessing the Accuracy of the Tracer Dilution Method with Atmospheric Dispersion Modeling

    NASA Astrophysics Data System (ADS)

    Taylor, D.; Delkash, M.; Chow, F. K.; Imhoff, P. T.

    2015-12-01

    Landfill methane emissions are difficult to estimate due to limited observations and data uncertainty. The mobile tracer dilution method is a widely used and cost-effective approach for predicting landfill methane emissions. The method uses a tracer gas released on the surface of the landfill and measures the concentrations of both methane and the tracer gas downwind. Mobile measurements are conducted with a gas analyzer mounted on a vehicle to capture transects of both gas plumes. The idea behind the method is that if the measurements are performed far enough downwind, the methane plume from the large area source of the landfill and the tracer plume from a small number of point sources will be sufficiently well-mixed to behave similarly, and the ratio between the concentrations will be a good estimate of the ratio between the two emissions rates. The mobile tracer dilution method is sensitive to different factors of the setup such as placement of the tracer release locations and distance from the landfill to the downwind measurements, which have not been thoroughly examined. In this study, numerical modeling is used as an alternative to field measurements to study the sensitivity of the tracer dilution method and provide estimates of measurement accuracy. Using topography and wind conditions for an actual landfill, a landfill emissions rate is prescribed in the model and compared against the emissions rate predicted by application of the tracer dilution method. Two different methane emissions scenarios are simulated: homogeneous emissions over the entire surface of the landfill, and heterogeneous emissions with a hot spot containing 80% of the total emissions where the daily cover area is located. Numerical modeling of the tracer dilution method is a useful tool for evaluating the method without having the expense and labor commitment of multiple field campaigns. Factors tested include number of tracers, distance between tracers, distance from landfill to transect

  1. Ground Truth Sampling and LANDSAT Accuracy Assessment

    NASA Technical Reports Server (NTRS)

    Robinson, J. W.; Gunther, F. J.; Campbell, W. J.

    1982-01-01

    It is noted that the key factor in any accuracy assessment of remote sensing data is the method used for determining the ground truth, independent of the remote sensing data itself. The sampling and accuracy procedures developed for nuclear power plant siting study are described. The purpose of the sampling procedure was to provide data for developing supervised classifications for two study sites and for assessing the accuracy of that and the other procedures used. The purpose of the accuracy assessment was to allow the comparison of the cost and accuracy of various classification procedures as applied to various data types.

  2. A diagnostic tool for determining the quality of accuracy validation. Assessing the method for determination of nitrate in drinking water.

    PubMed

    Escuder-Gilabert, L; Bonet-Domingo, E; Medina-Hernández, M J; Sagrado, S

    2007-01-01

    Realistic internal validation of a method implies the performance validation experiments under intermediate precision conditions. The validation results can be organized in an X (NrxNs) (replicates x runs) data matrix, analysis of which enables assessment of the accuracy of the method. By means of Monte Carlo simulation, uncertainty in the estimates of bias and precision can be assessed. A bivariate plot is presented for assessing whether the uncertainty intervals for the bias (E +/- U(E)) and intermediate precision (RSDi +/- U(RSDi) are included in prefixed limits (requirements for the method). As a case study, a method for determining the concentration of nitrate in drinking water at the official level set by 98/83/EC Directive is assessed by use of the proposed plot.

  3. Accuracy assessment on the crop area estimating method based on RS sampling at national scale: a case study of China's rice area estimation assessment

    NASA Astrophysics Data System (ADS)

    Qian, Yonglan; Yang, Bangjie; Jiao, Xianfeng; Pei, Zhiyuan; Li, Xuan

    2008-08-01

    Remote Sensing technology has been used in agricultural statistics since early 1970s in developed countries and since late 1970s in China. It has greatly improved the efficiency with its accurate, timingly and credible information. But agricultural monitoring using remote sensing has not yet been assessed with credible data in China and its accuracy seems not consistent and reliable to many users. The paper reviews different methods and the corresponding assessments of agricultural monitoring using remote sensing in developed countries and China, then assesses the crop area estimating method using Landsat TM remotely sensed data as sampling area in Northeast China. The ground truth is ga-thered with global positioning system and 40 sampling areas are used to assess the classification accu-racy. The error matrix is constructed from which the accuracy is calculated. The producer accuracy, the user accuracy and total accuracy are 89.53%, 95.37% and 87.02% respectively and the correlation coefficient between the ground truth and classification results is 0.96. A new error index δ is introduced and the average δ of rice area estimation to the truth data is 0.084. δ measures how much the RS classification result is positive or negative apart from the truth data.

  4. Accuracy of ELISA detection methods for gluten and reference materials: a realistic assessment.

    PubMed

    Diaz-Amigo, Carmen; Popping, Bert

    2013-06-19

    The determination of prolamins by ELISA and subsequent conversion of the resulting concentration to gluten content in food appears to be a comparatively simple and straightforward process with which many laboratories have years-long experience. At the end of the process, a value of gluten, expressed in mg/kg or ppm, is obtained. This value often is the basis for the decision if a product can be labeled gluten-free or not. On the basis of currently available scientific information, the accuracy of the obtained values with commonly used commercial ELISA kits has to be questioned. Although recently several multilaboratory studies have been conducted in an attempt to emphasize and ensure the accuracy of the results, data suggest that it was the precision of these assays, not the accuracy, that was confirmed because some of the underlying assumptions for calculating the gluten content lack scientific data support as well as appropriate reference materials for comparison. This paper discusses the issues of gluten determination and quantification with respect to antibody specificity, extraction procedures, reference materials, and their commutability.

  5. Assessment of the accuracy of plasma shape reconstruction by the Cauchy condition surface method in JT-60SA

    SciTech Connect

    Miyata, Y.; Suzuki, T.; Takechi, M.; Urano, H.; Ide, S.

    2015-07-15

    For the purpose of stable plasma equilibrium control and detailed analysis, it is essential to reconstruct an accurate plasma boundary on the poloidal cross section in tokamak devices. The Cauchy condition surface (CCS) method is a numerical approach for calculating the spatial distribution of the magnetic flux outside a hypothetical surface and reconstructing the plasma boundary from the magnetic measurements located outside the plasma. The accuracy of the plasma shape reconstruction has been assessed by comparing the CCS method and an equilibrium calculation in JT-60SA with a high elongation and triangularity of plasma shape. The CCS, on which both Dirichlet and Neumann conditions are unknown, is defined as a hypothetical surface located inside the real plasma region. The accuracy of the plasma shape reconstruction is sensitive to the CCS free parameters such as the number of unknown parameters and the shape in JT-60SA. It is found that the optimum number of unknown parameters and the size of the CCS that minimizes errors in the reconstructed plasma shape are in proportion to the plasma size. Furthermore, it is shown that the accuracy of the plasma shape reconstruction is greatly improved using the optimum number of unknown parameters and shape of the CCS, and the reachable reconstruction errors in plasma shape and locations of strike points are within the target ranges in JT-60SA.

  6. Assessment of the accuracy of plasma shape reconstruction by the Cauchy condition surface method in JT-60SA

    NASA Astrophysics Data System (ADS)

    Miyata, Y.; Suzuki, T.; Takechi, M.; Urano, H.; Ide, S.

    2015-07-01

    For the purpose of stable plasma equilibrium control and detailed analysis, it is essential to reconstruct an accurate plasma boundary on the poloidal cross section in tokamak devices. The Cauchy condition surface (CCS) method is a numerical approach for calculating the spatial distribution of the magnetic flux outside a hypothetical surface and reconstructing the plasma boundary from the magnetic measurements located outside the plasma. The accuracy of the plasma shape reconstruction has been assessed by comparing the CCS method and an equilibrium calculation in JT-60SA with a high elongation and triangularity of plasma shape. The CCS, on which both Dirichlet and Neumann conditions are unknown, is defined as a hypothetical surface located inside the real plasma region. The accuracy of the plasma shape reconstruction is sensitive to the CCS free parameters such as the number of unknown parameters and the shape in JT-60SA. It is found that the optimum number of unknown parameters and the size of the CCS that minimizes errors in the reconstructed plasma shape are in proportion to the plasma size. Furthermore, it is shown that the accuracy of the plasma shape reconstruction is greatly improved using the optimum number of unknown parameters and shape of the CCS, and the reachable reconstruction errors in plasma shape and locations of strike points are within the target ranges in JT-60SA.

  7. Assessment of the accuracy of plasma shape reconstruction by the Cauchy condition surface method in JT-60SA.

    PubMed

    Miyata, Y; Suzuki, T; Takechi, M; Urano, H; Ide, S

    2015-07-01

    For the purpose of stable plasma equilibrium control and detailed analysis, it is essential to reconstruct an accurate plasma boundary on the poloidal cross section in tokamak devices. The Cauchy condition surface (CCS) method is a numerical approach for calculating the spatial distribution of the magnetic flux outside a hypothetical surface and reconstructing the plasma boundary from the magnetic measurements located outside the plasma. The accuracy of the plasma shape reconstruction has been assessed by comparing the CCS method and an equilibrium calculation in JT-60SA with a high elongation and triangularity of plasma shape. The CCS, on which both Dirichlet and Neumann conditions are unknown, is defined as a hypothetical surface located inside the real plasma region. The accuracy of the plasma shape reconstruction is sensitive to the CCS free parameters such as the number of unknown parameters and the shape in JT-60SA. It is found that the optimum number of unknown parameters and the size of the CCS that minimizes errors in the reconstructed plasma shape are in proportion to the plasma size. Furthermore, it is shown that the accuracy of the plasma shape reconstruction is greatly improved using the optimum number of unknown parameters and shape of the CCS, and the reachable reconstruction errors in plasma shape and locations of strike points are within the target ranges in JT-60SA.

  8. Disease severity estimates - effects of rater accuracy and assessments methods for comparing treatments

    Technology Transfer Automated Retrieval System (TEKTRAN)

    Assessment of disease is fundamental to the discipline of plant pathology, and estimates of severity are often made visually. However, it is established that visual estimates can be inaccurate and unreliable. In this study estimates of Septoria leaf blotch on leaves of winter wheat from non-treated ...

  9. Assessing accuracy of measurements for a Wingate Test using the Taguchi method.

    PubMed

    Franklin, Kathryn L; Gordon, Rae S; Davies, Bruce; Baker, Julien S

    2008-01-01

    The purpose of this study was to establish the effects of four variables on the results obtained for a Wingate Anaerobic Test (WAnT). This study used a 30 second WAnT and compared data collection and analysed in different ways in order to form conclusions as to the relative importance of the variables on the results. Data was collected simultaneously by a commercially available software correction system manufactured by Cranlea Ltd., (Birmingham, England) system and an alternative method of data collection which involves the direct measurement of the flywheel velocity and the brake force. Data was compared using a design of experiments technique, the Taguchi method. Four variables were examined - flywheel speed, braking force, moment of inertia of the flywheel, and time intervals over which the work and power were calculated. The choice of time interval was identified as the most influential variable on the results. While the other factors have an influence on the results, the decreased time interval over which the data is averaged gave 9.8% increase in work done, 40.75% increase in peak power and 13.1% increase in mean power.

  10. Accuracy and Usefulness of Select Methods for Assessing Complete Collection of 24-Hour Urine: A Systematic Review.

    PubMed

    John, Katherine A; Cogswell, Mary E; Campbell, Norm R; Nowson, Caryl A; Legetic, Branka; Hennis, Anselm J M; Patel, Sheena M

    2016-05-01

    Twenty-four-hour urine collection is the recommended method for estimating sodium intake. To investigate the strengths and limitations of methods used to assess completion of 24-hour urine collection, the authors systematically reviewed the literature on the accuracy and usefulness of methods vs para-aminobenzoic acid (PABA) recovery (referent). The percentage of incomplete collections, based on PABA, was 6% to 47% (n=8 studies). The sensitivity and specificity for identifying incomplete collection using creatinine criteria (n=4 studies) was 6% to 63% and 57% to 99.7%, respectively. The most sensitive method for removing incomplete collections was a creatinine index <0.7. In pooled analysis (≥2 studies), mean urine creatinine excretion and volume were higher among participants with complete collection (P<.05); whereas, self-reported collection time did not differ by completion status. Compared with participants with incomplete collection, mean 24-hour sodium excretion was 19.6 mmol higher (n=1781 specimens, 5 studies) in patients with complete collection. Sodium excretion may be underestimated by inclusion of incomplete 24-hour urine collections. None of the current approaches reliably assess completion of 24-hour urine collection.

  11. Dynamic Accuracy of GPS Receivers for Use in Health Research: A Novel Method to Assess GPS Accuracy in Real-World Settings

    PubMed Central

    Schipperijn, Jasper; Kerr, Jacqueline; Duncan, Scott; Madsen, Thomas; Klinker, Charlotte Demant; Troelsen, Jens

    2014-01-01

    The emergence of portable global positioning system (GPS) receivers over the last 10 years has provided researchers with a means to objectively assess spatial position in free-living conditions. However, the use of GPS in free-living conditions is not without challenges and the aim of this study was to test the dynamic accuracy of a portable GPS device under real-world environmental conditions, for four modes of transport, and using three data collection intervals. We selected four routes on different bearings, passing through a variation of environmental conditions in the City of Copenhagen, Denmark, to test the dynamic accuracy of the Qstarz BT-Q1000XT GPS device. Each route consisted of a walk, bicycle, and vehicle lane in each direction. The actual width of each walking, cycling, and vehicle lane was digitized as accurately as possible using ultra-high-resolution aerial photographs as background. For each trip, we calculated the percentage that actually fell within the lane polygon, and within the 2.5, 5, and 10 m buffers respectively, as well as the mean and median error in meters. Our results showed that 49.6% of all ≈68,000 GPS points fell within 2.5 m of the expected location, 78.7% fell within 10 m and the median error was 2.9 m. The median error during walking trips was 3.9, 2.0 m for bicycle trips, 1.5 m for bus, and 0.5 m for car. The different area types showed considerable variation in the median error: 0.7 m in open areas, 2.6 m in half-open areas, and 5.2 m in urban canyons. The dynamic spatial accuracy of the tested device is not perfect, but we feel that it is within acceptable limits for larger population studies. Longer recording periods, for a larger population are likely to reduce the potentially negative effects of measurement inaccuracy. Furthermore, special care should be taken when the environment in which the study takes place could compromise the GPS signal. PMID:24653984

  12. Dynamic Accuracy of GPS Receivers for Use in Health Research: A Novel Method to Assess GPS Accuracy in Real-World Settings.

    PubMed

    Schipperijn, Jasper; Kerr, Jacqueline; Duncan, Scott; Madsen, Thomas; Klinker, Charlotte Demant; Troelsen, Jens

    2014-01-01

    The emergence of portable global positioning system (GPS) receivers over the last 10 years has provided researchers with a means to objectively assess spatial position in free-living conditions. However, the use of GPS in free-living conditions is not without challenges and the aim of this study was to test the dynamic accuracy of a portable GPS device under real-world environmental conditions, for four modes of transport, and using three data collection intervals. We selected four routes on different bearings, passing through a variation of environmental conditions in the City of Copenhagen, Denmark, to test the dynamic accuracy of the Qstarz BT-Q1000XT GPS device. Each route consisted of a walk, bicycle, and vehicle lane in each direction. The actual width of each walking, cycling, and vehicle lane was digitized as accurately as possible using ultra-high-resolution aerial photographs as background. For each trip, we calculated the percentage that actually fell within the lane polygon, and within the 2.5, 5, and 10 m buffers respectively, as well as the mean and median error in meters. Our results showed that 49.6% of all ≈68,000 GPS points fell within 2.5 m of the expected location, 78.7% fell within 10 m and the median error was 2.9 m. The median error during walking trips was 3.9, 2.0 m for bicycle trips, 1.5 m for bus, and 0.5 m for car. The different area types showed considerable variation in the median error: 0.7 m in open areas, 2.6 m in half-open areas, and 5.2 m in urban canyons. The dynamic spatial accuracy of the tested device is not perfect, but we feel that it is within acceptable limits for larger population studies. Longer recording periods, for a larger population are likely to reduce the potentially negative effects of measurement inaccuracy. Furthermore, special care should be taken when the environment in which the study takes place could compromise the GPS signal.

  13. Skinfold Assessment: Accuracy and Application

    ERIC Educational Resources Information Center

    Ball, Stephen; Swan, Pamela D.; Altena, Thomas S.

    2006-01-01

    Although not perfect, skinfolds (SK), or the measurement of fat under the skin, remains the most popular and practical method available to assess body composition on a large scale (Kuczmarski, Flegal, Campbell, & Johnson, 1994). Even for practitioners who have been using SK for years and are highly proficient at locating the correct anatomical…

  14. When Does Choice of Accuracy Measure Alter Imputation Accuracy Assessments?

    PubMed Central

    Ramnarine, Shelina; Zhang, Juan; Chen, Li-Shiun; Culverhouse, Robert; Duan, Weimin; Hancock, Dana B.; Hartz, Sarah M.; Johnson, Eric O.; Olfson, Emily; Schwantes-An, Tae-Hwi; Saccone, Nancy L.

    2015-01-01

    Imputation, the process of inferring genotypes for untyped variants, is used to identify and refine genetic association findings. Inaccuracies in imputed data can distort the observed association between variants and a disease. Many statistics are used to assess accuracy; some compare imputed to genotyped data and others are calculated without reference to true genotypes. Prior work has shown that the Imputation Quality Score (IQS), which is based on Cohen’s kappa statistic and compares imputed genotype probabilities to true genotypes, appropriately adjusts for chance agreement; however, it is not commonly used. To identify differences in accuracy assessment, we compared IQS with concordance rate, squared correlation, and accuracy measures built into imputation programs. Genotypes from the 1000 Genomes reference populations (AFR N = 246 and EUR N = 379) were masked to match the typed single nucleotide polymorphism (SNP) coverage of several SNP arrays and were imputed with BEAGLE 3.3.2 and IMPUTE2 in regions associated with smoking behaviors. Additional masking and imputation was conducted for sequenced subjects from the Collaborative Genetic Study of Nicotine Dependence and the Genetic Study of Nicotine Dependence in African Americans (N = 1,481 African Americans and N = 1,480 European Americans). Our results offer further evidence that concordance rate inflates accuracy estimates, particularly for rare and low frequency variants. For common variants, squared correlation, BEAGLE R2, IMPUTE2 INFO, and IQS produce similar assessments of imputation accuracy. However, for rare and low frequency variants, compared to IQS, the other statistics tend to be more liberal in their assessment of accuracy. IQS is important to consider when evaluating imputation accuracy, particularly for rare and low frequency variants. PMID:26458263

  15. Positional Accuracy Assessment of the Openstreetmap Buildings Layer Through Automatic Homologous Pairs Detection: the Method and a Case Study

    NASA Astrophysics Data System (ADS)

    Brovelli, M. A.; Minghini, M.; Molinari, M. E.; Zamboni, G.

    2016-06-01

    OpenStreetMap (OSM) is currently the largest openly licensed collection of geospatial data. Being OSM increasingly exploited in a variety of applications, research has placed great attention on the assessment of its quality. This work focuses on assessing the quality of OSM buildings. While most of the studies available in literature are limited to the evaluation of OSM building completeness, this work proposes an original approach to assess the positional accuracy of OSM buildings based on comparison with a reference dataset. The comparison relies on a quasi-automated detection of homologous pairs on the two datasets. Based on the homologous pairs found, warping algorithms like e.g. affine transformations and multi-resolution splines can be applied to the OSM buildings to generate a new version having an optimal local match to the reference layer. A quality assessment of the OSM buildings of Milan Municipality (Northern Italy), having an area of about 180 km2, is then presented. After computing some measures of completeness, the algorithm based on homologous points is run using the building layer of the official vector cartography of Milan Municipality as the reference dataset. Approximately 100000 homologous points are found, which show a systematic translation of about 0.4 m on both the X and Y directions and a mean distance of about 0.8 m between the datasets. Besides its efficiency and high degree of automation, the algorithm generates a warped version of OSM buildings which, having by definition a closest match to the reference buildings, can be eventually integrated in the OSM database.

  16. Arizona Vegetation Resource Inventory (AVRI) accuracy assessment

    USGS Publications Warehouse

    Szajgin, John; Pettinger, L.R.; Linden, D.S.; Ohlen, D.O.

    1982-01-01

    A quantitative accuracy assessment was performed for the vegetation classification map produced as part of the Arizona Vegetation Resource Inventory (AVRI) project. This project was a cooperative effort between the Bureau of Land Management (BLM) and the Earth Resources Observation Systems (EROS) Data Center. The objective of the accuracy assessment was to estimate (with a precision of ?10 percent at the 90 percent confidence level) the comission error in each of the eight level II hierarchical vegetation cover types. A stratified two-phase (double) cluster sample was used. Phase I consisted of 160 photointerpreted plots representing clusters of Landsat pixels, and phase II consisted of ground data collection at 80 of the phase I cluster sites. Ground data were used to refine the phase I error estimates by means of a linear regression model. The classified image was stratified by assigning each 15-pixel cluster to the stratum corresponding to the dominant cover type within each cluster. This method is known as stratified plurality sampling. Overall error was estimated to be 36 percent with a standard error of 2 percent. Estimated error for individual vegetation classes ranged from a low of 10 percent ?6 percent for evergreen woodland to 81 percent ?7 percent for cropland and pasture. Total cost of the accuracy assessment was $106,950 for the one-million-hectare study area. The combination of the stratified plurality sampling (SPS) method of sample allocation with double sampling provided the desired estimates within the required precision levels. The overall accuracy results confirmed that highly accurate digital classification of vegetation is difficult to perform in semiarid environments, due largely to the sparse vegetation cover. Nevertheless, these techniques show promise for providing more accurate information than is presently available for many BLM-administered lands.

  17. Presentation accuracy of Web animation methods.

    PubMed

    Schmidt, W C

    2001-05-01

    Several Web animation methods were independently assessed on fast and slow systems running two popular Web browsers under MacOS and Windows. The methods assessed included those requiring programming (Authorware, Java, Javascript/Jscript), browser extensions (Flash and Authorware), or neither (animated GIF). The number of raster scans that an image in an animation was presented for was counted. This was used as an estimate of the minimum presentation time for the image when the software was set to update the animation as quickly as possible. In a second condition, the image was set to be displayed for 100 msec, and differences between observed and expected presentations were used to assess accuracy. In general, all the methods except Java deteriorated as a function of the speed of the computer system, with the poorest temporal resolutions and greatest variability occurring on slower systems. For some animation methods, poor performance was dependent on browser, operating system, system speed, or combinations of these.

  18. Assessing the Accuracy of CME Speed and Trajectory Estimates from STEREO Observations Through a Comparison of Independent Methods

    NASA Astrophysics Data System (ADS)

    Davis, C. J.; Kennedy, J.; Davies, J. A.

    2010-05-01

    We have estimated the speed and direction of propagation of a number of Coronal Mass Ejections (CMEs) using single-spacecraft data from the STEREO Heliospheric Imager (HI) wide-field cameras. In general, these values are in good agreement with those predicted by Thernisien, Vourlidas, and Howard in Solar Phys. 256, 111 - 130 (2009) using a forward modelling method to fit CMEs imaged by the STEREO COR2 coronagraphs. The directions of the CMEs predicted by both techniques are in good agreement despite the fact that many of the CMEs under study travel in directions that cause them to fade rapidly in the HI images. The velocities estimated from both techniques are in general agreement although there are some interesting differences that may provide evidence for the influence of the ambient solar wind on the speed of CMEs. The majority of CMEs with a velocity estimated to be below 400 km s-1 in the COR2 field of view have higher estimated velocities in the HI field of view, while, conversely, those with COR2 velocities estimated to be above 400 km s-1 have lower estimated HI velocities. We interpret this as evidence for the deceleration of fast CMEs and the acceleration of slower CMEs by interaction with the ambient solar wind beyond the COR2 field of view. We also show that the uncertainties in our derived parameters are influenced by the range of elongations over which each CME can be tracked. In order to reduce the uncertainty in the predicted arrival time of a CME at 1 Astronomical Unit (AU) to within six hours, the CME needs to be tracked out to at least 30 degrees elongation. This is in good agreement with predictions of the accuracy of our technique based on Monte Carlo simulations. Within the set of studied CMEs, there are two clear events that were predicted from the HI data to travel over another spacecraft; in-situ measurements at these other spacecraft confirm the accuracy of these predictions. The ability of the HI cameras to image Corotating Interaction

  19. Accuracy of quantitative visual soil assessment

    NASA Astrophysics Data System (ADS)

    van Leeuwen, Maricke; Heuvelink, Gerard; Stoorvogel, Jetse; Wallinga, Jakob; de Boer, Imke; van Dam, Jos; van Essen, Everhard; Moolenaar, Simon; Verhoeven, Frank; Stoof, Cathelijne

    2016-04-01

    Visual soil assessment (VSA) is a method to assess soil quality visually, when standing in the field. VSA is increasingly used by farmers, farm organisations and companies, because it is rapid and cost-effective, and because looking at soil provides understanding about soil functioning. Often VSA is regarded as subjective, so there is a need to verify VSA. Also, many VSAs have not been fine-tuned for contrasting soil types. This could lead to wrong interpretation of soil quality and soil functioning when contrasting sites are compared to each other. We wanted to assess accuracy of VSA, while taking into account soil type. The first objective was to test whether quantitative visual field observations, which form the basis in many VSAs, could be validated with standardized field or laboratory measurements. The second objective was to assess whether quantitative visual field observations are reproducible, when used by observers with contrasting backgrounds. For the validation study, we made quantitative visual observations at 26 cattle farms. Farms were located at sand, clay and peat soils in the North Friesian Woodlands, the Netherlands. Quantitative visual observations evaluated were grass cover, number of biopores, number of roots, soil colour, soil structure, number of earthworms, number of gley mottles and soil compaction. Linear regression analysis showed that four out of eight quantitative visual observations could be well validated with standardized field or laboratory measurements. The following quantitative visual observations correlated well with standardized field or laboratory measurements: grass cover with classified images of surface cover; number of roots with root dry weight; amount of large structure elements with mean weight diameter; and soil colour with soil organic matter content. Correlation coefficients were greater than 0.3, from which half of the correlations were significant. For the reproducibility study, a group of 9 soil scientists and 7

  20. Classification accuracy of actuarial risk assessment instruments.

    PubMed

    Neller, Daniel J; Frederick, Richard I

    2013-01-01

    Users of commonly employed actuarial risk assessment instruments (ARAIs) hope to generate numerical probability statements about risk; however, ARAI manuals often do not explicitly report data that are essential for understanding the classification accuracy of the instruments. In addition, ARAI manuals often contain data that have the potential for misinterpretation. The authors of the present article address the accurate generation of probability statements. First, they illustrate how the reporting of numerical probability statements based on proportions rather than predictive values can mislead users of ARAIs. Next, they report essential test characteristics that, to date, have gone largely unreported in ARAI manuals. Then they discuss a graphing method that can enhance the practice of clinicians who communicate risk via numerical probability statements. After the authors review several strategies for selecting optimal cut-off scores, they show how the graphing method can be used to estimate positive predictive values for each cut-off score of commonly used ARAIs, across all possible base rates. They also show how the graphing method can be used to estimate base rates of violent recidivism in local samples.

  1. Assessment of the Thematic Accuracy of Land Cover Maps

    NASA Astrophysics Data System (ADS)

    Höhle, J.

    2015-08-01

    Several land cover maps are generated from aerial imagery and assessed by different approaches. The test site is an urban area in Europe for which six classes (`building', `hedge and bush', `grass', `road and parking lot', `tree', `wall and car port') had to be derived. Two classification methods were applied (`Decision Tree' and `Support Vector Machine') using only two attributes (height above ground and normalized difference vegetation index) which both are derived from the images. The assessment of the thematic accuracy applied a stratified design and was based on accuracy measures such as user's and producer's accuracy, and kappa coefficient. In addition, confidence intervals were computed for several accuracy measures. The achieved accuracies and confidence intervals are thoroughly analysed and recommendations are derived from the gained experiences. Reliable reference values are obtained using stereovision, false-colour image pairs, and positioning to the checkpoints with 3D coordinates. The influence of the training areas on the results is studied. Cross validation has been tested with a few reference points in order to derive approximate accuracy measures. The two classification methods perform equally for five classes. Trees are classified with a much better accuracy and a smaller confidence interval by means of the decision tree method. Buildings are classified by both methods with an accuracy of 99% (95% CI: 95%-100%) using independent 3D checkpoints. The average width of the confidence interval of six classes was 14% of the user's accuracy.

  2. The accuracy of the 24-h activity recall method for assessing sedentary behaviour: the physical activity measurement survey (PAMS) project.

    PubMed

    Kim, Youngwon; Welk, Gregory J

    2017-02-01

    Sedentary behaviour (SB) has emerged as a modifiable risk factor, but little is known about measurement errors of SB. The purpose of this study was to determine the validity of 24-h Physical Activity Recall (24PAR) relative to SenseWear Armband (SWA) for assessing SB. Each participant (n = 1485) undertook a series of data collection procedures on two randomly selected days: wearing a SWA for full 24-h, and then completing the telephone-administered 24PAR the following day to recall the past 24-h activities. Estimates of total sedentary time (TST) were computed without the inclusion of reported or recorded sleep time. Equivalence testing was used to compare estimates of TST. Analyses from equivalence testing showed no significant equivalence of 24PAR for TST (90% CI: 443.0 and 457.6 min · day(-1)) relative to SWA (equivalence zone: 580.7 and 709.8 min · day(-1)). Bland-Altman plots indicated individuals that were extremely or minimally sedentary provided relatively comparable sedentary time between 24PAR and SWA. Overweight/obese and/or older individuals were more likely to under-estimate sedentary time than normal weight and/or younger individuals. Measurement errors of 24PAR varied by the level of sedentary time and demographic indicators. This evidence informs future work to develop measurement error models to correct for errors of self-reports.

  3. Evaluating the Accuracy of Pharmacy Students' Self-Assessment Skills

    PubMed Central

    Gregory, Paul A. M.

    2007-01-01

    Objectives To evaluate the accuracy of self-assessment skills of senior-level bachelor of science pharmacy students. Methods A method proposed by Kruger and Dunning involving comparisons of pharmacy students' self-assessment with weighted average assessments of peers, standardized patients, and pharmacist-instructors was used. Results Eighty students participated in the study. Differences between self-assessment and external assessments were found across all performance quartiles. These differences were particularly large and significant in the third and fourth (lowest) quartiles and particularly marked in the areas of empathy, and logic/focus/coherence of interviewing. Conclusions The quality and accuracy of pharmacy students' self-assessment skills were not as strong as expected, particularly given recent efforts to include self-assessment in the curriculum. Further work is necessary to ensure this important practice competency and life skill is at the level expected for professional practice and continuous professional development. PMID:17998986

  4. Bullet trajectory reconstruction - Methods, accuracy and precision.

    PubMed

    Mattijssen, Erwin J A T; Kerkhoff, Wim

    2016-05-01

    Based on the spatial relation between a primary and secondary bullet defect or on the shape and dimensions of the primary bullet defect, a bullet's trajectory prior to impact can be estimated for a shooting scene reconstruction. The accuracy and precision of the estimated trajectories will vary depending on variables such as, the applied method of reconstruction, the (true) angle of incidence, the properties of the target material and the properties of the bullet upon impact. This study focused on the accuracy and precision of estimated bullet trajectories when different variants of the probing method, ellipse method, and lead-in method are applied on bullet defects resulting from shots at various angles of incidence on drywall, MDF and sheet metal. The results show that in most situations the best performance (accuracy and precision) is seen when the probing method is applied. Only for the lowest angles of incidence the performance was better when either the ellipse or lead-in method was applied. The data provided in this paper can be used to select the appropriate method(s) for reconstruction and to correct for systematic errors (accuracy) and to provide a value of the precision, by means of a confidence interval of the specific measurement.

  5. Positional Accuracy Assessment of Googleearth in Riyadh

    NASA Astrophysics Data System (ADS)

    Farah, Ashraf; Algarni, Dafer

    2014-06-01

    Google Earth is a virtual globe, map and geographical information program that is controlled by Google corporation. It maps the Earth by the superimposition of images obtained from satellite imagery, aerial photography and GIS 3D globe. With millions of users all around the globe, GoogleEarth® has become the ultimate source of spatial data and information for private and public decision-support systems besides many types and forms of social interactions. Many users mostly in developing countries are also using it for surveying applications, the matter that raises questions about the positional accuracy of the Google Earth program. This research presents a small-scale assessment study of the positional accuracy of GoogleEarth® Imagery in Riyadh; capital of Kingdom of Saudi Arabia (KSA). The results show that the RMSE of the GoogleEarth imagery is 2.18 m and 1.51 m for the horizontal and height coordinates respectively.

  6. Method for measuring centroid algorithm accuracy

    NASA Technical Reports Server (NTRS)

    Klein, S.; Liewer, K.

    2002-01-01

    This paper will describe such a method for measuring the accuracy of centroid algorithms using a relatively inexpensive setup consisting of a white light source, lenses, a CCD camea, an electro-strictive actuator, and a DAC (Digital-to-Analog Converter), and employing embedded PowerPC, VxWorks, and Solaris based software.

  7. Bayesian Methods for Medical Test Accuracy

    PubMed Central

    Broemeling, Lyle D.

    2011-01-01

    Bayesian methods for medical test accuracy are presented, beginning with the basic measures for tests with binary scores: true positive fraction, false positive fraction, positive predictive values, and negative predictive value. The Bayesian approach is taken because of its efficient use of prior information, and the analysis is executed with a Bayesian software package WinBUGS®. The ROC (receiver operating characteristic) curve gives the intrinsic accuracy of medical tests that have ordinal or continuous scores, and the Bayesian approach is illustrated with many examples from cancer and other diseases. Medical tests include X-ray, mammography, ultrasound, computed tomography, magnetic resonance imaging, nuclear medicine and tests based on biomarkers, such as blood glucose values for diabetes. The presentation continues with more specialized methods suitable for measuring the accuracies of clinical studies that have verification bias, and medical tests without a gold standard. Lastly, the review is concluded with Bayesian methods for measuring the accuracy of the combination of two or more tests. PMID:26859485

  8. Assessing the accuracy and reliability of ultrasonographic three-dimensional parathyroid volume measurement in a patient with secondary hyperparathyroidism: a comparison with the two-dimensional conventional method

    PubMed Central

    2017-01-01

    Purpose The purpose of this study was to investigate the accuracy and reliability of the semi-automated ultrasonographic volume measurement tool, virtual organ computer-aided analysis (VOCAL), for measuring the volume of parathyroid glands. Methods Volume measurements for 40 parathyroid glands were performed in patients with secondary hyperparathyroidism caused by chronic renal failure. The volume of the parathyroid glands was measured twice by experienced radiologists by two-dimensional (2D) and three-dimensional (3D) methods using conventional sonograms and the VOCAL with 30°angle increments before parathyroidectomy. The specimen volume was also measured postoperatively. Intraclass correlation coefficients (ICCs) and the absolute percentage error were used for estimating the reproducibility and accuracy of the two different methods. Results The ICC value between two measurements of the 2D method and the 3D method was 0.956 and 0.999, respectively. The mean absolute percentage error of the 2D method and the 3D VOCAL technique was 29.56% and 5.78%, respectively. For accuracy and reliability, the plots of the 3D method showed a more compact distribution than those of the 2D method on the Bland-Altman graph. Conclusion The rotational VOCAL method for measuring the parathyroid gland is more accurate and reliable than the conventional 2D measurement. This VOCAL method could be used as a more reliable follow-up imaging modality in a patient with hyperparathyroidism. PMID:27457337

  9. Accuracy Assessment of Altimeter Derived Geostrophic Velocities

    NASA Astrophysics Data System (ADS)

    Leben, R. R.; Powell, B. S.; Born, G. H.; Guinasso, N. L.

    2002-12-01

    Along track sea surface height anomaly gradients are proportional to cross track geostrophic velocity anomalies allowing satellite altimetry to provide much needed satellite observations of changes in the geostrophic component of surface ocean currents. Often, surface height gradients are computed from altimeter data archives that have been corrected to give the most accurate absolute sea level, a practice that may unnecessarily increase the error in the cross track velocity anomalies and thereby require excessive smoothing to mitigate noise. Because differentiation along track acts as a high-pass filter, many of the path length corrections applied to altimeter data for absolute height accuracy are unnecessary for the corresponding gradient calculations. We report on a study to investigate appropriate altimetric corrections and processing techniques for improving geostrophic velocity accuracy. Accuracy is assessed by comparing cross track current measurements from two moorings placed along the descending TOPEX/POSEIDON ground track number 52 in the Gulf of Mexico to the corresponding altimeter velocity estimates. The buoys are deployed and maintained by the Texas Automated Buoy System (TABS) under Interagency Contracts with Texas A&M University. The buoys telemeter observations in near real-time via satellite to the TABS station located at the Geochemical and Environmental Research Group (GERG) at Texas A&M. Buoy M is located in shelf waters of 57 m depth with a second, Buoy N, 38 km away on the shelf break at 105 m depth. Buoy N has been operational since the beginning of 2002 and has a current meter at 2m depth providing in situ measurements of surface velocities coincident with Jason and TOPEX/POSEIDON altimeter over flights. This allows one of the first detailed comparisons of shallow water near surface current meter time series to coincident altimetry.

  10. Quantifying optimal accuracy of local primary sequence bioinformatics methods

    NASA Astrophysics Data System (ADS)

    Aalberts, Daniel

    2005-03-01

    Traditional bioinformatics methods scan primary sequences for local patterns. It is important to assess how accurate local primary sequence methods can be. We study the problem of donor pre-mRNA splice site recognition, where the sequence overlaps between real and decoy data sets can be quantified, exposing the intrinsic limitations of the performance of local primary sequence methods. We assess the accuracy of local primary sequence methods generally by studying how they scale with dataset size and demonstrate that our new Primary Sequence Ranking methods have superior performance. Our Primary Sequence Ranking analysis tools are available at tt http://rna.williams.edu/

  11. [Navigation in implantology: Accuracy assessment regarding the literature].

    PubMed

    Barrak, Ibrahim Ádám; Varga, Endre; Piffko, József

    2016-06-01

    Our objective was to assess the literature regarding the accuracy of the different static guided systems. After applying electronic literature search we found 661 articles. After reviewing 139 articles, the authors chose 52 articles for full-text evaluation. 24 studies involved accuracy measurements. Fourteen of our selected references were clinical and ten of them were in vitro (modell or cadaver). Variance-analysis (Tukey's post-hoc test; p < 0.05) was conducted to summarize the selected publications. Regarding 2819 results the average mean error at the entry point was 0.98 mm. At the level of the apex the average deviation was 1.29 mm while the mean of the angular deviation was 3,96 degrees. Significant difference could be observed between the two methods of implant placement (partially and fully guided sequence) in terms of deviation at the entry point, apex and angular deviation. Different levels of quality and quantity of evidence were available for assessing the accuracy of the different computer-assisted implant placement. The rapidly evolving field of digital dentistry and the new developments will further improve the accuracy of guided implant placement. In the interest of being able to draw dependable conclusions and for the further evaluation of the parameters used for accuracy measurements, randomized, controlled single or multi-centered clinical trials are necessary.

  12. Clinical assessment of intraarterial blood gas monitor accuracy

    NASA Astrophysics Data System (ADS)

    Aziz, Salim; Spiess, R.; Roby, Paul; Kenny, Margaret

    1993-08-01

    The accuracy of intraarterial blood gas monitoring (IABGM) devices is challenging to assess under routine clinical conditions. When comparing discrete measurements by blood gas analyzer (BGA) to IABGM values, it is important that the BGA determinations (reference method) be as accurate as possible. In vitro decay of gas tensions caused by delay in BGA analysis is particularly problematic for specimens with high arterial oxygen tension (PaO2) values. Clinical instability of blood gases in the acutely ill patient may cause disagreement between BGA and IABGM values because of IABGM response time lag, particularly in the measurement of arterial blood carbon dioxide tension (PaCO2). We recommend that clinical assessments of IABGM accuracy by comparison with BGA use multiple bedside BGA instruments, and that blood sampling only occur during periods when IABGM values appear stable.

  13. Scalable Methods for Uncertainty Quantification, Data Assimilation and Target Accuracy Assessment for Multi-Physics Advanced Simulation of Light Water Reactors

    NASA Astrophysics Data System (ADS)

    Khuwaileh, Bassam

    High fidelity simulation of nuclear reactors entails large scale applications characterized with high dimensionality and tremendous complexity where various physics models are integrated in the form of coupled models (e.g. neutronic with thermal-hydraulic feedback). Each of the coupled modules represents a high fidelity formulation of the first principles governing the physics of interest. Therefore, new developments in high fidelity multi-physics simulation and the corresponding sensitivity/uncertainty quantification analysis are paramount to the development and competitiveness of reactors achieved through enhanced understanding of the design and safety margins. Accordingly, this dissertation introduces efficient and scalable algorithms for performing efficient Uncertainty Quantification (UQ), Data Assimilation (DA) and Target Accuracy Assessment (TAA) for large scale, multi-physics reactor design and safety problems. This dissertation builds upon previous efforts for adaptive core simulation and reduced order modeling algorithms and extends these efforts towards coupled multi-physics models with feedback. The core idea is to recast the reactor physics analysis in terms of reduced order models. This can be achieved via identifying the important/influential degrees of freedom (DoF) via the subspace analysis, such that the required analysis can be recast by considering the important DoF only. In this dissertation, efficient algorithms for lower dimensional subspace construction have been developed for single physics and multi-physics applications with feedback. Then the reduced subspace is used to solve realistic, large scale forward (UQ) and inverse problems (DA and TAA). Once the elite set of DoF is determined, the uncertainty/sensitivity/target accuracy assessment and data assimilation analysis can be performed accurately and efficiently for large scale, high dimensional multi-physics nuclear engineering applications. Hence, in this work a Karhunen-Loeve (KL

  14. Accuracy Assessment for AG500, Electromagnetic Articulograph

    ERIC Educational Resources Information Center

    Yunusova, Yana; Green, Jordan R.; Mefferd, Antje

    2009-01-01

    Purpose: The goal of this article was to evaluate the accuracy and reliability of the AG500 (Carstens Medizinelectronik, Lenglern, Germany), an electromagnetic device developed recently to register articulatory movements in three dimensions. This technology seems to have unprecedented capabilities to provide rich information about time-varying…

  15. Accuracy assessment of fluoroscopy-transesophageal echocardiography registration

    NASA Astrophysics Data System (ADS)

    Lang, Pencilla; Seslija, Petar; Bainbridge, Daniel; Guiraudon, Gerard M.; Jones, Doug L.; Chu, Michael W.; Holdsworth, David W.; Peters, Terry M.

    2011-03-01

    This study assesses the accuracy of a new transesophageal (TEE) ultrasound (US) fluoroscopy registration technique designed to guide percutaneous aortic valve replacement. In this minimally invasive procedure, a valve is inserted into the aortic annulus via a catheter. Navigation and positioning of the valve is guided primarily by intra-operative fluoroscopy. Poor anatomical visualization of the aortic root region can result in incorrect positioning, leading to heart valve embolization, obstruction of the coronary ostia and acute kidney injury. The use of TEE US images to augment intra-operative fluoroscopy provides significant improvements to image-guidance. Registration is achieved using an image-based TEE probe tracking technique and US calibration. TEE probe tracking is accomplished using a single-perspective pose estimation algorithm. Pose estimation from a single image allows registration to be achieved using only images collected in standard OR workflow. Accuracy of this registration technique is assessed using three models: a point target phantom, a cadaveric porcine heart with implanted fiducials, and in-vivo porcine images. Results demonstrate that registration can be achieved with an RMS error of less than 1.5mm, which is within the clinical accuracy requirements of 5mm. US-fluoroscopy registration based on single-perspective pose estimation demonstrates promise as a method for providing guidance to percutaneous aortic valve replacement procedures. Future work will focus on real-time implementation and a visualization system that can be used in the operating room.

  16. Assessment of accuracy of the structure-factor-size-estimator method in determining red blood cell aggregate size from ultrasound spectral backscatter coefficient.

    PubMed

    Saha, Ratan K; Franceschini, Emilie; Cloutier, Guy

    2011-04-01

    A computer simulation study to produce ultrasonic backscatter coefficients (BSCs) from red blood cell (RBC) clusters is discussed. The simulation algorithm is suitable for generating non-overlapping, isotropic, and fairly identical RBC clusters. RBCs were stacked following the hexagonal close packing (HCP) structure to form a compact spherical aggregate. Such an aggregate was repeated and placed randomly under non-overlapping condition in the three-dimensional space to mimic an aggregated blood sample. BSCs were computed between 750 KHz and 200 MHz for samples of various cluster sizes at different hematocrits. Magnitudes of BSCs increased with mean aggregate sizes at low frequencies (<20 MHz). The accuracy of the structure-factor-size-estimator (SFSE) method in determining mean aggregate size and packing factor was also examined. A good correlation (R(2) ≥ 0.94) between the mean size of aggregates predicted by the SFSE and true size was found for each hematocrit. This study shows that for spherical aggregates there exists a region for each hematocrit where SFSE works most accurately. Typically, error of SFSE in estimating mean cluster size was <20% for dimensions between 14 and 17 μm at 40% hematocrit. This study suggests that the theoretical framework of SFSE is valid under the assumption of isotropic aggregates.

  17. Accuracy assessment of NLCD 2006 land cover and impervious surface

    USGS Publications Warehouse

    Wickham, James D.; Stehman, Stephen V.; Gass, Leila; Dewitz, Jon; Fry, Joyce A.; Wade, Timothy G.

    2013-01-01

    Release of NLCD 2006 provides the first wall-to-wall land-cover change database for the conterminous United States from Landsat Thematic Mapper (TM) data. Accuracy assessment of NLCD 2006 focused on four primary products: 2001 land cover, 2006 land cover, land-cover change between 2001 and 2006, and impervious surface change between 2001 and 2006. The accuracy assessment was conducted by selecting a stratified random sample of pixels with the reference classification interpreted from multi-temporal high resolution digital imagery. The NLCD Level II (16 classes) overall accuracies for the 2001 and 2006 land cover were 79% and 78%, respectively, with Level II user's accuracies exceeding 80% for water, high density urban, all upland forest classes, shrubland, and cropland for both dates. Level I (8 classes) accuracies were 85% for NLCD 2001 and 84% for NLCD 2006. The high overall and user's accuracies for the individual dates translated into high user's accuracies for the 2001–2006 change reporting themes water gain and loss, forest loss, urban gain, and the no-change reporting themes for water, urban, forest, and agriculture. The main factor limiting higher accuracies for the change reporting themes appeared to be difficulty in distinguishing the context of grass. We discuss the need for more research on land-cover change accuracy assessment.

  18. Alaska national hydrography dataset positional accuracy assessment study

    USGS Publications Warehouse

    Arundel, Samantha; Yamamoto, Kristina H.; Constance, Eric; Mantey, Kim; Vinyard-Houx, Jeremy

    2013-01-01

    Initial visual assessments Wide range in the quality of fit between features in NHD and these new image sources. No statistical analysis has been performed to actually quantify accuracy Determining absolute accuracy is cost prohibitive (must collect independent, well defined test points) Quantitative analysis of relative positional error is feasible.

  19. Inertial Measures of Motion for Clinical Biomechanics: Comparative Assessment of Accuracy under Controlled Conditions – Changes in Accuracy over Time

    PubMed Central

    Lebel, Karina; Boissy, Patrick; Hamel, Mathieu; Duval, Christian

    2015-01-01

    Background Interest in 3D inertial motion tracking devices (AHRS) has been growing rapidly among the biomechanical community. Although the convenience of such tracking devices seems to open a whole new world of possibilities for evaluation in clinical biomechanics, its limitations haven’t been extensively documented. The objectives of this study are: 1) to assess the change in absolute and relative accuracy of multiple units of 3 commercially available AHRS over time; and 2) to identify different sources of errors affecting AHRS accuracy and to document how they may affect the measurements over time. Methods This study used an instrumented Gimbal table on which AHRS modules were carefully attached and put through a series of velocity-controlled sustained motions including 2 minutes motion trials (2MT) and 12 minutes multiple dynamic phases motion trials (12MDP). Absolute accuracy was assessed by comparison of the AHRS orientation measurements to those of an optical gold standard. Relative accuracy was evaluated using the variation in relative orientation between modules during the trials. Findings Both absolute and relative accuracy decreased over time during 2MT. 12MDP trials showed a significant decrease in accuracy over multiple phases, but accuracy could be enhanced significantly by resetting the reference point and/or compensating for initial Inertial frame estimation reference for each phase. Interpretation The variation in AHRS accuracy observed between the different systems and with time can be attributed in part to the dynamic estimation error, but also and foremost, to the ability of AHRS units to locate the same Inertial frame. Conclusions Mean accuracies obtained under the Gimbal table sustained conditions of motion suggest that AHRS are promising tools for clinical mobility assessment under constrained conditions of use. However, improvement in magnetic compensation and alignment between AHRS modules are desirable in order for AHRS to reach their

  20. Assessing Predictive Accuracy in Discriminant Analysis.

    ERIC Educational Resources Information Center

    Huberty, Carl J.; And Others

    1987-01-01

    Three estimates of the probabilities of correct classification in predictive discriminant analysis were computed using mathematical formulas, resubstitution, and external analyses: (1) optimal hit rate; (2) actual hit rate; and (3) expected actual hit rate. Methods were compared using Monte Carlo sampling from two data sets. (Author/GDC)

  1. Thematic Accuracy Assessment of the 2011 National Land ...

    EPA Pesticide Factsheets

    Accuracy assessment is a standard protocol of National Land Cover Database (NLCD) mapping. Here we report agreement statistics between map and reference labels for NLCD 2011, which includes land cover for ca. 2001, ca. 2006, and ca. 2011. The two main objectives were assessment of agreement between map and reference labels for the three, single-date NLCD land cover products at Level II and Level I of the classification hierarchy, and agreement for 17 land cover change reporting themes based on Level I classes (e.g., forest loss; forest gain; forest, no change) for three change periods (2001–2006, 2006–2011, and 2001–2011). The single-date overall accuracies were 82%, 83%, and 83% at Level II and 88%, 89%, and 89% at Level I for 2011, 2006, and 2001, respectively. Many class-specific user's accuracies met or exceeded a previously established nominal accuracy benchmark of 85%. Overall accuracies for 2006 and 2001 land cover components of NLCD 2011 were approximately 4% higher (at Level II and Level I) than the overall accuracies for the same components of NLCD 2006. The high Level I overall, user's, and producer's accuracies for the single-date eras in NLCD 2011 did not translate into high class-specific user's and producer's accuracies for many of the 17 change reporting themes. User's accuracies were high for the no change reporting themes, commonly exceeding 85%, but were typically much lower for the reporting themes that represented change. Only forest l

  2. On the accuracy assessment of Laplacian models in MPS

    NASA Astrophysics Data System (ADS)

    Ng, K. C.; Hwang, Y. H.; Sheu, T. W. H.

    2014-10-01

    From the basis of the Gauss divergence theorem applied on a circular control volume that was put forward by Isshiki (2011) in deriving the MPS-based differential operators, a more general Laplacian model is further deduced from the current work which involves the proposal of an altered kernel function. The Laplacians of several functions are evaluated and the accuracies of various MPS Laplacian models in solving the Poisson equation that is subjected to both Dirichlet and Neumann boundary conditions are assessed. For regular grids, the Laplacian model with smaller N is generally more accurate, owing to the reduction of leading errors due to those higher-order derivatives appearing in the modified equation. For irregular grids, an optimal N value does exist in ensuring better global accuracy, in which this optimal value of N will increase when cases employing highly irregular grids are computed. Finally, the accuracies of these MPS Laplacian models are assessed in an incompressible flow problem.

  3. ASSESSING ACCURACY OF NET CHANGE DERIVED FROM LAND COVER MAPS

    EPA Science Inventory

    Net change derived from land-cover maps provides important descriptive information for environmental monitoring and is often used as an input or explanatory variable in environmental models. The sampling design and analysis for assessing net change accuracy differ from traditio...

  4. Estimating Classification Consistency and Accuracy for Cognitive Diagnostic Assessment

    ERIC Educational Resources Information Center

    Cui, Ying; Gierl, Mark J.; Chang, Hua-Hua

    2012-01-01

    This article introduces procedures for the computation and asymptotic statistical inference for classification consistency and accuracy indices specifically designed for cognitive diagnostic assessments. The new classification indices can be used as important indicators of the reliability and validity of classification results produced by…

  5. Evaluating clinical accuracy of continuous glucose monitoring devices: other methods.

    PubMed

    Wentholt, Iris M E; Hart, August A; Hoekstra, Joost B L; DeVries, J Hans

    2008-08-01

    With more and more continuous glucose monitoring devices entering the market, the importance of adequate accuracy assessment grows. This review discusses pros and cons of Regression Analysis and Correlation Coefficient, Relative Difference measures, Bland Altman plot, ISO criteria, combined curve fitting, and epidemiological analyses, the latter including sensitivity, specificity and positive predictive value for hypoglycaemia. Finally, recommendations for much needed head-to-head studies are given. This paper is a revised and adapted version of How to assess and compare the accuracy of continuous glucose monitors?, Diabetes Technology and Therapeutics 2007, in press, published with permission of the editor.

  6. Bilingual Language Assessment: A Meta-Analysis of Diagnostic Accuracy

    ERIC Educational Resources Information Center

    Dollaghan, Christine A.; Horner, Elizabeth A.

    2011-01-01

    Purpose: To describe quality indicators for appraising studies of diagnostic accuracy and to report a meta-analysis of measures for diagnosing language impairment (LI) in bilingual Spanish-English U.S. children. Method: The authors searched electronically and by hand to locate peer-reviewed English-language publications meeting inclusion criteria;…

  7. Assessing genomic selection prediction accuracy in a dynamic barley breeding

    Technology Transfer Automated Retrieval System (TEKTRAN)

    Genomic selection is a method to improve quantitative traits in crops and livestock by estimating breeding values of selection candidates using phenotype and genome-wide marker data sets. Prediction accuracy has been evaluated through simulation and cross-validation, however validation based on prog...

  8. Standardized accuracy assessment of the calypso wireless transponder tracking system

    NASA Astrophysics Data System (ADS)

    Franz, A. M.; Schmitt, D.; Seitel, A.; Chatrasingh, M.; Echner, G.; Oelfke, U.; Nill, S.; Birkfellner, W.; Maier-Hein, L.

    2014-11-01

    Electromagnetic (EM) tracking allows localization of small EM sensors in a magnetic field of known geometry without line-of-sight. However, this technique requires a cable connection to the tracked object. A wireless alternative based on magnetic fields, referred to as transponder tracking, has been proposed by several authors. Although most of the transponder tracking systems are still in an early stage of development and not ready for clinical use yet, Varian Medical Systems Inc. (Palo Alto, California, USA) presented the Calypso system for tumor tracking in radiation therapy which includes transponder technology. But it has not been used for computer-assisted interventions (CAI) in general or been assessed for accuracy in a standardized manner, so far. In this study, we apply a standardized assessment protocol presented by Hummel et al (2005 Med. Phys. 32 2371-9) to the Calypso system for the first time. The results show that transponder tracking with the Calypso system provides a precision and accuracy below 1 mm in ideal clinical environments, which is comparable with other EM tracking systems. Similar to other systems the tracking accuracy was affected by metallic distortion, which led to errors of up to 3.2 mm. The potential of the wireless transponder tracking technology for use in many future CAI applications can be regarded as extremely high.

  9. A Framework for the Objective Assessment of Registration Accuracy

    PubMed Central

    Simonetti, Flavio; Foroni, Roberto Israel

    2014-01-01

    Validation and accuracy assessment are the main bottlenecks preventing the adoption of image processing algorithms in the clinical practice. In the classical approach, a posteriori analysis is performed through objective metrics. In this work, a different approach based on Petri nets is proposed. The basic idea consists in predicting the accuracy of a given pipeline based on the identification and characterization of the sources of inaccuracy. The concept is demonstrated on a case study: intrasubject rigid and affine registration of magnetic resonance images. Both synthetic and real data are considered. While synthetic data allow the benchmarking of the performance with respect to the ground truth, real data enable to assess the robustness of the methodology in real contexts as well as to determine the suitability of the use of synthetic data in the training phase. Results revealed a higher correlation and a lower dispersion among the metrics for simulated data, while the opposite trend was observed for pathologic ones. Results show that the proposed model not only provides a good prediction performance but also leads to the optimization of the end-to-end chain in terms of accuracy and robustness, setting the ground for its generalization to different and more complex scenarios. PMID:24659997

  10. Assessing the accuracy of self-reported self-talk

    PubMed Central

    Brinthaupt, Thomas M.; Benson, Scott A.; Kang, Minsoo; Moore, Zaver D.

    2015-01-01

    As with most kinds of inner experience, it is difficult to assess actual self-talk frequency beyond self-reports, given the often hidden and subjective nature of the phenomenon. The Self-Talk Scale (STS; Brinthaupt et al., 2009) is a self-report measure of self-talk frequency that has been shown to possess acceptable reliability and validity. However, no research using the STS has examined the accuracy of respondents’ self-reports. In the present paper, we report a series of studies directly examining the measurement of self-talk frequency and functions using the STS. The studies examine ways to validate self-reported self-talk by (1) comparing STS responses from 6 weeks earlier to recent experiences that might precipitate self-talk, (2) using experience sampling methods to determine whether STS scores are related to recent reports of self-talk over a period of a week, and (3) comparing self-reported STS scores to those provided by a significant other who rated the target on the STS. Results showed that (1) overall self-talk scores, particularly self-critical and self-reinforcing self-talk, were significantly related to reports of context-specific self-talk; (2) high STS scorers reported talking to themselves significantly more often during recent events compared to low STS scorers, and, contrary to expectations, (3) friends reported less agreement than strangers in their self-other self-talk ratings. Implications of the results for the validity of the STS and for measuring self-talk are presented. PMID:25999887

  11. Accuracy of activPAL Self-Attachment Methods

    ERIC Educational Resources Information Center

    Kringen, Nina L.; Healy, Genevieve N.; Winkler, Elisabeth A. H.; Clark, Bronwyn K.

    2016-01-01

    This study examined the accuracy of self-attachment of the activPAL activity monitor. A convenience sample of 50 participants self-attached the monitor after being presented with written material only (WMO) and then written and video (WV) instructions; and completed a questionnaire regarding the acceptability of the instructional methods.…

  12. The Social Accuracy Model of Interpersonal Perception: Assessing Individual Differences in Perceptive and Expressive Accuracy

    ERIC Educational Resources Information Center

    Biesanz, Jeremy C.

    2010-01-01

    The social accuracy model of interpersonal perception (SAM) is a componential model that estimates perceiver and target effects of different components of accuracy across traits simultaneously. For instance, Jane may be generally accurate in her perceptions of others and thus high in "perceptive accuracy"--the extent to which a particular…

  13. Effect of casting methods on accuracy of peridental restorations.

    PubMed

    Finger, W; Kota, K

    1982-06-01

    The present study has shown that the accuracy of peridental gold alloy castings depends 1) on the type of casting machine used, 2) on the diameter of the casting sprue, and 3) on the strength properties of the investment material. The dependence between the accuracy and the three factors mentioned is based on erosion of the investment mold by the inflow of the liquid casting alloy. The vacuum casting technique proved to be a more gentle casting method than centrifugal and vacuum/pressure techniques.

  14. Comparative assessment of thematic accuracy of GLC maps for specific applications using existing reference data

    NASA Astrophysics Data System (ADS)

    Tsendbazar, N. E.; de Bruin, S.; Mora, B.; Schouten, L.; Herold, M.

    2016-02-01

    Inputs to various applications and models, current global land cover (GLC) maps are based on different data sources and methods. Therefore, comparing GLC maps is challenging. Statistical comparison of GLC maps is further complicated by the lack of a reference dataset that is suitable for validating multiple maps. This study utilizes the existing Globcover-2005 reference dataset to compare thematic accuracies of three GLC maps for the year 2005 (Globcover, LC-CCI and MODIS). We translated and reinterpreted the LCCS (land cover classification system) classifier information of the reference dataset into the different map legends. The three maps were evaluated for a variety of applications, i.e., general circulation models, dynamic global vegetation models, agriculture assessments, carbon estimation and biodiversity assessments, using weighted accuracy assessment. Based on the impact of land cover confusions on the overall weighted accuracy of the GLC maps, we identified map improvement priorities. Overall accuracies were 70.8 ± 1.4%, 71.4 ± 1.3%, and 61.3 ± 1.5% for LC-CCI, MODIS, and Globcover, respectively. Weighted accuracy assessments produced increased overall accuracies (80-93%) since not all class confusion errors are important for specific applications. As a common denominator for all applications, the classes mixed trees, shrubs, grasses, and cropland were identified as improvement priorities. The results demonstrate the necessity of accounting for dissimilarities in the importance of map classification errors for different user application. To determine the fitness of use of GLC maps, accuracy of GLC maps should be assessed per application; there is no single-figure accuracy estimate expressing map fitness for all purposes.

  15. Effect of calibration method on Tekscan sensor accuracy.

    PubMed

    Brimacombe, Jill M; Wilson, David R; Hodgson, Antony J; Ho, Karen C T; Anglin, Carolyn

    2009-03-01

    Tekscan pressure sensors are used in biomechanics research to measure joint contact loads. While the overall accuracy of these sensors has been reported previously, the effects of different calibration algorithms on sensor accuracy have not been compared. The objectives of this validation study were to determine the most appropriate calibration method supplied in the Tekscan program software and to compare its accuracy to the accuracy obtained with two user-defined calibration protocols. We evaluated the calibration accuracies for test loads within the low range, high range, and full range of the sensor. Our experimental setup used materials representing those found in standard prosthetic joints, i.e., metal against plastic. The Tekscan power calibration was the most accurate of the algorithms provided with the system software, with an overall rms error of 2.7% of the tested sensor range, whereas the linear calibrations resulted in an overall rms error of up to 24% of the tested range. The user-defined ten-point cubic calibration was almost five times more accurate, on average, than the power calibration over the full range, with an overall rms error of 0.6% of the tested range. The user-defined three-point quadratic calibration was almost twice as accurate as the Tekscan power calibration, but was sensitive to the calibration loads used. We recommend that investigators design their own calibration curves not only to improve accuracy but also to understand the range(s) of highest error and to choose the optimal points within the expected sensing range for calibration. Since output and sensor nonlinearity depend on the experimental protocol (sensor type, interface shape and materials, sensor range in use, loading method, etc.), sensor behavior should be investigated for each different application.

  16. Accuracy Assessment of a Uav-Based Landslide Monitoring System

    NASA Astrophysics Data System (ADS)

    Peppa, M. V.; Mills, J. P.; Moore, P.; Miller, P. E.; Chambers, J. E.

    2016-06-01

    Landslides are hazardous events with often disastrous consequences. Monitoring landslides with observations of high spatio-temporal resolution can help mitigate such hazards. Mini unmanned aerial vehicles (UAVs) complemented by structure-from-motion (SfM) photogrammetry and modern per-pixel image matching algorithms can deliver a time-series of landslide elevation models in an automated and inexpensive way. This research investigates the potential of a mini UAV, equipped with a Panasonic Lumix DMC-LX5 compact camera, to provide surface deformations at acceptable levels of accuracy for landslide assessment. The study adopts a self-calibrating bundle adjustment-SfM pipeline using ground control points (GCPs). It evaluates misalignment biases and unresolved systematic errors that are transferred through the SfM process into the derived elevation models. To cross-validate the research outputs, results are compared to benchmark observations obtained by standard surveying techniques. The data is collected with 6 cm ground sample distance (GSD) and is shown to achieve planimetric and vertical accuracy of a few centimetres at independent check points (ICPs). The co-registration error of the generated elevation models is also examined in areas of stable terrain. Through this error assessment, the study estimates that the vertical sensitivity to real terrain change of the tested landslide is equal to 9 cm.

  17. Accuracy assessment of seven global land cover datasets over China

    NASA Astrophysics Data System (ADS)

    Yang, Yongke; Xiao, Pengfeng; Feng, Xuezhi; Li, Haixing

    2017-03-01

    Land cover (LC) is the vital foundation to Earth science. Up to now, several global LC datasets have arisen with efforts of many scientific communities. To provide guidelines for data usage over China, nine LC maps from seven global LC datasets (IGBP DISCover, UMD, GLC, MCD12Q1, GLCNMO, CCI-LC, and GlobeLand30) were evaluated in this study. First, we compared their similarities and discrepancies in both area and spatial patterns, and analysed their inherent relations to data sources and classification schemes and methods. Next, five sets of validation sample units (VSUs) were collected to calculate their accuracy quantitatively. Further, we built a spatial analysis model and depicted their spatial variation in accuracy based on the five sets of VSUs. The results show that, there are evident discrepancies among these LC maps in both area and spatial patterns. For LC maps produced by different institutes, GLC 2000 and CCI-LC 2000 have the highest overall spatial agreement (53.8%). For LC maps produced by same institutes, overall spatial agreement of CCI-LC 2000 and 2010, and MCD12Q1 2001 and 2010 reach up to 99.8% and 73.2%, respectively; while more efforts are still needed if we hope to use these LC maps as time series data for model inputting, since both CCI-LC and MCD12Q1 fail to represent the rapid changing trend of several key LC classes in the early 21st century, in particular urban and built-up, snow and ice, water bodies, and permanent wetlands. With the highest spatial resolution, the overall accuracy of GlobeLand30 2010 is 82.39%. For the other six LC datasets with coarse resolution, CCI-LC 2010/2000 has the highest overall accuracy, and following are MCD12Q1 2010/2001, GLC 2000, GLCNMO 2008, IGBP DISCover, and UMD in turn. Beside that all maps exhibit high accuracy in homogeneous regions; local accuracies in other regions are quite different, particularly in Farming-Pastoral Zone of North China, mountains in Northeast China, and Southeast Hills. Special

  18. Diagnostic accuracy assessment of cytopathological examination of feline sporotrichosis.

    PubMed

    Jessica, N; Sonia, R L; Rodrigo, C; Isabella, D F; Tânia, M P; Jeferson, C; Anna, B F; Sandro, A

    2015-11-01

    Sporotrichosis is an implantation mycosis caused by pathogenic species of Sporothrix schenckii complex that affects humans and animals, especially cats. Its main forms of zoonotic transmission include scratching, biting and/or contact with the exudate from lesions of sick cats. In Brazil, epidemic involving humans, dogs and cats has occurred since 1998. The definitive diagnosis of sporotrichosis is obtained by the isolation of the fungus in culture; however, the result can take up to four weeks, which may delay the beginning of antifungal treatment in some cases. Cytopathological examination is often used in feline sporotrichosis diagnosis, but accuracy parameters have not been established yet. The aim of this study was to evaluate the accuracy and reliability of cytopathological examination in the diagnosis of feline sporotrichosis. The present study included 244 cats from the metropolitan region of Rio de Janeiro, mostly males in reproductive age with three or more lesions in non-adjacent anatomical places. To evaluate the inter-observer reliability, two different observers performed the microscopic examination of the slides blindly. Test sensitivity was 84.9%. The values of positive predictive value, negative predictive value, positive likelihood ratio, negative likelihood ratio and accuracy were 86.0, 24.4, 2.02, 0.26 and 82.8%, respectively. The reliability between the two observers was considered substantial. We conclude that the cytopathological examination is a sensitive, rapid and practical method to be used in feline sporotrichosis diagnosis in outbreaks of this mycosis.

  19. Accuracy assessment of gridded precipitation datasets in the Himalayas

    NASA Astrophysics Data System (ADS)

    Khan, A.

    2015-12-01

    Accurate precipitation data are vital for hydro-climatic modelling and water resources assessments. Based on mass balance calculations and Turc-Budyko analysis, this study investigates the accuracy of twelve widely used precipitation gridded datasets for sub-basins in the Upper Indus Basin (UIB) in the Himalayas-Karakoram-Hindukush (HKH) region. These datasets are: 1) Global Precipitation Climatology Project (GPCP), 2) Climate Prediction Centre (CPC) Merged Analysis of Precipitation (CMAP), 3) NCEP / NCAR, 4) Global Precipitation Climatology Centre (GPCC), 5) Climatic Research Unit (CRU), 6) Asian Precipitation Highly Resolved Observational Data Integration Towards Evaluation of Water Resources (APHRODITE), 7) Tropical Rainfall Measuring Mission (TRMM), 8) European Reanalysis (ERA) interim data, 9) PRINCETON, 10) European Reanalysis-40 (ERA-40), 11) Willmott and Matsuura, and 12) WATCH Forcing Data based on ERA interim (WFDEI). Precipitation accuracy and consistency was assessed by physical mass balance involving sum of annual measured flow, estimated actual evapotranspiration (average of 4 datasets), estimated glacier mass balance melt contribution (average of 4 datasets), and ground water recharge (average of 3 datasets), during 1999-2010. Mass balance assessment was complemented by Turc-Budyko non-dimensional analysis, where annual precipitation, measured flow and potential evapotranspiration (average of 5 datasets) data were used for the same period. Both analyses suggest that all tested precipitation datasets significantly underestimate precipitation in the Karakoram sub-basins. For the Hindukush and Himalayan sub-basins most datasets underestimate precipitation, except ERA-interim and ERA-40. The analysis indicates that for this large region with complicated terrain features and stark spatial precipitation gradients the reanalysis datasets have better consistency with flow measurements than datasets derived from records of only sparsely distributed climatic

  20. Accuracy of Revised and Traditional Parallel Analyses for Assessing Dimensionality with Binary Data

    ERIC Educational Resources Information Center

    Green, Samuel B.; Redell, Nickalus; Thompson, Marilyn S.; Levy, Roy

    2016-01-01

    Parallel analysis (PA) is a useful empirical tool for assessing the number of factors in exploratory factor analysis. On conceptual and empirical grounds, we argue for a revision to PA that makes it more consistent with hypothesis testing. Using Monte Carlo methods, we evaluated the relative accuracy of the revised PA (R-PA) and traditional PA…

  1. A note on the accuracy of spectral method applied to nonlinear conservation laws

    NASA Technical Reports Server (NTRS)

    Shu, Chi-Wang; Wong, Peter S.

    1994-01-01

    Fourier spectral method can achieve exponential accuracy both on the approximation level and for solving partial differential equations if the solutions are analytic. For a linear partial differential equation with a discontinuous solution, Fourier spectral method produces poor point-wise accuracy without post-processing, but still maintains exponential accuracy for all moments against analytic functions. In this note we assess the accuracy of Fourier spectral method applied to nonlinear conservation laws through a numerical case study. We find that the moments with respect to analytic functions are no longer very accurate. However the numerical solution does contain accurate information which can be extracted by a post-processing based on Gegenbauer polynomials.

  2. Accuracy assessment of EPA protocol gases purchased in 1991

    SciTech Connect

    Coppedge, E.A.; Logan, T.J.; Midgett, M.R.; Shores, R.C.; Messner, M.J.

    1992-12-01

    The U.S. Environmental Protection Agency (EPA) has established quality assurance procedures for air pollution measurement systems that are intended to reduce the uncertainty in environmental measurements. The compressed gas standards of the program are used for calibration and audits of continuous emission monitoring systems. EPA's regulations require that the certified values for these standards be traceable to National Institute of Standards and Technology (NIST) Standard Reference Materials or to NIST/EPA-approved Certified Reference Materials via either of two traceability protocols. The manufacturer assessment was conducted to: (1) document the accuracy of the compressed gas standards' certified concentrations; and (2) ensure that the compressed gas standards' written certification reports met the documentation requirements of the protocol. All available sources were contacted and the following gas mixtures were acquired: (1) 300-ppm SO2 and 400-ppm NO in N2; and (2) 1500-ppm SO2 and 900-ppm NO in N2.

  3. High accuracy operon prediction method based on STRING database scores.

    PubMed

    Taboada, Blanca; Verde, Cristina; Merino, Enrique

    2010-07-01

    We present a simple and highly accurate computational method for operon prediction, based on intergenic distances and functional relationships between the protein products of contiguous genes, as defined by STRING database (Jensen,L.J., Kuhn,M., Stark,M., Chaffron,S., Creevey,C., Muller,J., Doerks,T., Julien,P., Roth,A., Simonovic,M. et al. (2009) STRING 8-a global view on proteins and their functional interactions in 630 organisms. Nucleic Acids Res., 37, D412-D416). These two parameters were used to train a neural network on a subset of experimentally characterized Escherichia coli and Bacillus subtilis operons. Our predictive model was successfully tested on the set of experimentally defined operons in E. coli and B. subtilis, with accuracies of 94.6 and 93.3%, respectively. As far as we know, these are the highest accuracies ever obtained for predicting bacterial operons. Furthermore, in order to evaluate the predictable accuracy of our model when using an organism's data set for the training procedure, and a different organism's data set for testing, we repeated the E. coli operon prediction analysis using a neural network trained with B. subtilis data, and a B. subtilis analysis using a neural network trained with E. coli data. Even for these cases, the accuracies reached with our method were outstandingly high, 91.5 and 93%, respectively. These results show the potential use of our method for accurately predicting the operons of any other organism. Our operon predictions for fully-sequenced genomes are available at http://operons.ibt.unam.mx/OperonPredictor/.

  4. Increasing accuracy in the assessment of motion sickness: A construct methodology

    NASA Technical Reports Server (NTRS)

    Stout, Cynthia S.; Cowings, Patricia S.

    1993-01-01

    The purpose is to introduce a new methodology that should improve the accuracy of the assessment of motion sickness. This construct methodology utilizes both subjective reports of motion sickness and objective measures of physiological correlates to assess motion sickness. Current techniques and methods used in the framework of a construct methodology are inadequate. Current assessment techniques for diagnosing motion sickness and space motion sickness are reviewed, and attention is called to the problems with the current methods. Further, principles of psychophysiology that when applied will probably resolve some of these problems are described in detail.

  5. Estimated Accuracy of Three Common Trajectory Statistical Methods

    NASA Technical Reports Server (NTRS)

    Kabashnikov, Vitaliy P.; Chaikovsky, Anatoli P.; Kucsera, Tom L.; Metelskaya, Natalia S.

    2011-01-01

    Three well-known trajectory statistical methods (TSMs), namely concentration field (CF), concentration weighted trajectory (CWT), and potential source contribution function (PSCF) methods were tested using known sources and artificially generated data sets to determine the ability of TSMs to reproduce spatial distribution of the sources. In the works by other authors, the accuracy of the trajectory statistical methods was estimated for particular species and at specified receptor locations. We have obtained a more general statistical estimation of the accuracy of source reconstruction and have found optimum conditions to reconstruct source distributions of atmospheric trace substances. Only virtual pollutants of the primary type were considered. In real world experiments, TSMs are intended for application to a priori unknown sources. Therefore, the accuracy of TSMs has to be tested with all possible spatial distributions of sources. An ensemble of geographical distributions of virtual sources was generated. Spearman s rank order correlation coefficient between spatial distributions of the known virtual and the reconstructed sources was taken to be a quantitative measure of the accuracy. Statistical estimates of the mean correlation coefficient and a range of the most probable values of correlation coefficients were obtained. All the TSMs that were considered here showed similar close results. The maximum of the ratio of the mean correlation to the width of the correlation interval containing the most probable correlation values determines the optimum conditions for reconstruction. An optimal geographical domain roughly coincides with the area supplying most of the substance to the receptor. The optimal domain s size is dependent on the substance decay time. Under optimum reconstruction conditions, the mean correlation coefficients can reach 0.70 0.75. The boundaries of the interval with the most probable correlation values are 0.6 0.9 for the decay time of 240 h

  6. Accuracy assessment of a surface electromyogram decomposition system in human first dorsal interosseus muscle

    NASA Astrophysics Data System (ADS)

    Hu, Xiaogang; Rymer, William Z.; Suresh, Nina L.

    2014-04-01

    Objective. The aim of this study is to assess the accuracy of a surface electromyogram (sEMG) motor unit (MU) decomposition algorithm during low levels of muscle contraction. Approach. A two-source method was used to verify the accuracy of the sEMG decomposition system, by utilizing simultaneous intramuscular and surface EMG recordings from the human first dorsal interosseous muscle recorded during isometric trapezoidal force contractions. Spike trains from each recording type were decomposed independently utilizing two different algorithms, EMGlab and dEMG decomposition algorithms. The degree of agreement of the decomposed spike timings was assessed for three different segments of the EMG signals, corresponding to specified regions in the force task. A regression analysis was performed to examine whether certain properties of the sEMG and force signal can predict the decomposition accuracy. Main results. The average accuracy of successful decomposition among the 119 MUs that were common to both intramuscular and surface records was approximately 95%, and the accuracy was comparable between the different segments of the sEMG signals (i.e., force ramp-up versus steady state force versus combined). The regression function between the accuracy and properties of sEMG and force signals revealed that the signal-to-noise ratio of the action potential and stability in the action potential records were significant predictors of the surface decomposition accuracy. Significance. The outcomes of our study confirm the accuracy of the sEMG decomposition algorithm during low muscle contraction levels and provide confidence in the overall validity of the surface dEMG decomposition algorithm.

  7. Method for improving accuracy in full evaporation headspace analysis.

    PubMed

    Xie, Wei-Qi; Chai, Xin-Sheng

    2017-03-21

    We report a new headspace analytical method in which multiple headspace extraction is incorporated with the full evaporation technique. The pressure uncertainty caused by the solid content change in the samples has a great impact to the measurement accuracy in the conventional full evaporation headspace analysis. The results (using ethanol solution as the model sample) showed that the present technique is effective to minimize such a problem. The proposed full evaporation multiple headspace extraction analysis technique is also automated and practical, and which could greatly broaden the applications of the full-evaporation-based headspace analysis. This article is protected by copyright. All rights reserved.

  8. Airborne Topographic Mapper Calibration Procedures and Accuracy Assessment

    NASA Technical Reports Server (NTRS)

    Martin, Chreston F.; Krabill, William B.; Manizade, Serdar S.; Russell, Rob L.; Sonntag, John G.; Swift, Robert N.; Yungel, James K.

    2012-01-01

    Description of NASA Airborn Topographic Mapper (ATM) lidar calibration procedures including analysis of the accuracy and consistancy of various ATM instrument parameters and the resulting influence on topographic elevation measurements. The ATM elevations measurements from a nominal operating altitude 500 to 750 m above the ice surface was found to be: Horizontal Accuracy 74 cm, Horizontal Precision 14 cm, Vertical Accuracy 6.6 cm, Vertical Precision 3 cm.

  9. Researches on High Accuracy Prediction Methods of Earth Orientation Parameters

    NASA Astrophysics Data System (ADS)

    Xu, X. Q.

    2015-09-01

    The Earth rotation reflects the coupling process among the solid Earth, atmosphere, oceans, mantle, and core of the Earth on multiple spatial and temporal scales. The Earth rotation can be described by the Earth's orientation parameters, which are abbreviated as EOP (mainly including two polar motion components PM_X and PM_Y, and variation in the length of day ΔLOD). The EOP is crucial in the transformation between the terrestrial and celestial reference systems, and has important applications in many areas such as the deep space exploration, satellite precise orbit determination, and astrogeodynamics. However, the EOP products obtained by the space geodetic technologies generally delay by several days to two weeks. The growing demands for modern space navigation make high-accuracy EOP prediction be a worthy topic. This thesis is composed of the following three aspects, for the purpose of improving the EOP forecast accuracy. (1) We analyze the relation between the length of the basic data series and the EOP forecast accuracy, and compare the EOP prediction accuracy for the linear autoregressive (AR) model and the nonlinear artificial neural network (ANN) method by performing the least squares (LS) extrapolations. The results show that the high precision forecast of EOP can be realized by appropriate selection of the basic data series length according to the required time span of EOP prediction: for short-term prediction, the basic data series should be shorter, while for the long-term prediction, the series should be longer. The analysis also showed that the LS+AR model is more suitable for the short-term forecasts, while the LS+ANN model shows the advantages in the medium- and long-term forecasts. (2) We develop for the first time a new method which combines the autoregressive model and Kalman filter (AR+Kalman) in short-term EOP prediction. The equations of observation and state are established using the EOP series and the autoregressive coefficients

  10. Novel method for high accuracy figure measurement of optical flat

    NASA Astrophysics Data System (ADS)

    E, Kewei; Li, Dahai; Yang, Lijie; Guo, Guangrao; Li, Mengyang; Wang, Xuemin; Zhang, Tao; Xiong, Zhao

    2017-01-01

    Phase Measuring Deflectometry (PMD) is a non-contact, high dynamic-range and full-field metrology which becomes a serious competitor to interferometry. However, the accuracy of deflectometry metrology is strongly influenced by the level of the calibrations, including test geometry, imaging pin-hole camera and digital display. In this paper, we propose a novel method that can measure optical flat surface figure to a high accuracy. We first calibrate the camera using a checker pattern shown on a LCD display at six different orientations, and the last orientation is aligned at the same position as the test optical flat. By using this method, lens distortions and the mapping relationship between the CCD pixels and the subaperture coordinates on the test optical flat can be determined at the same time. To further reduce the influence of the calibration errors on measurements, a reference optical flat with a high quality surface is measured, and then the system errors in our PMD setup can be eliminated by subtracting the figure of the reference flat from the figure of the test flat. Although any expensive coordinates measuring machine, such as laser tracker and coordinates measuring machine are not applied in our measurement, our experimental results of optical flat figure from low to high order aberrations still show a good agreement with that from the Fizeau interferometer.

  11. Assessment Of Accuracies Of Remote-Sensing Maps

    NASA Technical Reports Server (NTRS)

    Card, Don H.; Strong, Laurence L.

    1992-01-01

    Report describes study of accuracies of classifications of picture elements in map derived by digital processing of Landsat-multispectral-scanner imagery of coastal plain of Arctic National Wildlife Refuge. Accuracies of portions of map analyzed with help of statistical sampling procedure called "stratified plurality sampling", in which all picture elements in given cluster classified in stratum to which plurality of them belong.

  12. Assessing and Ensuring GOES-R Magnetometer Accuracy

    NASA Technical Reports Server (NTRS)

    Carter, Delano R.; Todirita, Monica; Kronenwetter, Jeffrey; Chu, Donald

    2016-01-01

    The GOES-R magnetometer subsystem accuracy requirement is 1.7 nanoteslas (nT). During quiet times (100 nT), accuracy is defined as absolute mean plus 3 sigma. During storms (300 nT), accuracy is defined as absolute mean plus 2 sigma. Error comes both from outside the magnetometers, e.g. spacecraft fields and misalignments, as well as inside, e.g. zero offset and scale factor errors. Because zero offset and scale factor drift over time, it will be necessary to perform annual calibration maneuvers. To predict performance before launch, we have used Monte Carlo simulations and covariance analysis. Both behave as expected, and their accuracy predictions agree within 30%. With the proposed calibration regimen, both suggest that the GOES-R magnetometer subsystem will meet its accuracy requirements.

  13. Assessing and Ensuring GOES-R Magnetometer Accuracy

    NASA Technical Reports Server (NTRS)

    Kronenwetter, Jeffrey; Carter, Delano R.; Todirita, Monica; Chu, Donald

    2016-01-01

    The GOES-R magnetometer accuracy requirement is 1.7 nanoteslas (nT). During quiet times (100 nT), accuracy is defined as absolute mean plus 3 sigma. During storms (300 nT), accuracy is defined as absolute mean plus 2 sigma. To achieve this, the sensor itself has better than 1 nT accuracy. Because zero offset and scale factor drift over time, it is also necessary to perform annual calibration maneuvers. To predict performance, we used covariance analysis and attempted to corroborate it with simulations. Although not perfect, the two generally agree and show the expected behaviors. With the annual calibration regimen, these predictions suggest that the magnetometers will meet their accuracy requirements.

  14. After Detection: The Improved Accuracy of Lung Cancer Assessment Using Radiologic Computer-aided Diagnosis

    PubMed Central

    Amir, Guy J.; Lehmann, Harold P.

    2015-01-01

    Rationale and Objectives The aim of this study was to evaluate the improved accuracy of radiologic assessment of lung cancer afforded by computer-aided diagnosis (CADx). Materials and Methods Inclusion/exclusion criteria were formulated, and a systematic inquiry of research databases was conducted. Following title and abstract review, an in-depth review of 149 surviving articles was performed with accepted articles undergoing a Quality Assessment of Diagnostic Accuracy Studies (QUADAS)-based quality review and data abstraction. Results A total of 14 articles, representing 1868 scans, passed the review. Increases in the receiver operating characteristic (ROC) area under the curve of .8 or higher were seen in all nine studies that reported it, except for one that employed subspecialized radiologists. Conclusions This systematic review demonstrated improved accuracy of lung cancer assessment using CADx over manual review, in eight high-quality observer-performance studies. The improved accuracy afforded by radiologic lung-CADx suggests the need to explore its use in screening and regular clinical workflow. PMID:26616209

  15. Accuracy Assessment of Coastal Topography Derived from Uav Images

    NASA Astrophysics Data System (ADS)

    Long, N.; Millescamps, B.; Pouget, F.; Dumon, A.; Lachaussée, N.; Bertin, X.

    2016-06-01

    To monitor coastal environments, Unmanned Aerial Vehicle (UAV) is a low-cost and easy to use solution to enable data acquisition with high temporal frequency and spatial resolution. Compared to Light Detection And Ranging (LiDAR) or Terrestrial Laser Scanning (TLS), this solution produces Digital Surface Model (DSM) with a similar accuracy. To evaluate the DSM accuracy on a coastal environment, a campaign was carried out with a flying wing (eBee) combined with a digital camera. Using the Photoscan software and the photogrammetry process (Structure From Motion algorithm), a DSM and an orthomosaic were produced. Compared to GNSS surveys, the DSM accuracy is estimated. Two parameters are tested: the influence of the methodology (number and distribution of Ground Control Points, GCPs) and the influence of spatial image resolution (4.6 cm vs 2 cm). The results show that this solution is able to reproduce the topography of a coastal area with a high vertical accuracy (< 10 cm). The georeferencing of the DSM require a homogeneous distribution and a large number of GCPs. The accuracy is correlated with the number of GCPs (use 19 GCPs instead of 10 allows to reduce the difference of 4 cm); the required accuracy should be dependant of the research problematic. Last, in this particular environment, the presence of very small water surfaces on the sand bank does not allow to improve the accuracy when the spatial resolution of images is decreased.

  16. Forecasting space weather: Can new econometric methods improve accuracy?

    NASA Astrophysics Data System (ADS)

    Reikard, Gordon

    2011-06-01

    Space weather forecasts are currently used in areas ranging from navigation and communication to electric power system operations. The relevant forecast horizons can range from as little as 24 h to several days. This paper analyzes the predictability of two major space weather measures using new time series methods, many of them derived from econometrics. The data sets are the A p geomagnetic index and the solar radio flux at 10.7 cm. The methods tested include nonlinear regressions, neural networks, frequency domain algorithms, GARCH models (which utilize the residual variance), state transition models, and models that combine elements of several techniques. While combined models are complex, they can be programmed using modern statistical software. The data frequency is daily, and forecasting experiments are run over horizons ranging from 1 to 7 days. Two major conclusions stand out. First, the frequency domain method forecasts the A p index more accurately than any time domain model, including both regressions and neural networks. This finding is very robust, and holds for all forecast horizons. Combining the frequency domain method with other techniques yields a further small improvement in accuracy. Second, the neural network forecasts the solar flux more accurately than any other method, although at short horizons (2 days or less) the regression and net yield similar results. The neural net does best when it includes measures of the long-term component in the data.

  17. Improving the Accuracy of High-Order Nodal Transport Methods

    SciTech Connect

    Azmy, Y.Y.; Buscaglia, G.C.; Zamonsky, O.M.

    1999-09-27

    This paper outlines some recent advances towards improving the accuracy of neutron transport calculations using the Arbitrarily High Order Transport-Nodal (AHOT-N) Method. These advances consist of several contributions: (a) A formula for the spatial weights that allows for the polynomial order to be raised arbitrarily high without suffering adverse effects from round-off error; (b) A reconstruction technique for the angular flux, based upon a recursive formula, that reduces the pointwise error by one ordeq (c) An a posterior error indicator that estimates the true error and its distribution throughout the domain, so that it can be used for adaptively refining the approximation. Present results are mainly for ID, extension to 2D-3D is in progress.

  18. Improving the Accuracy of High-Order Nodal Transport Methods

    SciTech Connect

    Azmy, Y.Y.; Buscaglia, G.C.; Zamonsky, O.M.

    1999-09-27

    This paper outlines some recent advances towards improving the accuracy of neutron calculations using the Arbitrarily High Order Transport-Nodal (AHOT-N) Method. These transport advances consist of several contributions: (a) A formula for the spatial weights that allows for the polynomial order to be raised arbitrarily high without suffering from pollution from round-off, error; (b) A reconstruction technique for the angular flux, based upon a recursive formula, that reduces the pointwise error by one order; (c) An a posterior error indicator that estimates the true error and its distribution throughout the domain, so that it can be used for adaptively reftig the approximation. Present results are mainly for ID, extension to 2D-3D is in progress.

  19. Assessing and ensuring GOES-R magnetometer accuracy

    NASA Astrophysics Data System (ADS)

    Carter, Delano; Todirita, Monica; Kronenwetter, Jeffrey; Dahya, Melissa; Chu, Donald

    2016-05-01

    The GOES-R magnetometer subsystem accuracy requirement is 1.7 nanoteslas (nT). During quiet times (100 nT), accuracy is defined as absolute mean plus 3 sigma error per axis. During storms (300 nT), accuracy is defined as absolute mean plus 2 sigma error per axis. Error comes both from outside the magnetometers, e.g. spacecraft fields and misalignments, as well as inside, e.g. zero offset and scale factor errors. Because zero offset and scale factor drift over time, it will be necessary to perform annual calibration maneuvers. To predict performance before launch, we have used Monte Carlo simulations and covariance analysis. With the proposed calibration regimen, both suggest that the magnetometer subsystem will meet its accuracy requirements.

  20. Assessment of the Accuracy of Close Distance Photogrammetric JRC Data

    NASA Astrophysics Data System (ADS)

    Kim, Dong Hyun; Poropat, George; Gratchev, Ivan; Balasubramaniam, Arumugam

    2016-11-01

    By using close range photogrammetry, this article investigates the accuracy of the photogrammetric estimation of rock joint roughness coefficients (JRC), a measure of the degree of roughness of rock joint surfaces. This methodology has proven to be convenient both in laboratory and in site conditions. However, the accuracy and precision of roughness profiles obtained from photogrammetric 3D images have not been properly established due to the variances caused by factors such as measurement errors and systematic errors in photogrammetry. In this study, the influences of camera-to-object distance, focal length and profile orientation on the accuracy of JRC values are investigated using several photogrammetry field surveys. Directional photogrammetric JRC data are compared with data derived from the measured profiles, so as to determine their accuracy. The extent of the accuracy of JRC values was examined based on the error models which were previously developed from laboratory tests and revised for better estimation in this study. The results show that high-resolution 3D images (point interval ≤1 mm) can reduce the JRC errors obtained from field photogrammetric surveys. Using the high-resolution images, the photogrammetric JRC values in the range of high oblique camera angles are highly consistent with the revised error models. Therefore, the analysis indicates that the revised error models facilitate the verification of the accuracy of photogrammetric JRC values.

  1. The ADI-FDTD method for high accuracy electrophysics applications

    NASA Astrophysics Data System (ADS)

    Haeri Kermani, Mohammad

    The Finite-Difference Time-Domain (FDTD) is a dependable method to simulate a wide range of problems from acoustics, to electromagnetics, and to photonics, amongst others. The execution time of an FDTD simulation is inversely proportional to the time-step size. Since the FDTD method is explicit, its time-step size is limited by the well-known Courant-Friedrich-Levy (CFL) stability limit. The CFL stability limit can render the simulation inefficient for very fine structures. The Alternating Direction Implicit FDTD (ADI-FDTD) method has been introduced as an unconditionally stable implicit method. Numerous works have shown that the ADI-FDTD method is stable even when the CFL stability limit is exceeded. Therefore, the ADI-FDTD method can be considered an efficient method for special classes of problems with very fine structures or high gradient fields. Whenever the ADI-FDTD method is used to simulate open-region radiation or scattering problems, the implementation of a mesh-truncation scheme or absorbing boundary condition becomes an integral part of the simulation. These truncation techniques represent, in essence, differential operators that are discretized using a distinct differencing scheme which can potentially affect the stability of the scheme used for the interior region. In this work, we show that the ADI-FDTD method can be rendered unstable when higher-order mesh truncation techniques such as Higdon's Absorbing Boundary Condition (ABC) or Complementary Derivatives Method (COM) are used. When having large field gradients within a limited volume, a non-uniform grid can reduce the computational domain and, therefore, it decreases the computational cost of the FDTD method. However, for high-accuracy problems, different grid sizes increase the truncation error at the boundary of domains having different grid sizes. To address this problem, we introduce the Complementary Derivatives Method (CDM), a second-order accurate interpolation scheme. The CDM theory is

  2. Standardizing the protocol for hemispherical photographs: accuracy assessment of binarization algorithms.

    PubMed

    Glatthorn, Jonas; Beckschäfer, Philip

    2014-01-01

    Hemispherical photography is a well-established method to optically assess ecological parameters related to plant canopies; e.g. ground-level light regimes and the distribution of foliage within the crown space. Interpreting hemispherical photographs involves classifying pixels as either sky or vegetation. A wide range of automatic thresholding or binarization algorithms exists to classify the photographs. The variety in methodology hampers ability to compare results across studies. To identify an optimal threshold selection method, this study assessed the accuracy of seven binarization methods implemented in software currently available for the processing of hemispherical photographs. Therefore, binarizations obtained by the algorithms were compared to reference data generated through a manual binarization of a stratified random selection of pixels. This approach was adopted from the accuracy assessment of map classifications known from remote sensing studies. Percentage correct (Pc) and kappa-statistics (K) were calculated. The accuracy of the algorithms was assessed for photographs taken with automatic exposure settings (auto-exposure) and photographs taken with settings which avoid overexposure (histogram-exposure). In addition, gap fraction values derived from hemispherical photographs were compared with estimates derived from the manually classified reference pixels. All tested algorithms were shown to be sensitive to overexposure. Three of the algorithms showed an accuracy which was high enough to be recommended for the processing of histogram-exposed hemispherical photographs: "Minimum" (Pc 98.8%; K 0.952), "Edge Detection" (Pc 98.1%; K 0.950), and "Minimum Histogram" (Pc 98.1%; K 0.947). The Minimum algorithm overestimated gap fraction least of all (11%). The overestimation by the algorithms Edge Detection (63%) and Minimum Histogram (67%) were considerably larger. For the remaining four evaluated algorithms (IsoData, Maximum Entropy, MinError, and Otsu

  3. Improving the accuracy of weight status assessment in infancy research.

    PubMed

    Dixon, Wallace E; Dalton, William T; Berry, Sarah M; Carroll, Vincent A

    2014-08-01

    Both researchers and primary care providers vary in their methods for assessing weight status in infants. The purpose of the present investigation was to compare standing-height-derived to recumbent-length-derived weight-for-length standardized (WLZ) scores, using the WHO growth curves, in a convenience sample of infants who visited the lab at 18 and 21 months of age. Fifty-eight primarily White, middle class infants (25 girls) from a semi-rural region of southern Appalachia visited the lab at 18 months, with 45 infants returning 3 months later. We found that recumbent-length-derived WLZ scores were significantly higher at 18 months than corresponding standing-height-derived WLZ scores. We also found that recumbent-length-derived WLZ scores, but not those derived from standing height measures, decreased significantly from 18 to 21 months. Although these differential results are attributable to the WHO database data entry syntax, which automatically corrects standing height measurements by adding 0.7 cm, they suggest that researchers proceed cautiously when using standing-height derived measures when calculating infant BMI z-scores. Our results suggest that for practical purposes, standing height measurements may be preferred, so long as they are entered into the WHO database as recumbent length measurements. We also encourage basic science infancy researchers to include BMI assessments as part of their routine assessment protocols, to serve as potential outcome measures for other basic science variables of theoretical interest.

  4. Pixels, Blocks of Pixels, and Polygons: Choosing a Spatial Unit for Thematic Accuracy Assessment

    EPA Science Inventory

    Pixels, polygons, and blocks of pixels are all potentially viable spatial assessment units for conducting an accuracy assessment. We develop a statistical population-based framework to examine how the spatial unit chosen affects the outcome of an accuracy assessment. The populati...

  5. Accuracy of Students' Self-Assessment and Their Beliefs about Its Utility

    ERIC Educational Resources Information Center

    Lew, Magdeleine D. N.; Alwis, W. A. M.; Schmidt, Henk G.

    2010-01-01

    The purpose of the two studies presented here was to evaluate the accuracy of students' self-assessment ability, to examine whether this ability improves over time and to investigate whether self-assessment is more accurate if students believe that it contributes to improving learning. To that end, the accuracy of the self-assessments of 3588…

  6. Does it Make a Difference? Investigating the Assessment Accuracy of Teacher Tutors and Student Tutors

    ERIC Educational Resources Information Center

    Herppich, Stephanie; Wittwer, Jorg; Nuckles, Matthias; Renkl, Alexander

    2013-01-01

    Tutors often have difficulty with accurately assessing a tutee's understanding. However, little is known about whether the professional expertise of tutors influences their assessment accuracy. In this study, the authors examined the accuracy with which 21 teacher tutors and 25 student tutors assessed a tutee's understanding of the human…

  7. The effect of different Global Navigation Satellite System methods on positioning accuracy in elite alpine skiing.

    PubMed

    Gilgien, Matthias; Spörri, Jörg; Limpach, Philippe; Geiger, Alain; Müller, Erich

    2014-10-03

    In sport science, Global Navigation Satellite Systems (GNSS) are frequently applied to capture athletes' position, velocity and acceleration. Application of GNSS includes a large range of different GNSS technologies and methods. To date no study has comprehensively compared the different GNSS methods applied. Therefore, the aim of the current study was to investigate the effect of differential and non-differential solutions, different satellite systems and different GNSS signal frequencies on position accuracy. Twelve alpine ski racers were equipped with high-end GNSS devices while performing runs on a giant slalom course. The skiers' GNSS antenna positions were calculated in three satellite signal obstruction conditions using five different GNSS methods. The GNSS antenna positions were compared to a video-based photogrammetric reference system over one turn and against the most valid GNSS method over the entire run. Furthermore, the time for acquisitioning differential GNSS solutions was assessed for four differential methods. The only GNSS method that consistently yielded sub-decimetre position accuracy in typical alpine skiing conditions was a differential method using American (GPS) and Russian (GLONASS) satellite systems and the satellite signal frequencies L1 and L2. Under conditions of minimal satellite signal obstruction, valid results were also achieved when either the satellite system GLONASS or the frequency L2 was dropped from the best configuration. All other methods failed to fulfill the accuracy requirements needed to detect relevant differences in the kinematics of alpine skiers, even in conditions favorable for GNSS measurements. The methods with good positioning accuracy had also the shortest times to compute differential solutions. This paper highlights the importance to choose appropriate methods to meet the accuracy requirements for sport applications.

  8. The Effect of Different Global Navigation Satellite System Methods on Positioning Accuracy in Elite Alpine Skiing

    PubMed Central

    Gilgien, Matthias; Spörri, Jörg; Limpach, Philippe; Geiger, Alain; Müller, Erich

    2014-01-01

    In sport science, Global Navigation Satellite Systems (GNSS) are frequently applied to capture athletes' position, velocity and acceleration. Application of GNSS includes a large range of different GNSS technologies and methods. To date no study has comprehensively compared the different GNSS methods applied. Therefore, the aim of the current study was to investigate the effect of differential and non-differential solutions, different satellite systems and different GNSS signal frequencies on position accuracy. Twelve alpine ski racers were equipped with high-end GNSS devices while performing runs on a giant slalom course. The skiers' GNSS antenna positions were calculated in three satellite signal obstruction conditions using five different GNSS methods. The GNSS antenna positions were compared to a video-based photogrammetric reference system over one turn and against the most valid GNSS method over the entire run. Furthermore, the time for acquisitioning differential GNSS solutions was assessed for four differential methods. The only GNSS method that consistently yielded sub-decimetre position accuracy in typical alpine skiing conditions was a differential method using American (GPS) and Russian (GLONASS) satellite systems and the satellite signal frequencies L1 and L2. Under conditions of minimal satellite signal obstruction, valid results were also achieved when either the satellite system GLONASS or the frequency L2 was dropped from the best configuration. All other methods failed to fulfill the accuracy requirements needed to detect relevant differences in the kinematics of alpine skiers, even in conditions favorable for GNSS measurements. The methods with good positioning accuracy had also the shortest times to compute differential solutions. This paper highlights the importance to choose appropriate methods to meet the accuracy requirements for sport applications. PMID:25285461

  9. Assessing the accuracy and reproducibility of modality independent elastography in a murine model of breast cancer

    PubMed Central

    Weis, Jared A.; Flint, Katelyn M.; Sanchez, Violeta; Yankeelov, Thomas E.; Miga, Michael I.

    2015-01-01

    Abstract. Cancer progression has been linked to mechanics. Therefore, there has been recent interest in developing noninvasive imaging tools for cancer assessment that are sensitive to changes in tissue mechanical properties. We have developed one such method, modality independent elastography (MIE), that estimates the relative elastic properties of tissue by fitting anatomical image volumes acquired before and after the application of compression to biomechanical models. The aim of this study was to assess the accuracy and reproducibility of the method using phantoms and a murine breast cancer model. Magnetic resonance imaging data were acquired, and the MIE method was used to estimate relative volumetric stiffness. Accuracy was assessed using phantom data by comparing to gold-standard mechanical testing of elasticity ratios. Validation error was <12%. Reproducibility analysis was performed on animal data, and within-subject coefficients of variation ranged from 2 to 13% at the bulk level and 32% at the voxel level. To our knowledge, this is the first study to assess the reproducibility of an elasticity imaging metric in a preclinical cancer model. Our results suggest that the MIE method can reproducibly generate accurate estimates of the relative mechanical stiffness and provide guidance on the degree of change needed in order to declare biological changes rather than experimental error in future therapeutic studies. PMID:26158120

  10. Accuracy of virtual models in the assessment of maxillary defects

    PubMed Central

    Kurşun, Şebnem; Kılıç, Cenk; Özen, Tuncer

    2015-01-01

    Purpose This study aimed to assess the reliability of measurements performed on three-dimensional (3D) virtual models of maxillary defects obtained using cone-beam computed tomography (CBCT) and 3D optical scanning. Materials and Methods Mechanical cavities simulating maxillary defects were prepared on the hard palate of nine cadavers. Images were obtained using a CBCT unit at three different fields-of-views (FOVs) and voxel sizes: 1) 60×60 mm FOV, 0.125 mm3 (FOV60); 2) 80×80 mm FOV, 0.160 mm3 (FOV80); and 3) 100×100 mm FOV, 0.250 mm3 (FOV100). Superimposition of the images was performed using software called VRMesh Design. Automated volume measurements were conducted, and differences between surfaces were demonstrated. Silicon impressions obtained from the defects were also scanned with a 3D optical scanner. Virtual models obtained using VRMesh Design were compared with impressions obtained by scanning silicon models. Gold standard volumes of the impression models were then compared with CBCT and 3D scanner measurements. Further, the general linear model was used, and the significance was set to p=0.05. Results A comparison of the results obtained by the observers and methods revealed the p values to be smaller than 0.05, suggesting that the measurement variations were caused by both methods and observers along with the different cadaver specimens used. Further, the 3D scanner measurements were closer to the gold standard measurements when compared to the CBCT measurements. Conclusion In the assessment of artificially created maxillary defects, the 3D scanner measurements were more accurate than the CBCT measurements. PMID:25793180

  11. Comparative Accuracy Assessment of Global Land Cover Datasets Using Existing Reference Data

    NASA Astrophysics Data System (ADS)

    Tsendbazar, N. E.; de Bruin, S.; Mora, B.; Herold, M.

    2014-12-01

    Land cover is a key variable to monitor the impact of human and natural processes on the biosphere. As one of the Essential Climate Variables, land cover observations are used for climate models and several other applications. Remote sensing technologies have enabled the generation of several global land cover (GLC) products that are based on different data sources and methods (e.g. legends). Moreover, the reported map accuracies result from varying validation strategies. Such differences make the comparison of the GLC products challenging and create confusion on selecting suitable datasets for different applications. This study aims to conduct comparative accuracy assessment of GLC datasets (LC-CCI 2005, MODIS 2005, and Globcover 2005) using the Globcover 2005 reference data which can represent the thematic differences of these GLC maps. This GLC reference dataset provides LCCS classifier information for 3 main land cover types for each sample plot. The LCCS classifier information was translated according to the legends of the GLC maps analysed. The preliminary analysis showed some challenges in LCCS classifier translation arising from missing important classifier information, differences in class definition between the legends and absence of class proportion of main land cover types. To overcome these issues, we consolidated the entire reference data (i.e. 3857 samples distributed at global scale). Then the GLC maps and the reference dataset were harmonized into 13 general classes to perform the comparative accuracy assessments. To help users on selecting suitable GLC dataset(s) for their application, we conducted the map accuracy assessments considering different users' perspectives: climate modelling, bio-diversity assessments, agriculture monitoring, and map producers. This communication will present the method and the results of this study and provide a set of recommendations to the GLC map producers and users with the aim to facilitate the use of GLC maps.

  12. New High-Accuracy Methods for Automatically Detecting & Tracking CMEs

    NASA Astrophysics Data System (ADS)

    Byrne, Jason; Morgan, H.; Habbal, S. R.

    2012-05-01

    With the large amounts of CME image data available from the SOHO and STEREO coronagraphs, manual cataloguing of events can be tedious and subject to user bias. Therefore automated catalogues, such as CACTus and SEEDS, have been developed in an effort to produce a robust method of detection and analysis of events. Here we present the development of a new CORIMP (coronal image processing) CME detection and tracking technique that overcomes many of the drawbacks of previous methods. It works by first employing a dynamic CME separation technique to remove the static background, and then characterizing CMEs via a multiscale edge-detection algorithm. This allows the inherent structure of the CMEs to be revealed in each image, which is usually prone to spatiotemporal crosstalk as a result of traditional image-differencing techniques. Thus the kinematic and morphological information on each event is resolved with higher accuracy than previous catalogues, revealing CME acceleration and expansion profiles otherwise undetected, and enabling a determination of the varying speeds attained across the span of the CME. The potential for a 3D characterization of the internal structure of CMEs is also demonstrated.

  13. ASSESSING THE ACCURACY OF NATIONAL LAND COVER DATASET AREA ESTIMATES AT MULTIPLE SPATIAL EXTENTS

    EPA Science Inventory

    Site specific accuracy assessments provide fine-scale evaluation of the thematic accuracy of land use/land cover (LULC) datasets; however, they provide little insight into LULC accuracy across varying spatial extents. Additionally, LULC data are typically used to describe lands...

  14. Constraint on Absolute Accuracy of Metacomprehension Assessments: The Anchoring and Adjustment Model vs. the Standards Model

    ERIC Educational Resources Information Center

    Kwon, Heekyung

    2011-01-01

    The objective of this study is to provide a systematic account of three typical phenomena surrounding absolute accuracy of metacomprehension assessments: (1) the absolute accuracy of predictions is typically quite low; (2) there exist individual differences in absolute accuracy of predictions as a function of reading skill; and (3) postdictions…

  15. Assessing accuracy in citizen science-based plant phenology monitoring

    NASA Astrophysics Data System (ADS)

    Fuccillo, Kerissa K.; Crimmins, Theresa M.; de Rivera, Catherine E.; Elder, Timothy S.

    2015-07-01

    In the USA, thousands of volunteers are engaged in tracking plant and animal phenology through a variety of citizen science programs for the purpose of amassing spatially and temporally comprehensive datasets useful to scientists and resource managers. The quality of these observations and their suitability for scientific analysis, however, remains largely unevaluated. We aimed to evaluate the accuracy of plant phenology observations collected by citizen scientist volunteers following protocols designed by the USA National Phenology Network (USA-NPN). Phenology observations made by volunteers receiving several hours of formal training were compared to those collected independently by a professional ecologist. Approximately 11,000 observations were recorded by 28 volunteers over the course of one field season. Volunteers consistently identified phenophases correctly (91 % overall) for the 19 species observed. Volunteers demonstrated greatest overall accuracy identifying unfolded leaves, ripe fruits, and open flowers. Transitional accuracy decreased for some species/phenophase combinations (70 % average), and accuracy varied significantly by phenophase and species ( p < 0.0001). Volunteers who submitted fewer observations over the period of study did not exhibit a higher error rate than those who submitted more total observations. Overall, these results suggest that volunteers with limited training can provide reliable observations when following explicit, standardized protocols. Future studies should investigate different observation models (i.e., group/individual, online/in-person training) over subsequent seasons with multiple expert comparisons to further substantiate the ability of these monitoring programs to supply accurate broadscale datasets capable of answering pressing ecological questions about global change.

  16. Accuracy of Wind Prediction Methods in the California Sea Breeze

    NASA Astrophysics Data System (ADS)

    Sumers, B. D.; Dvorak, M. J.; Ten Hoeve, J. E.; Jacobson, M. Z.

    2010-12-01

    In this study, we investigate the accuracy of measure-correlate-predict (MCP) algorithms and log law/power law scaling using data from two tall towers in coastal environments. We find that MCP algorithms accurately predict sea breeze winds and that log law/power law scaling methods struggle to predict 50-meter wind speeds. MCP methods have received significant attention as the wind industry has grown and the ability to accurately characterize the wind resource has become valuable. These methods are used to produce longer-term wind speed records from short-term measurement campaigns. A correlation is developed between the “target site,” where the developer is interested in building wind turbines, and a “reference site,” where long-term wind data is available. Up to twenty years of prior wind speeds are then are predicted. In this study, two existing MCP methods - linear regression and Mortimer’s method - are applied to predict 50-meter wind speeds at sites in the Salinas Valley and Redwood City, CA. The predictions are then verified with tall tower data. It is found that linear regression is poorly suited to MCP applications as the process produces inaccurate estimates of the cube of the wind speed at 50 meters. Meanwhile, Mortimer’s method, which bins data by direction and speed, is found to accurately predict the cube of the wind speed in both sea breeze and non-sea breeze conditions. We also find that log and power law are unstable predictors of wind speeds. While these methods produced accurate estimates of the average 50-meter wind speed at both sites, they predicted an average cube of the wind speed that was between 1.3 and 1.18 times the observed value. Inspection of time-series error reveals increased error in the mid-afternoon of the summer. This suggests that the cold sea breeze may disrupt the vertical temperature profile, create a stable atmosphere and violate the assumptions that allow log law scaling to work.

  17. Accuracy of Multiple Pour Cast from Various Elastomer Impression Methods.

    PubMed

    Haralur, Satheesh B; Saad Toman, Majed; Ali Al-Shahrani, Abdullah; Ali Al-Qarni, Abdullah

    2016-01-01

    The accurate duplicate cast obtained from a single impression reduces the profession clinical time, patient inconvenience, and extra material cost. The stainless steel working cast model assembly consisting of two abutments and one pontic area was fabricated. Two sets of six each custom aluminum trays were fabricated, with five mm spacer and two mm spacer. The impression methods evaluated during the study were additional silicone putty reline (two steps), heavy-light body (one step), monophase (one step), and polyether (one step). Type IV gypsum casts were poured at the interval of one hour, 12 hours, 24 hours, and 48 hours. The resultant cast was measured with traveling microscope for the comparative dimensional accuracy. The data obtained were subjected to Analysis of Variance test at significance level <0.05. The die obtained from two-step putty reline impression techniques had the percentage of variation for the height -0.36 to -0.97%, while diameter was increased by 0.40-0.90%. The values for one-step heavy-light body impression dies, additional silicone monophase impressions, and polyether were -0.73 to -1.21%, -1.34%, and -1.46% for the height and 0.50-0.80%, 1.20%, and -1.30% for the width, respectively.

  18. Accuracy of Multiple Pour Cast from Various Elastomer Impression Methods

    PubMed Central

    Saad Toman, Majed; Ali Al-Shahrani, Abdullah; Ali Al-Qarni, Abdullah

    2016-01-01

    The accurate duplicate cast obtained from a single impression reduces the profession clinical time, patient inconvenience, and extra material cost. The stainless steel working cast model assembly consisting of two abutments and one pontic area was fabricated. Two sets of six each custom aluminum trays were fabricated, with five mm spacer and two mm spacer. The impression methods evaluated during the study were additional silicone putty reline (two steps), heavy-light body (one step), monophase (one step), and polyether (one step). Type IV gypsum casts were poured at the interval of one hour, 12 hours, 24 hours, and 48 hours. The resultant cast was measured with traveling microscope for the comparative dimensional accuracy. The data obtained were subjected to Analysis of Variance test at significance level <0.05. The die obtained from two-step putty reline impression techniques had the percentage of variation for the height −0.36 to −0.97%, while diameter was increased by 0.40–0.90%. The values for one-step heavy-light body impression dies, additional silicone monophase impressions, and polyether were −0.73 to −1.21%, −1.34%, and −1.46% for the height and 0.50–0.80%, 1.20%, and −1.30% for the width, respectively. PMID:28096815

  19. Accuracy of Estrogen and Progesterone Receptor Assessment in Core Needle Biopsy Specimens of Breast Cancer

    PubMed Central

    Omranipour, Ramesh; Alipour, Sadaf; Hadji, Maryam; Fereidooni, Forouzandeh; Jahanzad, Issa; Bagheri, Khojasteh

    2013-01-01

    Background Diagnosis of breast cancer is completed through core needle biopsy (CNB) of the tumors but there is controversy on the accuracy of hormone receptor results on CNB specimens. Objectives We undertook this study to compare the results of hormone receptor assessment in CNB and surgical samples on our patients. Patients and Methods Hormone receptor status was determined in CNB and surgical samples in breast cancer patients whose CNB and operation had been performed in this institute from 2009 to 2011 and had not undergone neoadjuvant chemotherapy. Results About 350 patients, 60 cases met all the criteria for entering the study. The mean age was 49.8 years. Considering a confidence interval (CI) of 95%, the sensitivity of ER and PR assessment in CNB was 92.9% and 81%, respectively and the specificity of both was 100%. The Accuracy of CNB was 98% for ER and 93% for PR. Conclusions Our results confirm the acceptable accuracy of ER assessment on CNB. The subject needs further investigation in developing countries where omission of the test in surgical samples can be cost and time-saving. PMID:24349751

  20. Assessment of relative accuracy of AHN-2 laser scanning data using planar features.

    PubMed

    van der Sande, Corné; Soudarissanane, Sylvie; Khoshelham, Kourosh

    2010-01-01

    AHN-2 is the second part of the Actueel Hoogtebestand Nederland project, which concerns the acquisition of high-resolution altimetry data over the entire Netherlands using airborne laser scanning. The accuracy assessment of laser altimetry data usually relies on comparing corresponding tie elements, often points or lines, in the overlapping strips. This paper proposes a new approach to strip adjustment and accuracy assessment of AHN-2 data by using planar features. In the proposed approach a transformation is estimated between two overlapping strips by minimizing the distances between points in one strip and their corresponding planes in the other. The planes and the corresponding points are extracted in an automated segmentation process. The point-to-plane distances are used as observables in an estimation model, whereby the parameters of a transformation between the two strips and their associated quality measures are estimated. We demonstrate the performance of the method for the accuracy assessment of the AHN-2 dataset over Zeeland province of The Netherlands. The results show vertical offsets of up to 4 cm between the overlapping strips, and horizontal offsets ranging from 2 cm to 34 cm.

  1. Accuracy of the correlation method of the thermal neutron absorption cross-section determination for rocks

    NASA Astrophysics Data System (ADS)

    Krynicka, Ewa

    1995-08-01

    The influence of various random errors on the accuracy of thermal neutron absorption cross-sections determined by a correlation method is discussed. It is considered either as an absolute accuracy, when all experimental errors arc taken into account, or as an experimental assay accuracy, when the reference moderator parameters are assumed as the invariant data fixed for all experiments. The estimated accuracy is compared with the accuracy of results obtained for the same rock sample by Czubek's measurement method.

  2. Assessment of accuracy of CFD simulations through quantification of a numerical dissipation rate

    NASA Astrophysics Data System (ADS)

    Domaradzki, J. A.; Sun, G.; Xiang, X.; Chen, K. K.

    2016-11-01

    The accuracy of CFD simulations is typically assessed through a time consuming process of multiple runs and comparisons with available benchmark data. We propose that the accuracy can be assessed in the course of actual runs using a simpler method based on a numerical dissipation rate which is computed at each time step for arbitrary sub-domains using only information provided by the code in question (Schranner et al., 2015; Castiglioni and Domaradzki, 2015). Here, the method has been applied to analyze numerical simulation results obtained using OpenFOAM software for a flow around a sphere at Reynolds number of 1000. Different mesh resolutions were used in the simulations. For the coarsest mesh the ratio of the numerical dissipation to the viscous dissipation downstream of the sphere varies from 4.5% immediately behind the sphere to 22% further away. For the finest mesh this ratio varies from 0.4% behind the sphere to 6% further away. The large numerical dissipation in the former case is a direct indicator that the simulation results are inaccurate, e.g., the predicted Strouhal number is 16% lower than the benchmark. Low numerical dissipation in the latter case is an indicator of an acceptable accuracy, with the Strouhal number in the simulations matching the benchmark. Supported by NSF.

  3. Assessing accuracy in citizen science-based plant phenology monitoring.

    PubMed

    Fuccillo, Kerissa K; Crimmins, Theresa M; de Rivera, Catherine E; Elder, Timothy S

    2015-07-01

    In the USA, thousands of volunteers are engaged in tracking plant and animal phenology through a variety of citizen science programs for the purpose of amassing spatially and temporally comprehensive datasets useful to scientists and resource managers. The quality of these observations and their suitability for scientific analysis, however, remains largely unevaluated. We aimed to evaluate the accuracy of plant phenology observations collected by citizen scientist volunteers following protocols designed by the USA National Phenology Network (USA-NPN). Phenology observations made by volunteers receiving several hours of formal training were compared to those collected independently by a professional ecologist. Approximately 11,000 observations were recorded by 28 volunteers over the course of one field season. Volunteers consistently identified phenophases correctly (91% overall) for the 19 species observed. Volunteers demonstrated greatest overall accuracy identifying unfolded leaves, ripe fruits, and open flowers. Transitional accuracy decreased for some species/phenophase combinations (70% average), and accuracy varied significantly by phenophase and species (p < 0.0001). Volunteers who submitted fewer observations over the period of study did not exhibit a higher error rate than those who submitted more total observations. Overall, these results suggest that volunteers with limited training can provide reliable observations when following explicit, standardized protocols. Future studies should investigate different observation models (i.e., group/individual, online/in-person training) over subsequent seasons with multiple expert comparisons to further substantiate the ability of these monitoring programs to supply accurate broadscale datasets capable of answering pressing ecological questions about global change.

  4. Method of visualisation influences accuracy of measurements in cone-beam computed tomography.

    PubMed

    Patcas, Raphael; Angst, Christine; Kellenberger, Christian J; Schätzle, Marc A; Ullrich, Oliver; Markic, Goran

    2015-09-01

    This study evaluated the potential impact of different visualisation methods of cone-beam computed tomography (CBCT) on the accuracy of linear measurements of calcified structures, and assessed their interchangeability. High resolution (0.125 mm voxel) CBCT scans were obtained from eight cadaveric heads. The distance between the alveolar bone ridge and the incisal edge was determined for all mandibular incisors and canines, both anatomically and with measurements based on the following five CBCT visualisation methods: isosurface, direct volume rendering, multiplanar reformatting (MPR), maximum intensity projection of the volume of interest (VOIMIP), and average intensity projection of the volume of interest (VOIAvIP). All radiological methods were tested for repeatability and compared with anatomical results for accuracy, and limits of agreement were established. Interchangeability was evaluated by reviewing disparities between the methods and disclosing deterministic differences. Fine intra- and inter-observer repeatability was asserted for all visualisation methods (intraclass correlation coefficient ≤0.81). Measurements were most accurate when performed on MPR images and performed most disappointingly on isosurface-based images. Direct volume rendering, VOIMIP and VOIAvIP achieved acceptable results. It can be concluded that visualisation methods influence the accuracy of CBCT measurements. The isosurface viewing method is not recommended, and multiplanar reformatted images should be favoured for linear measurements of calcified structures.

  5. Procedural Documentation and Accuracy Assessment of Bathymetric Maps and Area/Capacity Tables for Small Reservoirs

    USGS Publications Warehouse

    Wilson, Gary L.; Richards, Joseph M.

    2006-01-01

    Because of the increasing use and importance of lakes for water supply to communities, a repeatable and reliable procedure to determine lake bathymetry and capacity is needed. A method to determine the accuracy of the procedure will help ensure proper collection and use of the data and resulting products. It is important to clearly define the intended products and desired accuracy before conducting the bathymetric survey to ensure proper data collection. A survey-grade echo sounder and differential global positioning system receivers were used to collect water-depth and position data in December 2003 at Sugar Creek Lake near Moberly, Missouri. Data were collected along planned transects, with an additional set of quality-assurance data collected for use in accuracy computations. All collected data were imported into a geographic information system database. A bathymetric surface model, contour map, and area/capacity tables were created from the geographic information system database. An accuracy assessment was completed on the collected data, bathymetric surface model, area/capacity table, and contour map products. Using established vertical accuracy standards, the accuracy of the collected data, bathymetric surface model, and contour map product was 0.67 foot, 0.91 foot, and 1.51 feet at the 95 percent confidence level. By comparing results from different transect intervals with the quality-assurance transect data, it was determined that a transect interval of 1 percent of the longitudinal length of Sugar Creek Lake produced nearly as good results as 0.5 percent transect interval for the bathymetric surface model, area/capacity table, and contour map products.

  6. 12 CFR 620.3 - Accuracy of reports and assessment of internal control over financial reporting.

    Code of Federal Regulations, 2010 CFR

    2010-01-01

    ... 12 Banks and Banking 6 2010-01-01 2010-01-01 false Accuracy of reports and assessment of internal control over financial reporting. 620.3 Section 620.3 Banks and Banking FARM CREDIT ADMINISTRATION FARM CREDIT SYSTEM DISCLOSURE TO SHAREHOLDERS General § 620.3 Accuracy of reports and assessment of...

  7. Evaluation of precision and accuracy assessment of different 3-D surface imaging systems for biomedical purposes.

    PubMed

    Eder, Maximilian; Brockmann, Gernot; Zimmermann, Alexander; Papadopoulos, Moschos A; Schwenzer-Zimmerer, Katja; Zeilhofer, Hans Florian; Sader, Robert; Papadopulos, Nikolaos A; Kovacs, Laszlo

    2013-04-01

    Three-dimensional (3-D) surface imaging has gained clinical acceptance, especially in the field of cranio-maxillo-facial and plastic, reconstructive, and aesthetic surgery. Six scanners based on different scanning principles (Minolta Vivid 910®, Polhemus FastSCAN™, GFM PRIMOS®, GFM TopoCAM®, Steinbichler Comet® Vario Zoom 250, 3dMD DSP 400®) were used to measure five sheep skulls of different sizes. In three areas with varying anatomical complexity (areas, 1 = high; 2 = moderate; 3 = low), 56 distances between 20 landmarks are defined on each skull. Manual measurement (MM), coordinate machine measurements (CMM) and computer tomography (CT) measurements were used to define a reference method for further precision and accuracy evaluation of different 3-D scanning systems. MM showed high correlation to CMM and CT measurements (both r = 0.987; p < 0.001) and served as the reference method. TopoCAM®, Comet® and Vivid 910® showed highest measurement precision over all areas of complexity; Vivid 910®, the Comet® and the DSP 400® demonstrated highest accuracy over all areas with Vivid 910® being most accurate in areas 1 and 3, and the DSP 400® most accurate in area 2. In accordance to the measured distance length, most 3-D devices present higher measurement precision and accuracy for large distances and lower degrees of precision and accuracy for short distances. In general, higher degrees of complexity are associated with lower 3-D assessment accuracy, suggesting that for optimal results, different types of scanners should be applied to specific clinical applications and medical problems according to their special construction designs and characteristics.

  8. An accuracy assessment of Magellan Very Long Baseline Interferometry (VLBI)

    NASA Technical Reports Server (NTRS)

    Engelhardt, D. B.; Kronschnabl, G. R.; Border, J. S.

    1990-01-01

    Very Long Baseline Interferometry (VLBI) measurements of the Magellan spacecraft's angular position and velocity were made during July through September, 1989, during the spacecraft's heliocentric flight to Venus. The purpose of this data acquisition and reduction was to verify this data type for operational use before Magellan is inserted into Venus orbit, in August, 1990. The accuracy of these measurements are shown to be within 20 nanoradians in angular position, and within 5 picoradians/sec in angular velocity. The media effects and their calibrations are quantified; the wet fluctuating troposphere is the dominant source of measurement error for angular velocity. The charged particle effect is completely calibrated with S- and X-Band dual-frequency calibrations. Increasing the accuracy of the Earth platform model parameters, by using VLBI-derived tracking station locations consistent with the planetary ephemeris frame, and by including high frequency Earth tidal terms in the Earth rotation model, add a few nanoradians improvement to the angular position measurements. Angular velocity measurements were insensitive to these Earth platform modelling improvements.

  9. Assessing expected accuracy of probe vehicle travel time reports

    SciTech Connect

    Hellinga, B.; Fu, L.

    1999-12-01

    The use of probe vehicles to provide estimates of link travel times has been suggested as a means of obtaining travel times within signalized networks for use in advanced travel information systems. Past research in the literature has proved contradictory conclusions regarding the expected accuracy of these probe-based estimates, and consequently has estimated different levels of market penetration of probe vehicles required to sustain accurate data within an advanced traveler information system. This paper examines the effect of sampling bias on the accuracy of the probe estimates. An analytical expression is derived on the basis of queuing theory to prove that bias in arrival time distributions and/or in the proportion of probes associated with each link departure turning movement will lead to a systematic bias in the sample estimate of the mean delay. Subsequently, the potential for and impact of sampling bias on a signalized link is examined by simulating an arterial corridor. The analytical derivation and the simulation analysis show that the reliability of probe-based average link travel times is highly affected by sampling bias. Furthermore, this analysis shows that the contradictory conclusions of previous research are directly related to the presence of absence of sample bias.

  10. Accuracy assessment of CKC high-density surface EMG decomposition in biceps femoris muscle

    NASA Astrophysics Data System (ADS)

    Marateb, H. R.; McGill, K. C.; Holobar, A.; Lateva, Z. C.; Mansourian, M.; Merletti, R.

    2011-10-01

    The aim of this study was to assess the accuracy of the convolution kernel compensation (CKC) method in decomposing high-density surface EMG (HDsEMG) signals from the pennate biceps femoris long-head muscle. Although the CKC method has already been thoroughly assessed in parallel-fibered muscles, there are several factors that could hinder its performance in pennate muscles. Namely, HDsEMG signals from pennate and parallel-fibered muscles differ considerably in terms of the number of detectable motor units (MUs) and the spatial distribution of the motor-unit action potentials (MUAPs). In this study, monopolar surface EMG signals were recorded from five normal subjects during low-force voluntary isometric contractions using a 92-channel electrode grid with 8 mm inter-electrode distances. Intramuscular EMG (iEMG) signals were recorded concurrently using monopolar needles. The HDsEMG and iEMG signals were independently decomposed into MUAP trains, and the iEMG results were verified using a rigorous a posteriori statistical analysis. HDsEMG decomposition identified from 2 to 30 MUAP trains per contraction. 3 ± 2 of these trains were also reliably detected by iEMG decomposition. The measured CKC decomposition accuracy of these common trains over a selected 10 s interval was 91.5 ± 5.8%. The other trains were not assessed. The significant factors that affected CKC decomposition accuracy were the number of HDsEMG channels that were free of technical artifact and the distinguishability of the MUAPs in the HDsEMG signal (P < 0.05). These results show that the CKC method reliably identifies at least a subset of MUAP trains in HDsEMG signals from low force contractions in pennate muscles.

  11. On Accuracy of Adaptive Grid Methods for Captured Shocks

    NASA Technical Reports Server (NTRS)

    Yamaleev, Nail K.; Carpenter, Mark H.

    2002-01-01

    The accuracy of two grid adaptation strategies, grid redistribution and local grid refinement, is examined by solving the 2-D Euler equations for the supersonic steady flow around a cylinder. Second- and fourth-order linear finite difference shock-capturing schemes, based on the Lax-Friedrichs flux splitting, are used to discretize the governing equations. The grid refinement study shows that for the second-order scheme, neither grid adaptation strategy improves the numerical solution accuracy compared to that calculated on a uniform grid with the same number of grid points. For the fourth-order scheme, the dominant first-order error component is reduced by the grid adaptation, while the design-order error component drastically increases because of the grid nonuniformity. As a result, both grid adaptation techniques improve the numerical solution accuracy only on the coarsest mesh or on very fine grids that are seldom found in practical applications because of the computational cost involved. Similar error behavior has been obtained for the pressure integral across the shock. A simple analysis shows that both grid adaptation strategies are not without penalties in the numerical solution accuracy. Based on these results, a new grid adaptation criterion for captured shocks is proposed.

  12. Vertical root fracture: Biological effects and accuracy of diagnostic imaging methods

    PubMed Central

    Baageel, Turki M.; Allah, Emad Habib; Bakalka, Ghaida T.; Jadu, Fatima; Yamany, Ibrahim; Jan, Ahmed M.; Bogari, Dania F.; Alhazzazi, Turki Y.

    2016-01-01

    This review assessed the most up-to-date literature on the accuracy of detecting vertical root fractures (VRFs] using the currently available diagnostic imaging methods. In addition, an overview of the biological and clinical aspects of VRFs will also be discussed. A systematic review of the literature was initiated in December of 2015 and then updated in May of 2016. The electronic databases searched included PubMed, Emabse, Ovid, and Google Scholar. An assessment of the methodological quality was performed using a modified version of the quality assessment of diagnostic accuracy studies tool. Twenty-two studies were included in this systematic review after applying specific inclusion and exclusion criteria. Of those, 12 favored using cone beam computed tomography (CBCT) for detecting VRF as compared to periapical radiographs, whereas 5 reported no differences between the two methods. The remaining 5 studies confirmed the advantages associated with using CBCT when diagnosing VRF and described the parameters and limitations associated with this method, but they were not comparative studies. In conclusion, overwhelming evidence suggests that the use of CBCT is a preferred method for detecting VRFs. Nevertheless, additional well controlled and high quality studies are needed to produce solid evidence and guidelines to support the routine use of CBCT in the diagnosis of VRFs as a standard of care. PMID:27652254

  13. Assessment of the accuracy and stability of ENSN sensors responses

    NASA Astrophysics Data System (ADS)

    Nofal, Hamed; Mohamed, Omar; Mohanna, Mahmoud; El-Gabry, Mohamed

    2015-06-01

    The Egyptian National Seismic Network (ENSN) is an advanced scientific tool used to investigate earth structure and seismic activity in Egypt. One of the main tasks of the engineering team of ENSN is to keep the accuracy and stability of the high performance seismic instruments as close as possible to the international standards used in international seismic network. To achieve this task, the seismometers are routinely calibrated. One of the final outcomes of the calibration process is a set of the actual poles and zeros of the seismometers. Due to the strategic importance of the High Dam, we present in this paper the results of the calibrating broad band (BB) seismometers type Trillium-40 (40 second). From these sets we computed both amplitude and phase responses as well as their deviations from the nominal responses of this particular seismometer type. The computed deviation of this sub-network is then statistically analyzed to obtain an overall estimate of the accuracy of measurements recorded by it. Such analysis might also discover some stations which are far from the international standards. This test will be carried out regularly at periods of several months to find out how stable the seismometer response is. As a result, the values of the magnitude and phase errors are confined between 0% and 2% for about 90% of the calibrated seismometers. The average magnitude error was found to be 5% from the nominal and 4% for average phase error. In order to eliminate any possible error in the measured data, the measured (true) poles and zeroes are used in the response files to replace the nominal values.

  14. Accuracy and repeatability of a new method for measuring facet loads in the lumbar spine.

    PubMed

    Wilson, Derek C; Niosi, Christina A; Zhu, Qingan A; Oxland, Thomas R; Wilson, David R

    2006-01-01

    We assessed the repeatability and accuracy of a relatively new, resistance-based sensor (Tekscan 6900) for measuring lumbar spine facet loads, pressures, and contact areas in cadaver specimens. Repeatability of measurements in the natural facet joint was determined for five trials of four specimens loaded in pure moment (+/- 7.5 N m) flexibility tests in axial rotation and flexion-extension. Accuracy of load measurements in four joints was assessed by applying known compressive loads of 25, 50, and 100 N to the natural facet joint in a materials testing machine and comparing the known applied load to the measured load. Measurements of load were obtained using two different calibration approaches: linear and two-point calibrations. Repeatability for force, pressure, and area (average of standard deviation as a percentage of the mean for all trials over all specimens) was 4-6% for axial rotation and 7-10% for extension. Peak resultant force in axial rotation was 30% smaller when calculated using the linear calibration method. The Tekscan sensor overestimated the applied force by 18 +/- 9% (mean+/-standard deviation), 35 +/- 7% and 50 +/- 9% for compressive loads of 100, 50, and 25 N, respectively. The two-point method overestimated the loads by 35 +/- 16%, 45 +/- 7%, and 56 +/- 10% for the same three loads. Our results show that the Tekscan sensor is repeatable. However, the sensor measurement range is not optimal for the small loads transmitted by the facets and measurement accuracy is highly dependent on calibration protocol.

  15. Accuracy testing of steel and electric groundwater-level measuring tapes: Test method and in-service tape accuracy

    USGS Publications Warehouse

    Fulford, Janice M.; Clayton, Christopher S.

    2015-10-09

    The calibration device and proposed method were used to calibrate a sample of in-service USGS steel and electric groundwater tapes. The sample of in-service groundwater steel tapes were in relatively good condition. All steel tapes, except one, were accurate to ±0.01 ft per 100 ft over their entire length. One steel tape, which had obvious damage in the first hundred feet, was marginally outside the accuracy of ±0.01 ft per 100 ft by 0.001 ft. The sample of in-service groundwater-level electric tapes were in a range of conditions—from like new, with cosmetic damage, to nonfunctional. The in-service electric tapes did not meet the USGS accuracy recommendation of ±0.01 ft. In-service electric tapes, except for the nonfunctional tape, were accurate to about ±0.03 ft per 100 ft. A comparison of new with in-service electric tapes found that steel-core electric tapes maintained their length and accuracy better than electric tapes without a steel core. The in-service steel tapes could be used as is and achieve USGS accuracy recommendations for groundwater-level measurements. The in-service electric tapes require tape corrections to achieve USGS accuracy recommendations for groundwater-level measurement.

  16. Precision and accuracy of visual foliar injury assessments

    SciTech Connect

    Gumpertz, M.L.; Tingey, D.T.; Hogsett, W.E.

    1982-07-01

    The study compared three measures of foliar injury: (i) mean percent leaf area injured of all leaves on the plant, (ii) mean percent leaf area injured of the three most injured leaves, and (iii) the proportion of injured leaves to total number of leaves. For the first measure, the variation caused by reader biases and day-to-day variations were compared with the innate plant-to-plant variation. Bean (Phaseolus vulgaris 'Pinto'), pea (Pisum sativum 'Little Marvel'), radish (Rhaphanus sativus 'Cherry Belle'), and spinach (Spinacia oleracea 'Northland') plants were exposed to either 3 ..mu..L L/sup -1/ SO/sub 2/ or 0.3 ..mu..L L/sup -1/ ozone for 2 h. Three leaf readers visually assessed the percent injury on every leaf of each plant while a fourth reader used a transparent grid to make an unbiased assessment for each plant. The mean leaf area injured of the three most injured leaves was highly correlated with all leaves on the plant only if the three most injured leaves were <100% injured. The proportion of leaves injured was not highly correlated with percent leaf area injured of all leaves on the plant for any species in this study. The largest source of variation in visual assessments was plant-to-plant variation, which ranged from 44 to 97% of the total variance, followed by variation among readers (0-32% of the variance). Except for radish exposed to ozone, the day-to-day variation accounted for <18% of the total. Reader bias in assessment of ozone injury was significant but could be adjusted for each reader by a simple linear regression (R/sup 2/ = 0.89-0.91) of the visual assessments against the grid assessments.

  17. Inertial Measures of Motion for Clinical Biomechanics: Comparative Assessment of Accuracy under Controlled Conditions - Effect of Velocity

    PubMed Central

    Lebel, Karina; Boissy, Patrick; Hamel, Mathieu; Duval, Christian

    2013-01-01

    Background Inertial measurement of motion with Attitude and Heading Reference Systems (AHRS) is emerging as an alternative to 3D motion capture systems in biomechanics. The objectives of this study are: 1) to describe the absolute and relative accuracy of multiple units of commercially available AHRS under various types of motion; and 2) to evaluate the effect of motion velocity on the accuracy of these measurements. Methods The criterion validity of accuracy was established under controlled conditions using an instrumented Gimbal table. AHRS modules were carefully attached to the center plate of the Gimbal table and put through experimental static and dynamic conditions. Static and absolute accuracy was assessed by comparing the AHRS orientation measurement to those obtained using an optical gold standard. Relative accuracy was assessed by measuring the variation in relative orientation between modules during trials. Findings Evaluated AHRS systems demonstrated good absolute static accuracy (mean error < 0.5o) and clinically acceptable absolute accuracy under condition of slow motions (mean error between 0.5o and 3.1o). In slow motions, relative accuracy varied from 2o to 7o depending on the type of AHRS and the type of rotation. Absolute and relative accuracy were significantly affected (p<0.05) by velocity during sustained motions. The extent of that effect varied across AHRS. Interpretation Absolute and relative accuracy of AHRS are affected by environmental magnetic perturbations and conditions of motions. Relative accuracy of AHRS is mostly affected by the ability of all modules to locate the same global reference coordinate system at all time. Conclusions Existing AHRS systems can be considered for use in clinical biomechanics under constrained conditions of use. While their individual capacity to track absolute motion is relatively consistent, the use of multiple AHRS modules to compute relative motion between rigid bodies needs to be optimized according to

  18. Methods for evaluating the predictive accuracy of structural dynamic models

    NASA Technical Reports Server (NTRS)

    Hasselman, Timothy K.; Chrostowski, Jon D.

    1991-01-01

    Modeling uncertainty is defined in terms of the difference between predicted and measured eigenvalues and eigenvectors. Data compiled from 22 sets of analysis/test results was used to create statistical databases for large truss-type space structures and both pretest and posttest models of conventional satellite-type space structures. Modeling uncertainty is propagated through the model to produce intervals of uncertainty on frequency response functions, both amplitude and phase. This methodology was used successfully to evaluate the predictive accuracy of several structures, including the NASA CSI Evolutionary Structure tested at Langley Research Center. Test measurements for this structure were within + one-sigma intervals of predicted accuracy for the most part, demonstrating the validity of the methodology and computer code.

  19. Accuracy of Four Dental Age Estimation Methods in Southern Indian Children

    PubMed Central

    Sanghvi, Praveen; Perumalla, Kiran Kumar; Srinivasaraju, D.; Srinivas, Jami; Kalyan, U. Siva; Rasool, SK. Md. Iftekhar

    2015-01-01

    Introduction: For various forensic investigations of both living and dead individuals, the knowledge of the actual age or date of birth of the subject is of utmost importance. In recent years, age estimation has gained importance for a variety of reasons, including identifying criminal and legal responsibility, and for many other social events such as birth certificate, marriage, beginning a job, joining the army and retirement. Developing teeth are used to assess maturity and estimate age in number of disciplines; however the accuracy of different methods has not been assessed systematically. The aim of this study was to determine the accuracy of four dental age estimation methods. Materials and Methods: Digital Orthopantomographs (OPGS) of South Indian children between the ages of 6 and 16 y who visited the department of Department of Oral medicine and Radiology of GITAM Dental College, Visakhapatnam, Andhra Pradesh, India with similar ethnic origin were assessed. Dental age was calculated using Demirjian, Willems, Nolla, and adopted Haavikko methods and the difference between estimated dental age and chronological age were compared with paired t-test and Wilcoxon signed rank test. Results: An overestimation of the dental age was observed by using Demirjian and Nolla methods (0.1±1.63, 0.47±0.83 years in total sample respectively) and an underestimation of dental age was observed by using Willems and Haavikko methods (-0.4±1.53, -2.9±1.41 years respectively in total sample). Conclusion: Nolla’s method was more accurate in estimating dental age compared to other methods. Moreover, all the four methods were found to be reliable in estimating age of individuals of unknown chronological age in South Indian children. PMID:25738008

  20. Mapping with Small UAS: A Point Cloud Accuracy Assessment

    NASA Astrophysics Data System (ADS)

    Toth, Charles; Jozkow, Grzegorz; Grejner-Brzezinska, Dorota

    2015-12-01

    Interest in using inexpensive Unmanned Aerial System (UAS) technology for topographic mapping has recently significantly increased. Small UAS platforms equipped with consumer grade cameras can easily acquire high-resolution aerial imagery allowing for dense point cloud generation, followed by surface model creation and orthophoto production. In contrast to conventional airborne mapping systems, UAS has limited ground coverage due to low flying height and limited flying time, yet it offers an attractive alternative to high performance airborne systems, as the cost of the sensors and platform, and the flight logistics, is relatively low. In addition, UAS is better suited for small area data acquisitions and to acquire data in difficult to access areas, such as urban canyons or densely built-up environments. The main question with respect to the use of UAS is whether the inexpensive consumer sensors installed in UAS platforms can provide the geospatial data quality comparable to that provided by conventional systems. This study aims at the performance evaluation of the current practice of UAS-based topographic mapping by reviewing the practical aspects of sensor configuration, georeferencing and point cloud generation, including comparisons between sensor types and processing tools. The main objective is to provide accuracy characterization and practical information for selecting and using UAS solutions in general mapping applications. The analysis is based on statistical evaluation as well as visual examination of experimental data acquired by a Bergen octocopter with three different image sensor configurations, including a GoPro HERO3+ Black Edition, a Nikon D800 DSLR and a Velodyne HDL-32. In addition, georeferencing data of varying quality were acquired and evaluated. The optical imagery was processed by using three commercial point cloud generation tools. Comparing point clouds created by active and passive sensors by using different quality sensors, and finally

  1. A retrospective study to validate an intraoperative robotic classification system for assessing the accuracy of kirschner wire (K-wire) placements with postoperative computed tomography classification system for assessing the accuracy of pedicle screw placements.

    PubMed

    Tsai, Tai-Hsin; Wu, Dong-Syuan; Su, Yu-Feng; Wu, Chieh-Hsin; Lin, Chih-Lung

    2016-09-01

    This purpose of this retrospective study is validation of an intraoperative robotic grading classification system for assessing the accuracy of Kirschner-wire (K-wire) placements with the postoperative computed tomography (CT)-base classification system for assessing the accuracy of pedicle screw placements.We conducted a retrospective review of prospectively collected data from 35 consecutive patients who underwent 176 robotic assisted pedicle screws instrumentation at Kaohsiung Medical University Hospital from September 2014 to November 2015. During the operation, we used a robotic grading classification system for verifying the intraoperative accuracy of K-wire placements. Three months after surgery, we used the common CT-base classification system to assess the postoperative accuracy of pedicle screw placements. The distributions of accuracy between the intraoperative robot-assisted and various postoperative CT-based classification systems were compared using kappa statistics of agreement.The intraoperative accuracies of K-wire placements before and after repositioning were classified as excellent (131/176, 74.4% and 133/176, 75.6%, respectively), satisfactory (36/176, 20.5% and 41/176, 23.3%, respectively), and malpositioned (9/176, 5.1% and 2/176, 1.1%, respectively)In postoperative CT-base classification systems were evaluated. No screw placements were evaluated as unacceptable under any of these systems. Kappa statistics revealed no significant differences between the proposed system and the aforementioned classification systems (P <0.001).Our results revealed no significant differences between the intraoperative robotic grading system and various postoperative CT-based grading systems. The robotic grading classification system is a feasible method for evaluating the accuracy of K-wire placements. Using the intraoperative robot grading system to classify the accuracy of K-wire placements enables predicting the postoperative accuracy of pedicle screw

  2. Evaluation of TDRSS-user orbit determination accuracy using batch least-squares and sequential methods

    NASA Technical Reports Server (NTRS)

    Oza, D. H.; Jones, T. L.; Hodjatzadeh, M.; Samii, M. V.; Doll, C. E.; Hart, R. C.; Mistretta, G. D.

    1991-01-01

    The development of the Real-Time Orbit Determination/Enhanced (RTOD/E) system as a prototype system for sequential orbit determination on a Disk Operating System (DOS) based Personal Computer (PC) is addressed. The results of a study to compare the orbit determination accuracy of a Tracking and Data Relay Satellite System (TDRSS) user spacecraft obtained using RTOD/E with the accuracy of an established batch least squares system, the Goddard Trajectory Determination System (GTDS), is addressed. Independent assessments were made to examine the consistencies of results obtained by the batch and sequential methods. Comparisons were made between the forward filtered RTOD/E orbit solutions and definitive GTDS orbit solutions for the Earth Radiation Budget Satellite (ERBS); the maximum solution differences were less than 25 m after the filter had reached steady state.

  3. Assessing the accuracy of image tracking algorithms on visible and thermal imagery using a deep restricted Boltzmann machine

    NASA Astrophysics Data System (ADS)

    Won, Stephen; Young, S. Susan

    2012-06-01

    Image tracking algorithms are critical to many applications including image super-resolution and surveillance. However, there exists no method to independently verify the accuracy of the tracking algorithm without a supplied control or visual inspection. This paper proposes an image tracking framework that uses deep restricted Boltzmann machines trained without external databases to quantify the accuracy of image tracking algorithms without the use of ground truths. In this paper, the tracking algorithm is comprised of the combination of flux tensor segmentation with four image registration methods, including correlation, Horn-Schunck optical flow, Lucas-Kanade optical flow, and feature correspondence methods. The robustness of the deep restricted Boltzmann machine is assessed by comparing between results from training with trusted and not-trusted data. Evaluations show that the deep restricted Boltzmann machine is a valid mechanism to assess the accuracy of a tracking algorithm without the use of ground truths.

  4. Using Accuracy of Self-Estimated Interest Type as a Sign of Career Choice Readiness in Career Assessment of Secondary Students

    ERIC Educational Resources Information Center

    Hirschi, Andreas; Lage, Damian

    2008-01-01

    A frequent applied method in career assessment to elicit clients' self-concepts is asking them to predict their interest assessment results. Accuracy in estimating one's interest type is commonly taken as a sign of more self-awareness and career choice readiness. The study evaluated the empirical relation of accuracy of self-estimation to career…

  5. The influence of sampling interval on the accuracy of trail impact assessment

    USGS Publications Warehouse

    Leung, Y.-F.; Marion, J.L.

    1999-01-01

    Trail impact assessment and monitoring (IA&M) programs have been growing in importance and application in recreation resource management at protected areas. Census-based and sampling-based approaches have been developed in such programs, with systematic point sampling being the most common survey design. This paper examines the influence of sampling interval on the accuracy of estimates for selected trail impact problems. A complete census of four impact types on 70 trails in Great Smoky Mountains National Park was utilized as the base data set for the analyses. The census data were resampled at increasing intervals to create a series of simulated point data sets. Estimates of frequency of occurrence and lineal extent for the four impact types were compared with the census data set. The responses of accuracy loss on lineal extent estimates to increasing sampling intervals varied across different impact types, while the responses on frequency of occurrence estimates were consistent, approximating an inverse asymptotic curve. These findings suggest that systematic point sampling may be an appropriate method for estimating the lineal extent but not the frequency of trail impacts. Sample intervals of less than 100 m appear to yield an excellent level of accuracy for the four impact types evaluated. Multiple regression analysis results suggest that appropriate sampling intervals are more likely to be determined by the type of impact in question rather than the length of trail. The census-based trail survey and the resampling-simulation method developed in this study can be a valuable first step in establishing long-term trail IA&M programs, in which an optimal sampling interval range with acceptable accuracy is determined before investing efforts in data collection.

  6. Evaluation of the accuracy of land-use based ecosystem service assessments for different thematic resolutions.

    PubMed

    Van der Biest, K; Vrebos, D; Staes, J; Boerema, A; Bodí, M B; Fransen, E; Meire, P

    2015-06-01

    The demand for pragmatic tools for mapping ecosystem services (ES) has led to the widespread application of land-use based proxy methods, mostly using coarse thematic resolution classification systems. Although various studies have demonstrated the limited reliability of land use as an indicator of service delivery, this does not prevent the method from being frequently applied on different institutional levels. It has recently been argued that a more detailed land use classification system may increase the accuracy of this approach. This research statistically compares maps of predicted ES delivery based on land use scoring for three different thematic resolutions (number of classes) with maps of ES delivery produced by biophysical models. Our results demonstrate that using a more detailed land use classification system does not significantly increase the accuracy of land-use based ES assessments for the majority of the considered ES. Correlations between land-use based assessments and biophysical model outcomes are relatively strong for provisioning services, independent of the classification system. However, large discrepancies occur frequently between the score and the model-based estimate. We conclude that land use, as a simple indicator, is not effective enough to be used in environmental management as it cannot capture differences in abiotic conditions and ecological processes that explain differences in service delivery. Using land use as a simple indicator will therefore result in inappropriate management decisions, even if a highly detailed land use classification system is used.

  7. Assessing the accuracy of quantitative molecular microbial profiling.

    PubMed

    O'Sullivan, Denise M; Laver, Thomas; Temisak, Sasithon; Redshaw, Nicholas; Harris, Kathryn A; Foy, Carole A; Studholme, David J; Huggett, Jim F

    2014-11-21

    The application of high-throughput sequencing in profiling microbial communities is providing an unprecedented ability to investigate microbiomes. Such studies typically apply one of two methods: amplicon sequencing using PCR to target a conserved orthologous sequence (typically the 16S ribosomal RNA gene) or whole (meta)genome sequencing (WGS). Both methods have been used to catalog the microbial taxa present in a sample and quantify their respective abundances. However, a comparison of the inherent precision or bias of the different sequencing approaches has not been performed. We previously developed a metagenomic control material (MCM) to investigate error when performing different sequencing strategies. Amplicon sequencing using four different primer strategies and two 16S rRNA regions was examined (Roche 454 Junior) and compared to WGS (Illumina HiSeq). All sequencing methods generally performed comparably and in good agreement with organism specific digital PCR (dPCR); WGS notably demonstrated very high precision. Where discrepancies between relative abundances occurred they tended to differ by less than twofold. Our findings suggest that when alternative sequencing approaches are used for microbial molecular profiling they can perform with good reproducibility, but care should be taken when comparing small differences between distinct methods. This work provides a foundation for future work comparing relative differences between samples and the impact of extraction methods. We also highlight the value of control materials when conducting microbial profiling studies to benchmark methods and set appropriate thresholds.

  8. Assessing the Accuracy of Quantitative Molecular Microbial Profiling

    PubMed Central

    O’Sullivan, Denise M.; Laver, Thomas; Temisak, Sasithon; Redshaw, Nicholas; Harris, Kathryn A.; Foy, Carole A.; Studholme, David J.; Huggett, Jim F.

    2014-01-01

    The application of high-throughput sequencing in profiling microbial communities is providing an unprecedented ability to investigate microbiomes. Such studies typically apply one of two methods: amplicon sequencing using PCR to target a conserved orthologous sequence (typically the 16S ribosomal RNA gene) or whole (meta)genome sequencing (WGS). Both methods have been used to catalog the microbial taxa present in a sample and quantify their respective abundances. However, a comparison of the inherent precision or bias of the different sequencing approaches has not been performed. We previously developed a metagenomic control material (MCM) to investigate error when performing different sequencing strategies. Amplicon sequencing using four different primer strategies and two 16S rRNA regions was examined (Roche 454 Junior) and compared to WGS (Illumina HiSeq). All sequencing methods generally performed comparably and in good agreement with organism specific digital PCR (dPCR); WGS notably demonstrated very high precision. Where discrepancies between relative abundances occurred they tended to differ by less than twofold. Our findings suggest that when alternative sequencing approaches are used for microbial molecular profiling they can perform with good reproducibility, but care should be taken when comparing small differences between distinct methods. This work provides a foundation for future work comparing relative differences between samples and the impact of extraction methods. We also highlight the value of control materials when conducting microbial profiling studies to benchmark methods and set appropriate thresholds. PMID:25421243

  9. Attribute-Level and Pattern-Level Classification Consistency and Accuracy Indices for Cognitive Diagnostic Assessment

    ERIC Educational Resources Information Center

    Wang, Wenyi; Song, Lihong; Chen, Ping; Meng, Yaru; Ding, Shuliang

    2015-01-01

    Classification consistency and accuracy are viewed as important indicators for evaluating the reliability and validity of classification results in cognitive diagnostic assessment (CDA). Pattern-level classification consistency and accuracy indices were introduced by Cui, Gierl, and Chang. However, the indices at the attribute level have not yet…

  10. Accuracy, precision, usability, and cost of portable silver test methods for ceramic filter factories.

    PubMed

    Meade, Rhiana D; Murray, Anna L; Mittelman, Anjuliee M; Rayner, Justine; Lantagne, Daniele S

    2017-02-01

    Locally manufactured ceramic water filters are one effective household drinking water treatment technology. During manufacturing, silver nanoparticles or silver nitrate are applied to prevent microbiological growth within the filter and increase bacterial removal efficacy. Currently, there is no recommendation for manufacturers to test silver concentrations of application solutions or filtered water. We identified six commercially available silver test strips, kits, and meters, and evaluated them by: (1) measuring in quintuplicate six samples from 100 to 1,000 mg/L (application range) and six samples from 0.0 to 1.0 mg/L (effluent range) of silver nanoparticles and silver nitrate to determine accuracy and precision; (2) conducting volunteer testing to assess ease-of-use; and (3) comparing costs. We found no method accurately detected silver nanoparticles, and accuracy ranged from 4 to 91% measurement error for silver nitrate samples. Most methods were precise, but only one method could test both application and effluent concentration ranges of silver nitrate. Volunteers considered test strip methods easiest. The cost for 100 tests ranged from 36 to 1,600 USD. We found no currently available method accurately and precisely measured both silver types at reasonable cost and ease-of-use, thus these methods are not recommended to manufacturers. We recommend development of field-appropriate methods that accurately and precisely measure silver nanoparticle and silver nitrate concentrations.

  11. Does DFT-SAPT method provide spectroscopic accuracy?

    SciTech Connect

    Shirkov, Leonid; Makarewicz, Jan

    2015-02-14

    Ground state potential energy curves for homonuclear and heteronuclear dimers consisting of noble gas atoms from He to Kr were calculated within the symmetry adapted perturbation theory based on the density functional theory (DFT-SAPT). These potentials together with spectroscopic data derived from them were compared to previous high-precision coupled cluster with singles and doubles including the connected triples theory calculations (or better if available) as well as to experimental data used as the benchmark. The impact of midbond functions on DFT-SAPT results was tested to study the convergence of the interaction energies. It was shown that, for most of the complexes, DFT-SAPT potential calculated at the complete basis set (CBS) limit is lower than the corresponding benchmark potential in the region near its minimum and hence, spectroscopic accuracy cannot be achieved. The influence of the residual term δ(HF) on the interaction energy was also studied. As a result, we have found that this term improves the agreement with the benchmark in the repulsive region for the dimers considered, but leads to even larger overestimation of potential depth D{sub e}. Although the standard hybrid exchange-correlation (xc) functionals with asymptotic correction within the second order DFT-SAPT do not provide the spectroscopic accuracy at the CBS limit, it is possible to adjust empirically basis sets yielding highly accurate results.

  12. Assessment of RFID Read Accuracy for ISS Water Kit

    NASA Technical Reports Server (NTRS)

    Chu, Andrew

    2011-01-01

    The Space Life Sciences Directorate/Medical Informatics and Health Care Systems Branch (SD4) is assessing the benefits Radio Frequency Identification (RFID) technology for tracking items flown onboard the International Space Station (ISS). As an initial study, the Avionic Systems Division Electromagnetic Systems Branch (EV4) is collaborating with SD4 to affix RFID tags to a water kit supplied by SD4 and studying the read success rate of the tagged items. The tagged water kit inside a Cargo Transfer Bag (CTB) was inventoried using three different RFID technologies, including the Johnson Space Center Building 14 Wireless Habitat Test Bed RFID portal, an RFID hand-held reader being targeted for use on board the ISS, and an RFID enclosure designed and prototyped by EV4.

  13. Accuracy assessment, using stratified plurality sampling, of portions of a LANDSAT classification of the Arctic National Wildlife Refuge Coastal Plain

    NASA Technical Reports Server (NTRS)

    Card, Don H.; Strong, Laurence L.

    1989-01-01

    An application of a classification accuracy assessment procedure is described for a vegetation and land cover map prepared by digital image processing of LANDSAT multispectral scanner data. A statistical sampling procedure called Stratified Plurality Sampling was used to assess the accuracy of portions of a map of the Arctic National Wildlife Refuge coastal plain. Results are tabulated as percent correct classification overall as well as per category with associated confidence intervals. Although values of percent correct were disappointingly low for most categories, the study was useful in highlighting sources of classification error and demonstrating shortcomings of the plurality sampling method.

  14. 12 CFR 630.5 - Accuracy of reports and assessment of internal control over financial reporting.

    Code of Federal Regulations, 2010 CFR

    2010-01-01

    ... CREDIT SYSTEM General § 630.5 Accuracy of reports and assessment of internal control over financial... assessment of internal control over financial reporting. (1) Annual reports must include a report by the Funding Corporation's management assessing the effectiveness of the internal control over...

  15. Evaluating the Effect of Learning Style and Student Background on Self-Assessment Accuracy

    ERIC Educational Resources Information Center

    Alaoutinen, Satu

    2012-01-01

    This study evaluates a new taxonomy-based self-assessment scale and examines factors that affect assessment accuracy and course performance. The scale is based on Bloom's Revised Taxonomy and is evaluated by comparing students' self-assessment results with course performance in a programming course. Correlation has been used to reveal possible…

  16. Efficiency and Accuracy Verification of the Explicit Numerical Manifold Method for Dynamic Problems

    NASA Astrophysics Data System (ADS)

    Qu, X. L.; Wang, Y.; Fu, G. Y.; Ma, G. W.

    2015-05-01

    The original numerical manifold method (NMM) employs an implicit time integration scheme to achieve higher computational accuracy, but its efficiency is relatively low, especially when the open-close iterations of contact are involved. To improve its computational efficiency, a modified version of the NMM based on an explicit time integration algorithm is proposed in this study. The lumped mass matrix, internal force and damping vectors are derived for the proposed explicit scheme. A calibration study on P-wave propagation along a rock bar is conducted to investigate the efficiency and accuracy of the developed explicit numerical manifold method (ENMM) for wave propagation problems. Various considerations in the numerical simulations are discussed, and parametric studies are carried out to obtain an insight into the influencing factors on the efficiency and accuracy of wave propagation. To further verify the capability of the proposed ENMM, dynamic stability assessment for a fractured rock slope under seismic effect is analysed. It is shown that, compared to the original NMM, the computational efficiency of the proposed ENMM can be significantly improved.

  17. Assessing Sensor Accuracy for Non-Adjunct Use of Continuous Glucose Monitoring

    PubMed Central

    Patek, Stephen D.; Ortiz, Edward Andrew; Breton, Marc D.

    2015-01-01

    Abstract Background: The level of continuous glucose monitoring (CGM) accuracy needed for insulin dosing using sensor values (i.e., the level of accuracy permitting non-adjunct CGM use) is a topic of ongoing debate. Assessment of this level in clinical experiments is virtually impossible because the magnitude of CGM errors cannot be manipulated and related prospectively to clinical outcomes. Materials and Methods: A combination of archival data (parallel CGM, insulin pump, self-monitoring of blood glucose [SMBG] records, and meals for 56 pump users with type 1 diabetes) and in silico experiments was used to “replay” real-life treatment scenarios and relate sensor error to glycemic outcomes. Nominal blood glucose (BG) traces were extracted using a mathematical model, yielding 2,082 BG segments each initiated by insulin bolus and confirmed by SMBG. These segments were replayed at seven sensor accuracy levels (mean absolute relative differences [MARDs] of 3–22%) testing six scenarios: insulin dosing using sensor values, threshold, and predictive alarms, each without or with considering CGM trend arrows. Results: In all six scenarios, the occurrence of hypoglycemia (frequency of BG levels ≤50 mg/dL and BG levels ≤39 mg/dL) increased with sensor error, displaying an abrupt slope change at MARD =10%. Similarly, hyperglycemia (frequency of BG levels ≥250 mg/dL and BG levels ≥400 mg/dL) increased and displayed an abrupt slope change at MARD=10%. When added to insulin dosing decisions, information from CGM trend arrows, threshold, and predictive alarms resulted in improvement in average glycemia by 1.86, 8.17, and 8.88 mg/dL, respectively. Conclusions: Using CGM for insulin dosing decisions is feasible below a certain level of sensor error, estimated in silico at MARD=10%. In our experiments, further accuracy improvement did not contribute substantively to better glycemic outcomes. PMID:25436913

  18. Conformity assessment of the measurement accuracy in testing laboratories using a software application

    NASA Astrophysics Data System (ADS)

    Diniţă, A.

    2017-02-01

    This article presents a method for assessing the accuracy of the measurements obtained at different tests conducted in laboratories by implementing the interlaboratory comparison method (organization, performance and evaluation of measurements of tests on the same or similar items by two or more laboratories under predetermined conditions). The program (independent software application), realised by the author and described in this paper, analyses the measurement accuracy and performance of testing laboratory by comparing the results obtained from different tests, using the modify Youden diagram, helping identify different types of errors that can occur in measurement, according to ISO 13528:2015, Statistical methods for use in proficiency testing by interlaboratory comparison. A case study is presented in the article by determining the chemical composition of identical samples from five different laboratories. The Youden diagram obtained from this study case was used to identify errors in the laboratory testing equipment. This paper was accepted for publication in Proceedings after double peer reviewing process but was not presented at the Conference ROTRIB’16

  19. Accuracy of subjective assessment of fever by Nigerian mothers in under-5 children

    PubMed Central

    Odinaka, Kelechi Kenneth; Edelu, Benedict O.; Nwolisa, Emeka Charles; Amamilo, Ifeyinwa B.; Okolo, Seline N.

    2014-01-01

    Background: Many mothers still rely on palpation to determine if their children have fever at home before deciding to seek medical attention or administer self-medications. This study was carried out to determine the accuracy of subjective assessment of fever by Nigerian mothers in Under-5 Children. Patients and Methods: Each eligible child had a tactile assessment of fever by the mother after which the axillary temperature was measured. Statistical analysis was done using SPSS version 19 (IBM Inc. Chicago Illinois, USA, 2010). Result: A total of 113 mother/child pairs participated in the study. Palpation overestimates fever by 24.6%. Irrespective of the surface of the hand used for palpation, the sensitivity, specificity, positive predictive value (PPV) and negative predictive value (NPV) of tactile assessment were 82.4%, 37.1%, 51.9% and 71.9%, respectively. The use of the palmer surface of the hand had a better sensitivity (95.2%) than the dorsum of the hand (69.2%). The use of multiple sites had better sensitivity (86.7%) than the use of single site (76.2%). Conclusion: Tactile assessment of childhood fevers by mothers is still a relevant screening tool for the presence or absence fever. Palpation with the palmer surface of the hand using multiple sites improves the reliability of tactile assessment of fever. PMID:25114371

  20. Reproducibility and accuracy of optic nerve sheath diameter assessment using ultrasound compared to magnetic resonance imaging

    PubMed Central

    2013-01-01

    Background Quantification of the optic nerve sheath diameter (ONSD) by transbulbar sonography is a promising non-invasive technique for the detection of altered intracranial pressure. In order to establish this method as follow-up tool in diseases with intracranial hyper- or hypotension scan-rescan reproducibility and accuracy need to be systematically investigated. Methods The right ONSD of 15 healthy volunteers (mean age 24.5 ± 0.8 years) were measured by both transbulbar sonography (9 – 3 MHz) and 3 Tesla MRI (half-Fourier acquisition single-shot turbo spin-echo sequences, HASTE) 3 and 5 mm behind papilla. All volunteers underwent repeated ultrasound and MRI examinations in order to assess scan-rescan reproducibility and accuracy. Moreover, inter- and intra-observer variabilities were calculated for both techniques. Results Scan-rescan reproducibility was robust for ONSD quantification by sonography and MRI at both depths (r > 0.75, p ≤ 0.001, mean differences < 2%). Comparing ultrasound- and MRI-derived ONSD values, we found acceptable agreement between both methods for measurements at a depth of 3 mm (r = 0.72, p = 0.002, mean difference < 5%). Further analyses revealed good inter- and intra-observer reliability for sonographic measurements 3 mm behind the papilla and for MRI at 3 and 5 mm (r > 0.82, p < 0.001, mean differences < 5%). Conclusions Sonographic ONSD quantification 3 mm behind the papilla can be performed with good reproducibility, measurement accuracy and observer agreement. Thus, our findings emphasize the feasibility of this technique as a non-invasive bedside tool for longitudinal ONSD measurements. PMID:24289136

  1. Geometric calibration and accuracy assessment of a multispectral imager on UAVs

    NASA Astrophysics Data System (ADS)

    Zheng, Fengjie; Yu, Tao; Chen, Xingfeng; Chen, Jiping; Yuan, Guoti

    2012-11-01

    The increasing developments in Unmanned Aerial Vehicles (UAVs) platforms and associated sensing technologies have widely promoted UAVs remote sensing application. UAVs, especially low-cost UAVs, limit the sensor payload in weight and dimension. Mostly, cameras on UAVs are panoramic, fisheye lens, small-format CCD planar array camera, unknown intrinsic parameters and lens optical distortion will cause serious image aberrations, even leading a few meters or tens of meters errors in ground per pixel. However, the characteristic of high spatial resolution make accurate geolocation more critical to UAV quantitative remote sensing research. A method for MCC4-12F Multispectral Imager designed to load on UAVs has been developed and implemented. Using multi-image space resection algorithm to assess geometric calibration parameters of random position and different photogrammetric altitudes in 3D test field, which is suitable for multispectral cameras. Both theoretical and practical accuracy assessments were selected. The results of theoretical strategy, resolving object space and image point coordinate differences by space intersection, showed that object space RMSE were 0.2 and 0.14 pixels in X direction and in Y direction, image space RMSE were superior to 0.5 pixels. In order to verify the accuracy and reliability of the calibration parameters,practical study was carried out in Tianjin UAV flight experiments, the corrected accuracy validated by ground checkpoints was less than 0.3m. Typical surface reflectance retrieved on the basis of geo-rectified data was compared with ground ASD measurement resulting 4% discrepancy. Hence, the approach presented here was suitable for UAV multispectral imager.

  2. Forecasting method in multilateration accuracy based on laser tracker measurement

    NASA Astrophysics Data System (ADS)

    Aguado, Sergio; Santolaria, Jorge; Samper, David; José Aguilar, Juan

    2017-02-01

    Multilateration based on a laser tracker (LT) requires the measurement of a set of points from three or more positions. Although the LTs’ angular information is not used, multilateration produces a volume of measurement uncertainty. This paper presents two new coefficients from which to determine whether the measurement of a set of points, before performing the necessary measurements, will improve or worsen the accuracy of the multilateration results, avoiding unnecessary measurement, and reducing the time and economic cost required. The first specific coefficient measurement coefficient (MCLT) is unique for each laser tracker. It determines the relationship between the radial and angular laser tracker measurement noise. Similarly, the second coefficient is related with specific conditions of measurement β. It is related with the spatial angle between the laser tracker positions α and its effect on error reduction. Both parameters MCLT and β are linked in error reduction limits. Beside these, a new methodology to determine the multilateration reduction limit according to the multilateration technique of an ideal laser tracker distribution and a random one are presented. It provides general rules and advice from synthetic tests that are validated through a real test carried out in a coordinate measurement machine.

  3. Alternative Confidence Interval Methods Used in the Diagnostic Accuracy Studies.

    PubMed

    Erdoğan, Semra; Gülhan, Orekıcı Temel

    2016-01-01

    Background/Aim. It is necessary to decide whether the newly improved methods are better than the standard or reference test or not. To decide whether the new diagnostics test is better than the gold standard test/imperfect standard test, the differences of estimated sensitivity/specificity are calculated with the help of information obtained from samples. However, to generalize this value to the population, it should be given with the confidence intervals. The aim of this study is to evaluate the confidence interval methods developed for the differences between the two dependent sensitivity/specificity values on a clinical application. Materials and Methods. In this study, confidence interval methods like Asymptotic Intervals, Conditional Intervals, Unconditional Interval, Score Intervals, and Nonparametric Methods Based on Relative Effects Intervals are used. Besides, as clinical application, data used in diagnostics study by Dickel et al. (2010) has been taken as a sample. Results. The results belonging to the alternative confidence interval methods for Nickel Sulfate, Potassium Dichromate, and Lanolin Alcohol are given as a table. Conclusion. While preferring the confidence interval methods, the researchers have to consider whether the case to be compared is single ratio or dependent binary ratio differences, the correlation coefficient between the rates in two dependent ratios and the sample sizes.

  4. Study of accuracy of precipitation measurements using simulation method

    NASA Astrophysics Data System (ADS)

    Nagy, Zoltán; Lajos, Tamás; Morvai, Krisztián

    2013-04-01

    of wind shield improve the accuracy of precipitation measurements? · Try to find the source of the error that can be detected at tipping bucket raingauge in winter time because of use of heating power? On our poster we would like to present the answers to the questions listed above.

  5. Thematic Accuracy Assessment of the 2011 National Land Cover Database (NLCD)

    EPA Science Inventory

    Accuracy assessment is a standard protocol of National Land Cover Database (NLCD) mapping. Here we report agreement statistics between map and reference labels for NLCD 2011, which includes land cover for ca. 2001, ca. 2006, and ca. 2011. The two main objectives were assessment o...

  6. Examining rating quality in writing assessment: rater agreement, error, and accuracy.

    PubMed

    Wind, Stefanie A; Engelhard, George

    2012-01-01

    The use of performance assessments in which human raters evaluate student achievement has become increasingly prevalent in high-stakes assessment systems such as those associated with recent policy initiatives (e.g., Race to the Top). In this study, indices of rating quality are compared between two measurement perspectives. Within the context of a large-scale writing assessment, this study focuses on the alignment between indices of rater agreement, error, and accuracy based on traditional and Rasch measurement theory perspectives. Major empirical findings suggest that Rasch-based indices of model-data fit for ratings provide information about raters that is comparable to direct measures of accuracy. The use of easily obtained approximations of direct accuracy measures holds significant implications for monitoring rating quality in large-scale rater-mediated performance assessments.

  7. Beyond the Correlation Coefficient in Studies of Self-Assessment Accuracy: Commentary on Zell & Krizan (2014).

    PubMed

    Dunning, David; Helzer, Erik G

    2014-03-01

    Zell and Krizan (2014, this issue) provide a comprehensive yet incomplete portrait of the factors influencing accurate self-assessment. This is no fault of their own. Much work on self-accuracy focuses on the correlation coefficient as the measure of accuracy, but it is not the only way self-accuracy can be measured. As such, its use can provide an incomplete and potentially misleading story. We urge researchers to explore measures of bias as well as correlation, because there are indirect hints that each respond to a different psychological dynamic. We further entreat researchers to develop other creative measures of accuracy and not to forget that self-accuracy may come not only from personal knowledge but also from insight about human nature more generally.

  8. High Order Accuracy Methods for Supersonic Reactive Flows

    DTIC Science & Technology

    2008-06-25

    bubble interactions [12], the supersonic cavity flows [11], and etc. The difficulty of implement - ing the spectral method to these complex fluid...enhance the convergence properties of the approximation via a filter function σ(η) [30] with the following properties σ(η)= σ(−η), σ(±1) = 0, σ(0)= 1, σ... implementation of the Hybrid method depends on the ability to obtain accurate information on the smoothness of a function . In this work, we employ the Multi

  9. Accuracy assessment of Kinect for Xbox One in point-based tracking applications

    NASA Astrophysics Data System (ADS)

    Goral, Adrian; Skalski, Andrzej

    2015-12-01

    We present the accuracy assessment of a point-based tracking system built on Kinect v2. In our approach, color, IR and depth data were used to determine the positions of spherical markers. To accomplish this task, we calibrated the depth/infrared and color cameras using a custom method. As a reference tool we used Polaris Spectra optical tracking system. The mean error obtained within the range from 0.9 to 2.9 m was 61.6 mm. Although the depth component of the error turned out to be the largest, the random error of depth estimation was only 1.24 mm on average. Our Kinect-based system also allowed for reliable angular measurements within the range of ±20° from the sensor's optical axis.

  10. Integrated three-dimensional digital assessment of accuracy of anterior tooth movement using clear aligners

    PubMed Central

    Zhang, Xiao-Juan; He, Li; Tian, Jie; Bai, Yu-Xing; Li, Song

    2015-01-01

    Objective To assess the accuracy of anterior tooth movement using clear aligners in integrated three-dimensional digital models. Methods Cone-beam computed tomography was performed before and after treatment with clear aligners in 32 patients. Plaster casts were laser-scanned for virtual setup and aligner fabrication. Differences in predicted and achieved root and crown positions of anterior teeth were compared on superimposed maxillofacial digital images and virtual models and analyzed by Student's t-test. Results The mean discrepancies in maxillary and mandibular crown positions were 0.376 ± 0.041 mm and 0.398 ± 0.037 mm, respectively. Maxillary and mandibular root positions differed by 2.062 ± 0.128 mm and 1.941 ± 0.154 mm, respectively. Conclusions Crowns but not roots of anterior teeth can be moved to designated positions using clear aligners, because these appliances cause tooth movement by tilting motion. PMID:26629473

  11. Accuracy assessment of the integration of GNSS and a MEMS IMU in a terrestrial platform.

    PubMed

    Madeira, Sergio; Yan, Wenlin; Bastos, Luísa; Gonçalves, José A

    2014-11-04

    MEMS Inertial Measurement Units are available at low cost and can replace expensive units in mobile mapping platforms which need direct georeferencing. This is done through the integration with GNSS measurements in order to achieve a continuous positioning solution and to obtain orientation angles. This paper presents the results of the assessment of the accuracy of a system that integrates GNSS and a MEMS IMU in a terrestrial platform. We describe the methodology used and the tests realized where the accuracy of the positions and orientation parameters were assessed using an independent photogrammetric technique employing cameras that integrate the mobile mapping system developed by the authors. Results for the accuracy of attitude angles and coordinates show that accuracies better than a decimeter in positions, and under a degree in angles, can be achieved even considering that the terrestrial platform is operating in less than favorable environments.

  12. Enhancing the accuracy of knowledge discovery: a supervised learning method

    PubMed Central

    2014-01-01

    Background The amount of biomedical literature available is growing at an explosive speed, but a large amount of useful information remains undiscovered in it. Researchers can make informed biomedical hypotheses through mining this literature. Unfortunately, popular mining methods based on co-occurrence produce too many target concepts, leading to the declining relevance ranking of the potential target concepts. Methods This paper presents a new method for selecting linking concepts which exploits statistical and textual features to represent each linking concept, and then classifies them as relevant or irrelevant to the starting concepts. Relevant linking concepts are then used to discover target concepts. Results Through an evaluation it is observed textual features improve the results obtained with only statistical features. We successfully replicate Swanson's two classic discoveries and find the rankings of potentially relevant target concepts are relatively high. Conclusions The number of target concepts is greatly reduced and potentially relevant target concepts gain higher ranking by adopting only relevant linking concepts. Thus, the proposed method has the potential to help biomedical experts find the most useful and valuable target concepts effectively. PMID:25474584

  13. Thematic accuracy assessment of the 2011 National Land Cover Database (NLCD)

    USGS Publications Warehouse

    Wickham, James; Stehman, Stephen V.; Gass, Leila; Dewitz, Jon; Sorenson, Daniel G.; Granneman, Brian J.; Poss, Richard V.; Baer, Lori Anne

    2017-01-01

    Accuracy assessment is a standard protocol of National Land Cover Database (NLCD) mapping. Here we report agreement statistics between map and reference labels for NLCD 2011, which includes land cover for ca. 2001, ca. 2006, and ca. 2011. The two main objectives were assessment of agreement between map and reference labels for the three, single-date NLCD land cover products at Level II and Level I of the classification hierarchy, and agreement for 17 land cover change reporting themes based on Level I classes (e.g., forest loss; forest gain; forest, no change) for three change periods (2001–2006, 2006–2011, and 2001–2011). The single-date overall accuracies were 82%, 83%, and 83% at Level II and 88%, 89%, and 89% at Level I for 2011, 2006, and 2001, respectively. Many class-specific user's accuracies met or exceeded a previously established nominal accuracy benchmark of 85%. Overall accuracies for 2006 and 2001 land cover components of NLCD 2011 were approximately 4% higher (at Level II and Level I) than the overall accuracies for the same components of NLCD 2006. The high Level I overall, user's, and producer's accuracies for the single-date eras in NLCD 2011 did not translate into high class-specific user's and producer's accuracies for many of the 17 change reporting themes. User's accuracies were high for the no change reporting themes, commonly exceeding 85%, but were typically much lower for the reporting themes that represented change. Only forest loss, forest gain, and urban gain had user's accuracies that exceeded 70%. Lower user's accuracies for the other change reporting themes may be attributable to the difficulty in determining the context of grass (e.g., open urban, grassland, agriculture) and between the components of the forest-shrubland-grassland gradient at either the mapping phase, reference label assignment phase, or both. NLCD 2011 user's accuracies for forest loss, forest gain, and urban gain compare favorably with results from other

  14. The Eye Phone Study: reliability and accuracy of assessing Snellen visual acuity using smartphone technology

    PubMed Central

    Perera, C; Chakrabarti, R; Islam, F M A; Crowston, J

    2015-01-01

    Purpose Smartphone-based Snellen visual acuity charts has become popularized; however, their accuracy has not been established. This study aimed to evaluate the equivalence of a smartphone-based visual acuity chart with a standard 6-m Snellen visual acuity (6SVA) chart. Methods First, a review of available Snellen chart applications on iPhone was performed to determine the most accurate application based on optotype size. Subsequently, a prospective comparative study was performed by measuring conventional 6SVA and then iPhone visual acuity using the ‘Snellen' application on an Apple iPhone 4. Results Eleven applications were identified, with accuracy of optotype size ranging from 4.4–39.9%. Eighty-eight patients from general medical and surgical wards in a tertiary hospital took part in the second part of the study. The mean difference in logMAR visual acuity between the two charts was 0.02 logMAR (95% limit of agreement −0.332, 0.372 logMAR). The largest mean difference in logMAR acuity was noted in the subgroup of patients with 6SVA worse than 6/18 (n=5), who had a mean difference of two Snellen visual acuity lines between the charts (0.276 logMAR). Conclusion We did not identify a Snellen visual acuity app at the time of study, which could predict a patients standard Snellen visual acuity within one line. There was considerable variability in the optotype accuracy of apps. Further validation is required for assessment of acuity in patients with severe vision impairment. PMID:25931170

  15. Assessing the accuracy and performance of implicit solvent models for drug molecules: conformational ensemble approaches.

    PubMed

    Kolář, Michal; Fanfrlík, Jindřich; Lepšík, Martin; Forti, Flavio; Luque, F Javier; Hobza, Pavel

    2013-05-16

    The accuracy and performance of implicit solvent methods for solvation free energy calculations were assessed on a set of 20 neutral drug molecules. Molecular dynamics (MD) provided ensembles of conformations in water and water-saturated octanol. The solvation free energies were calculated by popular implicit solvent models based on quantum mechanical (QM) electronic densities (COSMO-RS, MST, SMD) as well as on molecular mechanical (MM) point-charge models (GB, PB). The performance of the implicit models was tested by a comparison with experimental water-octanol transfer free energies (ΔG(ow)) by using single- and multiconformation approaches. MD simulations revealed difficulties in a priori estimation of the flexibility features of the solutes from simple structural descriptors, such as the number of rotatable bonds. An increasing accuracy of the calculated ΔG(ow) was observed in the following order: GB1 ~ PB < GB7 ≪ MST < SMD ~ COSMO-RS with a clear distinction identified between MM- and QM-based models, although for the set excluding three largest molecules, the differences among COSMO-RS, MST, and SMD were negligible. It was shown that the single-conformation approach applied to crystal geometries provides a rather accurate estimate of ΔG(ow) for rigid molecules yet fails completely for the flexible ones. The multiconformation approaches improved the performance, but only when the deformation contribution was ignored. It was revealed that for large-scale calculations on small molecules a recent GB model, GB7, provided a reasonable accuracy/speed ratio. In conclusion, the study contributes to the understanding of solvation free energy calculations for physical and medicinal chemistry applications.

  16. Accuracy of velocity and power determination by the Doppler method

    NASA Technical Reports Server (NTRS)

    Rottger, J.

    1984-01-01

    When designing a Mesosphere-Stratosphere-Troposphere (MST) radar antenna one has to trade between the choices to optimize the effective aperture or to optimize the sidelobe suppression. An optimization of the aperture increases the sensitivity. Suppression of side-lobes by tapering attenuates undesirable signals which spoil the estimates of reflectivity and velocity. Generally, any sidelobe effects are equivalent to a broadening of the antenna beam. The return signal is due to a product of the antenna pattern with the varying atmospheric reflectivity structures. Thus, knowing the antenna pattern, it is in principle possible to find the signal spectra, which, however, may be a tedious computational and ambiguous procedure. For vertically pointing main beams the sidelobe effects are efficiently suppressed because of the aspect sensitivity. It follows that sidelobes are a minor problem for spaced antenna methods. However, they can be crucial for Doppler methods, which need off-vertical beams. If a sidelobe is pointing towards the zenith a larger power may be received from the vertical than off-vertical directions, but quantitative estimates of this effect are not yet known. To get an error estimate of sidelobe effects with an off-vertical main beam a 1-dimensional example is considered.

  17. Methods for evaluating the predictive accuracy of structural dynamic models

    NASA Technical Reports Server (NTRS)

    Hasselman, T. K.; Chrostowski, Jon D.

    1990-01-01

    Uncertainty of frequency response using the fuzzy set method and on-orbit response prediction using laboratory test data to refine an analytical model are emphasized with respect to large space structures. Two aspects of the fuzzy set approach were investigated relative to its application to large structural dynamics problems: (1) minimizing the number of parameters involved in computing possible intervals; and (2) the treatment of extrema which may occur in the parameter space enclosed by all possible combinations of the important parameters of the model. Extensive printer graphics were added to the SSID code to help facilitate model verification, and an application of this code to the LaRC Ten Bay Truss is included in the appendix to illustrate this graphics capability.

  18. An accuracy measurement method for star trackers based on direct astronomic observation

    PubMed Central

    Sun, Ting; Xing, Fei; Wang, Xiaochu; You, Zheng; Chu, Daping

    2016-01-01

    Star tracker is one of the most promising optical attitude measurement devices and it is widely used in spacecraft for its high accuracy. However, how to realize and verify such an accuracy remains a crucial but unsolved issue until now. The authenticity of the accuracy measurement method of a star tracker will eventually determine the satellite performance. A new and robust accuracy measurement method for a star tracker based on the direct astronomical observation is proposed here. In comparison with the conventional method with simulated stars, this method utilizes real navigation stars as observation targets which makes the measurement results more authoritative and authentic. Transformations between different coordinate systems are conducted on the account of the precision movements of the Earth, and the error curves of directional vectors are obtained along the three axes. Based on error analysis and accuracy definitions, a three-axis accuracy evaluation criterion has been proposed in this paper, which could determine pointing and rolling accuracy of a star tracker directly. Experimental measurements confirm that this method is effective and convenient to implement. Such a measurement environment is close to the in-orbit conditions and it can satisfy the stringent requirement for high-accuracy star trackers. PMID:26948412

  19. Accuracy of Optimized Branched Algorithms to Assess Activity-Specific PAEE

    PubMed Central

    Edwards, Andy G.; Hill, James O.; Byrnes, William C.; Browning, Raymond C.

    2009-01-01

    PURPOSE To assess the activity-specific accuracy achievable by branched algorithm (BA) analysis of simulated daily-living physical activity energy expenditure (PAEE) within a sedentary population. METHODS Sedentary men (n=8) and women (n=8) first performed a treadmill calibration protocol, during which heart rate (HR), accelerometry (ACC), and PAEE were measured in 1-minute epochs. From these data, HR-PAEE, and ACC-PAEE regressions were constructed and used in each of six analytic models to predict PAEE from ACC and HR data collected during a subsequent simulated daily-living protocol. Criterion PAEE was measured during both protocols via indirect calorimetry. The accuracy achieved by each model was assessed by the root mean square of the difference between model-predicted daily–living PAEE and the criterion daily-living PAEE (expressed here as % of mean daily living PAEE). RESULTS Across the range of activities an unconstrained post hoc optimized branched algorithm best predicted criterion PAEE. Estimates using individual calibration were generally more accurate than those using group calibration (14 vs. 16 % error, respectively). These analyses also performed well within each of the six daily-living activities, but systematic errors appeared for several of those activities, which may be explained by an inability of the algorithm to simultaneously accommodate a heterogeneous range of activities. Analyses of between mean square error by subject and activity suggest that optimization involving minimization of RMS for total daily-living PAEE is associated with decreased error between subjects but increased error between activities. CONCLUSION The performance of post hoc optimized branched algorithms may be limited by heterogeneity in the daily-living activities being performed. PMID:19952842

  20. Assessment of the Accuracy of Pharmacy Students’ Compounded Solutions Using Vapor Pressure Osmometry

    PubMed Central

    McPherson, Timothy B.

    2013-01-01

    Objective. To assess the effectiveness of using a vapor pressure osmometer to measure the accuracy of pharmacy students’ compounding skills. Design. Students calculated the theoretical osmotic pressure (mmol/kg) of a solution as a pre-laboratory exercise, compared their calculations with actual values, and then attempted to determine the cause of any errors found. Assessment. After the introduction of the vapor pressure osmometer, the first-time pass rate for solution compounding has varied from 85% to 100%. Approximately 85% of students surveyed reported that the instrument was valuable as a teaching tool because it objectively assessed their work and provided immediate formative assessment. Conclusions. This simple technique of measuring compounding accuracy using a vapor pressure osmometer allowed students to see the importance of quality control and assessment in practice for both pharmacists and technicians. PMID:23610476

  1. Assessment of the accuracy of pharmacy students' compounded solutions using vapor pressure osmometry.

    PubMed

    Kolling, William M; McPherson, Timothy B

    2013-04-12

    OBJECTIVE. To assess the effectiveness of using a vapor pressure osmometer to measure the accuracy of pharmacy students' compounding skills. DESIGN. Students calculated the theoretical osmotic pressure (mmol/kg) of a solution as a pre-laboratory exercise, compared their calculations with actual values, and then attempted to determine the cause of any errors found. ASSESSMENT. After the introduction of the vapor pressure osmometer, the first-time pass rate for solution compounding has varied from 85% to 100%. Approximately 85% of students surveyed reported that the instrument was valuable as a teaching tool because it objectively assessed their work and provided immediate formative assessment. CONCLUSIONS. This simple technique of measuring compounding accuracy using a vapor pressure osmometer allowed students to see the importance of quality control and assessment in practice for both pharmacists and technicians.

  2. Automated Analysis of Counselor Style and Effects: The Development and Evaluation of Methods and Materials to Assess the Stylistic Accuracy and Outcome Effectiveness of Counselor Verbal Behavior. Final Report.

    ERIC Educational Resources Information Center

    Pepyne, Edward W.

    This project attempts to develop, evaluate and implement methods and materials for the automated analysis of the stylistic characteristics of counselor verbal behavior and its effects on client verbal behavior within the counseling interview. To achieve this purpose, the project designed a system of computer programs, the DISCOURSE ANALYSIS…

  3. A laboratory assessment of the measurement accuracy of weighing type rainfall intensity gauges

    NASA Astrophysics Data System (ADS)

    Colli, M.; Chan, P. W.; Lanza, L. G.; La Barbera, P.

    2012-04-01

    In recent years the WMO Commission for Instruments and Methods of Observation (CIMO) fostered noticeable advancements in the accuracy of precipitation measurement issue by providing recommendations on the standardization of equipment and exposure, instrument calibration and data correction as a consequence of various comparative campaigns involving manufacturers and national meteorological services from the participating countries (Lanza et al., 2005; Vuerich et al., 2009). Extreme events analysis is proven to be highly affected by the on-site RI measurement accuracy (see e.g. Molini et al., 2004) and the time resolution of the available RI series certainly constitutes another key-factor in constructing hyetographs that are representative of real rain events. The OTT Pluvio2 weighing gauge (WG) and the GEONOR T-200 vibrating-wire precipitation gauge demonstrated very good performance under previous constant flow rate calibration efforts (Lanza et al., 2005). Although WGs do provide better performance than more traditional Tipping Bucket Rain gauges (TBR) under continuous and constant reference intensity, dynamic effects seem to affect the accuracy of WG measurements under real world/time varying rainfall conditions (Vuerich et al., 2009). The most relevant is due to the response time of the acquisition system and the derived systematic delay of the instrument in assessing the exact weight of the bin containing cumulated precipitation. This delay assumes a relevant role in case high resolution rain intensity time series are sought from the instrument, as is the case of many hydrologic and meteo-climatic applications. This work reports the laboratory evaluation of Pluvio2 and T-200 rainfall intensity measurements accuracy. Tests are carried out by simulating different artificial precipitation events, namely non-stationary rainfall intensity, using a highly accurate dynamic rainfall generator. Time series measured by an Ogawa drop counter (DC) at a field test site

  4. Accuracy and precision of the three-dimensional assessment of the facial surface using a 3-D laser scanner.

    PubMed

    Kovacs, L; Zimmermann, A; Brockmann, G; Baurecht, H; Schwenzer-Zimmerer, K; Papadopulos, N A; Papadopoulos, M A; Sader, R; Biemer, E; Zeilhofer, H F

    2006-06-01

    Three-dimensional (3-D) recording of the surface of the human body or anatomical areas has gained importance in many medical specialties. Thus, it is important to determine scanner precision and accuracy in defined medical applications and to establish standards for the recording procedure. Here we evaluated the precision and accuracy of 3-D assessment of the facial area with the Minolta Vivid 910 3D Laser Scanner. We also investigated the influence of factors related to the recording procedure and the processing of scanner data on final results. These factors include lighting, alignment of scanner and object, the examiner, and the software used to convert measurements into virtual images. To assess scanner accuracy, we compared scanner data to those obtained by manual measurements on a dummy. Less than 7% of all results with the scanner method were outside a range of error of 2 mm when compared to corresponding reference measurements. Accuracy, thus, proved to be good enough to satisfy requirements for numerous clinical applications. Moreover, the experiments completed with the dummy yielded valuable information for optimizing recording parameters for best results. Thus, under defined conditions, precision and accuracy of surface models of the human face recorded with the Minolta Vivid 910 3D Scanner presumably can also be enhanced. Future studies will involve verification of our findings using test persons. The current findings indicate that the Minolta Vivid 910 3D Scanner might be used with benefit in medicine when recording the 3-D surface structures of the face.

  5. Methods & Strategies: Deep Assessment

    ERIC Educational Resources Information Center

    Haas, Alison; Hollimon, Shameka; Lee, Okhee

    2015-01-01

    The "Next Generation Science Standards" ("NGSS") push students to have "a deeper understanding of content" (NGSS Lead States 2013, Appendix A, p. 4). However, with the reality of high-stakes assessments that rely primarily on multiple-choice questions, how can a science teacher analyze students' written responses…

  6. Improving the accuracy of the discrete gradient method in the one-dimensional case.

    PubMed

    Cieśliński, Jan L; Ratkiewicz, Bogusław

    2010-01-01

    We present two numerical schemes of high accuracy for one-dimensional dynamical systems. They are modifications of the discrete gradient method and keep its advantages, including stability and conservation of the energy integral. However, their accuracy is higher by several orders of magnitude.

  7. Quality Assessment of Comparative Diagnostic Accuracy Studies: Our Experience Using a Modified Version of the QUADAS-2 Tool

    ERIC Educational Resources Information Center

    Wade, Ros; Corbett, Mark; Eastwood, Alison

    2013-01-01

    Assessing the quality of included studies is a vital step in undertaking a systematic review. The recently revised Quality Assessment of Diagnostic Accuracy Studies (QUADAS) tool (QUADAS-2), which is the only validated quality assessment tool for diagnostic accuracy studies, does not include specific criteria for assessing comparative studies. As…

  8. Accuracy of dental age estimation in Venezuelan children: comparison of Demirjian and Willems methods.

    PubMed

    Medina, Aída C; Blanco, Lucila

    2014-01-01

    Dental age is a somatic maturity indicator with importance in clinical and forensic dentistry. The purpose of this study is to compare the applicability of the Demirjian and Willems methods for dental age estimation in a group of Venezuelan children. Panoramic radiographs of 238 Venezuelan children aged 5-13 years were used to assess dental age using the methods described by Demirjian and Willems. Children with unclear panoramic radiographs, dental agenesis, and premature loss of primary teeth were excluded. Mean differences between dental age and chronological age by gender and age groups were estimated (ANOVA, Student tests p = 0.05). For the Demirjian method, the mean difference between dental age and chronological age was 0.62 +/- 0.93 years, statistically significant. The mean overestimation was lower for females than for males (females 0.56 +/- 0.96 years, males 0.67 +/- 0.93 years). For the Willems method, the mean difference between dental age and chronological age was 0.15 +/- 0.97 years, not statistically significant. Accuracy was significantly different between genders, performing best for females (females 0.01 +/- 0.96 years, males 0.29 +/- 0.96 years). The Willems method for age estimation was found to be more accurate than the Demirjian method in this sample of Venezuelan children.

  9. A method for improved accuracy in three dimensions for determining wheel/rail contact points

    NASA Astrophysics Data System (ADS)

    Yang, Xinwen; Gu, Shaojie; Zhou, Shunhua; Zhou, Yu; Lian, Songliang

    2015-11-01

    Searching for the contact points between wheels and rails is important because these points represent the points of exerted contact forces. In order to obtain an accurate contact point and an in-depth description of the wheel/rail contact behaviours on a curved track or in a turnout, a method with improved accuracy in three dimensions is proposed to determine the contact points and the contact patches between the wheel and the rail when considering the effect of the yaw angle and the roll angle on the motion of the wheel set. The proposed method, with no need of the curve fitting of the wheel and rail profiles, can accurately, directly, and comprehensively determine the contact interface distances between the wheel and the rail. The range iteration algorithm is used to improve the computation efficiency and reduce the calculation required. The present computation method is applied for the analysis of the contact of rails of CHINA (CHN) 75 kg/m and wheel sets of wearing type tread of China's freight cars. In addition, it can be proved that the results of the proposed method are consistent with that of Kalker's program CONTACT, and the maximum deviation from the wheel/rail contact patch area of this two methods is approximately 5%. The proposed method, can also be used to investigate static wheel/rail contact. Some wheel/rail contact points and contact patch distributions are discussed and assessed, wheel and rail non-worn and worn profiles included.

  10. Assessing map accuracy in a remotely sensed, ecoregion-scale cover map

    USGS Publications Warehouse

    Edwards, T.C.; Moisen, G.G.; Cutler, D.R.

    1998-01-01

    Landscape- and ecoregion-based conservation efforts increasingly use a spatial component to organize data for analysis and interpretation. A challenge particular to remotely sensed cover maps generated from these efforts is how best to assess the accuracy of the cover maps, especially when they can exceed 1000 s/km2 in size. Here we develop and describe a methodological approach for assessing the accuracy of large-area cover maps, using as a test case the 21.9 million ha cover map developed for Utah Gap Analysis. As part of our design process, we first reviewed the effect of intracluster correlation and a simple cost function on the relative efficiency of cluster sample designs to simple random designs. Our design ultimately combined clustered and subsampled field data stratified by ecological modeling unit and accessibility (hereafter a mixed design). We next outline estimation formulas for simple map accuracy measures under our mixed design and report results for eight major cover types and the three ecoregions mapped as part of the Utah Gap Analysis. Overall accuracy of the map was 83.2% (SE=1.4). Within ecoregions, accuracy ranged from 78.9% to 85.0%. Accuracy by cover type varied, ranging from a low of 50.4% for barren to a high of 90.6% for man modified. In addition, we examined gains in efficiency of our mixed design compared with a simple random sample approach. In regard to precision, our mixed design was more precise than a simple random design, given fixed sample costs. We close with a discussion of the logistical constraints facing attempts to assess the accuracy of large-area, remotely sensed cover maps.

  11. Evaluating accuracy of structural geometry by DXA methods with an anthropometric proximal femur phantom.

    PubMed

    Khoo, B C C; Beck, T J; Brown, K; Price, R I

    2013-09-01

    DXA-derived bone structural geometry has been reported extensively but lacks an accuracy standard. In this study, we describe a novel anthropometric structural geometry phantom that simulates the proximal femur for use in assessing accuracy of geometry measurements by DXA or other X-ray methods. The phantom consists of seven different interchangeable neck modules with geometries that span the range of dimensions in an adult human proximal femur, including those representing osteoporosis. Ten repeated hip scans of each neck module using two current DXA scanner models were performed without repositioning. After scanner specific calibration, hip structure analysis was used to derive structural geometry. Scanner performance was similar for the two manufacturers. DXA-derived HSA geometric measurements were highly correlated with values derived directly from phantom geometry and position; R² between DXA and phantom measures were greater than 94% for all parameters, while precision error ranged between 0.3 and 3.9%. Despite high R² there were some systematic geometry errors for both scanners that were small for outer diameter, but increasing with complexity of geometrical parameter; e.g. buckling ratio. In summary, the anthropometric phantom and its fabrication concept were shown to be appropriate for evaluating proximal femoral structural geometry in two different DXA systems.

  12. Comparison of ERBS orbit determination accuracy using batch least-squares and sequential methods

    NASA Technical Reports Server (NTRS)

    Oza, D. H.; Jones, T. L.; Fabien, S. M.; Mistretta, G. D.; Hart, R. C.; Doll, C. E.

    1991-01-01

    The Flight Dynamics Div. (FDD) at NASA-Goddard commissioned a study to develop the Real Time Orbit Determination/Enhanced (RTOD/E) system as a prototype system for sequential orbit determination of spacecraft on a DOS based personal computer (PC). An overview is presented of RTOD/E capabilities and the results are presented of a study to compare the orbit determination accuracy for a Tracking and Data Relay Satellite System (TDRSS) user spacecraft obtained using RTOS/E on a PC with the accuracy of an established batch least squares system, the Goddard Trajectory Determination System (GTDS), operating on a mainframe computer. RTOD/E was used to perform sequential orbit determination for the Earth Radiation Budget Satellite (ERBS), and the Goddard Trajectory Determination System (GTDS) was used to perform the batch least squares orbit determination. The estimated ERBS ephemerides were obtained for the Aug. 16 to 22, 1989, timeframe, during which intensive TDRSS tracking data for ERBS were available. Independent assessments were made to examine the consistencies of results obtained by the batch and sequential methods. Comparisons were made between the forward filtered RTOD/E orbit solutions and definitive GTDS orbit solutions for ERBS; the solution differences were less than 40 meters after the filter had reached steady state.

  13. Evaluation of Landsat-4 orbit determination accuracy using batch least-squares and sequential methods

    NASA Technical Reports Server (NTRS)

    Oza, D. H.; Jones, T. L.; Feiertag, R.; Samii, M. V.; Doll, C. E.; Mistretta, G. D.; Hart, R. C.

    1993-01-01

    The Goddard Space Flight Center (GSFC) Flight Dynamics Division (FDD) commissioned Applied Technology Associates, Incorporated, to develop the Real-Time Orbit Determination/Enhanced (RTOD/E) system on a Disk Operating System (DOS)-based personal computer (PC) as a prototype system for sequential orbit determination of spacecraft. This paper presents the results of a study to compare the orbit determination accuracy for a Tracking and Data Relay Satellite (TDRS) System (TDRSS) user spacecraft, Landsat-4, obtained using RTOD/E, operating on a PC, with the accuracy of an established batch least-squares system, the Goddard Trajectory Determination System (GTDS), operating on a mainframe computer. The results of Landsat-4 orbit determination will provide useful experience for the Earth Observing System (EOS) series of satellites. The Landsat-4 ephemerides were estimated for the May 18-24, 1992, timeframe, during which intensive TDRSS tracking data for Landsat-4 were available. During this period, there were two separate orbit-adjust maneuvers on one of the TDRSS spacecraft (TDRS-East) and one small orbit-adjust maneuver for Landsat-4. Independent assessments were made of the consistencies (overlap comparisons for the batch case and covariances and the first measurement residuals for the sequential case) of solutions produced by the batch and sequential methods. The forward-filtered RTOD/E orbit solutions were compared with the definitive GTDS orbit solutions for Landsat-4; the solution differences were generally less than 30 meters after the filter had reached steady state.

  14. Mapping stream habitats with a global positioning system: Accuracy, precision, and comparison with traditional methods

    USGS Publications Warehouse

    Dauwalter, D.C.; Fisher, W.L.; Belt, K.C.

    2006-01-01

    We tested the precision and accuracy of the Trimble GeoXT??? global positioning system (GPS) handheld receiver on point and area features and compared estimates of stream habitat dimensions (e.g., lengths and areas of riffles and pools) that were made in three different Oklahoma streams using the GPS receiver and a tape measure. The precision of differentially corrected GPS (DGPS) points was not affected by the number of GPS position fixes (i.e., geographic location estimates) averaged per DGPS point. Horizontal error of points ranged from 0.03 to 2.77 m and did not differ with the number of position fixes per point. The error of area measurements ranged from 0.1% to 110.1% but decreased as the area increased. Again, error was independent of the number of position fixes averaged per polygon corner. The estimates of habitat lengths, widths, and areas did not differ when measured using two methods of data collection (GPS and a tape measure), nor did the differences among methods change at three stream sites with contrasting morphologies. Measuring features with a GPS receiver was up to 3.3 times faster on average than using a tape measure, although signal interference from high streambanks or overhanging vegetation occasionally limited satellite signal availability and prolonged measurements with a GPS receiver. There were also no differences in precision of habitat dimensions when mapped using a continuous versus a position fix average GPS data collection method. Despite there being some disadvantages to using the GPS in stream habitat studies, measuring stream habitats with a GPS resulted in spatially referenced data that allowed the assessment of relative habitat position and changes in habitats over time, and was often faster than using a tape measure. For most spatial scales of interest, the precision and accuracy of DGPS data are adequate and have logistical advantages when compared to traditional methods of measurement. ?? 2006 Springer Science+Business Media

  15. Gender Differences in Structured Risk Assessment: Comparing the Accuracy of Five Instruments

    ERIC Educational Resources Information Center

    Coid, Jeremy; Yang, Min; Ullrich, Simone; Zhang, Tianqiang; Sizmur, Steve; Roberts, Colin; Farrington, David P.; Rogers, Robert D.

    2009-01-01

    Structured risk assessment should guide clinical risk management, but it is uncertain which instrument has the highest predictive accuracy among men and women. In the present study, the authors compared the Psychopathy Checklist-Revised (PCL-R; R. D. Hare, 1991, 2003); the Historical, Clinical, Risk Management-20 (HCR-20; C. D. Webster, K. S.…

  16. 12 CFR 620.3 - Accuracy of reports and assessment of internal control over financial reporting.

    Code of Federal Regulations, 2011 CFR

    2011-01-01

    ... in accordance with all applicable statutory or regulatory requirements, and (3) The information is true, accurate, and complete to the best of signatories' knowledge and belief. (d) Management... CREDIT SYSTEM DISCLOSURE TO SHAREHOLDERS General § 620.3 Accuracy of reports and assessment of...

  17. Assessing the Accuracy of MODIS-NDVI Derived Land-Cover Across the Great Lakes Basin

    EPA Science Inventory

    This research describes the accuracy assessment process for a land-cover dataset developed for the Great Lakes Basin (GLB). This land-cover dataset was developed from the 2007 MODIS Normalized Difference Vegetation Index (NDVI) 16-day composite (MOD13Q) 250 m time-series data. Tr...

  18. The Word Writing CAFE: Assessing Student Writing for Complexity, Accuracy, and Fluency

    ERIC Educational Resources Information Center

    Leal, Dorothy J.

    2005-01-01

    The Word Writing CAFE is a new assessment tool designed for teachers to evaluate objectively students' word-writing ability for fluency, accuracy, and complexity. It is designed to be given to the whole class at one time. This article describes the development of the CAFE and provides directions for administering and scoring it. The author also…

  19. A PIXEL COMPOSITION-BASED REFERENCE DATA SET FOR THEMATIC ACCURACY ASSESSMENT

    EPA Science Inventory

    Developing reference data sets for accuracy assessment of land-cover classifications derived from coarse spatial resolution sensors such as MODIS can be difficult due to the large resolution differences between the image data and available reference data sources. Ideally, the spa...

  20. Modifications to the accuracy assessment analysis routine MLTCRP to produce an output file

    NASA Technical Reports Server (NTRS)

    Carnes, J. G.

    1978-01-01

    Modifications are described that were made to the analysis program MLTCRP in the accuracy assessment software system to produce a disk output file. The output files produced by this modified program are used to aggregate data for regions greater than a single segment.

  1. Assessing Observer Accuracy in Continuous Recording of Rate and Duration: Three Algorithms Compared

    ERIC Educational Resources Information Center

    Mudford, Oliver C.; Martin, Neil T.; Hui, Jasmine K. Y.; Taylor, Sarah Ann

    2009-01-01

    The three algorithms most frequently selected by behavior-analytic researchers to compute interobserver agreement with continuous recording were used to assess the accuracy of data recorded from video samples on handheld computers by 12 observers. Rate and duration of responding were recorded for three samples each. Data files were compared with…

  2. In the Right Ballpark? Assessing the Accuracy of Net Price Calculators

    ERIC Educational Resources Information Center

    Anthony, Aaron M.; Page, Lindsay C.; Seldin, Abigail

    2016-01-01

    Large differences often exist between a college's sticker price and net price after accounting for financial aid. Net price calculators (NPCs) were designed to help students more accurately estimate their actual costs to attend a given college. This study assesses the accuracy of information provided by net price calculators. Specifically, we…

  3. Gaining Precision and Accuracy on Microprobe Trace Element Analysis with the Multipoint Background Method

    NASA Astrophysics Data System (ADS)

    Allaz, J. M.; Williams, M. L.; Jercinovic, M. J.; Donovan, J. J.

    2014-12-01

    Electron microprobe trace element analysis is a significant challenge, but can provide critical data when high spatial resolution is required. Due to the low peak intensity, the accuracy and precision of such analyses relies critically on background measurements, and on the accuracy of any pertinent peak interference corrections. A linear regression between two points selected at appropriate off-peak positions is a classical approach for background characterization in microprobe analysis. However, this approach disallows an accurate assessment of background curvature (usually exponential). Moreover, if present, background interferences can dramatically affect the results if underestimated or ignored. The acquisition of a quantitative WDS scan over the spectral region of interest is still a valuable option to determine the background intensity and curvature from a fitted regression of background portions of the scan, but this technique retains an element of subjectivity as the analyst has to select areas in the scan, which appear to represent background. We present here a new method, "Multi-Point Background" (MPB), that allows acquiring up to 24 off-peak background measurements from wavelength positions around the peaks. This method aims to improve the accuracy, precision, and objectivity of trace element analysis. The overall efficiency is amended because no systematic WDS scan needs to be acquired in order to check for the presence of possible background interferences. Moreover, the method is less subjective because "true" backgrounds are selected by the statistical exclusion of erroneous background measurements, reducing the need for analyst intervention. This idea originated from efforts to refine EPMA monazite U-Th-Pb dating, where it was recognised that background errors (peak interference or background curvature) could result in errors of several tens of million years on the calculated age. Results obtained on a CAMECA SX-100 "UltraChron" using monazite

  4. Accuracy Evaluation of a Mobile Mapping System with Advanced Statistical Methods

    NASA Astrophysics Data System (ADS)

    Toschi, I.; Rodríguez-Gonzálvez, P.; Remondino, F.; Minto, S.; Orlandini, S.; Fuller, A.

    2015-02-01

    This paper discusses a methodology to evaluate the precision and the accuracy of a commercial Mobile Mapping System (MMS) with advanced statistical methods. So far, the metric potentialities of this emerging mapping technology have been studied in few papers, where generally the assumption that errors follow a normal distribution is made. In fact, this hypothesis should be carefully verified in advance, in order to test how well the Gaussian classic statistics can adapt to datasets that are usually affected by asymmetrical gross errors. The workflow adopted in this study relies on a Gaussian assessment, followed by an outlier filtering process. Finally, non-parametric statistical models are applied, in order to achieve a robust estimation of the error dispersion. Among the different MMSs available on the market, the latest solution provided by RIEGL is here tested, i.e. the VMX-450 Mobile Laser Scanning System. The test-area is the historic city centre of Trento (Italy), selected in order to assess the system performance in dealing with a challenging and historic urban scenario. Reference measures are derived from photogrammetric and Terrestrial Laser Scanning (TLS) surveys. All datasets show a large lack of symmetry that leads to the conclusion that the standard normal parameters are not adequate to assess this type of data. The use of non-normal statistics gives thus a more appropriate description of the data and yields results that meet the quoted a-priori errors.

  5. An interpolation method for stream habitat assessments

    USGS Publications Warehouse

    Sheehan, Kenneth R.; Welsh, Stuart A.

    2015-01-01

    Interpolation of stream habitat can be very useful for habitat assessment. Using a small number of habitat samples to predict the habitat of larger areas can reduce time and labor costs as long as it provides accurate estimates of habitat. The spatial correlation of stream habitat variables such as substrate and depth improves the accuracy of interpolated data. Several geographical information system interpolation methods (natural neighbor, inverse distance weighted, ordinary kriging, spline, and universal kriging) were used to predict substrate and depth within a 210.7-m2 section of a second-order stream based on 2.5% and 5.0% sampling of the total area. Depth and substrate were recorded for the entire study site and compared with the interpolated values to determine the accuracy of the predictions. In all instances, the 5% interpolations were more accurate for both depth and substrate than the 2.5% interpolations, which achieved accuracies up to 95% and 92%, respectively. Interpolations of depth based on 2.5% sampling attained accuracies of 49–92%, whereas those based on 5% percent sampling attained accuracies of 57–95%. Natural neighbor interpolation was more accurate than that using the inverse distance weighted, ordinary kriging, spline, and universal kriging approaches. Our findings demonstrate the effective use of minimal amounts of small-scale data for the interpolation of habitat over large areas of a stream channel. Use of this method will provide time and cost savings in the assessment of large sections of rivers as well as functional maps to aid the habitat-based management of aquatic species.

  6. Accuracy assessment of satellite altimetry over central East Antarctica by kinematic GNSS and crossover analysis

    NASA Astrophysics Data System (ADS)

    Schröder, Ludwig; Richter, Andreas; Fedorov, Denis; Knöfel, Christoph; Ewert, Heiko; Dietrich, Reinhard; Matveev, Aleksey Yu.; Scheinert, Mirko; Lukin, Valery

    2014-05-01

    Satellite altimetry is a unique technique to observe the contribution of the Antarctic ice sheet to global sea-level change. To fulfill the high quality requirements for its application, the respective products need to be validated against independent data like ground-based measurements. Kinematic GNSS provides a powerful method to acquire precise height information along the track of a vehicle. Within a collaboration of TU Dresden and Russian partners during the Russian Antarctic Expeditions in the seasons from 2001 to 2013 we recorded several such profiles in the region of the subglacial Lake Vostok, East Antarctica. After 2006 these datasets also include observations along seven continental traverses with a length of about 1600km each between the Antarctic coast and the Russian research station Vostok (78° 28' S, 106° 50' E). After discussing some special issues concerning the processing of the kinematic GNSS profiles under the very special conditions of the interior of the Antarctic ice sheet, we will show their application for the validation of NASA's laser altimeter satellite mission ICESat and of ESA's ice mission CryoSat-2. Analysing the height differences at crossover points, we can get clear insights into the height regime at the subglacial Lake Vostok. Thus, these profiles as well as the remarkably flat lake surface itself can be used to investigate the accuracy and possible error influences of these missions. We will show how the transmit-pulse reference selection correction (Gaussian vs. centroid, G-C) released in January 2013 helped to further improve the release R633 ICESat data and discuss the height offsets and other effects of the CryoSat-2 radar data. In conclusion we show that only a combination of laser and radar altimetry can provide both, a high precision and a good spatial coverage. An independent validation with ground-based observations is crucial for a thorough accuracy assessment.

  7. Accuracy Assessment and Correction of Vaisala RS92 Radiosonde Water Vapor Measurements

    NASA Technical Reports Server (NTRS)

    Whiteman, David N.; Miloshevich, Larry M.; Vomel, Holger; Leblanc, Thierry

    2008-01-01

    Relative humidity (RH) measurements from Vaisala RS92 radiosondes are widely used in both research and operational applications, although the measurement accuracy is not well characterized as a function of its known dependences on height, RH, and time of day (or solar altitude angle). This study characterizes RS92 mean bias error as a function of its dependences by comparing simultaneous measurements from RS92 radiosondes and from three reference instruments of known accuracy. The cryogenic frostpoint hygrometer (CFH) gives the RS92 accuracy above the 700 mb level; the ARM microwave radiometer gives the RS92 accuracy in the lower troposphere; and the ARM SurTHref system gives the RS92 accuracy at the surface using 6 RH probes with NIST-traceable calibrations. These RS92 assessments are combined using the principle of Consensus Referencing to yield a detailed estimate of RS92 accuracy from the surface to the lowermost stratosphere. An empirical bias correction is derived to remove the mean bias error, yielding corrected RS92 measurements whose mean accuracy is estimated to be +/-3% of the measured RH value for nighttime soundings and +/-4% for daytime soundings, plus an RH offset uncertainty of +/-0.5%RH that is significant for dry conditions. The accuracy of individual RS92 soundings is further characterized by the 1-sigma "production variability," estimated to be +/-1.5% of the measured RH value. The daytime bias correction should not be applied to cloudy daytime soundings, because clouds affect the solar radiation error in a complicated and uncharacterized way.

  8. A novel method for crosstalk analysis of biological networks: improving accuracy of pathway annotation

    PubMed Central

    Ogris, Christoph; Guala, Dimitri; Helleday, Thomas; Sonnhammer, Erik L. L.

    2017-01-01

    Analyzing gene expression patterns is a mainstay to gain functional insights of biological systems. A plethora of tools exist to identify significant enrichment of pathways for a set of differentially expressed genes. Most tools analyze gene overlap between gene sets and are therefore severely hampered by the current state of pathway annotation, yet at the same time they run a high risk of false assignments. A way to improve both true positive and false positive rates (FPRs) is to use a functional association network and instead look for enrichment of network connections between gene sets. We present a new network crosstalk analysis method BinoX that determines the statistical significance of network link enrichment or depletion between gene sets, using the binomial distribution. This is a much more appropriate statistical model than previous methods have employed, and as a result BinoX yields substantially better true positive and FPRs than was possible before. A number of benchmarks were performed to assess the accuracy of BinoX and competing methods. We demonstrate examples of how BinoX finds many biologically meaningful pathway annotations for gene sets from cancer and other diseases, which are not found by other methods. BinoX is available at http://sonnhammer.org/BinoX. PMID:27664219

  9. An automated method for the evaluation of the pointing accuracy of sun-tracking devices

    NASA Astrophysics Data System (ADS)

    Baumgartner, Dietmar J.; Rieder, Harald E.; Pötzi, Werner; Freislich, Heinrich; Strutzmann, Heinz

    2016-04-01

    The accuracy of measurements of solar radiation (direct and diffuse radiation) depends significantly on the accuracy of the operational sun-tracking device. Thus rigid targets for instrument performance and operation are specified for international monitoring networks, such as e.g., the Baseline Surface Radiation Network (BSRN) operating under the auspices of the World Climate Research Program (WCRP). Sun-tracking devices fulfilling these accuracy targets are available from various instrument manufacturers, however none of the commercially available systems comprises a secondary accuracy control system, allowing platform operators to independently validate the pointing accuracy of sun-tracking sensors during operation. Here we present KSO-STREAMS (KSO-SunTRackEr Accuracy Monitoring System), a fully automated, system independent and cost-effective method for evaluating the pointing accuracy of sun-tracking devices. We detail the monitoring system setup, its design and specifications and results from its application to the sun-tracking system operated at the Austrian RADiation network (ARAD) site Kanzelhöhe Observatory (KSO). Results from KSO-STREAMS (for mid-March to mid-June 2015) show that the tracking accuracy of the device operated at KSO lies well within BSRN specifications (i.e. 0.1 degree accuracy). We contrast results during clear-sky and partly cloudy conditions documenting sun-tracking performance at manufacturer specified accuracies for active tracking (0.02 degrees) and highlight accuracies achieved during passive tracking i.e. periods with less than 300 W m-2 direct radiation. Furthermore we detail limitations to tracking surveillance during overcast conditions and periods of partial solar limb coverage by clouds.

  10. Statistical downscaling of precipitation using local regression and high accuracy surface modeling method

    NASA Astrophysics Data System (ADS)

    Zhao, Na; Yue, Tianxiang; Zhou, Xun; Zhao, Mingwei; Liu, Yu; Du, Zhengping; Zhang, Lili

    2016-03-01

    Downscaling precipitation is required in local scale climate impact studies. In this paper, a statistical downscaling scheme was presented with a combination of geographically weighted regression (GWR) model and a recently developed method, high accuracy surface modeling method (HASM). This proposed method was compared with another downscaling method using the Coupled Model Intercomparison Project Phase 5 (CMIP5) database and ground-based data from 732 stations across China for the period 1976-2005. The residual which was produced by GWR was modified by comparing different interpolators including HASM, Kriging, inverse distance weighted method (IDW), and Spline. The spatial downscaling from 1° to 1-km grids for period 1976-2005 and future scenarios was achieved by using the proposed downscaling method. The prediction accuracy was assessed at two separate validation sites throughout China and Jiangxi Province on both annual and seasonal scales, with the root mean square error (RMSE), mean relative error (MRE), and mean absolute error (MAE). The results indicate that the developed model in this study outperforms the method that builds transfer function using the gauge values. There is a large improvement in the results when using a residual correction with meteorological station observations. In comparison with other three classical interpolators, HASM shows better performance in modifying the residual produced by local regression method. The success of the developed technique lies in the effective use of the datasets and the modification process of the residual by using HASM. The results from the future climate scenarios show that precipitation exhibits overall increasing trend from T1 (2011-2040) to T2 (2041-2070) and T2 to T3 (2071-2100) in RCP2.6, RCP4.5, and RCP8.5 emission scenarios. The most significant increase occurs in RCP8.5 from T2 to T3, while the lowest increase is found in RCP2.6 from T2 to T3, increased by 47.11 and 2.12 mm, respectively.

  11. Accuracy Assessment of Using Rapid Prototyping Drill Templates for Atlantoaxial Screw Placement: A Cadaver Study

    PubMed Central

    Guo, Shuai; Lu, Teng; Hu, Qiaolong; Yang, Baohui; He, Xijing

    2016-01-01

    Purpose. To preliminarily evaluate the feasibility and accuracy of using rapid prototyping drill templates (RPDTs) for C1 lateral mass screw (C1-LMS) and C2 pedicle screw (C2-PS) placement. Methods. 23 formalin-fixed craniocervical cadaver specimens were randomly divided into two groups. In the conventional method group, intraoperative fluoroscopy was used to assist the screw placement. In the RPDT navigation group, specific RPDTs were constructed for each specimen and were used intraoperatively for screw placement navigation. The screw position, the operating time, and the fluoroscopy time for each screw placement were compared between the 2 groups. Results. Compared with the conventional method, the RPDT technique significantly increased the placement accuracy of the C2-PS (p < 0.05). In the axial plane, using RPDTs also significantly increased C1-LMS placement accuracy (p < 0.05). In the sagittal plane, although using RPDTs had a very high accuracy rate (100%) in C1-LMS placement, it was not statistically significant compared with the conventional method (p > 0.05). Moreover, the RPDT technique significantly decreased the operating and fluoroscopy times. Conclusion. Using RPDTs significantly increases the accuracy of C1-LMS and C2-PS placement while decreasing the screw placement time and the radiation exposure. Due to these advantages, this approach is worth promoting for use in the Harms technique. PMID:28004004

  12. Subglacial bedform orientation, one-dimensional size, and directional shape measurement method accuracy

    NASA Astrophysics Data System (ADS)

    Jorge, Marco G.; Brennand, Tracy A.

    2016-04-01

    This study is an assessment of previously reported automated methods and of a new method for measuring longitudinal subglacial bedform (LSB) morphometry. It evaluates the adequacy (accuracy and precision) of orientation, length and longitudinal asymmetry data derived from the longest straight line (LSL) enclosed by the LSB's footprint, the footprint's minimum bounding rectangle longitudinal axis (RLA) and the footprint's standard deviational ellipse (SDE) longitudinal axis (LA) (new method), and the adequacy of length based on an ellipse fitted to the area and perimeter of the footprint (elliptical length). Tests are based on 100 manually mapped drumlins and mega-scale glacial lineations representing the size and shape range of LSBs in the Puget Lowland drumlin field, WA, USA. Data from manually drawn LAs are used as reference for method evaluation. With the exception of elliptical length, errors decrease rapidly with increasing footprint elongation (decreasing potential angular divergence between LAs). For LSBs with elongation <5 and excluding the 5% largest errors (outliers), 1) the LSL, RLA and SDE methods had very small mean absolute error (MAE) in all measures (e.g., MAE <5° in orientation and <5 m in length); they can be confidently used to characterize the central tendency of LSB samples. 2) When analyzing data spatially at large cartographic scales, the LSL method should be avoided for orientation (36% of the errors were larger than 5°). 3) Elliptical length was the least accurate of all methods (MAE of 56.1 m and 15% of the errors larger than 5%); its use should be discontinued. 4) The relative adequacy of the LSL and RLA depends on footprint shape; SDE computed with the footprint's structural vertices is relatively shape-independent and is the preferred method. This study is significant also for negative-relief, and fluvial and aeolian bedforms.

  13. Accuracy of Assessment of Eligibility for Early Medical Abortion by Community Health Workers in Ethiopia, India and South Africa

    PubMed Central

    Nguyen, My Huong; Habib, Ndema; Afework, Mesganaw Fantahun; Harries, Jane; Iyengar, Kirti; Moodley, Jennifer; Constant, Deborah; Sen, Swapnaleen

    2016-01-01

    Objective To assess the accuracy of assessment of eligibility for early medical abortion by community health workers using a simple checklist toolkit. Design Diagnostic accuracy study. Setting Ethiopia, India and South Africa. Methods Two hundred seventeen women in Ethiopia, 258 in India and 236 in South Africa were enrolled into the study. A checklist toolkit to determine eligibility for early medical abortion was validated by comparing results of clinician and community health worker assessment of eligibility using the checklist toolkit with the reference standard exam. Results Accuracy was over 90% and the negative likelihood ratio <0.1 at all three sites when used by clinician assessors. Positive likelihood ratios were 4.3 in Ethiopia, 5.8 in India and 6.3 in South Africa. When used by community health workers the overall accuracy of the toolkit was 92% in Ethiopia, 80% in India and 77% in South Africa negative likelihood ratios were 0.08 in Ethiopia, 0.25 in India and 0.22 in South Africa and positive likelihood ratios were 5.9 in Ethiopia and 2.0 in India and South Africa. Conclusion The checklist toolkit, as used by clinicians, was excellent at ruling out participants who were not eligible, and moderately effective at ruling in participants who were eligible for medical abortion. Results were promising when used by community health workers particularly in Ethiopia where they had more prior experience with use of diagnostic aids and longer professional training. The checklist toolkit assessments resulted in some participants being wrongly assessed as eligible for medical abortion which is an area of concern. Further research is needed to streamline the components of the tool, explore optimal duration and content of training for community health workers, and test feasibility and acceptability. PMID:26731176

  14. Accuracy assessment: The statistical approach to performance evaluation in LACIE. [Great Plains corridor, United States

    NASA Technical Reports Server (NTRS)

    Houston, A. G.; Feiveson, A. H.; Chhikara, R. S.; Hsu, E. M. (Principal Investigator)

    1979-01-01

    A statistical methodology was developed to check the accuracy of the products of the experimental operations throughout crop growth and to determine whether the procedures are adequate to accomplish the desired accuracy and reliability goals. It has allowed the identification and isolation of key problems in wheat area yield estimation, some of which have been corrected and some of which remain to be resolved. The major unresolved problem in accuracy assessment is that of precisely estimating the bias of the LACIE production estimator. Topics covered include: (1) evaluation techniques; (2) variance and bias estimation for the wheat production estimate; (3) the 90/90 evaluation; (4) comparison of the LACIE estimate with reference standards; and (5) first and second order error source investigations.

  15. Assessment of the genomic prediction accuracy for feed efficiency traits in meat-type chickens

    PubMed Central

    Wang, Jie; Ma, Jie; Shu, Dingming; Lund, Mogens Sandø; Su, Guosheng; Qu, Hao

    2017-01-01

    Feed represents the major cost of chicken production. Selection for improving feed utilization is a feasible way to reduce feed cost and greenhouse gas emissions. The objectives of this study were to investigate the efficiency of genomic prediction for feed conversion ratio (FCR), residual feed intake (RFI), average daily gain (ADG) and average daily feed intake (ADFI) and to assess the impact of selection for feed efficiency traits FCR and RFI on eviscerating percentage (EP), breast muscle percentage (BMP) and leg muscle percentage (LMP) in meat-type chickens. Genomic prediction was assessed using a 4-fold cross-validation for two validation scenarios. The first scenario was a random family sampling validation (CVF), and the second scenario was a random individual sampling validation (CVR). Variance components were estimated based on the genomic relationship built with single nucleotide polymorphism markers. Genomic estimated breeding values (GEBV) were predicted using a genomic best linear unbiased prediction model. The accuracies of GEBV were evaluated in two ways: the correlation between GEBV and corrected phenotypic value divided by the square root of heritability, i.e., the correlation-based accuracy, and model-based theoretical accuracy. Breeding values were also predicted using a conventional pedigree-based best linear unbiased prediction model in order to compare accuracies of genomic and conventional predictions. The heritability estimates of FCR and RFI were 0.29 and 0.50, respectively. The heritability estimates of ADG, ADFI, EP, BMP and LMP ranged from 0.34 to 0.53. In the CVF scenario, the correlation-based accuracy and the theoretical accuracy of genomic prediction for FCR were slightly higher than those for RFI. The correlation-based accuracies for FCR, RFI, ADG and ADFI were 0.360, 0.284, 0.574 and 0.520, respectively, and the model-based theoretical accuracies were 0.420, 0.414, 0.401 and 0.382, respectively. In the CVR scenario, the correlation

  16. The suitability of common metrics for assessing parotid and larynx autosegmentation accuracy.

    PubMed

    Beasley, William J; McWilliam, Alan; Aitkenhead, Adam; Mackay, Ranald I; Rowbottom, Carl G

    2016-03-08

    Contouring structures in the head and neck is time-consuming, and automatic seg-mentation is an important part of an adaptive radiotherapy workflow. Geometric accuracy of automatic segmentation algorithms has been widely reported, but there is no consensus as to which metrics provide clinically meaningful results. This study investigated whether geometric accuracy (as quantified by several commonly used metrics) was associated with dosimetric differences for the parotid and larynx, comparing automatically generated contours against manually drawn ground truth contours. This enabled the suitability of different commonly used metrics to be assessed for measuring automatic segmentation accuracy of the parotid and larynx. Parotid and larynx structures for 10 head and neck patients were outlined by five clinicians to create ground truth structures. An automatic segmentation algorithm was used to create automatically generated normal structures, which were then used to create volumetric-modulated arc therapy plans. The mean doses to the automatically generated structures were compared with those of the corresponding ground truth structures, and the relative difference in mean dose was calculated for each structure. It was found that this difference did not correlate with the geometric accuracy provided by several metrics, notably the Dice similarity coefficient, which is a commonly used measure of spatial overlap. Surface-based metrics provided stronger correlation and are, therefore, more suitable for assessing automatic seg-mentation of the parotid and larynx.

  17. Electronic and Courier Methods of Information Dissemination: A Test of Accuracy.

    ERIC Educational Resources Information Center

    DeWine, Sue; And Others

    As part of a larger endeavor to evaluate the impact of communication technology on organizations, this study assesses the accuracy of information diffusion via electronic-mail and courier-mail systems in two large organizations which have implemented electronic-mail systems in the last three years. Data were obtained through the use of…

  18. Evaluating the effect of learning style and student background on self-assessment accuracy

    NASA Astrophysics Data System (ADS)

    Alaoutinen, Satu

    2012-06-01

    This study evaluates a new taxonomy-based self-assessment scale and examines factors that affect assessment accuracy and course performance. The scale is based on Bloom's Revised Taxonomy and is evaluated by comparing students' self-assessment results with course performance in a programming course. Correlation has been used to reveal possible connections between student information and both self-assessment and course performance. The results show that students can place their knowledge along the taxonomy-based scale quite well and the scale seems to fit engineering students' learning style. Advanced students assess themselves more accurately than novices. The results also show that reflective students were better in programming than active. The scale used in this study gives a more objective picture of students' knowledge than general scales and with modifications it can be used in other classes than programming.

  19. QUADAS-2: a revised tool for the quality assessment of diagnostic accuracy studies.

    PubMed

    Whiting, Penny F; Rutjes, Anne W S; Westwood, Marie E; Mallett, Susan; Deeks, Jonathan J; Reitsma, Johannes B; Leeflang, Mariska M G; Sterne, Jonathan A C; Bossuyt, Patrick M M

    2011-10-18

    In 2003, the QUADAS tool for systematic reviews of diagnostic accuracy studies was developed. Experience, anecdotal reports, and feedback suggested areas for improvement; therefore, QUADAS-2 was developed. This tool comprises 4 domains: patient selection, index test, reference standard, and flow and timing. Each domain is assessed in terms of risk of bias, and the first 3 domains are also assessed in terms of concerns regarding applicability. Signalling questions are included to help judge risk of bias. The QUADAS-2 tool is applied in 4 phases: summarize the review question, tailor the tool and produce review-specific guidance, construct a flow diagram for the primary study, and judge bias and applicability. This tool will allow for more transparent rating of bias and applicability of primary diagnostic accuracy studies.

  20. Methods for Aquatic Resource Assessment

    EPA Science Inventory

    The Methods for Aquatic Resource Assessment (MARA) project consists of three main activities in support of assessing the conditions of the nation’s aquatic resources: 1) scientific support for EPA Office of Water’s national aquatic resource surveys; 2) spatial predications of riv...

  1. Assessment Methods in Medical Education

    ERIC Educational Resources Information Center

    Norcini, John J.; McKinley, Danette W.

    2007-01-01

    Since the 1950s, there has been rapid and extensive change in the way assessment is conducted in medical education. Several new methods of assessment have been developed and implemented over this time and they have focused on clinical skills (taking a history from a patient and performing a physical examination), communication skills, procedural…

  2. Accuracy assessment of topographic mapping using UAV image integrated with satellite images

    NASA Astrophysics Data System (ADS)

    Azmi, S. M.; Ahmad, Baharin; Ahmad, Anuar

    2014-02-01

    Unmanned Aerial Vehicle or UAV is extensively applied in various fields such as military applications, archaeology, agriculture and scientific research. This study focuses on topographic mapping and map updating. UAV is one of the alternative ways to ease the process of acquiring data with lower operating costs, low manufacturing and operational costs, plus it is easy to operate. Furthermore, UAV images will be integrated with QuickBird images that are used as base maps. The objective of this study is to make accuracy assessment and comparison between topographic mapping using UAV images integrated with aerial photograph and satellite image. The main purpose of using UAV image is as a replacement for cloud covered area which normally exists in aerial photograph and satellite image, and for updating topographic map. Meanwhile, spatial resolution, pixel size, scale, geometric accuracy and correction, image quality and information contents are important requirements needed for the generation of topographic map using these kinds of data. In this study, ground control points (GCPs) and check points (CPs) were established using real time kinematic Global Positioning System (RTK-GPS) technique. There are two types of analysis that are carried out in this study which are quantitative and qualitative assessments. Quantitative assessment is carried out by calculating root mean square error (RMSE). The outputs of this study include topographic map and orthophoto. From this study, the accuracy of UAV image is ± 0.460 m. As conclusion, UAV image has the potential to be used for updating of topographic maps.

  3. Three-Dimensional Accuracy of Facial Scan for Facial Deformities in Clinics: A New Evaluation Method for Facial Scanner Accuracy

    PubMed Central

    Zhao, Yi-jiao; Xiong, Yu-xue; Wang, Yong

    2017-01-01

    In this study, the practical accuracy (PA) of optical facial scanners for facial deformity patients in oral clinic was evaluated. Ten patients with a variety of facial deformities from oral clinical were included in the study. For each patient, a three-dimensional (3D) face model was acquired, via a high-accuracy industrial “line-laser” scanner (Faro), as the reference model and two test models were obtained, via a “stereophotography” (3dMD) and a “structured light” facial scanner (FaceScan) separately. Registration based on the iterative closest point (ICP) algorithm was executed to overlap the test models to reference models, and “3D error” as a new measurement indicator calculated by reverse engineering software (Geomagic Studio) was used to evaluate the 3D global and partial (upper, middle, and lower parts of face) PA of each facial scanner. The respective 3D accuracy of stereophotography and structured light facial scanners obtained for facial deformities was 0.58±0.11 mm and 0.57±0.07 mm. The 3D accuracy of different facial partitions was inconsistent; the middle face had the best performance. Although the PA of two facial scanners was lower than their nominal accuracy (NA), they all met the requirement for oral clinic use. PMID:28056044

  4. Three-Dimensional Accuracy of Facial Scan for Facial Deformities in Clinics: A New Evaluation Method for Facial Scanner Accuracy.

    PubMed

    Zhao, Yi-Jiao; Xiong, Yu-Xue; Wang, Yong

    2017-01-01

    In this study, the practical accuracy (PA) of optical facial scanners for facial deformity patients in oral clinic was evaluated. Ten patients with a variety of facial deformities from oral clinical were included in the study. For each patient, a three-dimensional (3D) face model was acquired, via a high-accuracy industrial "line-laser" scanner (Faro), as the reference model and two test models were obtained, via a "stereophotography" (3dMD) and a "structured light" facial scanner (FaceScan) separately. Registration based on the iterative closest point (ICP) algorithm was executed to overlap the test models to reference models, and "3D error" as a new measurement indicator calculated by reverse engineering software (Geomagic Studio) was used to evaluate the 3D global and partial (upper, middle, and lower parts of face) PA of each facial scanner. The respective 3D accuracy of stereophotography and structured light facial scanners obtained for facial deformities was 0.58±0.11 mm and 0.57±0.07 mm. The 3D accuracy of different facial partitions was inconsistent; the middle face had the best performance. Although the PA of two facial scanners was lower than their nominal accuracy (NA), they all met the requirement for oral clinic use.

  5. Positioning accuracy assessment for the 4GEO/5IGSO/2MEO constellation of COMPASS

    NASA Astrophysics Data System (ADS)

    Zhou, ShanShi; Cao, YueLing; Zhou, JianHua; Hu, XiaoGong; Tang, ChengPan; Liu, Li; Guo, Rui; He, Feng; Chen, JunPing; Wu, Bin

    2012-12-01

    Determined to become a new member of the well-established GNSS family, COMPASS (or BeiDou-2) is developing its capabilities to provide high accuracy positioning services. Two positioning modes are investigated in this study to assess the positioning accuracy of COMPASS' 4GEO/5IGSO/2MEO constellation. Precise Point Positioning (PPP) for geodetic users and real-time positioning for common navigation users are utilized. To evaluate PPP accuracy, coordinate time series repeatability and discrepancies with GPS' precise positioning are computed. Experiments show that COMPASS PPP repeatability for the east, north and up components of a receiver within mainland China is better than 2 cm, 2 cm and 5 cm, respectively. Apparent systematic offsets of several centimeters exist between COMPASS precise positioning and GPS precise positioning, indicating errors remaining in the treatments of COMPASS measurement and dynamic models and reference frame differences existing between two systems. For common positioning users, COMPASS provides both open and authorized services with rapid differential corrections and integrity information available to authorized users. Our assessment shows that in open service positioning accuracy of dual-frequency and single-frequency users is about 5 m and 6 m (RMS), respectively, which may be improved to about 3 m and 4 m (RMS) with the addition of differential corrections. Less accurate Signal In Space User Ranging Error (SIS URE) and Geometric Dilution of Precision (GDOP) contribute to the relatively inferior accuracy of COMPASS as compared to GPS. Since the deployment of the remaining 1 GEO and 2 MEO is not able to significantly improve GDOP, the performance gap could only be overcome either by the use of differential corrections or improvement of the SIS URE, or both.

  6. Improved accuracy for finite element structural analysis via a new integrated force method

    NASA Technical Reports Server (NTRS)

    Patnaik, Surya N.; Hopkins, Dale A.; Aiello, Robert A.; Berke, Laszlo

    1992-01-01

    A comparative study was carried out to determine the accuracy of finite element analyses based on the stiffness method, a mixed method, and the new integrated force and dual integrated force methods. The numerical results were obtained with the following software: MSC/NASTRAN and ASKA for the stiffness method; an MHOST implementation method for the mixed method; and GIFT for the integrated force methods. The results indicate that on an overall basis, the stiffness and mixed methods present some limitations. The stiffness method generally requires a large number of elements in the model to achieve acceptable accuracy. The MHOST method tends to achieve a higher degree of accuracy for course models than does the stiffness method implemented by MSC/NASTRAN and ASKA. The two integrated force methods, which bestow simultaneous emphasis on stress equilibrium and strain compatibility, yield accurate solutions with fewer elements in a model. The full potential of these new integrated force methods remains largely unexploited, and they hold the promise of spawning new finite element structural analysis tools.

  7. Improved accuracy for finite element structural analysis via an integrated force method

    NASA Technical Reports Server (NTRS)

    Patnaik, S. N.; Hopkins, D. A.; Aiello, R. A.; Berke, L.

    1992-01-01

    A comparative study was carried out to determine the accuracy of finite element analyses based on the stiffness method, a mixed method, and the new integrated force and dual integrated force methods. The numerical results were obtained with the following software: MSC/NASTRAN and ASKA for the stiffness method; an MHOST implementation method for the mixed method; and GIFT for the integrated force methods. The results indicate that on an overall basis, the stiffness and mixed methods present some limitations. The stiffness method generally requires a large number of elements in the model to achieve acceptable accuracy. The MHOST method tends to achieve a higher degree of accuracy for course models than does the stiffness method implemented by MSC/NASTRAN and ASKA. The two integrated force methods, which bestow simultaneous emphasis on stress equilibrium and strain compatibility, yield accurate solutions with fewer elements in a model. The full potential of these new integrated force methods remains largely unexploited, and they hold the promise of spawning new finite element structural analysis tools.

  8. Improvement of Accuracy in Environmental Dosimetry by TLD Cards Using Three-dimensional Calibration Method

    PubMed Central

    HosseiniAliabadi, S. J.; Hosseini Pooya, S. M.; Afarideh, H.; Mianji, F.

    2015-01-01

    Introduction The angular dependency of response for TLD cards may cause deviation from its true value on the results of environmental dosimetry, since TLDs may be exposed to radiation at different angles of incidence from the surrounding area. Objective A 3D setting of TLD cards has been calibrated isotropically in a standard radiation field to evaluate the improvement of the accuracy of measurement for environmental dosimetry. Method Three personal TLD cards were rectangularly placed in a cylindrical holder, and calibrated using 1D and 3D calibration methods. Then, the dosimeter has been used simultaneously with a reference instrument in a real radiation field measuring the accumulated dose within a time interval. Result The results show that the accuracy of measurement has been improved by 6.5% using 3D calibration factor in comparison with that of normal 1D calibration method. Conclusion This system can be utilized in large scale environmental monitoring with a higher accuracy. PMID:26157729

  9. Results of 17 Independent Geopositional Accuracy Assessments of Earth Satellite Corporation's GeoCover Landsat Thematic Mapper Imagery. Geopositional Accuracy Validation of Orthorectified Landsat TM Imagery: Northeast Asia

    NASA Technical Reports Server (NTRS)

    Smith, Charles M.

    2003-01-01

    This report provides results of an independent assessment of the geopositional accuracy of the Earth Satellite (EarthSat) Corporation's GeoCover, Orthorectified Landsat Thematic Mapper (TM) imagery over Northeast Asia. This imagery was purchased through NASA's Earth Science Enterprise (ESE) Scientific Data Purchase (SDP) program.

  10. Assessing accuracy and precision for field and laboratory data: a perspective in ecosystem restoration

    USGS Publications Warehouse

    Stapanian, Martin A.; Lewis, Timothy E; Palmer, Craig J.; Middlebrook Amos, Molly

    2016-01-01

    Unlike most laboratory studies, rigorous quality assurance/quality control (QA/QC) procedures may be lacking in ecosystem restoration (“ecorestoration”) projects, despite legislative mandates in the United States. This is due, in part, to ecorestoration specialists making the false assumption that some types of data (e.g. discrete variables such as species identification and abundance classes) are not subject to evaluations of data quality. Moreover, emergent behavior manifested by complex, adapting, and nonlinear organizations responsible for monitoring the success of ecorestoration projects tend to unconsciously minimize disorder, QA/QC being an activity perceived as creating disorder. We discuss similarities and differences in assessing precision and accuracy for field and laboratory data. Although the concepts for assessing precision and accuracy of ecorestoration field data are conceptually the same as laboratory data, the manner in which these data quality attributes are assessed is different. From a sample analysis perspective, a field crew is comparable to a laboratory instrument that requires regular “recalibration,” with results obtained by experts at the same plot treated as laboratory calibration standards. Unlike laboratory standards and reference materials, the “true” value for many field variables is commonly unknown. In the laboratory, specific QA/QC samples assess error for each aspect of the measurement process, whereas field revisits assess precision and accuracy of the entire data collection process following initial calibration. Rigorous QA/QC data in an ecorestoration project are essential for evaluating the success of a project, and they provide the only objective “legacy” of the dataset for potential legal challenges and future uses.

  11. Integrating Landsat and California pesticide exposure estimation at aggregated analysis scales: Accuracy assessment of rurality

    NASA Astrophysics Data System (ADS)

    Vopham, Trang Minh

    Pesticide exposure estimation in epidemiologic studies can be constrained to analysis scales commonly available for cancer data - census tracts and ZIP codes. Research goals included (1) demonstrating the feasibility of modifying an existing geographic information system (GIS) pesticide exposure method using California Pesticide Use Reports (PURs) and land use surveys to incorporate Landsat remote sensing and to accommodate aggregated analysis scales, and (2) assessing the accuracy of two rurality metrics (quality of geographic area being rural), Rural-Urban Commuting Area (RUCA) codes and the U.S. Census Bureau urban-rural system, as surrogates for pesticide exposure when compared to the GIS gold standard. Segments, derived from 1985 Landsat NDVI images, were classified using a crop signature library (CSL) created from 1990 Landsat NDVI images via a sum of squared differences (SSD) measure. Organochlorine, organophosphate, and carbamate Kern County PUR applications (1974-1990) were matched to crop fields using a modified three-tier approach. Annual pesticide application rates (lb/ac), and sensitivity and specificity of each rurality metric were calculated. The CSL (75 land use classes) classified 19,752 segments [median SSD 0.06 NDVI]. Of the 148,671 PUR records included in the analysis, Landsat contributed 3,750 (2.5%) additional tier matches. ZIP Code Tabulation Area (ZCTA) rates ranged between 0 and 1.36 lb/ac and census tract rates between 0 and 1.57 lb/ac. Rurality was a mediocre pesticide exposure surrogate; higher rates were observed among urban areal units. ZCTA-level RUCA codes offered greater specificity (39.1-60%) and sensitivity (25-42.9%). The U.S. Census Bureau metric offered greater specificity (92.9-97.5%) at the census tract level; sensitivity was low (≤6%). The feasibility of incorporating Landsat into a modified three-tier GIS approach was demonstrated. Rurality accuracy is affected by rurality metric, areal aggregation, pesticide chemical

  12. Analysis of reliability, accuracy, sensitivity and predictive value of a subjective method to classify facial pattern in adults

    PubMed Central

    Queiroz, Gilberto Vilanova; Rino, José; de Paiva, João Batista; Capelozza, Leopoldino

    2016-01-01

    ABSTRACT Introduction: Craniofacial pattern diagnosis is vital in Orthodontics, as it influences decision-making regarding treatment options and prognosis. Capelozza Filho proposed a subjective method for facial classification comprising five patterns: I, II, III, Long Face and Short Face. Objective: To investigate the accuracy of a subjective classification method of facial patterns applied to adults. Methods: A sample consisting of 52 adults was used for this study. Frontal and lateral view photographs were taken with subjects at rest position, including frontal smile. Lateral cephalometric radiographs were organized in a PowerPoint® presentation and submitted to 20 raters. Method performance was assessed by examining reproducibility with Kappa test and calculating accuracy, sensitivity and positive predictive values, for which 70% was set as critical value. The gold standard of the classification was personally set by the author of the method. Results: Reproducibility was considered moderate (Kappa = 0.501); while accuracy, sensitivity and positive predictive values yielded similar results, but below 70%. Conclusions: The subjective method of facial classification employed in the present study still needs to have its morphological criteria improved in order to be used to discriminate the five facial patterns. PMID:28125141

  13. Improving the accuracy of convexity splitting methods for gradient flow equations

    NASA Astrophysics Data System (ADS)

    Glasner, Karl; Orizaga, Saulo

    2016-06-01

    This paper introduces numerical time discretization methods which significantly improve the accuracy of the convexity-splitting approach of Eyre (1998) [7], while retaining the same numerical cost and stability properties. A first order method is constructed by iteration of a semi-implicit method based upon decomposing the energy into convex and concave parts. A second order method is also presented based on backwards differentiation formulas. Several extrapolation procedures for iteration initialization are proposed. We show that, under broad circumstances, these methods have an energy decreasing property, leading to good numerical stability. The new schemes are tested using two evolution equations commonly used in materials science: the Cahn-Hilliard equation and the phase field crystal equation. We find that our methods can increase accuracy by many orders of magnitude in comparison to the original convexity-splitting algorithm. In addition, the optimal methods require little or no iteration, making their computation cost similar to the original algorithm.

  14. Assessing the accuracy of an inter-institutional automated patient-specific health problem list

    PubMed Central

    2010-01-01

    Background Health problem lists are a key component of electronic health records and are instrumental in the development of decision-support systems that encourage best practices and optimal patient safety. Most health problem lists require initial clinical information to be entered manually and few integrate information across care providers and institutions. This study assesses the accuracy of a novel approach to create an inter-institutional automated health problem list in a computerized medical record (MOXXI) that integrates three sources of information for an individual patient: diagnostic codes from medical services claims from all treating physicians, therapeutic indications from electronic prescriptions, and single-indication drugs. Methods Data for this study were obtained from 121 general practitioners and all medical services provided for 22,248 of their patients. At the opening of a patient's file, all health problems detected through medical service utilization or single-indication drug use were flagged to the physician in the MOXXI system. Each new arising health problem were presented as 'potential' and physicians were prompted to specify if the health problem was valid (Y) or not (N) or if they preferred to reassess its validity at a later time. Results A total of 263,527 health problems, representing 891 unique problems, were identified for the group of 22,248 patients. Medical services claims contributed to the majority of problems identified (77%), followed by therapeutic indications from electronic prescriptions (14%), and single-indication drugs (9%). Physicians actively chose to assess 41.7% (n = 106,950) of health problems. Overall, 73% of the problems assessed were considered valid; 42% originated from medical service diagnostic codes, 11% from single indication drugs, and 47% from prescription indications. Twelve percent of problems identified through other treating physicians were considered valid compared to 28% identified through study

  15. How Can We Evaluate the Accuracy of Small Stream Maps? -Focusing on Sampling Method and Statistical Analysis -

    NASA Astrophysics Data System (ADS)

    Park, J.

    2010-12-01

    The Washington State Department of Natural Resources’ (DNR) Forest Practices Habitat Conservation Plan (FPHCP) requires establishment of riparian management zones (RMZs) or equipment limitation zones (ELZs). In order to establish RMZs and ELZs, the DNR is required to update GIS-based stream maps showing the locations of type Ns (Non-fish seasonal) streams as well as type S (Shorelines of the state), type F (Fish habitat), and type Np (Non-fish perennial) streams. While there are few disputes over the positional accuracy of large streams, the representation of small streams such as Ns and small type S or F streams (less than 10’ width) have been considered to need more improvement of their positional accuracy. Numerous remotely sensed stream-mapping methods have been developed in the last several decades that use an array of remote sensing data such as aerial photography, satellite optical imagery, and Digital Elevation Model (DEM) topographic data. While the positional accuracy of the final stream map products has been considered essential to determine the map quality, the estimation or comparison of the positional accuracy of small stream map products has not been well studied, and rarely attempted by remotely sensed stream map developers. Assessments of the positional accuracy of stream maps are not covered properly because it is not easy to acquire the field reference data, especially for small streams under the canopy located in remote forest areas. More importantly, as of this writing, we are not aware of any prominent method to estimate or compare the positional accuracy of stream maps. Since general positional accuracy assessment methods for remotely sensed map products are designed for at least two dimensional features, they are not suitable for linear features such as streams. Due to the difficulties inherent in stream features, estimation methods for stream maps' accuracy have not dealt with the positional accuracy itself but the hydrological

  16. Studies of the accuracy of time integration methods for reaction-diffusion equations

    NASA Astrophysics Data System (ADS)

    Ropp, David L.; Shadid, John N.; Ober, Curtis C.

    2004-03-01

    In this study we present numerical experiments of time integration methods applied to systems of reaction-diffusion equations. Our main interest is in evaluating the relative accuracy and asymptotic order of accuracy of the methods on problems which exhibit an approximate balance between the competing component time scales. Nearly balanced systems can produce a significant coupling of the physical mechanisms and introduce a slow dynamical time scale of interest. These problems provide a challenging test for this evaluation and tend to reveal subtle differences between the various methods. The methods we consider include first- and second-order semi-implicit, fully implicit, and operator-splitting techniques. The test problems include a prototype propagating nonlinear reaction-diffusion wave, a non-equilibrium radiation-diffusion system, a Brusselator chemical dynamics system and a blow-up example. In this evaluation we demonstrate a "split personality" for the operator-splitting methods that we consider. While operator-splitting methods often obtain very good accuracy, they can also manifest a serious degradation in accuracy due to stability problems.

  17. Accuracy assessment of modeling architectural structures and details using terrestrial laser scanning

    NASA Astrophysics Data System (ADS)

    Kedzierski, M.; Walczykowski, P.; Orych, A.; Czarnecka, P.

    2015-08-01

    One of the most important aspects when performing architectural documentation of cultural heritage structures is the accuracy of both the data and the products which are generated from these data: documentation in the form of 3D models or vector drawings. The paper describes an assessment of the accuracy of modelling data acquired using a terrestrial phase scanner in relation to the density of a point cloud representing the surface of different types of construction materials typical for cultural heritage structures. This analysis includes the impact of the scanning geometry: the incidence angle of the laser beam and the scanning distance. For the purposes of this research, a test field consisting of samples of different types of construction materials (brick, wood, plastic, plaster, a ceramic tile, sheet metal) was built. The study involved conducting measurements at different angles and from a range of distances for chosen scanning densities. Data, acquired in the form of point clouds, were then filtered and modelled. An accuracy assessment of the 3D model was conducted by fitting it with the point cloud. The reflection intensity of each type of material was also analyzed, trying to determine which construction materials have the highest reflectance coefficients, and which have the lowest reflection coefficients, and in turn how this variable changes for different scanning parameters. Additionally measurements were taken of a fragment of a building in order to compare the results obtained in laboratory conditions, with those taken in field conditions.

  18. Estimating Orientation Using Magnetic and Inertial Sensors and Different Sensor Fusion Approaches: Accuracy Assessment in Manual and Locomotion Tasks

    PubMed Central

    Bergamini, Elena; Ligorio, Gabriele; Summa, Aurora; Vannozzi, Giuseppe; Cappozzo, Aurelio; Sabatini, Angelo Maria

    2014-01-01

    Magnetic and inertial measurement units are an emerging technology to obtain 3D orientation of body segments in human movement analysis. In this respect, sensor fusion is used to limit the drift errors resulting from the gyroscope data integration by exploiting accelerometer and magnetic aiding sensors. The present study aims at investigating the effectiveness of sensor fusion methods under different experimental conditions. Manual and locomotion tasks, differing in time duration, measurement volume, presence/absence of static phases, and out-of-plane movements, were performed by six subjects, and recorded by one unit located on the forearm or the lower trunk, respectively. Two sensor fusion methods, representative of the stochastic (Extended Kalman Filter) and complementary (Non-linear observer) filtering, were selected, and their accuracy was assessed in terms of attitude (pitch and roll angles) and heading (yaw angle) errors using stereophotogrammetric data as a reference. The sensor fusion approaches provided significantly more accurate results than gyroscope data integration. Accuracy improved mostly for heading and when the movement exhibited stationary phases, evenly distributed 3D rotations, it occurred in a small volume, and its duration was greater than approximately 20 s. These results were independent from the specific sensor fusion method used. Practice guidelines for improving the outcome accuracy are provided. PMID:25302810

  19. Gender differences in structured risk assessment: comparing the accuracy of five instruments.

    PubMed

    Coid, Jeremy; Yang, Min; Ullrich, Simone; Zhang, Tianqiang; Sizmur, Steve; Roberts, Colin; Farrington, David P; Rogers, Robert D

    2009-04-01

    Structured risk assessment should guide clinical risk management, but it is uncertain which instrument has the highest predictive accuracy among men and women. In the present study, the authors compared the Psychopathy Checklist-Revised (PCL-R; R. D. Hare, 1991, 2003); the Historical, Clinical, Risk Management-20 (HCR-20; C. D. Webster, K. S. Douglas, D. Eaves, & S. D. Hart, 1997); the Risk Matrix 2000-Violence (RM2000[V]; D. Thornton et al., 2003); the Violence Risk Appraisal Guide (VRAG; V. L. Quinsey, G. T. Harris, M. E. Rice, & C. A. Cormier, 1998); the Offenders Group Reconviction Scale (OGRS; J. B. Copas & P. Marshall, 1998; R. Taylor, 1999); and the total previous convictions among prisoners, prospectively assessed prerelease. The authors compared predischarge measures with subsequent offending and instruments ranked using multivariate regression. Most instruments demonstrated significant but moderate predictive ability. The OGRS ranked highest for violence among men, and the PCL-R and HCR-20 H subscale ranked highest for violence among women. The OGRS and total previous acquisitive convictions demonstrated greatest accuracy in predicting acquisitive offending among men and women. Actuarial instruments requiring no training to administer performed as well as personality assessment and structured risk assessment and were superior among men for violence.

  20. Thermal radiation view factor: Methods, accuracy and computer-aided procedures

    NASA Technical Reports Server (NTRS)

    Kadaba, P. V.

    1982-01-01

    The computer aided thermal analysis programs which predicts the result of predetermined acceptable temperature range prior to stationing of these orbiting equipment in various attitudes with respect to the Sun and the Earth was examined. Complexity of the surface geometries suggests the use of numerical schemes for the determination of these viewfactors. Basic definitions and standard methods which form the basis for various digital computer methods and various numerical methods are presented. The physical model and the mathematical methods on which a number of available programs are built are summarized. The strength and the weaknesses of the methods employed, the accuracy of the calculations and the time required for computations are evaluated. The situations where accuracies are important for energy calculations are identified and methods to save computational times are proposed. Guide to best use of the available programs at several centers and the future choices for efficient use of digital computers are included in the recommendations.

  1. Assessing the accuracy of the International Classification of Diseases codes to identify abusive head trauma: a feasibility study

    PubMed Central

    Berger, Rachel P; Parks, Sharyn; Fromkin, Janet; Rubin, Pamela; Pecora, Peter J

    2016-01-01

    Objective To assess the accuracy of an International Classification of Diseases (ICD) code-based operational case definition for abusive head trauma (AHT). Methods Subjects were children <5 years of age evaluated for AHT by a hospital-based Child Protection Team (CPT) at a tertiary care paediatric hospital with a completely electronic medical record (EMR) system. Subjects were designated as non-AHT traumatic brain injury (TBI) or AHT based on whether the CPT determined that the injuries were due to AHT. The sensitivity and specificity of the ICD-based definition were calculated. Results There were 223 children evaluated for AHT: 117 AHT and 106 non-AHT TBI. The sensitivity and specificity of the ICD-based operational case definition were 92% (95% CI 85.8 to 96.2) and 96% (95% CI 92.3 to 99.7), respectively. All errors in sensitivity and three of the four specificity errors were due to coder error; one specificity error was a physician error. Conclusions In a paediatric tertiary care hospital with an EMR system, the accuracy of an ICD-based case definition for AHT was high. Additional studies are needed to assess the accuracy of this definition in all types of hospitals in which children with AHT are cared for. PMID:24167034

  2. Height Accuracy Based on Different Rtk GPS Method for Ultralight Aircraft Images

    NASA Astrophysics Data System (ADS)

    Tahar, K. N.

    2015-08-01

    Height accuracy is one of the important elements in surveying work especially for control point's establishment which requires an accurate measurement. There are many methods can be used to acquire height value such as tacheometry, leveling and Global Positioning System (GPS). This study has investigated the effect on height accuracy based on different observations which are single based and network based GPS methods. The GPS network is acquired from the local network namely Iskandar network. This network has been setup to provide real-time correction data to rover GPS station while the single network is based on the known GPS station. Nine ground control points were established evenly at the study area. Each ground control points were observed about two and ten minutes. It was found that, the height accuracy give the different result for each observation.

  3. [Assessment of precision and accuracy of digital surface photogrammetry with the DSP 400 system].

    PubMed

    Krimmel, M; Kluba, S; Dietz, K; Reinert, S

    2005-03-01

    The objective of the present study was to evaluate the precision and accuracy of facial anthropometric measurements obtained through digital 3-D surface photogrammetry with the DSP 400 system in comparison to traditional 2-D photogrammetry. Fifty plaster casts of cleft infants were imaged and 21 standard anthropometric measurements were obtained. For precision assessment the measurements were performed twice in a subsample. Accuracy was determined by comparison of direct measurements and indirect 2-D and 3-D image measurements. Precision of digital surface photogrammetry was almost as good as direct anthropometry and clearly better than 2-D photogrammetry. Measurements derived from 3-D images showed better congruence to direct measurements than from 2-D photos. Digital surface photogrammetry with the DSP 400 system is sufficiently precise and accurate for craniofacial anthropometric examinations.

  4. An accuracy assessment of Cartesian-mesh approaches for the Euler equations

    NASA Technical Reports Server (NTRS)

    Coirier, William J.; Powell, Kenneth G.

    1995-01-01

    A critical assessment of the accuracy of Cartesian-mesh approaches for steady, transonic solutions of the Euler equations of gas dynamics is made. An exact solution of the Euler equations (Ringleb's flow) is used not only to infer the order of the truncation error of the Cartesian-mesh approaches, but also to compare the magnitude of the discrete error directly to that obtained with a structured mesh approach. Uniformly and adaptively refined solutions using a Cartesian-mesh approach are obtained and compared to each other and to uniformly refined structured mesh results. The effect of cell merging is investigated as well as the use of two different K-exact reconstruction procedures. The solution methodology of the schemes is explained and tabulated results are presented to compare the solution accuracies.

  5. Accuracy assessment of a mobile terrestrial lidar survey at Padre Island National Seashore

    USGS Publications Warehouse

    Lim, Samsung; Thatcher, Cindy A.; Brock, John C.; Kimbrow, Dustin R.; Danielson, Jeffrey J.; Reynolds, B.J.

    2013-01-01

    The higher point density and mobility of terrestrial laser scanning (light detection and ranging (lidar)) is desired when extremely detailed elevation data are needed for mapping vertically orientated complex features such as levees, dunes, and cliffs, or when highly accurate data are needed for monitoring geomorphic changes. Mobile terrestrial lidar scanners have the capability for rapid data collection on a larger spatial scale compared with tripod-based terrestrial lidar, but few studies have examined the accuracy of this relatively new mapping technology. For this reason, we conducted a field test at Padre Island National Seashore of a mobile lidar scanner mounted on a sport utility vehicle and integrated with a position and orientation system. The purpose of the study was to assess the vertical and horizontal accuracy of data collected by the mobile terrestrial lidar system, which is georeferenced to the Universal Transverse Mercator coordinate system and the North American Vertical Datum of 1988. To accomplish the study objectives, independent elevation data were collected by conducting a high-accuracy global positioning system survey to establish the coordinates and elevations of 12 targets spaced throughout the 12 km transect. These independent ground control data were compared to the lidar scanner-derived elevations to quantify the accuracy of the mobile lidar system. The performance of the mobile lidar system was also tested at various vehicle speeds and scan density settings (e.g. field of view and linear point spacing) to estimate the optimal parameters for desired point density. After adjustment of the lever arm parameters, the final point cloud accuracy was 0.060 m (east), 0.095 m (north), and 0.053 m (height). The very high density of the resulting point cloud was sufficient to map fine-scale topographic features, such as the complex shape of the sand dunes.

  6. Calibration of ground-based microwave radiometers - Accuracy assessment and recommendations for network users

    NASA Astrophysics Data System (ADS)

    Pospichal, Bernhard; Küchler, Nils; Löhnert, Ulrich; Crewell, Susanne; Czekala, Harald; Güldner, Jürgen

    2016-04-01

    Ground-based microwave radiometers (MWR) are becoming widely used in atmospheric remote sensing and start to be routinely operated by national weather services and other institutions. However, common standards for calibration of these radiometers and a detailed knowledge about the error characteristics is needed, in order to assimilate the data into models. Intercomparisons of calibrations by different MWRs have rarely been done. Therefore, two calibration experiments in Lindenberg (2014) and Meckenheim (2015) were performed in the frame of TOPROF (Cost action ES1303) in order to assess uncertainties and differences between various instruments. In addition, a series of experiments were taken in Oklahoma in autumn 2014. The focus lay on the performance of the two main instrument types, which are currently used operationally. These are the MP-Profiler series by Radiometrics Corporation as well as the HATPRO series by Radiometer Physics GmbH (RPG). Both instrument types are operating in two frequency bands, one along the 22 GHz water vapour line, the other one at the lower wing of the 60 GHz oxygen absorption complex. The goal was to establish protocols for providing quality controlled (QC) MWR data and their uncertainties. To this end, standardized calibration procedures for MWR were developed and recommendations for radiometer users were compiled. We focus here mainly on data types, integration times and optimal settings for calibration intervals, both for absolute (liquid nitrogen, tipping curve) as well as relative (hot load, noise diode) calibrations. Besides the recommendations for ground-based MWR operators, we will present methods to determine the accuracy of the calibration as well as means for automatic data quality control. In addition, some results from the intercomparison of different radiometers will be discussed.

  7. Accuracy of Panoramic Radiograph in Assessment of the Relationship Between Mandibular Canal and Impacted Third Molars

    PubMed Central

    Tantanapornkul, Weeraya; Mavin, Darika; Prapaiphittayakun, Jaruthai; Phipatboonyarat, Natnicha; Julphantong, Wanchanok

    2016-01-01

    Background: The relationship between impacted mandibular third molar and mandibular canal is important for removal of this tooth. Panoramic radiography is one of the commonly used diagnostic tools for evaluating the relationship of these two structures. Objectives: To evaluate the accuracy of panoramic radiographic findings in predicting direct contact between mandibular canal and impacted third molars on 3D digital images, and to define panoramic criterion in predicting direct contact between the two structures. Methods: Two observers examined panoramic radiographs of 178 patients (256 impacted mandibular third molars). Panoramic findings of interruption of mandibular canal wall, isolated or with darkening of third molar root, diversion of mandibular canal and narrowing of third molar root were evaluated for 3D digital radiography. Direct contact between mandibular canal and impacted third molars on 3D digital images was then correlated with panoramic findings. Panoramic criterion was also defined in predicting direct contact between the two structures. Results: Panoramic findings of interruption of mandibular canal wall, isolated or with darkening of third molar root were statistically significantly correlated with direct contact between mandibular canal and impacted third molars on 3D digital images (p < 0.005), and were defined as panoramic criteria in predicting direct contact between the two structures. Conclusion: Interruption of mandibular canal wall, isolated or with darkening of third molar root observed on panoramic radiographs were effective in predicting direct contact between mandibular canal and impacted third molars on 3D digital images. Panoramic radiography is one of the efficient diagnostic tools for pre-operative assessment of impacted mandibular third molars. PMID:27398105

  8. Assessing the dosimetric accuracy of MR-generated synthetic CT images for focal brain VMAT radiotherapy

    PubMed Central

    Paradis, Eric; Cao, Yue; Lawrence, Theodore S.; Tsien, Christina; Vineberg, Karen; Balter, James M.

    2015-01-01

    Purpose To assess the dosimetric accuracy of synthetic CT volumes generated from MRI data for focal brain radiotherapy. Methods A study was conducted on 12 patients with gliomas who underwent both MR and CT imaging as part of their simulation for external beam treatment planning. Synthetic CT (MRCT) volumes were generated from the MR images. The patients’ clinical treatment planning directives were used to create 12 individual Volumetric Modulated Arc Therapy (VMAT) plans, which were then optimized 10 times on each of their respective CT and MRCT-derived electron density maps. Dose metrics derived from optimization criteria, as well as monitor units and gamma analyses, were evaluated to quantify differences between the imaging modalities. Results Mean differences between Planning Target Volume (PTV) doses on MRCT and CT plans across all patients were 0.0% (range −0.1 to 0.2%) for D95%, 0.0% (−0.7 to 0.6%) for D5%, and −0.2% (−1.0 to 0.2%) for Dmax. MRCT plans showed no significant change in monitor units (−0.4%) compared to CT plans. Organs at risk (OARs) had an average Dmax difference of 0.0 Gy (−2.2 to 1.9 Gy) over 85 structures across all 12 patients, with no significant differences when calculated doses approached planning constraints. Conclusions Focal brain VMAT plans optimized on MRCT images show excellent dosimetric agreement with standard CT-optimized plans. PTVs show equivalent coverage, and OARs do not show any overdose. These results indicate that MRI-derived synthetic CT volumes can be used to support treatment planning of most patients treated for intracranial lesions. PMID:26581151

  9. The effect of atmospheric and topographic correction methods on land cover classification accuracy

    NASA Astrophysics Data System (ADS)

    Vanonckelen, Steven; Lhermitte, Stefaan; Van Rompaey, Anton

    2013-10-01

    Mapping of vegetation in mountain areas based on remote sensing is obstructed by atmospheric and topographic distortions. A variety of atmospheric and topographic correction methods has been proposed to minimize atmospheric and topographic effects and should in principle lead to a better land cover classification. Only a limited number of atmospheric and topographic combinations has been tested and the effect on class accuracy and on different illumination conditions is not yet researched extensively. The purpose of this study was to evaluate the effect of coupled correction methods on land cover classification accuracy. Therefore, all combinations of three atmospheric (no atmospheric correction, dark object subtraction and correction based on transmittance functions) and five topographic corrections (no topographic correction, band ratioing, cosine correction, pixel-based Minnaert and pixel-based C-correction) were applied on two acquisitions (2009 and 2010) of a Landsat image in the Romanian Carpathian mountains. The accuracies of the fifteen resulting land cover maps were evaluated statistically based on two validation sets: a random validation set and a validation subset containing pixels present in the difference area between the uncorrected classification and one of the fourteen corrected classifications. New insights into the differences in classification accuracy were obtained. First, results showed that all corrected images resulted in higher overall classification accuracies than the uncorrected images. The highest accuracy for the full validation set was achieved after combination of an atmospheric correction based on transmittance functions and a pixel-based Minnaert topographic correction. Secondly, class accuracies of especially the coniferous and mixed forest classes were enhanced after correction. There was only a minor improvement for the other land cover classes (broadleaved forest, bare soil, grass and water). This was explained by the position

  10. Accuracy of methods to measure femoral head penetration within metal-backed acetabular components.

    PubMed

    Callary, Stuart A; Solomon, Lucian B; Holubowycz, Oksana T; Campbell, David G; Howie, Donald W

    2016-06-30

    A number of different software programs are used to investigate the in vivo wear of polyethylene bearings in total hip arthroplasty. With wear rates below 0.1 mm/year now commonly being reported for highly cross-linked polyethylene (XLPE) components, it is important to identify the accuracy of the methods used to measure such small movements. The aims of this study were to compare the accuracy of current software programs used to measure two-dimensional (2D) femoral head penetration (FHP) and to determine whether the accuracy is influenced by larger femoral heads or by different methods of representing the acetabular component within radiostereometric analysis (RSA). A hip phantom was used to compare known movements of the femoral head within a metal-backed acetabular component to FHP measured radiographically using RSA, Hip Analysis Suite (HAS), PolyWare, Ein Bild Roentgen Analyse (EBRA), and Roentgen Monographic Analysis Tool (ROMAN). RSA was significantly more accurate than the HAS, PolyWare, and ROMAN methods when measuring 2D FHP with a 28 mm femoral head. Femoral head size influenced the accuracy of HAS and ROMAN 2D FHP measurements, EBRA proximal measurements, and RSA measurements in the proximal and anterior direction. The use of different acetabular reference segments did not influence accuracy of RSA measurements. The superior accuracy and reduced variability of RSA wear measurements allow much smaller cohorts to be used in RSA clinical wear studies than those utilizing other software programs. © 2016 Orthopaedic Research Society. Published by Wiley Periodicals, Inc. J Orthop Res.

  11. ESA ExoMars: Pre-launch PanCam Geometric Modeling and Accuracy Assessment

    NASA Astrophysics Data System (ADS)

    Li, D.; Li, R.; Yilmaz, A.

    2014-08-01

    ExoMars is the flagship mission of the European Space Agency (ESA) Aurora Programme. The mobile scientific platform, or rover, will carry a drill and a suite of instruments dedicated to exobiology and geochemistry research. As the ExoMars rover is designed to travel kilometres over the Martian surface, high-precision rover localization and topographic mapping will be critical for traverse path planning and safe planetary surface operations. For such purposes, the ExoMars rover Panoramic Camera system (PanCam) will acquire images that are processed into an imagery network providing vision information for photogrammetric algorithms to localize the rover and generate 3-D mapping products. Since the design of the ExoMars PanCam will influence localization and mapping accuracy, quantitative error analysis of the PanCam design will improve scientists' awareness of the achievable level of accuracy, and enable the PanCam design team to optimize its design to achieve the highest possible level of localization and mapping accuracy. Based on photogrammetric principles and uncertainty propagation theory, we have developed a method to theoretically analyze how mapping and localization accuracy would be affected by various factors, such as length of stereo hard-baseline, focal length, and pixel size, etc.

  12. Prediction accuracy in estimating joint angle trajectories using a video posture coding method for sagittal lifting tasks.

    PubMed

    Chang, Chien-Chi; McGorry, Raymond W; Lin, Jia-Hua; Xu, Xu; Hsiang, Simon M

    2010-08-01

    This study investigated prediction accuracy of a video posture coding method for lifting joint trajectory estimation. From three filming angles, the coder selected four key snapshots, identified joint angles and then a prediction program estimated the joint trajectories over the course of a lift. Results revealed a limited range of differences of joint angles (elbow, shoulder, hip, knee, ankle) between the manual coding method and the electromagnetic motion tracking system approach. Lifting range significantly affected estimate accuracy for all joints and camcorder filming angle had a significant effect on all joints but the hip. Joint trajectory predictions were more accurate for knuckle-to-shoulder lifts than for floor-to-shoulder or floor-to-knuckle lifts with average root mean square errors (RMSE) of 8.65 degrees , 11.15 degrees and 11.93 degrees , respectively. Accuracy was also greater for the filming angles orthogonal to the participant's sagittal plane (RMSE = 9.97 degrees ) as compared to filming angles of 45 degrees (RMSE = 11.01 degrees ) or 135 degrees (10.71 degrees ). The effects of lifting speed and loading conditions were minimal. To further increase prediction accuracy, improved prediction algorithms and/or better posture matching methods should be investigated. STATEMENT OF RELEVANCE: Observation and classification of postures are common steps in risk assessment of manual materials handling tasks. The ability to accurately predict lifting patterns through video coding can provide ergonomists with greater resolution in characterising or assessing the lifting tasks than evaluation based solely on sampling with a single lifting posture event.

  13. Estimation of diagnostic test accuracy without full verification: a review of latent class methods

    PubMed Central

    Collins, John; Huynh, Minh

    2014-01-01

    The performance of a diagnostic test is best evaluated against a reference test that is without error. For many diseases, this is not possible, and an imperfect reference test must be used. However, diagnostic accuracy estimates may be biased if inaccurately verified status is used as the truth. Statistical models have been developed to handle this situation by treating disease as a latent variable. In this paper, we conduct a systematized review of statistical methods using latent class models for estimating test accuracy and disease prevalence in the absence of complete verification. PMID:24910172

  14. Accuracy and limitations of localized Green's function methods for materials science applications

    NASA Astrophysics Data System (ADS)

    Smirnov, A. V.; Johnson, D. D.

    2001-12-01

    We compare screened real-space and reciprocal-space implementations of Korringa-Kohn-Rostoker electronic-structure method for their applicability to largescale problems requiring various levels of accuracy. We show that real-space calculations in metals can become impractical to describe energies. We suggest a combined r- and k-space scheme as the most efficient and flexible strategy for accurate energy calculations. Our hybrid code is suitable for (parallel) large-scale calculations involving complex, multicomponent systems. We also discuss how details of numerical procedures can affect accuracy of such calculations.

  15. Estimation of diagnostic test accuracy without full verification: a review of latent class methods.

    PubMed

    Collins, John; Huynh, Minh

    2014-10-30

    The performance of a diagnostic test is best evaluated against a reference test that is without error. For many diseases, this is not possible, and an imperfect reference test must be used. However, diagnostic accuracy estimates may be biased if inaccurately verified status is used as the truth. Statistical models have been developed to handle this situation by treating disease as a latent variable. In this paper, we conduct a systematized review of statistical methods using latent class models for estimating test accuracy and disease prevalence in the absence of complete verification.

  16. Proposed Testing to Assess the Accuracy of Glass-To-Metal Seal Stress Analyses.

    SciTech Connect

    Chambers, Robert S.; Emery, John M; Tandon, Rajan; Antoun, Bonnie R.; Stavig, Mark E.; Newton, Clay S.; Gibson, Cory S; Bencoe, Denise N.

    2014-09-01

    The material characterization tests conducted on 304L VAR stainless steel and Schott 8061 glass have provided higher fidelity data for calibration of material models used in Glass - T o - Metal (GTM) seal analyses. Specifically, a Thermo - Multi - Linear Elastic Plastic ( thermo - MLEP) material model has be en defined for S S304L and the Simplified Potential Energy Clock nonlinear visc oelastic model has been calibrated for the S8061 glass. To assess the accuracy of finite element stress analyses of GTM seals, a suite of tests are proposed to provide data for comparison to mo del predictions.

  17. Qualitative methods for assessing risk

    SciTech Connect

    Mahn, J.A.; Hannaman, G.W.; Kryska, P.

    1995-03-01

    The purpose of this document is to describe a qualitative risk assessment process that supplements the requirements of DOE/AL 5481.1B. Although facility managers have a choice of assessing risk either quantitatively or qualitatively, trade offs are involved in making the most appropriate choice for a given application. The results that can be obtained from a quantitative risk assessment are significantly more robust than those results derived from a qualitative approach. However, the advantages derived from quantitative risk assessment are achieved at a greater expenditure of money, time and convenience. This document provides the elements of a framework for performing a much less costly qualitative risk assessment, while retaining the best attributes of quantitative methods. The approach discussed herein will; (1) provide facility managers with the tools to prepare consistent, site wide assessments, and (2) aid the reviewers who may be tasked to evaluate the assessments. Added cost/benefit measures of the qualitative methodology include the identification of mechanisms for optimally allocating resources for minimizing risk in an expeditious, and fiscally responsible manner.

  18. Enhancing the accuracy of the Fowler method for monitoring non-constant work functions

    NASA Astrophysics Data System (ADS)

    Friedl, R.

    2016-04-01

    The Fowler method is a prominent non-invasive technique to determine the absolute work function of a surface based on the photoelectric effect. The evaluation procedure relies on the correlation of the photocurrent with the incident photon energy hν which is mainly dependent on the surface work function χ. Applying Fowler's theory of the photocurrent, the measurements can be fitted by the theoretical curve near the threshold hν⪆χ yielding the work function χ and a parameter A. The straightforward experimental implementation of the Fowler method is to use several particular photon energies, e.g. via interference filters. However, with a realization like that the restriction hν ≈ χ can easily be violated, especially when the work function of the material is decreasing during the measurements as, for instance, with coating or adsorption processes. This can lead to an overestimation of the evaluated work function value of typically some 0.1 eV, reaching up to more than 0.5 eV in an unfavorable case. A detailed analysis of the Fowler theory now reveals the background of that effect and shows that the fit-parameter A can be used to assess the accuracy of the determined value of χ conveniently during the measurements. Moreover, a scheme is introduced to quantify a potential overestimation and to perform a correction to χ to a certain extent. The issues are demonstrated exemplarily at the monitoring of the work function reduction of a stainless steel sample surface due to caesiation.

  19. Assessing the accuracy and repeatability of automated photogrammetrically generated digital surface models from unmanned aerial system imagery

    NASA Astrophysics Data System (ADS)

    Chavis, Christopher

    Using commercial digital cameras in conjunction with Unmanned Aerial Systems (UAS) to generate 3-D Digital Surface Models (DSMs) and orthomosaics is emerging as a cost-effective alternative to Light Detection and Ranging (LiDAR). Powerful software applications such as Pix4D and APS can automate the generation of DSM and orthomosaic products from a handful of inputs. However, the accuracy of these models is relatively untested. The objectives of this study were to generate multiple DSM and orthomosaic pairs of the same area using Pix4D and APS from flights of imagery collected with a lightweight UAS. The accuracy of each individual DSM was assessed in addition to the consistency of the method to model one location over a period of time. Finally, this study determined if the DSMs automatically generated using lightweight UAS and commercial digital cameras could be used for detecting changes in elevation and at what scale. Accuracy was determined by comparing DSMs to a series of reference points collected with survey grade GPS. Other GPS points were also used as control points to georeference the products within Pix4D and APS. The effectiveness of the products for change detection was assessed through image differencing and observance of artificially induced, known elevation changes. The vertical accuracy with the optimal data and model is ≈ 25 cm and the highest consistency over repeat flights is a standard deviation of ≈ 5 cm. Elevation change detection based on such UAS imagery and DSM models should be viable for detecting infrastructure change in urban or suburban environments with little dense canopy vegetation.

  20. Analysis on accuracy improvement of rotor-stator rubbing localization based on acoustic emission beamforming method.

    PubMed

    He, Tian; Xiao, Denghong; Pan, Qiang; Liu, Xiandong; Shan, Yingchun

    2014-01-01

    This paper attempts to introduce an improved acoustic emission (AE) beamforming method to localize rotor-stator rubbing fault in rotating machinery. To investigate the propagation characteristics of acoustic emission signals in casing shell plate of rotating machinery, the plate wave theory is used in a thin plate. A simulation is conducted and its result shows the localization accuracy of beamforming depends on multi-mode, dispersion, velocity and array dimension. In order to reduce the effect of propagation characteristics on the source localization, an AE signal pre-process method is introduced by combining plate wave theory and wavelet packet transform. And the revised localization velocity to reduce effect of array size is presented. The accuracy of rubbing localization based on beamforming and the improved method of present paper are compared by the rubbing test carried on a test table of rotating machinery. The results indicate that the improved method can localize rub fault effectively.

  1. Assessing external cause of injury coding accuracy for transport injury hospitalizations.

    PubMed

    Bowman, Stephen M; Aitken, Mary E

    2011-01-01

    External cause of injury codes (E codes) capture circumstances surrounding injuries. While hospital discharge data are primarily collected for administrative/billing purposes, these data are secondarily used for injury surveillance. We assessed the accuracy and completeness of hospital discharge data for transport-related crashes using trauma registry data as the gold standard. We identified mechanisms of injury with significant disagreement and developed recommendations to improve the accuracy of E codes in administrative data. Overall, we linked 2,192 (99.9 percent) of the 2,195 discharge records to trauma registry records. General mechanism categories showed good agreement, with 84.7 percent of records coded consistently between registry and discharge data (Kappa 0.762, p < .001). However, agreement was lower for specific categories (e.g., ATV crashes), with discharge records capturing only 70.4 percent of cases identified in trauma registry records. Efforts should focus on systematically improving E-code accuracy and detail through training, education, and informatics such as automated data linkages to trauma registries.

  2. Application of a Monte Carlo accuracy assessment tool to TDRS and GPS

    NASA Technical Reports Server (NTRS)

    Pavloff, Michael S.

    1994-01-01

    In support of a NASA study on the application of radio interferometry to satellite orbit determination, MITRE developed a simulation tool for assessing interferometric tracking accuracy. Initially, the tool was applied to the problem of determining optimal interferometric station siting for orbit determination of the Tracking and Data Relay Satellite (TDRS). Subsequently, the Orbit Determination Accuracy Estimator (ODAE) was expanded to model the general batch maximum likelihood orbit determination algorithms of the Goddard Trajectory Determination System (GTDS) with measurement types including not only group and phase delay from radio interferometry, but also range, range rate, angular measurements, and satellite-to-satellite measurements. The user of ODAE specifies the statistical properties of error sources, including inherent observable imprecision, atmospheric delays, station location uncertainty, and measurement biases. Upon Monte Carlo simulation of the orbit determination process, ODAE calculates the statistical properties of the error in the satellite state vector and any other parameters for which a solution was obtained in the orbit determination. This paper presents results from ODAE application to two different problems: (1)determination of optimal geometry for interferometirc tracking of TDRS, and (2) expected orbit determination accuracy for Global Positioning System (GPS) tracking of low-earth orbit (LEO) satellites. Conclusions about optimal ground station locations for TDRS orbit determination by radio interferometry are presented, and the feasibility of GPS-based tracking for IRIDIUM, a LEO mobile satellite communications (MOBILSATCOM) system, is demonstrated.

  3. Inter-comparison and accuracy assessment of TRMM 3B42 products over Turkey

    NASA Astrophysics Data System (ADS)

    Amjad, Muhammad; Yilmaz, M. Tugrul

    2016-04-01

    Accurate estimation of precipitation, especially over complex topography, is impeded by many factors depending on the platform that it is acquired. Satellites have the advantage of providing spatially and temporally continuous and consistent datasets. However, utilizing satellite precipitation data in various applications requires its uncertainty estimation to be carried out robustly. In this study, accuracy of two Tropical Rainfall Measurement Mission (Version 3B42) products, TRMM 3B42 V6 and TRMM3B42 V7, are assessed for their accuracy by inter-comparing their monthly time series against ground observations obtained over 256 stations in Turkey. Errors are further analyzed for their seasonal and climate-dependent variability. Both V6 and V7 products show better performance during summers than winters. V6 product has dry bias over drier regions and V7 product has wet bias over wetter regions of the country. Moreover, rainfall measuring accuracies of both versions are much lower along coastal regions and at lower altitudes. Overall, the statistics of the monthly products confirm V7 product is an improved version compared to V6. (This study was supported by TUBITAK fund # 114Y676).

  4. Retrieval of Urban Boundary Layer Structures from Doppler Lidar Data. Part I: Accuracy Assessment

    SciTech Connect

    Xia, Quanxin; Lin, Ching Long; Calhoun, Ron; Newsom, Rob K.

    2008-01-01

    Two coherent Doppler lidars from the US Army Research Laboratory (ARL) and Arizona State University (ASU) were deployed in the Joint Urban 2003 atmospheric dispersion field experiment (JU2003) held in Oklahoma City. The dual lidar data are used to evaluate the accuracy of the four-dimensional variational data assimilation (4DVAR) method and identify the coherent flow structures in the urban boundary layer. The objectives of the study are three-fold. The first objective is to examine the effect of eddy viscosity models on the quality of retrieved velocity data. The second objective is to determine the fidelity of single-lidar 4DVAR and evaluate the difference between single- and dual-lidar retrievals. The third objective is to correlate the retrieved flow structures with the ground building data. It is found that the approach of treating eddy viscosity as part of control variables yields better results than the approach of prescribing viscosity. The ARL single-lidar 4DVAR is able to retrieve radial velocity fields with an accuracy of 98% in the along-beam direction and 80-90% in the cross-beam direction. For the dual-lidar 4DVAR, the accuracy of retrieved radial velocity in the ARL cross-beam direction improves to 90-94%. By using the dual-lidar retrieved data as a reference, the single-lidar 4DVAR is able to recover fluctuating velocity fields with 70-80% accuracy in the along-beam direction and 60-70% accuracy in the cross-beam direction. Large-scale convective roll structures are found in the vicinity of downtown airpark and parks. Vortical structures are identified near the business district. Strong updrafts and downdrafts are also found above a cluster of restaurants.

  5. Evaluating IRT- and CTT-Based Methods of Estimating Classification Consistency and Accuracy Indices from Single Administrations

    ERIC Educational Resources Information Center

    Deng, Nina

    2011-01-01

    Three decision consistency and accuracy (DC/DA) methods, the Livingston and Lewis (LL) method, LEE method, and the Hambleton and Han (HH) method, were evaluated. The purposes of the study were: (1) to evaluate the accuracy and robustness of these methods, especially when their assumptions were not well satisfied, (2) to investigate the "true"…

  6. Shuttle radar topography mission accuracy assessment and evaluation for hydrologic modeling

    NASA Astrophysics Data System (ADS)

    Mercuri, Pablo Alberto

    Digital Elevation Models (DEMs) are increasingly used even in low relief landscapes for multiple mapping applications and modeling approaches such as surface hydrology, flood risk mapping, agricultural suitability, and generation of topographic attributes. The National Aeronautics and Space Administration (NASA) has produced a nearly global database of highly accurate elevation data, the Shuttle Radar Topography Mission (SRTM) DEM. The main goals of this thesis were to investigate quality issues of SRTM, provide measures of vertical accuracy with emphasis on low relief areas, and to analyze the performance for the generation of physical boundaries and streams for watershed modeling and characterization. The absolute and relative accuracy of the two SRTM resolutions, at 1 and 3 arc-seconds, were investigated to generate information that can be used as a reference in areas with similar characteristics in other regions of the world. The absolute accuracy was obtained from accurate point estimates using the best available federal geodetic network in Indiana. The SRTM root mean square error for this area of the Midwest US surpassed data specifications. It was on the order of 2 meters for the 1 arc-second resolution in flat areas of the Midwest US. Estimates of error were smaller for the global coverage 3 arc-second data with very similar results obtained in the flat plains in Argentina. In addition to calculating the vertical accuracy, the impacts of physiography and terrain attributes, like slope, on the error magnitude were studied. The assessment also included analysis of the effects of land cover on vertical accuracy. Measures of local variability were described to identify the adjacency effects produced by surface features in the SRTM DEM, like forests and manmade features near the geodetic point. Spatial relationships among the bare-earth National Elevation Data and SRTM were also analyzed to assess the relative accuracy that was 2.33 meters in terms of the total

  7. Accuracy of teacher assessments of second-language students at risk for reading disability.

    PubMed

    Limbos, M M; Geva, E

    2001-01-01

    This study examined the accuracy of teacher assessments in screening for reading disabilities among students of English as a second language (ESL) and as a first language (L1). Academic and oral language tests were administered to 369 children (249 ESL, 120 L1) at the beginning of Grade 1 and at the end of Grade 2. Concurrently, 51 teachers nominated children at risk for reading failure and completed rating scales assessing academic and oral language skills. Scholastic records were reviewed for notation of concern or referral. The criterion measure was a standardized reading score based on phonological awareness, rapid naming, and word recognition. Results indicated that teacher rating scales and nominations had low sensitivity in identifying ESL and L1 students at risk for reading disability at the 1-year mark. Relative to other forms of screening, teacher-expressed concern had lower sensitivity. Finally, oral language proficiency contributed to misclassifications in the ESL group.

  8. Accuracy of knee range of motion assessment after total knee arthroplasty.

    PubMed

    Lavernia, Carlos; D'Apuzzo, Michele; Rossi, Mark D; Lee, David

    2008-09-01

    Measurement of knee joint range of motion (ROM) is important to assess after total knee arthroplasty. Our objective was to determine level of agreement and accuracy between observers with different knowledge on total ROM after total knee arthroplasty. Forty-one patients underwent x-ray of active and passive knee ROM (gold standard). Five different raters evaluated observed and measured ROM: orthopedic surgeon, clinical fellow, physician assistant, research fellow, and a physical therapist. A 1-way analysis of variance was used to determine differences in ROM between raters over both conditions. Limit of agreement for each rater for both active and passive total ROM under both conditions was calculated. Analysis of variance indicated a difference between raters for all conditions (range, P = .004 to P < or =.0001). The trend for all raters was to overestimate ROM at higher ranges. Assessment of ROM through direct observation without a goniometer provides inaccurate findings.

  9. Analyses of odontometric sexual dimorphism and sex assessment accuracy on a large sample.

    PubMed

    Angadi, Punnya V; Hemani, S; Prabhu, Sudeendra; Acharya, Ashith B

    2013-08-01

    Correct sex assessment of skeletonized human remains allows investigators to undertake a more focused search of missing persons' files to establish identity. Univariate and multivariate odontometric sex assessment has been explored in recent years on small sample sizes and have not used a test sample. Consequently, inconsistent results have been produced in terms of accuracy of sex allocation. This paper has derived data from a large sample of males and females, and applied logistic regression formulae on a test sample. Using a digital caliper, buccolingual and mesiodistal dimensions of all permanent teeth (except third molars) were measured on 600 dental casts (306 females, 294 males) of young adults (18-32 years), and the data subjected to univariate (independent samples' t-test) and multivariate statistics (stepwise logistic regression analysis, or LRA). The analyses revealed that canines were the most sexually dimorphic teeth followed by molars. All tooth variables were larger in males, with 51/56 (91.1%) being statistically larger (p < 0.05). When the stepwise LRA formulae were applied to a test sample of 69 subjects (40 females, 29 males) of the same age range, allocation accuracy of 68.1% for the maxillary teeth, 73.9% for the mandibular teeth, and 71% for teeth of both jaws combined, were obtained. The high univariate sexual dimorphism observed herein contrasts with some reports of low, and sometimes reverse, sexual dimorphism (the phenomenon of female tooth dimensions being larger than males'); the LRA results, too, are in contradiction to a previous report of virtually 100% sex allocation for a small heterogeneous sample. These reflect the importance of using a large sample to quantify sexual dimorphism in tooth dimensions and the application of the derived formulae on a test dataset to ascertain accuracy which, at best, is moderate in nature.

  10. Spall strength of liquid copper and accuracy of the acoustic method

    NASA Astrophysics Data System (ADS)

    Cai, Y.; Wu, H. A.; Luo, S. N.

    2017-03-01

    We investigate spallation in liquid copper at high strain rates induced by planar shock loading with classical molecular dynamics simulations. Spallation simulations are performed at different initial temperatures, shock durations, and shock strengths. Loading may have pronounced effects on spall strength. The acoustic method for deducing spall strength and strain rate from free surface velocity histories is discussed in detail and compared to direct simulations. The effects of temperature rise induced by shock wave, tension attenuation, sound speed, and density on the accuracy of the acoustic method are examined; the contributing factors to errors are identified; and the modifications to the choice of sound speed and density are proposed to improve the accuracy of the acoustic method.

  11. Biased binomial assessment of cross-validated estimation of classification accuracies illustrated in diagnosis predictions.

    PubMed

    Noirhomme, Quentin; Lesenfants, Damien; Gomez, Francisco; Soddu, Andrea; Schrouff, Jessica; Garraux, Gaëtan; Luxen, André; Phillips, Christophe; Laureys, Steven

    2014-01-01

    Multivariate classification is used in neuroimaging studies to infer brain activation or in medical applications to infer diagnosis. Their results are often assessed through either a binomial or a permutation test. Here, we simulated classification results of generated random data to assess the influence of the cross-validation scheme on the significance of results. Distributions built from classification of random data with cross-validation did not follow the binomial distribution. The binomial test is therefore not adapted. On the contrary, the permutation test was unaffected by the cross-validation scheme. The influence of the cross-validation was further illustrated on real-data from a brain-computer interface experiment in patients with disorders of consciousness and from an fMRI study on patients with Parkinson disease. Three out of 16 patients with disorders of consciousness had significant accuracy on binomial testing, but only one showed significant accuracy using permutation testing. In the fMRI experiment, the mental imagery of gait could discriminate significantly between idiopathic Parkinson's disease patients and healthy subjects according to the permutation test but not according to the binomial test. Hence, binomial testing could lead to biased estimation of significance and false positive or negative results. In our view, permutation testing is thus recommended for clinical application of classification with cross-validation.

  12. A new method for measuring the rotational accuracy of rolling element bearings

    NASA Astrophysics Data System (ADS)

    Chen, Ye; Zhao, Xiangsong; Gao, Weiguo; Hu, Gaofeng; Zhang, Shizhen; Zhang, Dawei

    2016-12-01

    The rotational accuracy of a machine tool spindle has critical influence upon the geometric shape and surface roughness of finished workpiece. The rotational performance of the rolling element bearings is a main factor which affects the spindle accuracy, especially in the ultra-precision machining. In this paper, a new method is developed to measure the rotational accuracy of rolling element bearings of machine tool spindles. Variable and measurable axial preload is applied to seat the rolling elements in the bearing races, which is used to simulate the operating conditions. A high-precision (radial error is less than 300 nm) and high-stiffness (radial stiffness is 600 N/μm) hydrostatic reference spindle is adopted to rotate the inner race of the test bearing. To prevent the outer race from rotating, a 2-degrees of freedom flexure hinge mechanism (2-DOF FHM) is designed. Correction factors by using stiffness analysis are adopted to eliminate the influences of 2-DOF FHM in the radial direction. Two capacitive displacement sensors with nano-resolution (the highest resolution is 9 nm) are used to measure the radial error motion of the rolling element bearing, without separating the profile error as the traditional rotational accuracy metrology of the spindle. Finally, experimental measurements are performed at different spindle speeds (100-4000 rpm) and axial preloads (75-780 N). Synchronous and asynchronous error motion values are evaluated to demonstrate the feasibility and repeatability of the developed method and instrument.

  13. Accuracy and stability in incompressible SPH (ISPH) based on the projection method and a new approach

    SciTech Connect

    Xu Rui Stansby, Peter; Laurence, Dominique

    2009-10-01

    The stability and accuracy of three methods which enforce either a divergence-free velocity field, density invariance, or their combination are tested here through the standard Taylor-Green and spin-down vortex problems. While various approaches to incompressible SPH (ISPH) have been proposed in the past decade, the present paper is restricted to the projection method for the pressure and velocity coupling. It is shown that the divergence-free ISPH method cannot maintain stability in certain situations although it is accurate before instability sets in. The density-invariant ISPH method is stable but inaccurate with random-noise like disturbances. The combined ISPH, combining advantages in divergence-free ISPH and density-invariant ISPH, can maintain accuracy and stability although at a higher computational cost. Redistribution of particles on a fixed uniform mesh is also shown to be effective but the attraction of a mesh-free method is lost. A new divergence-free ISPH approach is proposed here which maintains accuracy and stability while remaining mesh free without increasing computational cost by slightly shifting particles away from streamlines, although the necessary interpolation of hydrodynamic characteristics means the formulation ceases to be strictly conservative. This avoids the highly anisotropic particle spacing which eventually triggers instability. Importantly pressure fields are free from spurious oscillations, up to the highest Reynolds numbers tested.

  14. Accuracy and feasibility of video analysis for assessing hamstring flexibility and validity of the sit-and-reach test.

    PubMed

    Mier, Constance M

    2011-12-01

    The accuracy of video analysis of the passive straight-leg raise test (PSLR) and the validity of the sit-and-reach test (SR) were tested in 60 men and women. Computer software measured static hip-joint flexion accurately. High within-session reliability of the PSLR was demonstrated (R > .97). Test-retest (separate days) reliability for SR was high in men (R = .97) and women R = .98) moderate for PSLR in men (R = .79) and women (R = .89). SR validity (PSLR as criterion) was higher in women (Day 1, r = .69; Day 2, r = .81) than men (Day 1, r = .64; Day 2, r = .66). In conclusion, video analysis is accurate and feasible for assessing static joint angles, PSLR and SR tests are very reliable methods for assessing flexibility, and the SR validity for hamstring flexibility was found to be moderate in women and low in men.

  15. Accuracy of perceptual and acoustic methods for the detection of inspiratory loci in spontaneous speech.

    PubMed

    Wang, Yu-Tsai; Nip, Ignatius S B; Green, Jordan R; Kent, Ray D; Kent, Jane Finley; Ullman, Cara

    2012-12-01

    The present study investigates the accuracy of perceptually and acoustically determined inspiratory loci in spontaneous speech for the purpose of identifying breath groups. Sixteen participants were asked to talk about simple topics in daily life at a comfortable speaking rate and loudness while connected to a pneumotach and audio microphone. The locations of inspiratory loci were determined on the basis of the aerodynamic signal, which served as a reference for loci identified perceptually and acoustically. Signal detection theory was used to evaluate the accuracy of the methods. The results showed that the greatest accuracy in pause detection was achieved (1) perceptually, on the basis of agreement between at least two of three judges, and (2) acoustically, using a pause duration threshold of 300 ms. In general, the perceptually based method was more accurate than was the acoustically based method. Inconsistencies among perceptually determined, acoustically determined, and aerodynamically determined inspiratory loci for spontaneous speech should be weighed in selecting a method of breath group determination.

  16. A PRIOR EVALUATION OF TWO-STAGE CLUSTER SAMPLING FOR ACCURACY ASSESSMENT OF LARGE-AREA LAND-COVER MAPS

    EPA Science Inventory

    Two-stage cluster sampling reduces the cost of collecting accuracy assessment reference data by constraining sample elements to fall within a limited number of geographic domains (clusters). However, because classification error is typically positively spatially correlated, withi...

  17. Assessment of the accuracy of ABC/2 variations in traumatic epidural hematoma volume estimation: a retrospective study

    PubMed Central

    Hu, Tingting; Zhang, Zhen

    2016-01-01

    Background. The traumatic epidural hematoma (tEDH) volume is often used to assist in tEDH treatment planning and outcome prediction. ABC/2 is a well-accepted volume estimation method that can be used for tEDH volume estimation. Previous studies have proposed different variations of ABC/2; however, it is unclear which variation will provide a higher accuracy. Given the promising clinical contribution of accurate tEDH volume estimations, we sought to assess the accuracy of several ABC/2 variations in tEDH volume estimation. Methods. The study group comprised 53 patients with tEDH who had undergone non-contrast head computed tomography scans. For each patient, the tEDH volume was automatically estimated by eight ABC/2 variations (four traditional and four newly derived) with an in-house program, and results were compared to those from manual planimetry. Linear regression, the closest value, percentage deviation, and Bland-Altman plot were adopted to comprehensively assess accuracy. Results. Among all ABC/2 variations assessed, the traditional variations y = 0.5 × A1B1C1 (or A2B2C1) and the newly derived variations y = 0.65 × A1B1C1 (or A2B2C1) achieved higher accuracy than the other variations. No significant differences were observed between the estimated volume values generated by these variations and those of planimetry (p > 0.05). Comparatively, the former performed better than the latter in general, with smaller mean percentage deviations (7.28 ± 5.90% and 6.42 ± 5.74% versus 19.12 ± 6.33% and 21.28 ± 6.80%, respectively) and more values closest to planimetry (18/53 and 18/53 versus 2/53 and 0/53, respectively). Besides, deviations of most cases in the former fell within the range of <10% (71.70% and 84.91%, respectively), whereas deviations of most cases in the latter were in the range of 10–20% and >20% (90.57% and 96.23, respectively). Discussion. In the current study, we adopted an automatic approach to assess the accuracy of several ABC/2 variations

  18. The SCoRE residual: a quality index to assess the accuracy of joint estimations.

    PubMed

    Ehrig, Rainald M; Heller, Markus O; Kratzenstein, Stefan; Duda, Georg N; Trepczynski, Adam; Taylor, William R

    2011-04-29

    The determination of an accurate centre of rotation (CoR) from skin markers is essential for the assessment of abnormal gait patterns in clinical gait analysis. Despite the many functional approaches to estimate CoRs, no non-invasive analytical determination of the error in the reconstructed joint location is currently available. The purpose of this study was therefore to verify the residual of the symmetrical centre of rotation estimation (SCoRE) as a reliable indirect measure of the error of the computed joint centre. To evaluate the SCoRE residual, numerical simulations were performed to evaluate CoR estimations at different ranges of joint motion. A statistical model was developed and used to determine the theoretical relationships among the SCoRE residual, the magnitude of the skin marker artefact, the corrections to the marker positions, and the error of the CoR estimations to the known centre of rotation. We found that the equation err=0.5r(s) provides a reliable relationship among the CoR error, err, and the scaled SCoRE residual, r(s), providing that any skin marker artefact is first minimised using the optimal common shape technique (OCST). Measurements on six healthy volunteers showed a reduction of SCoRE residual from 11 to below 6mm and therefore demonstrated consistency of the theoretical considerations and numerical simulations with the in vivo data. This study also demonstrates the significant benefit of the OCST for reducing skin marker artefact and thus for predicting the accuracy of determining joint centre positions in functional gait analysis. For the first time, this understanding of the SCoRE residual allows a measure of error in the non-invasive assessment of joint centres. This measure now enables a rapid assessment of the accuracy of the CoR as well as an estimation of the reproducibility and repeatability of skeletal motion patterns.

  19. Dietary assessment methods: dietary records.

    PubMed

    Ortega, Rosa M; Pérez-Rodrigo, Carmen; López-Sobaler, Ana M

    2015-02-26

    Dietary records or food diaries can be highlighted among dietary assessment methods of the current diet for their interest and validity. It is a prospective, open-ended survey method collecting data about the foods and beverages consumed over a previously specified period of time. Dietary records can be used to estimate current diet of individuals and population groups, as well as to identify groups at risk of inadequacy. It is a dietary assessment method interesting for its use in epidemiological or in clinical studies. High validity and precision has been reported for the method when used following adequate procedures and considering the sufficient number of days. Thus, dietary records are often considered as a reference method in validation studies. Nevertheless, the method is affected by error and has limitations due mainly to the tendency of subjects to report food consumption close to those socially desirable. Additional problems are related to the high burden posed on respondents. The method can also influence food behavior in respondents in order to simplify the registration of food intake and some subjects can experience difficulties in writing down the foods and beverages consumed or in describing the portion sizes. Increasing the number of days observed reduces the quality of completed diet records. It should also be considered the high cost of coding and processing information collected in diet records. One of the main advantages of the method is the registration of the foods and beverages as consumed, thus reducing the problem of food omissions due to memory failure. Weighted food records provide more precise estimates of consumed portions. New Technologies can be helpful to improve and ease collaboration of respondents, as well as precision of the estimates, although it would be desirable to evaluate the advantages and limitations in order to optimize the implementation.

  20. Systolic Blood Pressure Accuracy Enhancement in the Electronic Palpation Method Using Pulse Waveform

    DTIC Science & Technology

    2007-11-02

    1 of 4 SYSTOLIC BLOOD PRESSURE ACCURACY ENHANCEMENT IN THE ELECTRONIC PALPATION METHOD USING PULSE WAVEFORM H. S. S. Sorvoja1, R. A. Myllylä1...systolic blood pressure measurements based on pulse waveform. A set of measurement was carried out with elderly cardiac surgery patients. The experiments... blood pressure . Systolic pressure errors were defined and correlations with other specific values, like pressure rise time, pulse wave velocity

  1. Measurement methods and accuracy analysis of Chang'E-5 Panoramic Camera installation parameters

    NASA Astrophysics Data System (ADS)

    Yan, Wei; Ren, Xin; Liu, Jianjun; Tan, Xu; Wang, Wenrui; Chen, Wangli; Zhang, Xiaoxia; Li, Chunlai

    2016-04-01

    Chang'E-5 (CE-5) is a lunar probe for the third phase of China Lunar Exploration Project (CLEP), whose main scientific objectives are to implement lunar surface sampling and to return the samples back to the Earth. To achieve these goals, investigation of lunar surface topography and geological structure within sampling area seems to be extremely important. The Panoramic Camera (PCAM) is one of the payloads mounted on CE-5 lander. It consists of two optical systems which installed on a camera rotating platform. Optical images of sampling area can be obtained by PCAM in the form of a two-dimensional image and a stereo images pair can be formed by left and right PCAM images. Then lunar terrain can be reconstructed based on photogrammetry. Installation parameters of PCAM with respect to CE-5 lander are critical for the calculation of exterior orientation elements (EO) of PCAM images, which is used for lunar terrain reconstruction. In this paper, types of PCAM installation parameters and coordinate systems involved are defined. Measurement methods combining camera images and optical coordinate observations are studied for this work. Then research contents such as observation program and specific solution methods of installation parameters are introduced. Parametric solution accuracy is analyzed according to observations obtained by PCAM scientifically validated experiment, which is used to test the authenticity of PCAM detection process, ground data processing methods, product quality and so on. Analysis results show that the accuracy of the installation parameters affects the positional accuracy of corresponding image points of PCAM stereo images within 1 pixel. So the measurement methods and parameter accuracy studied in this paper meet the needs of engineering and scientific applications. Keywords: Chang'E-5 Mission; Panoramic Camera; Installation Parameters; Total Station; Coordinate Conversion

  2. Comprehensive Numerical Analysis of Finite Difference Time Domain Methods for Improving Optical Waveguide Sensor Accuracy

    PubMed Central

    Samak, M. Mosleh E. Abu; Bakar, A. Ashrif A.; Kashif, Muhammad; Zan, Mohd Saiful Dzulkifly

    2016-01-01

    This paper discusses numerical analysis methods for different geometrical features that have limited interval values for typically used sensor wavelengths. Compared with existing Finite Difference Time Domain (FDTD) methods, the alternating direction implicit (ADI)-FDTD method reduces the number of sub-steps by a factor of two to three, which represents a 33% time savings in each single run. The local one-dimensional (LOD)-FDTD method has similar numerical equation properties, which should be calculated as in the previous method. Generally, a small number of arithmetic processes, which result in a shorter simulation time, are desired. The alternating direction implicit technique can be considered a significant step forward for improving the efficiency of unconditionally stable FDTD schemes. This comparative study shows that the local one-dimensional method had minimum relative error ranges of less than 40% for analytical frequencies above 42.85 GHz, and the same accuracy was generated by both methods.

  3. Assessment of the sources of error affecting the quantitative accuracy of SPECT imaging in small animals

    SciTech Connect

    Joint Graduate Group in Bioengineering, University of California, San Francisco and University of California, Berkeley; Department of Radiology, University of California; Gullberg, Grant T; Hwang, Andrew B.; Franc, Benjamin L.; Gullberg, Grant T.; Hasegawa, Bruce H.

    2008-02-15

    Small animal SPECT imaging systems have multiple potential applications in biomedical research. Whereas SPECT data are commonly interpreted qualitatively in a clinical setting, the ability to accurately quantify measurements will increase the utility of the SPECT data for laboratory measurements involving small animals. In this work, we assess the effect of photon attenuation, scatter and partial volume errors on the quantitative accuracy of small animal SPECT measurements, first with Monte Carlo simulation and then confirmed with experimental measurements. The simulations modeled the imaging geometry of a commercially available small animal SPECT system. We simulated the imaging of a radioactive source within a cylinder of water, and reconstructed the projection data using iterative reconstruction algorithms. The size of the source and the size of the surrounding cylinder were varied to evaluate the effects of photon attenuation and scatter on quantitative accuracy. We found that photon attenuation can reduce the measured concentration of radioactivity in a volume of interest in the center of a rat-sized cylinder of water by up to 50percent when imaging with iodine-125, and up to 25percent when imaging with technetium-99m. When imaging with iodine-125, the scatter-to-primary ratio can reach up to approximately 30percent, and can cause overestimation of the radioactivity concentration when reconstructing data with attenuation correction. We varied the size of the source to evaluate partial volume errors, which we found to be a strong function of the size of the volume of interest and the spatial resolution. These errors can result in large (>50percent) changes in the measured amount of radioactivity. The simulation results were compared with and found to agree with experimental measurements. The inclusion of attenuation correction in the reconstruction algorithm improved quantitative accuracy. We also found that an improvement of the spatial resolution through the

  4. Assessment of the sources of error affecting the quantitative accuracy of SPECT imaging in small animals

    NASA Astrophysics Data System (ADS)

    Hwang, Andrew B.; Franc, Benjamin L.; Gullberg, Grant T.; Hasegawa, Bruce H.

    2008-05-01

    Small animal SPECT imaging systems have multiple potential applications in biomedical research. Whereas SPECT data are commonly interpreted qualitatively in a clinical setting, the ability to accurately quantify measurements will increase the utility of the SPECT data for laboratory measurements involving small animals. In this work, we assess the effect of photon attenuation, scatter and partial volume errors on the quantitative accuracy of small animal SPECT measurements, first with Monte Carlo simulation and then confirmed with experimental measurements. The simulations modeled the imaging geometry of a commercially available small animal SPECT system. We simulated the imaging of a radioactive source within a cylinder of water, and reconstructed the projection data using iterative reconstruction algorithms. The size of the source and the size of the surrounding cylinder were varied to evaluate the effects of photon attenuation and scatter on quantitative accuracy. We found that photon attenuation can reduce the measured concentration of radioactivity in a volume of interest in the center of a rat-sized cylinder of water by up to 50% when imaging with iodine-125, and up to 25% when imaging with technetium-99m. When imaging with iodine-125, the scatter-to-primary ratio can reach up to approximately 30%, and can cause overestimation of the radioactivity concentration when reconstructing data with attenuation correction. We varied the size of the source to evaluate partial volume errors, which we found to be a strong function of the size of the volume of interest and the spatial resolution. These errors can result in large (>50%) changes in the measured amount of radioactivity. The simulation results were compared with and found to agree with experimental measurements. The inclusion of attenuation correction in the reconstruction algorithm improved quantitative accuracy. We also found that an improvement of the spatial resolution through the use of resolution

  5. Point Cloud Derived Fromvideo Frames: Accuracy Assessment in Relation to Terrestrial Laser Scanningand Digital Camera Data

    NASA Astrophysics Data System (ADS)

    Delis, P.; Zacharek, M.; Wierzbicki, D.; Grochala, A.

    2017-02-01

    The use of image sequences in the form of video frames recorded on data storage is very useful in especially when working with large and complex structures. Two cameras were used in this study: Sony NEX-5N (for the test object) and Sony NEX-VG10 E (for the historic building). In both cases, a Sony α f = 16 mm fixed focus wide-angle lens was used. Single frames with sufficient overlap were selected from the video sequence using an equation for automatic frame selection. In order to improve the quality of the generated point clouds, each video frame underwent histogram equalization and image sharpening. Point clouds were generated from the video frames using the SGM-like image matching algorithm. The accuracy assessment was based on two reference point clouds: the first from terrestrial laser scanning and the second generated based on images acquired using a high resolution camera, the NIKON D800. The performed research has shown, that highest accuracies are obtained for point clouds generated from video frames, for which a high pass filtration and histogram equalization had been performed. Studies have shown that to obtain a point cloud density comparable to TLS, an overlap between subsequent video frames must be 85 % or more. Based on the point cloud generated from video data, a parametric 3D model can be generated. This type of the 3D model can be used in HBIM construction.

  6. Linear combination methods to improve diagnostic/prognostic accuracy on future observations

    PubMed Central

    Kang, Le; Liu, Aiyi; Tian, Lili

    2014-01-01

    Multiple diagnostic tests or biomarkers can be combined to improve diagnostic accuracy. The problem of finding the optimal linear combinations of biomarkers to maximise the area under the receiver operating characteristic curve has been extensively addressed in the literature. The purpose of this article is threefold: (1) to provide an extensive review of the existing methods for biomarker combination; (2) to propose a new combination method, namely, the nonparametric stepwise approach; (3) to use leave-one-pair-out cross-validation method, instead of re-substitution method, which is overoptimistic and hence might lead to wrong conclusion, to empirically evaluate and compare the performance of different linear combination methods in yielding the largest area under receiver operating characteristic curve. A data set of Duchenne muscular dystrophy was analysed to illustrate the applications of the discussed combination methods. PMID:23592714

  7. Accuracy of least-squares methods for the Navier-Stokes equations

    NASA Technical Reports Server (NTRS)

    Bochev, Pavel B.; Gunzburger, Max D.

    1993-01-01

    Recently there has been substantial interest in least-squares finite element methods for velocity-vorticity-pressure formulations of the incompressible Navier-Stokes equations. The main cause for this interest is the fact that algorithms for the resulting discrete equations can be devised which require the solution of only symmetric, positive definite systems of algebraic equations. On the other hand, it is well-documented that methods using the vorticity as a primary variable often yield very poor approximations. Thus, here we study the accuracy of these methods through a series of computational experiments, and also comment on theoretical error estimates. It is found, despite the failure of standard methods for deriving error estimates, that computational evidence suggests that these methods are, at the least, nearly optimally accurate. Thus, in addition to the desirable matrix properties yielded by least-squares methods, one also obtains accurate approximations.

  8. The diagnostic accuracy of pharmacological stress echocardiography for the assessment of coronary artery disease: a meta-analysis

    PubMed Central

    Picano, Eugenio; Molinaro, Sabrina; Pasanisi, Emilio

    2008-01-01

    Background Recent American Heart Association/American College of Cardiology guidelines state that "dobutamine stress echo has substantially higher sensitivity than vasodilator stress echo for detection of coronary artery stenosis" while the European Society of Cardiology guidelines and the European Association of Echocardiography recommendations conclude that "the two tests have very similar applications". Who is right? Aim To evaluate the diagnostic accuracy of dobutamine versus dipyridamole stress echocardiography through an evidence-based approach. Methods From PubMed search, we identified all papers with coronary angiographic verification and head-to-head comparison of dobutamine stress echo (40 mcg/kg/min ± atropine) versus dipyridamole stress echo performed with state-of-the art protocols (either 0.84 mg/kg in 10' plus atropine, or 0.84 mg/kg in 6' without atropine). A total of 5 papers have been found. Pooled weight meta-analysis was performed. Results the 5 analyzed papers recruited 435 patients, 299 with and 136 without angiographically assessed coronary artery disease (quantitatively assessed stenosis > 50%). Dipyridamole and dobutamine showed similar accuracy (87%, 95% confidence intervals, CI, 83–90, vs. 84%, CI, 80–88, p = 0.48), sensitivity (85%, CI 80–89, vs. 86%, CI 78–91, p = 0.81) and specificity (89%, CI 82–94 vs. 86%, CI 75–89, p = 0.15). Conclusion When state-of-the art protocols are considered, dipyridamole and dobutamine stress echo have similar accuracy, specificity and – most importantly – sensitivity for detection of CAD. European recommendations concluding that "dobutamine and vasodilators (at appropriately high doses) are equally potent ischemic stressors for inducing wall motion abnormalities in presence of a critical coronary artery stenosis" are evidence-based. PMID:18565214

  9. Accuracy of actuarial procedures for assessment of sexual offender recidivism risk may vary across ethnicity.

    PubMed

    Långström, Niklas

    2004-04-01

    Little is known about whether the accuracy of tools for assessment of sexual offender recidivism risk holds across ethnic minority offenders. I investigated the predictive validity across ethnicity for the RRASOR and the Static-99 actuarial risk assessment procedures in a national cohort of all adult male sex offenders released from prison in Sweden 1993-1997. Subjects ordered out of Sweden upon release from prison were excluded and remaining subjects (N = 1303) divided into three subgroups based on citizenship. Eighty-three percent of the subjects were of Nordic ethnicity, and non-Nordic citizens were either of non-Nordic European (n = 49, hereafter called European) or African Asian descent (n = 128). The two tools were equally accurate among Nordic and European sexual offenders for the prediction of any sexual and any violent nonsexual recidivism. In contrast, neither measure could differentiate African Asian sexual or violent recidivists from nonrecidivists. Compared to European offenders, AfricanAsian offenders had more often sexually victimized a nonrelative or stranger, had higher Static-99 scores, were younger, more often single, and more often homeless. The results require replication, but suggest that the promising predictive validity seen with some risk assessment tools may not generalize across offender ethnicity or migration status. More speculatively, different risk factors or causal chains might be involved in the development or persistence of offending among minority or immigrant sexual abusers.

  10. Accuracy Assessment of GPS Buoy Sea Level Measurements for Coastal Applications

    NASA Astrophysics Data System (ADS)

    Chiu, S.; Cheng, K.

    2008-12-01

    The GPS buoy in this study contains a geodetic antenna and a compact floater with the GPS receiver and power supply tethered to a boat. The coastal applications using GPS include monitoring of sea level and its change, calibration of satellite altimeters, hydrological or geophysical parameters modeling, seafloor geodesy, and others. Among these applications, in order to understand the overall data or model quality, it is required to gain the knowledge of position accuracy of GPS buoys or GPS-equipped vessels. Despite different new GPS data processing techniques, e.g., Precise Point Positioning (PPP) and virtual reference station (VRS), that require a prioir information obtained from the a regional GPS network. While the required a prioir information can be implemented on land, it may not be available on the sea. Hence, in this study, the GPS buoy was positioned with respect to a onshore GPS reference station using the traditional double- difference technique. Since the atmosphere starts to decorrelate as the baseline, the distance between the buoy and the reference station, increases, the positioning accuracy consequently decreases. Therefore, this study aims to assess the buoy position accuracy as the baseline increases and in order to quantify the upper limit of sea level measured by the GPS buoy. A GPS buoy campaign was conducted by National Chung Cheng University in An Ping, Taiwan with a 8- hour GPS buoy data collection. In addition, a GPS network contains 4 Continuous GPS (CGPS) stations in Taiwan was established with the goal to enable baselines in different range for buoy data processing. A vector relation from the network was utilized in order to find the correct ambiguities, which were applied to the long-baseline solution to eliminate the position error caused by incorrect ambiguities. After this procedure, a 3.6-cm discrepancy was found in the mean sea level solution between the long (~80 km) and the short (~1.5 km) baselines. The discrepancy between a

  11. MRI-Based Computed Tomography Metal Artifact Correction Method for Improving Proton Range Calculation Accuracy

    SciTech Connect

    Park, Peter C.; Schreibmann, Eduard; Roper, Justin; Elder, Eric; Crocker, Ian; Fox, Tim; Zhu, X. Ronald; Dong, Lei; Dhabaan, Anees

    2015-03-15

    Purpose: Computed tomography (CT) artifacts can severely degrade dose calculation accuracy in proton therapy. Prompted by the recently increased popularity of magnetic resonance imaging (MRI) in the radiation therapy clinic, we developed an MRI-based CT artifact correction method for improving the accuracy of proton range calculations. Methods and Materials: The proposed method replaces corrupted CT data by mapping CT Hounsfield units (HU number) from a nearby artifact-free slice, using a coregistered MRI. MRI and CT volumetric images were registered with use of 3-dimensional (3D) deformable image registration (DIR). The registration was fine-tuned on a slice-by-slice basis by using 2D DIR. Based on the intensity of paired MRI pixel values and HU from an artifact-free slice, we performed a comprehensive analysis to predict the correct HU for the corrupted region. For a proof-of-concept validation, metal artifacts were simulated on a reference data set. Proton range was calculated using reference, artifactual, and corrected images to quantify the reduction in proton range error. The correction method was applied to 4 unique clinical cases. Results: The correction method resulted in substantial artifact reduction, both quantitatively and qualitatively. On respective simulated brain and head and neck CT images, the mean error was reduced from 495 and 370 HU to 108 and 92 HU after correction. Correspondingly, the absolute mean proton range errors of 2.4 cm and 1.7 cm were reduced to less than 2 mm in both cases. Conclusions: Our MRI-based CT artifact correction method can improve CT image quality and proton range calculation accuracy for patients with severe CT artifacts.

  12. Accuracy, Precision, Ease-Of-Use, and Cost of Methods to Test Ebola-Relevant Chlorine Solutions.

    PubMed

    Wells, Emma; Wolfe, Marlene K; Murray, Anna; Lantagne, Daniele

    2016-01-01

    To prevent transmission in Ebola Virus Disease (EVD) outbreaks, it is recommended to disinfect living things (hands and people) with 0.05% chlorine solution and non-living things (surfaces, personal protective equipment, dead bodies) with 0.5% chlorine solution. In the current West African EVD outbreak, these solutions (manufactured from calcium hypochlorite (HTH), sodium dichloroisocyanurate (NaDCC), and sodium hypochlorite (NaOCl)) have been widely used in both Ebola Treatment Unit and community settings. To ensure solution quality, testing is necessary, however test method appropriateness for these Ebola-relevant concentrations has not previously been evaluated. We identified fourteen commercially-available methods to test Ebola-relevant chlorine solution concentrations, including two titration methods, four DPD dilution methods, and six test strips. We assessed these methods by: 1) determining accuracy and precision by measuring in quintuplicate five different 0.05% and 0.5% chlorine solutions manufactured from NaDCC, HTH, and NaOCl; 2) conducting volunteer testing to assess ease-of-use; and, 3) determining costs. Accuracy was greatest in titration methods (reference-12.4% error compared to reference method), then DPD dilution methods (2.4-19% error), then test strips (5.2-48% error); precision followed this same trend. Two methods had an accuracy of <10% error across all five chlorine solutions with good precision: Hach digital titration for 0.05% and 0.5% solutions (recommended for contexts with trained personnel and financial resources), and Serim test strips for 0.05% solutions (recommended for contexts where rapid, inexpensive, and low-training burden testing is needed). Measurement error from test methods not including pH adjustment varied significantly across the five chlorine solutions, which had pH values 5-11. Volunteers found test strip easiest and titration hardest; costs per 100 tests were $14-37 for test strips and $33-609 for titration. Given the

  13. Accuracy, Precision, Ease-Of-Use, and Cost of Methods to Test Ebola-Relevant Chlorine Solutions

    PubMed Central

    Wells, Emma; Wolfe, Marlene K.; Murray, Anna; Lantagne, Daniele

    2016-01-01

    To prevent transmission in Ebola Virus Disease (EVD) outbreaks, it is recommended to disinfect living things (hands and people) with 0.05% chlorine solution and non-living things (surfaces, personal protective equipment, dead bodies) with 0.5% chlorine solution. In the current West African EVD outbreak, these solutions (manufactured from calcium hypochlorite (HTH), sodium dichloroisocyanurate (NaDCC), and sodium hypochlorite (NaOCl)) have been widely used in both Ebola Treatment Unit and community settings. To ensure solution quality, testing is necessary, however test method appropriateness for these Ebola-relevant concentrations has not previously been evaluated. We identified fourteen commercially-available methods to test Ebola-relevant chlorine solution concentrations, including two titration methods, four DPD dilution methods, and six test strips. We assessed these methods by: 1) determining accuracy and precision by measuring in quintuplicate five different 0.05% and 0.5% chlorine solutions manufactured from NaDCC, HTH, and NaOCl; 2) conducting volunteer testing to assess ease-of-use; and, 3) determining costs. Accuracy was greatest in titration methods (reference-12.4% error compared to reference method), then DPD dilution methods (2.4–19% error), then test strips (5.2–48% error); precision followed this same trend. Two methods had an accuracy of <10% error across all five chlorine solutions with good precision: Hach digital titration for 0.05% and 0.5% solutions (recommended for contexts with trained personnel and financial resources), and Serim test strips for 0.05% solutions (recommended for contexts where rapid, inexpensive, and low-training burden testing is needed). Measurement error from test methods not including pH adjustment varied significantly across the five chlorine solutions, which had pH values 5–11. Volunteers found test strip easiest and titration hardest; costs per 100 tests were $14–37 for test strips and $33–609 for titration

  14. LNG Safety Assessment Evaluation Methods

    SciTech Connect

    Muna, Alice Baca; LaFleur, Angela Christine

    2015-05-01

    Sandia National Laboratories evaluated published safety assessment methods across a variety of industries including Liquefied Natural Gas (LNG), hydrogen, land and marine transportation, as well as the US Department of Defense (DOD). All the methods were evaluated for their potential applicability for use in the LNG railroad application. After reviewing the documents included in this report, as well as others not included because of repetition, the Department of Energy (DOE) Hydrogen Safety Plan Checklist is most suitable to be adapted to the LNG railroad application. This report was developed to survey industries related to rail transportation for methodologies and tools that can be used by the FRA to review and evaluate safety assessments submitted by the railroad industry as a part of their implementation plans for liquefied or compressed natural gas storage ( on-board or tender) and engine fueling delivery systems. The main sections of this report provide an overview of various methods found during this survey. In most cases, the reference document is quoted directly. The final section provides discussion and a recommendation for the most appropriate methodology that will allow efficient and consistent evaluations to be made. The DOE Hydrogen Safety Plan Checklist was then revised to adapt it as a methodology for the Federal Railroad Administration’s use in evaluating safety plans submitted by the railroad industry.

  15. Complex shape product tolerance and accuracy control method for virtual assembly

    NASA Astrophysics Data System (ADS)

    Ma, Huiping; Jin, Yuanqiang; Zhang, Xiaoguang; Zhou, Hai

    2015-02-01

    The simulation of virtual assembly process for engineering design lacks of accuracy in the software of three-dimension CAD at present. Product modeling technology with tolerance, assembly precision preanalysis technique and precision control method are developed. To solve the problem of lack of precision information transmission in CAD, tolerance mathematical model of Small Displacement Torsor (SDT) is presented, which can bring about technology transfer and establishment of digital control function for geometric elements from the definition, description, specification to the actual inspection and evaluation process. Current tolerance optimization design methods for complex shape product are proposed for optimization of machining technology, effective cost control and assembly quality of the products.

  16. Breeding Jatropha curcas by genomic selection: A pilot assessment of the accuracy of predictive models

    PubMed Central

    de Azevedo Peixoto, Leonardo; Laviola, Bruno Galvêas; Alves, Alexandre Alonso; Rosado, Tatiana Barbosa; Bhering, Leonardo Lopes

    2017-01-01

    Genomic wide selection is a promising approach for improving the selection accuracy in plant breeding, particularly in species with long life cycles, such as Jatropha. Therefore, the objectives of this study were to estimate the genetic parameters for grain yield (GY) and the weight of 100 seeds (W100S) using restricted maximum likelihood (REML); to compare the performance of GWS methods to predict GY and W100S; and to estimate how many markers are needed to train the GWS model to obtain the maximum accuracy. Eight GWS models were compared in terms of predictive ability. The impact that the marker density had on the predictive ability was investigated using a varying number of markers, from 2 to 1,248. Because the genetic variance between evaluated genotypes was significant, it was possible to obtain selection gain. All of the GWS methods tested in this study can be used to predict GY and W100S in Jatropha. A training model fitted using 1,000 and 800 markers is sufficient to capture the maximum genetic variance and, consequently, maximum prediction ability of GY and W100S, respectively. This study demonstrated the applicability of genome-wide prediction to identify useful genetic sources of GY and W100S for Jatropha breeding. Further research is needed to confirm the applicability of the proposed approach to other complex traits. PMID:28296913

  17. Assessment of Classification Accuracies of SENTINEL-2 and LANDSAT-8 Data for Land Cover / Use Mapping

    NASA Astrophysics Data System (ADS)

    Hale Topaloğlu, Raziye; Sertel, Elif; Musaoğlu, Nebiye

    2016-06-01

    This study aims to compare classification accuracies of land cover/use maps created from Sentinel-2 and Landsat-8 data. Istanbul metropolitan city of Turkey, with a population of around 14 million, having different landscape characteristics was selected as study area. Water, forest, agricultural areas, grasslands, transport network, urban, airport- industrial units and barren land- mine land cover/use classes adapted from CORINE nomenclature were used as main land cover/use classes to identify. To fulfil the aims of this research, recently acquired dated 08/02/2016 Sentinel-2 and dated 22/02/2016 Landsat-8 images of Istanbul were obtained and image pre-processing steps like atmospheric and geometric correction were employed. Both Sentinel-2 and Landsat-8 images were resampled to 30m pixel size after geometric correction and similar spectral bands for both satellites were selected to create a similar base for these multi-sensor data. Maximum Likelihood (MLC) and Support Vector Machine (SVM) supervised classification methods were applied to both data sets to accurately identify eight different land cover/ use classes. Error matrix was created using same reference points for Sentinel-2 and Landsat-8 classifications. After the classification accuracy, results were compared to find out the best approach to create current land cover/use map of the region. The results of MLC and SVM classification methods were compared for both images.

  18. Breeding Jatropha curcas by genomic selection: A pilot assessment of the accuracy of predictive models.

    PubMed

    Azevedo Peixoto, Leonardo de; Laviola, Bruno Galvêas; Alves, Alexandre Alonso; Rosado, Tatiana Barbosa; Bhering, Leonardo Lopes

    2017-01-01

    Genomic wide selection is a promising approach for improving the selection accuracy in plant breeding, particularly in species with long life cycles, such as Jatropha. Therefore, the objectives of this study were to estimate the genetic parameters for grain yield (GY) and the weight of 100 seeds (W100S) using restricted maximum likelihood (REML); to compare the performance of GWS methods to predict GY and W100S; and to estimate how many markers are needed to train the GWS model to obtain the maximum accuracy. Eight GWS models were compared in terms of predictive ability. The impact that the marker density had on the predictive ability was investigated using a varying number of markers, from 2 to 1,248. Because the genetic variance between evaluated genotypes was significant, it was possible to obtain selection gain. All of the GWS methods tested in this study can be used to predict GY and W100S in Jatropha. A training model fitted using 1,000 and 800 markers is sufficient to capture the maximum genetic variance and, consequently, maximum prediction ability of GY and W100S, respectively. This study demonstrated the applicability of genome-wide prediction to identify useful genetic sources of GY and W100S for Jatropha breeding. Further research is needed to confirm the applicability of the proposed approach to other complex traits.

  19. High accuracy position response calibration method for a micro-channel plate ion detector

    NASA Astrophysics Data System (ADS)

    Hong, R.; Leredde, A.; Bagdasarova, Y.; Fléchard, X.; García, A.; Müller, P.; Knecht, A.; Liénard, E.; Kossin, M.; Sternberg, M. G.; Swanson, H. E.; Zumwalt, D. W.

    2016-11-01

    We have developed a position response calibration method for a micro-channel plate (MCP) detector with a delay-line anode position readout scheme. Using an in situ calibration mask, an accuracy of 8 μm and a resolution of 85 μm (FWHM) have been achieved for MeV-scale α particles and ions with energies of ∼10 keV. At this level of accuracy, the difference between the MCP position responses to high-energy α particles and low-energy ions is significant. The improved performance of the MCP detector can find applications in many fields of AMO and nuclear physics. In our case, it helps reducing systematic uncertainties in a high-precision nuclear β-decay experiment.

  20. Assessing the accuracy of the Second Military Survey for the Doren Landslide (Vorarlberg, Austria)

    NASA Astrophysics Data System (ADS)

    Zámolyi, András.; Székely, Balázs; Biszak, Sándor

    2010-05-01

    Reconstruction of the early and long-term evolution of landslide areas is especially important for determining the proportion of anthropogenic influence on the evolution of the region affected by mass movements. The recent geologic and geomorphological setting of the prominent Doren landslide in Vorarlberg (Western Austria) has been studied extensively by various research groups and civil engineering companies. Civil aerial imaging of the area dates back to the 1950's. Modern monitoring techniques include aerial imaging as well as airborne and terrestrial laser scanning (LiDAR) providing us with almost yearly assessment of the changing geomorphology of the area. However, initiation of the landslide occurred most probably earlier than the application of these methods, since there is evidence that the landslide was already active in the 1930's. For studying the initial phase of landslide formation one possibility is to get back on information recorded on historic photographs or historic maps. In this case study we integrated topographic information from the map sheets of the Second Military Survey of the Habsburg Empire that was conducted in Vorarlberg during the years 1816-1821 (Kretschmer et al., 2004) into a comprehensive GIS. The region of interest around the Doren landslide was georeferenced using the method of Timár et al. (2006) refined by Molnár (2009) thus providing a geodetically correct positioning and the possibility of matching the topographic features from the historic map with features recognized in the LiDAR DTM. The landslide of Doren is clearly visible in the historic map. Additionally, prominent geomorphologic features such as morphological scarps, rills and gullies, mass movement lobes and the course of the Weißach rivulet can be matched. Not only the shape and character of these elements can be recognized and matched, but also the positional accuracy is adequate for geomorphological studies. Since the settlement structure is very stable in the

  1. SU-E-J-117: Verification Method for the Detection Accuracy of Automatic Winston Lutz Test

    SciTech Connect

    Tang, A; Chan, K; Fee, F; Chau, R

    2014-06-01

    Purpose: Winston Lutz test (WLT) has been a standard QA procedure performed prior to SRS treatment, to verify the mechanical iso-center setup accuracy upon different Gantry/Couch movements. Several detection algorithms exist,for analyzing the ball-radiation field alignment automatically. However, the accuracy of these algorithms have not been fully addressed. Here, we reveal the possible errors arise from each step in WLT, and verify the software detection accuracy with the Rectilinear Phantom Pointer (RLPP), a tool commonly used for aligning treatment plan coordinate with mechanical iso-center. Methods: WLT was performed with the radio-opaque ball mounted on a MIS and irradiated onto EDR2 films. The films were scanned and processed with an in-house Matlab program for automatic iso-center detection. Tests were also performed to identify the errors arise from setup, film development and scanning process. The radioopaque ball was then mounted onto the RLPP, and offset laterally and longitudinally in 7 known positions ( 0, ±0.2, ±0.5, ±0.8 mm) manually for irradiations. The gantry and couch was set to zero degree for all irradiation. The same scanned images were processed repeatedly to check the repeatability of the software. Results: Miminal discrepancies (mean=0.05mm) were detected with 2 films overlapped and irradiated but developed separately. This reveals the error arise from film processor and scanner alone. Maximum setup errors were found to be around 0.2mm, by analyzing data collected from 10 irradiations over 2 months. For the known shift introduced using the RLPP, the results agree with the manual offset, and fit linearly (R{sup 2}>0.99) when plotted relative to the first ball with zero shift. Conclusion: We systematically reveal the possible errors arise from each step in WLT, and introduce a simple method to verify the detection accuracy of our in-house software using a clinically available tool.

  2. Accuracy evaluation of numerical methods used in state-of-the-art simulators for spiking neural networks.

    PubMed

    Henker, Stephan; Partzsch, Johannes; Schüffny, René

    2012-04-01

    With the various simulators for spiking neural networks developed in recent years, a variety of numerical solution methods for the underlying differential equations are available. In this article, we introduce an approach to systematically assess the accuracy of these methods. In contrast to previous investigations, our approach focuses on a completely deterministic comparison and uses an analytically solved model as a reference. This enables the identification of typical sources of numerical inaccuracies in state-of-the-art simulation methods. In particular, with our approach we can separate the error of the numerical integration from the timing error of spike detection and propagation, the latter being prominent in simulations with fixed timestep. To verify the correctness of the testing procedure, we relate the numerical deviations to theoretical predictions for the employed numerical methods. Finally, we give an example of the influence of simulation artefacts on network behaviour and spike-timing-dependent plasticity (STDP), underlining the importance of spike-time accuracy for the simulation of STDP.

  3. Accuracy and Reliability of a Novel Method for Fusion of Digital Dental Casts and Cone Beam Computed Tomography Scans

    PubMed Central

    Rangel, Frits A.; Maal, Thomas J. J.; Bronkhorst, Ewald M.; Breuning, K. Hero; Schols, Jan G. J. H.; Bergé, Stefaan J.; Kuijpers-Jagtman, Anne Marie

    2013-01-01

    Several methods have been proposed to integrate digital models into Cone Beam Computed Tomography scans. Since all these methods have some drawbacks such as radiation exposure, soft tissue deformation and time-consuming digital handling processes, we propose a new method to integrate digital dental casts into Cone Beam Computed Tomography scans. Plaster casts of 10 patients were randomly selected and 5 titanium markers were glued to the upper and lower plaster cast. The plaster models were scanned, impressions were taken from the plaster models and the impressions were also scanned. Linear measurements were performed on all three models, to assess accuracy and reproducibility. Besides that, matching of the scanned plaster models and scanned impressions was done, to assess the accuracy of the matching procedure. Results show that all measurement errors are smaller than 0.2 mm, and that 81% is smaller than 0.1 mm. Matching of the scanned plaster casts and scanned impressions show a mean error between the two surfaces of the upper arch of 0.14 mm and for the lower arch of 0.18 mm. The time needed for reconstructing the CBCT scans to a digital patient, where the impressions are integrated into the CBCT scan of the patient takes about 15 minutes, with little variance between patients. In conclusion, we can state that this new method is a reliable method to integrate digital dental casts into CBCT scans. As far as radiation exposure, soft tissue deformation and digital handling processes are concerned, it is a significant improvement compared to the previously published methods. PMID:23527111

  4. Quality assessment of comparative diagnostic accuracy studies: our experience using a modified version of the QUADAS-2 tool.

    PubMed

    Wade, Ros; Corbett, Mark; Eastwood, Alison

    2013-09-01

    Assessing the quality of included studies is a vital step in undertaking a systematic review. The recently revised Quality Assessment of Diagnostic Accuracy Studies (QUADAS) tool (QUADAS-2), which is the only validated quality assessment tool for diagnostic accuracy studies, does not include specific criteria for assessing comparative studies. As part of an assessment that included comparative diagnostic accuracy studies, we used a modified version of QUADAS-2 to assess study quality. We modified QUADAS-2 by duplicating questions relating to the index test, to assess the relevant potential sources of bias for both the index test and comparator test. We also added review-specific questions. We have presented our modified version of QUADAS-2 and outlined some key issues for consideration when assessing the quality of comparative diagnostic accuracy studies, to help guide other systematic reviewers conducting comparative diagnostic reviews. Until QUADAS is updated to incorporate assessment of comparative studies, QUADAS-2 can be used, although modification and careful thought is required. It is important to reflect upon whether aspects of study design and methodology favour one of the tests over another.

  5. Assessment of the Sensitivity, Specificity, and Accuracy of Thermography in Identifying Patients with TMD

    PubMed Central

    Woźniak, Krzysztof; Szyszka-Sommerfeld, Liliana; Trybek, Grzegorz; Piątkowska, Dagmara

    2015-01-01

    Background The purpose of the present study was to evaluate the sensitivity, specificity, and accuracy of thermography in identifying patients with temporomandibular dysfunction (TMD). Material/Methods The study sample consisted of 50 patients (27 women and 23 men) ages 19.2 to 24.5 years (mean age 22.43±1.04) with subjective symptoms of TMD (Ai II–III) and 50 patients (25 women and 25 men) ages 19.3 to 25.1 years (mean age 22.21±1.18) with no subjective symptoms of TMD (Ai I). The anamnestic interviews were conducted according to the three-point anamnestic index of temporomandibular dysfunction (Ai). The thermography was performed using a ThermaCAM TMSC500 (FLIR Systems AB, Sweden) independent thermal vision system. Thermography was closely combined with a 10-min chewing test. Results The results of our study indicated that the absolute difference in temperature between the right and left side (ΔT) has the highest diagnostic value. The diagnostic effectiveness of this parameter increased after the chewing test. The cut-off points for values of temperature differences between the right and left side and identifying 95.5% of subjects with no functional disorders according to the temporomandibular dysfunction index Di (specificity 95.5%) were 0.26°C (AUC=0.7422, sensitivity 44.3%, accuracy 52.4%) before the chewing test and 0.52°C (AUC=0.7920, sensitivity 46.4%, accuracy 56.3%) after it. Conclusions The evaluation of thermography demonstrated its diagnostic usefulness in identifying patients with TMD with limited effectiveness. The chewing test helped in increasing the diagnostic efficiency of thermography in identifying patients with TMD. PMID:26002613

  6. Creating a Standard Set of Metrics to Assess Accuracy of Solar Forecasts: Preliminary Results

    NASA Astrophysics Data System (ADS)

    Banunarayanan, V.; Brockway, A.; Marquis, M.; Haupt, S. E.; Brown, B.; Fowler, T.; Jensen, T.; Hamann, H.; Lu, S.; Hodge, B.; Zhang, J.; Florita, A.

    2013-12-01

    The U.S. Department of Energy (DOE) SunShot Initiative, launched in 2011, seeks to reduce the cost of solar energy systems by 75% from 2010 to 2020. In support of the SunShot Initiative, the DOE Office of Energy Efficiency and Renewable Energy (EERE) is partnering with the National Oceanic and Atmospheric Administration (NOAA) and solar energy stakeholders to improve solar forecasting. Through a funding opportunity announcement issued in the April, 2012, DOE is funding two teams - led by National Center for Atmospheric Research (NCAR), and by IBM - to perform three key activities in order to improve solar forecasts. The teams will: (1) With DOE and NOAA's leadership and significant stakeholder input, develop a standardized set of metrics to evaluate forecast accuracy, and determine the baseline and target values for these metrics; (2) Conduct research that yields a transformational improvement in weather models and methods for forecasting solar irradiance and power; and (3) Incorporate solar forecasts into the system operations of the electric power grid, and evaluate the impact of forecast accuracy on the economics and reliability of operations using the defined, standard metrics. This paper will present preliminary results on the first activity: the development of a standardized set of metrics, baselines and target values. The results will include a proposed framework for metrics development, key categories of metrics, descriptions of each of the proposed set of specific metrics to measure forecast accuracy, feedback gathered from a range of stakeholders on the metrics, and processes to determine baselines and target values for each metric. The paper will also analyze the temporal and spatial resolutions under which these metrics would apply, and conclude with a summary of the work in progress on solar forecasting activities funded by DOE.

  7. Accuracy of five implant impression technique: effect of splinting materials and methods

    PubMed Central

    Cho, Sung-Bum

    2011-01-01

    PURPOSE The aim of this study was to evaluate the effect of dimensional stability of splinting material on the accuracy of master casts. MATERIALS AND METHODS A stainless steel metal model with 6 implants embedded was used as a master model. Implant level impressions were made after square impression copings were splinted using 5 different techniques as follows. (1) Splinted with autopolymerizing resin and sectioned, reconnected to compensate polymerization shrinkage before the impression procedure. (2) Splinted with autopolymerizing resin just before impression procedure. (3) Primary impression made with impression plaster and secondary impression were made over with polyether impression material. (4) Splinted with impression plaster. (5) Splinted with VPS bite registration material. From master model, 5 impressions and 5 experimental casts, total 25 casts were made for each of 5 splinting methods. The distortion values of each splinting methods were measured using coordinate measuring machine, capable of recordings in the x-, y-, z-axes. A one-way analysis of variance (ANOVA) at a confidence level of 95% was used to evaluate the data and Tukey's studentized range test was used to determine significant differences between the groups. RESULTS Group 1 showed best accuracy followed by Group 3 & 4. Group 2 and 5 showed relatively larger distortion value than other groups. No significant difference was found between group 3, 4, 5 in x-axis, group 2, 3, 4 in y-axis and group 1, 3, 4, 5 in z-axis (P<.0001). CONCLUSION Both Splinting impression copings with autopolymerizing resin following compensation of polymerization shrinkage and splinting method with impression plaster can enhance the accuracy of master cast and impression plaster can be used simple and effective splinting material for implant impression procedure. PMID:22259700

  8. Application of Digital Image Correlation Method to Improve the Accuracy of Aerial Photo Stitching

    NASA Astrophysics Data System (ADS)

    Tung, Shih-Heng; Jhou, You-Liang; Shih, Ming-Hsiang; Hsiao, Han-Wei; Sung, Wen-Pei

    2016-04-01

    Satellite images and traditional aerial photos have been used in remote sensing for a long time. However, there are some problems with these images. For example, the resolution of satellite image is insufficient, the cost to obtain traditional images is relatively high and there is also human safety risk in traditional flight. These result in the application limitation of these images. In recent years, the control technology of unmanned aerial vehicle (UAV) is rapidly developed. This makes unmanned aerial vehicle widely used in obtaining aerial photos. Compared to satellite images and traditional aerial photos, these aerial photos obtained using UAV have the advantages of higher resolution, low cost. Because there is no crew in UAV, it is still possible to take aerial photos using UAV under unstable weather conditions. Images have to be orthorectified and their distortion must be corrected at first. Then, with the help of image matching technique and control points, these images can be stitched or used to establish DEM of ground surface. These images or DEM data can be used to monitor the landslide or estimate the volume of landslide. For the image matching, we can use such as Harris corner method, SIFT or SURF to extract and match feature points. However, the accuracy of these methods for matching is about pixel or sub-pixel level. The accuracy of digital image correlation method (DIC) during image matching can reach about 0.01pixel. Therefore, this study applies digital image correlation method to match extracted feature points. Then the stitched images are observed to judge the improvement situation. This study takes the aerial photos of a reservoir area. These images are stitched under the situations with and without the help of DIC. The results show that the misplacement situation in the stitched image using DIC to match feature points has been significantly improved. This shows that the use of DIC to match feature points can actually improve the accuracy of

  9. Comparative adaptation accuracy of acrylic denture bases evaluated by two different methods.

    PubMed

    Lee, Chung-Jae; Bok, Sung-Bem; Bae, Ji-Young; Lee, Hae-Hyoung

    2010-08-01

    This study examined the adaptation accuracy of acrylic denture base processed using fluid-resin (PERform), injection-moldings (SR-Ivocap, Success, Mak Press), and two compression-molding techniques. The adaptation accuracy was measured primarily by the posterior border gaps at the mid-palatal area using a microscope and subsequently by weighing of the weight of the impression material between the denture base and master cast using hand-mixed and automixed silicone. The correlation between the data measured using these two test methods was examined. The PERform and Mak Press produced significantly smaller maximum palatal gap dimensions than the other groups (p<0.05). Mak Press also showed a significantly smaller weight of automixed silicone material than the other groups (p<0.05), while SR-Ivocap and Success showed similar adaptation accuracy to the compression-molding denture. The correlationship between the magnitude of the posterior border gap and the weight of the silicone impression materials was affected by either the material or mixing variables.

  10. Novel Resistance Measurement Method: Analysis of Accuracy and Thermal Dependence with Applications in Fiber Materials.

    PubMed

    Casans, Silvia; Rosado-Muñoz, Alfredo; Iakymchuk, Taras

    2016-12-14

    Material resistance is important since different physicochemical properties can be extracted from it. This work describes a novel resistance measurement method valid for a wide range of resistance values up to 100 GΩ at a low powered, small sized, digitally controlled and wireless communicated device. The analog and digital circuits of the design are described, analysing the main error sources affecting the accuracy. Accuracy and extended uncertainty are obtained for a pattern decade box, showing a maximum of 1 % accuracy for temperatures below 30 ∘ C in the range from 1 MΩ to 100 GΩ. Thermal analysis showed stability up to 50 ∘ C for values below 10 GΩ and systematic deviations for higher values. Power supply V i applied to the measurement probes is also analysed, showing no differences in case of the pattern decade box, except for resistance values above 10 GΩ and temperatures above 35 ∘ C. To evaluate the circuit behaviour under fiber materials, an 11-day drying process in timber from four species (Oregon pine-Pseudotsuga menziesii, cedar-Cedrus atlantica, ash-Fraxinus excelsior, chestnut-Castanea sativa) was monitored. Results show that the circuit, as expected, provides different resistance values (they need individual conversion curves) for different species and the same ambient conditions. Additionally, it was found that, contrary to the decade box analysis, V i affects the resistance value due to material properties. In summary, the proposed circuit is able to accurately measure material resistance that can be further related to material properties.

  11. Novel Resistance Measurement Method: Analysis of Accuracy and Thermal Dependence with Applications in Fiber Materials

    PubMed Central

    Casans, Silvia; Rosado-Muñoz, Alfredo; Iakymchuk, Taras

    2016-01-01

    Material resistance is important since different physicochemical properties can be extracted from it. This work describes a novel resistance measurement method valid for a wide range of resistance values up to 100 GΩ at a low powered, small sized, digitally controlled and wireless communicated device. The analog and digital circuits of the design are described, analysing the main error sources affecting the accuracy. Accuracy and extended uncertainty are obtained for a pattern decade box, showing a maximum of 1% accuracy for temperatures below 30 ∘C in the range from 1 MΩ to 100 GΩ. Thermal analysis showed stability up to 50 ∘C for values below 10 GΩ and systematic deviations for higher values. Power supply Vi applied to the measurement probes is also analysed, showing no differences in case of the pattern decade box, except for resistance values above 10 GΩ and temperatures above 35 ∘C. To evaluate the circuit behaviour under fiber materials, an 11-day drying process in timber from four species (Oregon pine-Pseudotsuga menziesii, cedar-Cedrus atlantica, ash-Fraxinus excelsior, chestnut-Castanea sativa) was monitored. Results show that the circuit, as expected, provides different resistance values (they need individual conversion curves) for different species and the same ambient conditions. Additionally, it was found that, contrary to the decade box analysis, Vi affects the resistance value due to material properties. In summary, the proposed circuit is able to accurately measure material resistance that can be further related to material properties. PMID:27983652

  12. Improving the accuracy of CT dimensional metrology by a novel beam hardening correction method

    NASA Astrophysics Data System (ADS)

    Zhang, Xiang; Li, Lei; Zhang, Feng; Xi, Xiaoqi; Deng, Lin; Yan, Bin

    2015-01-01

    Its powerful nondestructive characteristics are attracting more and more research into the study of computed tomography (CT) for dimensional metrology, which offers a practical alternative to the common measurement methods. However, the inaccuracy and uncertainty severely limit the further utilization of CT for dimensional metrology due to many factors, among which the beam hardening (BH) effect plays a vital role. This paper mainly focuses on eliminating the influence of the BH effect in the accuracy of CT dimensional metrology. To correct the BH effect, a novel exponential correction model is proposed. The parameters of the model are determined by minimizing the gray entropy of the reconstructed volume. In order to maintain the consistency and contrast of the corrected volume, a punishment term is added to the cost function, enabling more accurate measurement results to be obtained by the simple global threshold method. The proposed method is efficient, and especially suited to the case where there is a large difference in gray value between material and background. Different spheres with known diameters are used to verify the accuracy of dimensional measurement. Both simulation and real experimental results demonstrate the improvement in measurement precision. Moreover, a more complex workpiece is also tested to show that the proposed method is of general feasibility.

  13. [Analysis on the accuracy of simple selection method of Fengshi (GB 31)].

    PubMed

    Li, Zhixing; Zhang, Haihua; Li, Suhe

    2015-12-01

    To explore the accuracy of simple selection method of Fengshi (GB 31). Through the study of the ancient and modern data,the analysis and integration of the acupuncture books,the comparison of the locations of Fengshi (GB 31) by doctors from all dynasties and the integration of modern anatomia, the modern simple selection method of Fengshi (GB 31) is definite, which is the same as the traditional way. It is believed that the simple selec tion method is in accord with the human-oriented thought of TCM. Treatment by acupoints should be based on the emerging nature and the individual difference of patients. Also, it is proposed that Fengshi (GB 31) should be located through the integration between the simple method and body surface anatomical mark.

  14. Improving the Accuracy of Urban Environmental Quality Assessment Using Geographically-Weighted Regression Techniques

    PubMed Central

    Faisal, Kamil; Shaker, Ahmed

    2017-01-01

    Urban Environmental Quality (UEQ) can be treated as a generic indicator that objectively represents the physical and socio-economic condition of the urban and built environment. The value of UEQ illustrates a sense of satisfaction to its population through assessing different environmental, urban and socio-economic parameters. This paper elucidates the use of the Geographic Information System (GIS), Principal Component Analysis (PCA) and Geographically-Weighted Regression (GWR) techniques to integrate various parameters and estimate the UEQ of two major cities in Ontario, Canada. Remote sensing, GIS and census data were first obtained to derive various environmental, urban and socio-economic parameters. The aforementioned techniques were used to integrate all of these environmental, urban and socio-economic parameters. Three key indicators, including family income, higher level of education and land value, were used as a reference to validate the outcomes derived from the integration techniques. The results were evaluated by assessing the relationship between the extracted UEQ results and the reference layers. Initial findings showed that the GWR with the spatial lag model represents an improved precision and accuracy by up to 20% with respect to those derived by using GIS overlay and PCA techniques for the City of Toronto and the City of Ottawa. The findings of the research can help the authorities and decision makers to understand the empirical relationships among environmental factors, urban morphology and real estate and decide for more environmental justice. PMID:28272334

  15. Accuracy Assessment of Immediate and Delayed Implant Placements Using CAD/CAM Surgical Guides.

    PubMed

    Alzoubi, Fawaz; Massoomi, Nima; Nattestad, Anders

    2016-10-01

    The aim of this study is to assess the accuracy of immediately placed implants using Anatomage Invivo5 computer-assisted design/computer-assisted manufacturing (CAD/CAM) surgical guides and compare the accuracy to delayed implant placement protocol. Patients who had implants placed using Anatomage Invivo5 CAD/CAM surgical guides during the period of 2012-2015 were evaluated retrospectively. Patients who received immediate implant placements and/or delayed implant placements replacing 1-2 teeth were included in this study. Pre- and postsurgical images were superimposed to evaluate deviations at the crest, apex, and angle. A total of 40 implants placed in 29 patients were included in this study. The overall mean deviations measured at the crest, apex, and angle were 0.86 mm, 1.25 mm, and 3.79°, respectively. The means for the immediate group deviations were: crest = 0.85 mm, apex = 1.10, and angle = 3.49°. The means for the delayed group deviations were: crest = 0.88 mm, apex = 1.59, and angle = 4.29°. No statistically significant difference was found at the crest and angle; however, there was a statistically significant difference between the immediate and delayed group at the apex, with the immediate group presenting more accurate placements at the apical point than the delayed group. CAD/CAM surgical guides can be reliable tools to accurately place implants immediately and/or in a delayed fashion. No statistically significant differences were found between the delayed and the immediate group at the crest and angle, however apical position was more accurate in the immediate group.

  16. Accuracy of Specific BIVA for the Assessment of Body Composition in the United States Population

    PubMed Central

    Buffa, Roberto; Saragat, Bruno; Cabras, Stefano; Rinaldi, Andrea C.; Marini, Elisabetta

    2013-01-01

    Background Bioelectrical impedance vector analysis (BIVA) is a technique for the assessment of hydration and nutritional status, used in the clinical practice. Specific BIVA is an analytical variant, recently proposed for the Italian elderly population, that adjusts bioelectrical values for body geometry. Objective Evaluating the accuracy of specific BIVA in the adult U.S. population, compared to the ‘classic’ BIVA procedure, using DXA as the reference technique, in order to obtain an interpretative model of body composition. Design A cross-sectional sample of 1590 adult individuals (836 men and 754 women, 21–49 years old) derived from the NHANES 2003–2004 was considered. Classic and specific BIVA were applied. The sensitivity and specificity in recognizing individuals below the 5th and above the 95th percentiles of percent fat (FMDXA%) and extracellular/intracellular water (ECW/ICW) ratio were evaluated by receiver operating characteristic (ROC) curves. Classic and specific BIVA results were compared by a probit multiple-regression. Results Specific BIVA was significantly more accurate than classic BIVA in evaluating FMDXA% (ROC areas: 0.84–0.92 and 0.49–0.61 respectively; p = 0.002). The evaluation of ECW/ICW was accurate (ROC areas between 0.83 and 0.96) and similarly performed by the two procedures (p = 0.829). The accuracy of specific BIVA was similar in the two sexes (p = 0.144) and in FMDXA% and ECW/ICW (p = 0.869). Conclusions Specific BIVA showed to be an accurate technique. The tolerance ellipses of specific BIVA can be used for evaluating FM% and ECW/ICW in the U.S. adult population. PMID:23484033

  17. Accuracy Assessment of GO Pro Hero 3 (black) Camera in Underwater Environment

    NASA Astrophysics Data System (ADS)

    Helmholz, , P.; Long, J.; Munsie, T.; Belton, D.

    2016-06-01

    Modern digital cameras are increasing in quality whilst decreasing in size. In the last decade, a number of waterproof consumer digital cameras (action cameras) have become available, which often cost less than 500. A possible application of such action cameras is in the field of Underwater Photogrammetry. Especially with respect to the fact that with the change of the medium to below water can in turn counteract the distortions present. The goal of this paper is to investigate the suitability of such action cameras for underwater photogrammetric applications focusing on the stability of the camera and the accuracy of the derived coordinates for possible photogrammetric applications. For this paper a series of image sequences was capture in a water tank. A calibration frame was placed in the water tank allowing the calibration of the camera and the validation of the measurements using check points. The accuracy assessment covered three test sets operating three GoPro sports cameras of the same model (Hero 3 black). The test set included the handling of the camera in a controlled manner where the camera was only dunked into the water tank using 7MP and 12MP resolution and a rough handling where the camera was shaken as well as being removed from the waterproof case using 12MP resolution. The tests showed that the camera stability was given with a maximum standard deviation of the camera constant σc of 0.0031mm for 7MB (for an average c of 2.720mm) and 0.0072 mm for 12MB (for an average c of 3.642mm). The residual test of the check points gave for the 7MB test series the largest rms value with only 0.450mm and the largest maximal residual of only 2.5 mm. For the 12MB test series the maximum rms value is 0. 653mm.

  18. On the convergence and accuracy of the FDTD method for nanoplasmonics.

    PubMed

    Lesina, Antonino Calà; Vaccari, Alessandro; Berini, Pierre; Ramunno, Lora

    2015-04-20

    Use of the Finite-Difference Time-Domain (FDTD) method to model nanoplasmonic structures continues to rise - more than 2700 papers have been published in 2014 on FDTD simulations of surface plasmons. However, a comprehensive study on the convergence and accuracy of the method for nanoplasmonic structures has yet to be reported. Although the method may be well-established in other areas of electromagnetics, the peculiarities of nanoplasmonic problems are such that a targeted study on convergence and accuracy is required. The availability of a high-performance computing system (a massively parallel IBM Blue Gene/Q) allows us to do this for the first time. We consider gold and silver at optical wavelengths along with three "standard" nanoplasmonic structures: a metal sphere, a metal dipole antenna and a metal bowtie antenna - for the first structure comparisons with the analytical extinction, scattering, and absorption coefficients based on Mie theory are possible. We consider different ways to set-up the simulation domain, we vary the mesh size to very small dimensions, we compare the simple Drude model with the Drude model augmented with two critical points correction, we compare single-precision to double-precision arithmetic, and we compare two staircase meshing techniques, per-component and uniform. We find that the Drude model with two critical points correction (at least) must be used in general. Double-precision arithmetic is needed to avoid round-off errors if highly converged results are sought. Per-component meshing increases the accuracy when complex geometries are modeled, but the uniform mesh works better for structures completely fillable by the Yee cell (e.g., rectangular structures). Generally, a mesh size of 0.25 nm is required to achieve convergence of results to ∼ 1%. We determine how to optimally setup the simulation domain, and in so doing we find that performing scattering calculations within the near-field does not necessarily produces large

  19. Accuracy Assessment of a Complex Building 3d Model Reconstructed from Images Acquired with a Low-Cost Uas

    NASA Astrophysics Data System (ADS)

    Oniga, E.; Chirilă, C.; Stătescu, F.

    2017-02-01

    Nowadays, Unmanned Aerial Systems (UASs) are a wide used technique for acquisition in order to create buildings 3D models, providing the acquisition of a high number of images at very high resolution or video sequences, in a very short time. Since low-cost UASs are preferred, the accuracy of a building 3D model created using this platforms must be evaluated. To achieve results, the dean's office building from the Faculty of "Hydrotechnical Engineering, Geodesy and Environmental Engineering" of Iasi, Romania, has been chosen, which is a complex shape building with the roof formed of two hyperbolic paraboloids. Seven points were placed on the ground around the building, three of them being used as GCPs, while the remaining four as Check points (CPs) for accuracy assessment. Additionally, the coordinates of 10 natural CPs representing the building characteristic points were measured with a Leica TCR 405 total station. The building 3D model was created as a point cloud which was automatically generated based on digital images acquired with the low-cost UASs, using the image matching algorithm and different software like 3DF Zephyr, Visual SfM, PhotoModeler Scanner and Drone2Map for ArcGIS. Except for the PhotoModeler Scanner software, the interior and exterior orientation parameters were determined simultaneously by solving a self-calibrating bundle adjustment. Based on the UAS point clouds, automatically generated by using the above mentioned software and GNSS data respectively, the parameters of the east side hyperbolic paraboloid were calculated using the least squares method and a statistical blunder detection. Then, in order to assess the accuracy of the building 3D model, several comparisons were made for the facades and the roof with reference data, considered with minimum errors: TLS mesh for the facades and GNSS mesh for the roof. Finally, the front facade of the building was created in 3D based on its characteristic points using the PhotoModeler Scanner

  20. Accuracy Validation of an Automated Method for Prostate Segmentation in Magnetic Resonance Imaging.

    PubMed

    Shahedi, Maysam; Cool, Derek W; Bauman, Glenn S; Bastian-Jordan, Matthew; Fenster, Aaron; Ward, Aaron D

    2017-03-24

    Three dimensional (3D) manual segmentation of the prostate on magnetic resonance imaging (MRI) is a laborious and time-consuming task that is subject to inter-observer variability. In this study, we developed a fully automatic segmentation algorithm for T2-weighted endorectal prostate MRI and evaluated its accuracy within different regions of interest using a set of complementary error metrics. Our dataset contained 42 T2-weighted endorectal MRI from prostate cancer patients. The prostate was manually segmented by one observer on all of the images and by two other observers on a subset of 10 images. The algorithm first coarsely localizes the prostate in the image using a template matching technique. Then, it defines the prostate surface using learned shape and appearance information from a set of training images. To evaluate the algorithm, we assessed the error metric values in the context of measured inter-observer variability and compared performance to that of our previously published semi-automatic approach. The automatic algorithm needed an average execution time of ∼60 s to segment the prostate in 3D. When compared to a single-observer reference standard, the automatic algorithm has an average mean absolute distance of 2.8 mm, Dice similarity coefficient of 82%, recall of 82%, precision of 84%, and volume difference of 0.5 cm(3) in the mid-gland. Concordant with other studies, accuracy was highest in the mid-gland and lower in the apex and base. Loss of accuracy with respect to the semi-automatic algorithm was less than the measured inter-observer variability in manual segmentation for the same task.

  1. [Enzymatic determination of total cholesterol in serum: accuracy and comparison with other methods (author's transl)].

    PubMed

    Stähler, F; Munz, E; Kattermann, R

    1975-04-18

    Cholesterol can be specifically measured without difficulty and without complex reagents by means of a newly developed enzymatic colour test. Intensive technical evaluation confirmed its accuracy. Manual use gave a day-to-day coefficient of variation of 2-3 per cent; the sensitivity at a cholesterol concentration of 200 mg/dl was E equals 0.153 (gamma equals 405 nm), with a linearity up to 1000 mg/dl. Recovery of added pure cholesterol solution was 100 plus or minus 2 percent. A quantitative study of 53 representative drugs, anti-coagulants and metabolites was performed both in test tube and on patients. Accuracy of the result was unaffected by any of the substances (alpha equals 0.05). Comparison with the multi-step extraction method used at present as a reference (Abell-Kendall) gave a regression equation of y equals 0.99 times plus 1.1 (x axis: extraction method; y axis: enzymatic colour test). The direct chemical method after Liebermann and Burchard gave, in part, markedly differing results because of considerable systematic and accidental errors.

  2. Assessment of Required Accuracy of Digital Elevation Data for Hydrologic Modeling

    NASA Technical Reports Server (NTRS)

    Kenward, T.; Lettenmaier, D. P.

    1997-01-01

    The effect of vertical accuracy of Digital Elevation Models (DEMs) on hydrologic models is evaluated by comparing three DEMs and resulting hydrologic model predictions applied to a 7.2 sq km USDA - ARS watershed at Mahantango Creek, PA. The high resolution (5 m) DEM was resempled to a 30 m resolution using method that constrained the spatial structure of the elevations to be comparable with the USGS and SIR-C DEMs. This resulting 30 m DEM was used as the reference product for subsequent comparisons. Spatial fields of directly derived quantities, such as elevation differences, slope, and contributing area, were compared to the reference product, as were hydrologic model output fields derived using each of the three DEMs at the common 30 m spatial resolution.

  3. Assessing the prediction accuracy of cure in the Cox proportional hazards cure model: an application to breast cancer data.

    PubMed

    Asano, Junichi; Hirakawa, Akihiro; Hamada, Chikuma

    2014-01-01

    A cure rate model is a survival model incorporating the cure rate with the assumption that the population contains both uncured and cured individuals. It is a powerful statistical tool for prognostic studies, especially in cancer. The cure rate is important for making treatment decisions in clinical practice. The proportional hazards (PH) cure model can predict the cure rate for each patient. This contains a logistic regression component for the cure rate and a Cox regression component to estimate the hazard for uncured patients. A measure for quantifying the predictive accuracy of the cure rate estimated by the Cox PH cure model is required, as there has been a lack of previous research in this area. We used the Cox PH cure model for the breast cancer data; however, the area under the receiver operating characteristic curve (AUC) could not be estimated because many patients were censored. In this study, we used imputation-based AUCs to assess the predictive accuracy of the cure rate from the PH cure model. We examined the precision of these AUCs using simulation studies. The results demonstrated that the imputation-based AUCs were estimable and their biases were negligibly small in many cases, although ordinary AUC could not be estimated. Additionally, we introduced the bias-correction method of imputation-based AUCs and found that the bias-corrected estimate successfully compensated the overestimation in the simulation studies. We also illustrated the estimation of the imputation-based AUCs using breast cancer data.

  4. Accuracy of pattern detection methods in the performance of golf putting.

    PubMed

    Couceiro, Micael S; Dias, Gonçalo; Mendes, Rui; Araújo, Duarte

    2013-01-01

    The authors present a comparison of the classification accuracy of 5 pattern detection methods in the performance of golf putting. The detection of the position of the golf club was performed using a computer vision technique followed by the estimation algorithm Darwinian particle swarm optimization to obtain a kinematical model of each trial. The estimated parameters of the models were subsequently used as sample of five classification algorithms: (a) linear discriminant analysis, (b) quadratic discriminant analysis, (c) naive Bayes with normal distribution, (d) naive Bayes with kernel smoothing density estimate, and (e) least squares support vector machines. Beyond testing the performance of each classification method, it was also possible to identify a putting signature that characterized each golf player. It may be concluded that these methods can be applied to the study of coordination and motor control on the putting performance, allowing for the analysis of the intra- and interpersonal variability of motor behavior in performance contexts.

  5. Interpolation methods and the accuracy of lattice-Boltzmann mesh refinement

    DOE PAGES

    Guzik, Stephen M.; Weisgraber, Todd H.; Colella, Phillip; ...

    2013-12-10

    A lattice-Boltzmann model to solve the equivalent of the Navier-Stokes equations on adap- tively refined grids is presented. A method for transferring information across interfaces between different grid resolutions was developed following established techniques for finite- volume representations. This new approach relies on a space-time interpolation and solving constrained least-squares problems to ensure conservation. The effectiveness of this method at maintaining the second order accuracy of lattice-Boltzmann is demonstrated through a series of benchmark simulations and detailed mesh refinement studies. These results exhibit smaller solution errors and improved convergence when compared with similar approaches relying only on spatial interpolation. Examplesmore » highlighting the mesh adaptivity of this method are also provided.« less

  6. Interpolation methods and the accuracy of lattice-Boltzmann mesh refinement

    SciTech Connect

    Guzik, Stephen M.; Weisgraber, Todd H.; Colella, Phillip; Alder, Berni J.

    2013-12-10

    A lattice-Boltzmann model to solve the equivalent of the Navier-Stokes equations on adap- tively refined grids is presented. A method for transferring information across interfaces between different grid resolutions was developed following established techniques for finite- volume representations. This new approach relies on a space-time interpolation and solving constrained least-squares problems to ensure conservation. The effectiveness of this method at maintaining the second order accuracy of lattice-Boltzmann is demonstrated through a series of benchmark simulations and detailed mesh refinement studies. These results exhibit smaller solution errors and improved convergence when compared with similar approaches relying only on spatial interpolation. Examples highlighting the mesh adaptivity of this method are also provided.

  7. Accuracy Assessment of Three-dimensional Surface Reconstructions of In vivo Teeth from Cone-beam Computed Tomography

    PubMed Central

    Sang, Yan-Hui; Hu, Hong-Cheng; Lu, Song-He; Wu, Yu-Wei; Li, Wei-Ran; Tang, Zhi-Hui

    2016-01-01

    Background: The accuracy of three-dimensional (3D) reconstructions from cone-beam computed tomography (CBCT) has been particularly important in dentistry, which will affect the effectiveness of diagnosis, treatment plan, and outcome in clinical practice. The aims of this study were to assess the linear, volumetric, and geometric accuracy of 3D reconstructions from CBCT and to investigate the influence of voxel size and CBCT system on the reconstructions results. Methods: Fifty teeth from 18 orthodontic patients were assigned to three groups as NewTom VG 0.15 mm group (NewTom VG; voxel size: 0.15 mm; n = 17), NewTom VG 0.30 mm group (NewTom VG; voxel size: 0.30 mm; n = 16), and VATECH DCTPRO 0.30 mm group (VATECH DCTPRO; voxel size: 0.30 mm; n = 17). The 3D reconstruction models of the teeth were segmented from CBCT data manually using Mimics 18.0 (Materialise Dental, Leuven, Belgium), and the extracted teeth were scanned by 3Shape optical scanner (3Shape A/S, Denmark). Linear and volumetric deviations were separately assessed by comparing the length and volume of the 3D reconstruction model with physical measurement by paired t-test. Geometric deviations were assessed by the root mean square value of the imposed 3D reconstruction and optical models by one-sample t-test. To assess the influence of voxel size and CBCT system on 3D reconstruction, analysis of variance (ANOVA) was used (α = 0.05). Results: The linear, volumetric, and geometric deviations were −0.03 ± 0.48 mm, −5.4 ± 2.8%, and 0.117 ± 0.018 mm for NewTom VG 0.15 mm group; −0.45 ± 0.42 mm, −4.5 ± 3.4%, and 0.116 ± 0.014 mm for NewTom VG 0.30 mm group; and −0.93 ± 0.40 mm, −4.8 ± 5.1%, and 0.194 ± 0.117 mm for VATECH DCTPRO 0.30 mm group, respectively. There were statistically significant differences between groups in terms of linear measurement (P < 0.001), but no significant difference in terms of volumetric measurement (P = 0.774). No statistically significant difference were

  8. A simple method for improving the time-stepping accuracy in atmosphere and ocean models

    NASA Astrophysics Data System (ADS)

    Williams, P. D.

    2012-12-01

    In contemporary numerical simulations of the atmosphere and ocean, evidence suggests that time-stepping errors may be a significant component of total model error, on both weather and climate time-scales. This presentation will review the available evidence, and will then suggest a simple but effective method for substantially improving the time-stepping numerics at no extra computational expense. A common time-stepping method in atmosphere and ocean models is the leapfrog scheme combined with the Robert-Asselin (RA) filter. This method is used in the following models (and many more): ECHAM, MAECHAM, MM5, CAM, MESO-NH, HIRLAM, KMCM, LIMA, SPEEDY, IGCM, PUMA, COSMO, FSU-GSM, FSU-NRSM, NCEP-GFS, NCEP-RSM, NSEAM, NOGAPS, RAMS, and CCSR/NIES-AGCM. Although the RA filter controls the time-splitting instability, it also introduces non-physical damping and reduces the accuracy. This presentation proposes a simple modification to the RA filter, which has become known as the RAW filter (Williams 2009, 2011). When used in conjunction with the leapfrog scheme, the RAW filter eliminates the non-physical damping and increases the amplitude accuracy by two orders, yielding third-order accuracy. (The phase accuracy remains second-order.) The RAW filter can easily be incorporated into existing models, typically via the insertion of just a single line of code. Better simulations are obtained at no extra computational expense. Results will be shown from recent implementations of the RAW filter in various models, including SPEEDY and COSMO. For example, in SPEEDY, the skill of weather forecasts is found to be significantly improved. In particular, in tropical surface pressure predictions, five-day forecasts made using the RAW filter have approximately the same skill as four-day forecasts made using the RA filter (Amezcua, Kalnay & Williams 2011). These improvements are encouraging for the use of the RAW filter in other atmosphere and ocean models. References PD Williams (2009) A

  9. Brief inhalation method to measure cerebral oxygen extraction fraction with PET: Accuracy determination under pathologic conditions

    SciTech Connect

    Altman, D.I.; Lich, L.L.; Powers, W.J. )

    1991-09-01

    The initial validation of the brief inhalation method to measure cerebral oxygen extraction fraction (OEF) with positron emission tomography (PET) was performed in non-human primates with predominantly normal cerebral oxygen metabolism (CMRO2). Sensitivity analysis by computer simulation, however, indicated that this method may be subject to increasing error as CMRO2 decreases. Accuracy of the method under pathologic conditions of reduced CMRO2 has not been determined. Since reduced CMRO2 values are observed frequently in newborn infants and in regions of ischemia and infarction in adults, we determined the accuracy of the brief inhalation method in non-human primates by comparing OEF measured with PET to OEF measured by arteriovenous oxygen difference (A-VO2) under pathologic conditions of reduced CMRO2 (0.27-2.68 ml 100g-1 min-1). A regression equation of OEF (PET) = 1.07 {times} OEF (A-VO2) + 0.017 (r = 0.99, n = 12) was obtained. The absolute error in oxygen extraction measured with PET was small (mean 0.03 {plus minus} 0.04, range -0.03 to 0.12) and was independent of cerebral blood flow, cerebral blood volume, CMRO2, or OEF. The percent error was higher (19 {plus minus} 37), particularly when OEF is below 0.15. These data indicate that the brief inhalation method can be used for measurement of cerebral oxygen extraction and cerebral oxygen metabolism under pathologic conditions of reduced cerebral oxygen metabolism, with these limitations borne in mind.

  10. A stencil penalty method for improving accuracy of constraint immersed boundary method

    NASA Astrophysics Data System (ADS)

    Bale, Rahul; Jansson, Niclas; Onishi, Keiji; Tsubokura, Makoto; Patankar, Neelesh

    2016-11-01

    The constraint based immersed boundary (cIB) method is known to be accurate for low and moderate Reynolds number (Re) flows. At high Re, we found that cIB is not able produce accurate results. High Re flows typically result in large pressure gradient across fluid-IB interface. This is especially pronounced when the IB is an interface with "zero-thickness." There is also a jump in pressure which leads to incorrect evaluation of pressure gradients near the fluid-IB interface. This error leads to inaccuracies in the boundary layer around the IB and can also lead to leakage of flow across the interface. We propose a novel IB formulation with a modified pressure gradient operator that calculates one-sided gradients on either side of the interface. This removes spurious gradients in pressure across the interface. The pressure gradient operator is modified using a WENO based stencil penalization scheme.

  11. Methods of geodiversity assessment and theirs application

    NASA Astrophysics Data System (ADS)

    Zwoliński, Zbigniew; Najwer, Alicja; Giardino, Marco

    2016-04-01

    analysis is not as simple as in the case of direct methods. Indirect methods offer the possibility to assessment and mapping of geodiversity of large and not easy accessible research areas. The foregoing examples mainly tend to refer to areas at the local or regional scale. These analyses are although possible for the implementation of large spatial units, such as territories of the country or state (Zwoliński 2007, Benito-Calvo et al. 2009, Pereira et al. 2013, 2015). A fundamental difference lies in the appropriate, corresponding to the spatial scale and the specification of the study areas, selection of the assessing criterion, and above all, the input geodata. In the geodiversity assessments, access to the data in the adequate resolution and accuracy is especially important. Acquisition and integration of the geodata often requires considerable financial and temporal outlay and not infrequently could be a serious limitation to perform some analyzes. The proposal of geomorphometry based landform geodiversity indirect assessment method and single, in addition, easy to obtain source data - digital elevation model, might create new opportunities for its broad implementation across numerous disciplines. The research on the assessment of geodiversity must be regarded to be at an initial stage at present. While the conception of geodiversity itself has a reliable theoretical foundation, no universal method of its assessment has been developed yet. It is only the adoption of a generally accepted and clear methodology of geodiversity evaluation that will make it possible to implement it widely in many fields of science, administration and the management of geospace. Then geodiversity can become as important an indicator as biodiversity is today.

  12. Accuracy of enthalpy and entropy determination using the kinetic method: are we approaching a consensus?

    PubMed

    Drahos, László; Peltz, Csaba; Vékey, Károly

    2004-09-01

    There is an emerging consensus regarding the applicability of the kinetic method. All parties acknowledge that it is an approximate quantitative technique, capable of yielding not only enthalpy, but also entropy values. Opinions differ mainly on the accuracy of the results but it is agreed that the energy (effective temperature) dependence of kinetic method plots needs to be checked in all but the simplest of cases. When the 'apparent basicity' is found to depend on collision energy (and hence effective temperature), the extended kinetic method must be used. We have performed a large-scale modeling study, involving thousands of randomly selected molecular systems and a variety of experimental conditions, using exact calculations and realistic data sets. The results show that when the measured entropy difference between the two competing reaction channels is less than approximately 35 J mol(-1) K(-1), overall errors (standard deviations) of DeltaH(298) determined by the kinetic method are +/-5 kJ mol(-1); those of DeltaS(298) are +/-10 J mol(-1) K(-1). These include not only inherent errors of the kinetic method, but also errors in ion abundance measurement (5%) and inaccurate knowledge of reference compound thermochemistry (+/-2 kJ mol(-1), on average). We recommend, in general, that these errors be reported in kinetic method studies. When the measured entropy difference between the two competing fragmentation channels is large (>35 J mol(-1) K(-1)), it is likely to be significantly underestimated and errors of the kinetic method increase significantly.

  13. Utilizing the Global Land Cover 2000 reference dataset for a comparative accuracy assessment of 1 km global land cover maps

    NASA Astrophysics Data System (ADS)

    Schultz, M.; Tsendbazazr, N. E.; Herold, M.; Jung, M.; Mayaux, P.; Goehman, H.

    2015-04-01

    Many investigators use global land cover (GLC) maps for different purposes, such as an input for global climate models. The current GLC maps used for such purposes are based on different remote sensing data, methodologies and legends. Consequently, comparison of GLC maps is difficult and information about their relative utility is limited. The objective of this study is to analyse and compare the thematic accuracies of GLC maps (i.e., IGBP-DISCover, UMD, MODIS, GLC2000 and SYNMAP) at 1 km resolutions by (a) re-analysing the GLC2000 reference dataset, (b) applying a generalized GLC legend and (c) comparing their thematic accuracies at different homogeneity levels. The accuracy assessment was based on the GLC2000 reference dataset with 1253 samples that were visually interpreted. The legends of the GLC maps and the reference datasets were harmonized into 11 general land cover classes. There results show that the map accuracy estimates vary up to 10-16% depending on the homogeneity of the reference point (HRP) for all the GLC maps. An increase of the HRP resulted in higher overall accuracies but reduced accuracy confidence for the GLC maps due to less number of accountable samples. The overall accuracy of the SYNMAP was the highest at any HRP level followed by the GLC2000. The overall accuracies of the maps also varied by up to 10% depending on the definition of agreement between the reference and map categories in heterogeneous landscape. A careful consideration of heterogeneous landscape is therefore recommended for future accuracy assessments of land cover maps.

  14. Assessing the impact of measurement frequency on accuracy and uncertainty of water quality data

    NASA Astrophysics Data System (ADS)

    Helm, Björn; Schiffner, Stefanie; Krebs, Peter

    2014-05-01

    Physico-chemical water quality is a major objective for the evaluation of the ecological state of a river water body. Physical and chemical water properties are measured to assess the river state, identify prevalent pressures and develop mitigating measures. Regularly water quality is assessed based on weekly to quarterly grab samples. The increasing availability of online-sensor data measured at a high frequency allows for an enhanced understanding of emission and transport dynamics, as well as the identification of typical and critical states. In this study we present a systematic approach to assess the impact of measurement frequency on the accuracy and uncertainty of derived aggregate indicators of environmental quality. High frequency measured (10 min-1 and 15 min-1) data on water temperature, pH, turbidity, electric conductivity and concentrations of dissolved oxygen nitrate, ammonia and phosphate are assessed in resampling experiments. The data is collected at 14 sites in eastern and northern Germany representing catchments between 40 km2 and 140 000 km2 of varying properties. Resampling is performed to create series of hourly to quarterly frequency, including special restrictions like sampling at working hours or discharge compensation. Statistical properties and their confidence intervals are determined in a bootstrapping procedure and evaluated along a gradient of sampling frequency. For all variables the range of the aggregate indicators increases largely in the bootstrapping realizations with decreasing sampling frequency. Mean values of electric conductivity, pH and water temperature obtained with monthly frequency differ in average less than five percent from the original data. Mean dissolved oxygen, nitrate and phosphate had in most stations less than 15 % bias. Ammonia and turbidity are most sensitive to the increase of sampling frequency with up to 30 % in average and 250 % maximum bias at monthly sampling frequency. A systematic bias is recognized

  15. Accuracy of Cameriere's cut-off value for third molar in assessing 18 years of age.

    PubMed

    De Luca, S; Biagi, R; Begnoni, G; Farronato, G; Cingolani, M; Merelli, V; Ferrante, L; Cameriere, R

    2014-02-01

    Due to increasingly numerous international migrations, estimating the age of unaccompanied minors is becoming of enormous significance for forensic professionals who are required to deliver expert opinions. The third molar tooth is one of the few anatomical sites available for estimating the age of individuals in late adolescence. This study verifies the accuracy of Cameriere's cut-off value of the third molar index (I3M) in assessing 18 years of age. For this purpose, a sample of orthopantomographs (OPTs) of 397 living subjects aged between 13 and 22 years (192 female and 205 male) was analyzed. Age distribution gradually decreases as I3M increases in both males and females. The results show that the sensitivity of the test was 86.6%, with a 95% confidence interval of (80.8%, 91.1%), and its specificity was 95.7%, with a 95% confidence interval of (92.1%, 98%). The proportion of correctly classified individuals was 91.4%. Estimated post-test probability, p was 95.6%, with a 95% confidence interval of (92%, 98%). Hence, the probability that a subject positive on the test (i.e., I3M<0.08) was 18 years of age or older was 95.6%.

  16. Assessment of dimensional accuracy of preadjusted metal injection molding orthodontic brackets

    PubMed Central

    Alavi, Shiva; Tajmirriahi, Farnaz

    2016-01-01

    Background: the aim of this study is to evaluate the dimensional accuracy of McLaughlin, Bennett, and Trevisi (MBT) brackets manufactured by two different companies (American Orthodontics and Ortho Organizers) and determine variations in incorporation of values in relation to tip and torque in these products. Materials and Methods: In the present analytical/descriptive study, 64 maxillary right central brackets manufactured by two companies (American Orthodontics and Ortho Organizers) were selected randomly and evaluated for the accuracy of the values in relation to torque and angulation presented by the manufacturers. They were placed in a video measuring machine using special revolvers under them and were positioned in a manner so that the light beams would be directed on the floor of the slot without the slot walls being seen. Then, the software program of the same machine was used to determine the values of each bracket type. The means of measurements were determined for each sample and were analyzed with independent t-test and one-sample t-test. Results: Based on the confidence interval, it can be concluded that at 95% probability, the means of tip angles of maxillary right central brackets of these two brands were 4.1–4.3° and the torque angles were 16.39–16.72°. The tips in these samples were at a range of 3.33–4.98°, and the torque was at a range of 15.22–18.48°. Conclusion: In the present study, there were no significant differences in the angulation incorporated into the brackets from the two companies; however, they were significantly different from the tiP values for the MBT prescription. In relation to torque, there was a significant difference between the American Orthodontic brackets exhibited significant differences with the reported 17°, too. PMID:27857770

  17. Method for improving terahertz band absorption spectrum measurement accuracy using noncontact sample thickness measurement.

    PubMed

    Li, Zhi; Zhang, Zhaohui; Zhao, Xiaoyan; Su, Haixia; Yan, Fang; Zhang, Han

    2012-07-10

    The terahertz absorption spectrum has a complex nonlinear relationship with sample thickness, which is normally measured mechanically with limited accuracy. As a result, the terahertz absorption spectrum is usually determined incorrectly. In this paper, an iterative algorithm is proposed to accurately determine sample thickness. This algorithm is independent of the initial value used and results in convergent calculations. Precision in sample thickness can be improved up to 0.1 μm. A more precise absorption spectrum can then be extracted. By comparing the proposed method with the traditional method based on mechanical thickness measurements, quantitative analysis experiments on a three-component amino acid mixture shows that the global error decreased from 0.0338 to 0.0301.

  18. Self Assessment in Schizophrenia: Accuracy of Evaluation of Cognition and Everyday Functioning

    PubMed Central

    Gould, Felicia; McGuire, Laura Stone; Durand, Dante; Sabbag, Samir; Larrauri, Carlos; Patterson, Thomas L.; Twamley, Elizabeth W.; Harvey, Philip D.

    2015-01-01

    Objective Self-assessment deficits, often referred to as impaired insight or unawareness of illness, are well established in people with schizophrenia. There are multiple levels of awareness, including awareness of symptoms, functional deficits, cognitive impairments, and the ability to monitor cognitive and functional performance in an ongoing manner. The present study aimed to evaluate the comparative predictive value of each aspect of awareness on the levels of everyday functioning in people with schizophrenia. Method We examined multiple aspects of self-assessment of functioning in 214 people with schizophrenia. We also collected information on everyday functioning rated by high contact clinicians and examined the importance of self-assessment for the prediction of real world functional outcomes. The relative impact of performance based measures of cognition, functional capacity, and metacognitive performance on everyday functioning was also examined. Results Misestimation of ability emerged as the strongest predictor of real world functioning and exceeded the influences of cognitive performance, functional capacity performance, and performance-based assessment of metacognitive monitoring. The relative contribution of the factors other than self-assessment varied according to which domain of everyday functioning was being examined, but in all cases, accounted for less predictive variance. Conclusions These results underscore the functional impact of misestimating one’s current functioning and relative level of ability. These findings are consistent with the use of insight-focused treatments and compensatory strategies designed to increase self-awareness in multiple functional domains. PMID:25643212

  19. Millimetre Level Accuracy GNSS Positioning with the Blind Adaptive Beamforming Method in Interference Environments.

    PubMed

    Daneshmand, Saeed; Marathe, Thyagaraja; Lachapelle, Gérard

    2016-10-31

    The use of antenna arrays in Global Navigation Satellite System (GNSS) applications is gaining significant attention due to its superior capability to suppress both narrowband and wideband interference. However, the phase distortions resulting from array processing may limit the applicability of these methods for high precision applications using carrier phase based positioning techniques. This paper studies the phase distortions occurring with the adaptive blind beamforming method in which satellite angle of arrival (AoA) information is not employed in the optimization problem. To cater to non-stationary interference scenarios, the array weights of the adaptive beamformer are continuously updated. The effects of these continuous updates on the tracking parameters of a GNSS receiver are analyzed. The second part of this paper focuses on reducing the phase distortions during the blind beamforming process in order to allow the receiver to perform carrier phase based positioning by applying a constraint on the structure of the array configuration and by compensating the array uncertainties. Limitations of the previous methods are studied and a new method is proposed that keeps the simplicity of the blind beamformer structure and, at the same time, reduces tracking degradations while achieving millimetre level positioning accuracy in interference environments. To verify the applicability of the proposed method and analyze the degradations, array signals corresponding to the GPS L1 band are generated using a combination of hardware and software simulators. Furthermore, the amount of degradation and performance of the proposed method under different conditions are evaluated based on Monte Carlo simulations.

  20. Millimetre Level Accuracy GNSS Positioning with the Blind Adaptive Beamforming Method in Interference Environments

    PubMed Central

    Daneshmand, Saeed; Marathe, Thyagaraja; Lachapelle, Gérard

    2016-01-01

    The use of antenna arrays in Global Navigation Satellite System (GNSS) applications is gaining significant attention due to its superior capability to suppress both narrowband and wideband interference. However, the phase distortions resulting from array processing may limit the applicability of these methods for high precision applications using carrier phase based positioning techniques. This paper studies the phase distortions occurring with the adaptive blind beamforming method in which satellite angle of arrival (AoA) information is not employed in the optimization problem. To cater to non-stationary interference scenarios, the array weights of the adaptive beamformer are continuously updated. The effects of these continuous updates on the tracking parameters of a GNSS receiver are analyzed. The second part of this paper focuses on reducing the phase distortions during the blind beamforming process in order to allow the receiver to perform carrier phase based positioning by applying a constraint on the structure of the array configuration and by compensating the array uncertainties. Limitations of the previous methods are studied and a new method is proposed that keeps the simplicity of the blind beamformer structure and, at the same time, reduces tracking degradations while achieving millimetre level positioning accuracy in interference environments. To verify the applicability of the proposed method and analyze the degradations, array signals corresponding to the GPS L1 band are generated using a combination of hardware and software simulators. Furthermore, the amount of degradation and performance of the proposed method under different conditions are evaluated based on Monte Carlo simulations. PMID:27809252

  1. A Comparative Investigation of Several Methods of Aiding College Freshmen to Achieve Grammatical Accuracy in Written Composition.

    ERIC Educational Resources Information Center

    Essary, William Howard

    Two problems were investigated in this study: (1) Which (if any) method of teaching freshmen composition is most effective in helping college students achieve grammatical accuracy? (2) Is improvement in grammatical accuracy paralleled or contrasted with improvement in content? Relatively weak students (low C high-school average and a mean SAT…

  2. 40 CFR 80.584 - What are the precision and accuracy criteria for approval of test methods for determining the...

    Code of Federal Regulations, 2011 CFR

    2011-07-01

    ... 40 Protection of Environment 16 2011-07-01 2011-07-01 false What are the precision and accuracy....584 What are the precision and accuracy criteria for approval of test methods for determining the sulfur content of motor vehicle diesel fuel, NRLM diesel fuel, and ECA marine fuel? (a) Precision....

  3. 40 CFR 80.584 - What are the precision and accuracy criteria for approval of test methods for determining the...

    Code of Federal Regulations, 2010 CFR

    2010-07-01

    ... 40 Protection of Environment 16 2010-07-01 2010-07-01 false What are the precision and accuracy....584 What are the precision and accuracy criteria for approval of test methods for determining the sulfur content of motor vehicle diesel fuel, NRLM diesel fuel, and ECA marine fuel? (a) Precision....

  4. QuickBird and OrbView-3 Geopositional Accuracy Assessment

    NASA Technical Reports Server (NTRS)

    Helder, Dennis; Ross, Kenton

    2006-01-01

    Objective: Compare vendor-provided image coordinates with known references visible in the imagery. Approach: Use multiple, well-characterized sites with >40 ground control points (GCPs); sites that are a) Well distributed; b) Accurately surveyed; and c) Easily found in imagery. Perform independent assessments with independent teams. Each team has slightly different measurement techniques and data processing methods. NASA Stennis Space Center. South Dakota State University.

  5. Assessment of Precipitation Forecast Accuracy over Eastern Black Sea Region using WRF-ARW

    NASA Astrophysics Data System (ADS)

    Bıyık, G.; Unal, Y.; Onol, B.

    2009-09-01

    Surface topography such as mountain barriers, existing water bodies and semi-permanent mountain glaciers changes large scale atmospheric patterns and creates a challenge for a reliable precipitation prediction. Eastern Black sea region of Turkey is an example. Black Sea Mountain chains lies west to east along the coastline with the average height of 2000 m and the highest point is 3973 m, and from the coastline to inland there is a very sharp topography change. For this project we select the Eastern Black Sea region of Turkey to assess precipitation forecast accuracy. This is a unique region of Turkey which receive both highest amount of precipitation and precipitation throughout whole year. Amount of rain and snow is important because they supply water to the main river systems of Turkey. Turkey is in general under the influence of both continental polar (Cp) and tropical air masses. Their interaction with the orography causes orographic precipitation being effective on the region. Also Caucasus Mountains, which is the highest point of Georgia, moderates the climate of the southern parts by not letting penetration of colder air masses from north. Southern part of the western Black Sea region has more continental climate because of the lee side effect of the mountains Therefore, precipitation forecast in the region is important for operational forecasters and researchers. Our aim in this project is to investigate WRF precipitation accuracy during 10 extreme precipitation, 10 normal precipitation and 10 no precipitation days by using forecast for two days ahead. Cases are selected in years between 2000 and 2003. Eleven Eastern Black Sea stations located along the coastline are used to determine 20 extreme and 10 average precipitation days. During project, three different resolutions with three nested domains are tested to determine the model sensivity to domain boundaries and resolution. As a result of our tests, 6 km resolution for finer domain was found suitable

  6. Assessing the Accuracy of Sentinel-3 SLSTR Sea-Surface Temperature Retrievals Using High Accuracy Infrared Radiiometers on Ships of Opportunity

    NASA Astrophysics Data System (ADS)

    Minnett, P. J.; Izaguirre, M. A.; Szcszodrak, M.; Williams, E.; Reynolds, R. M.

    2015-12-01

    The assessment of errors and uncertainties in satellite-derived SSTs can be achieved by comparisons with independent measurements of skin SST of high accuracy. Such validation measurements are provided by well-calibrated infrared radiometers mounted on ships. The second generation of Marine-Atmospheric Emitted Radiance Interferometers (M-AERIs) have recently been developed and two are now deployed on cruise ships of Royal Caribbean Cruise Lines that operate in the Caribbean Sea, North Atlantic and Mediterranean Sea. In addition, two Infrared SST Autonomous Radiometers (ISARs) are mounted alternately on a vehicle transporter of NYK Lines that crosses the Pacific Ocean between Japan and the USA. Both M-AERIs and ISARs are self-calibrating radiometers having two internal blackbody cavities to provide at-sea calibration of the measured radiances, and the accuracy of the internal calibration is periodically determined by measurements of a NIST-traceable blackbody cavity in the laboratory. This provides SI-traceability for the at-sea measurements. It is anticipated that these sensors will be deployed during the next several years and will be available for the validation of the SLSTRs on Sentinel-3a and -3b.

  7. A Comparative Accuracy Analysis of Classification Methods in Determination of Cultivated Lands with Spot 5 Satellite Imagery

    NASA Astrophysics Data System (ADS)

    kaya, S.; Alganci, U.; Sertel, E.; Ustundag, B.

    2013-12-01

    A Comparative Accuracy Analysis of Classification Methods in Determination of Cultivated Lands with Spot 5 Satellite Imagery Ugur ALGANCI1, Sinasi KAYA1,2, Elif SERTEL1,2,Berk USTUNDAG3 1 ITU, Center for Satellite Communication and Remote Sensing, 34469, Maslak-Istanbul,Turkey 2 ITU, Department of Geomatics, 34469, Maslak-Istanbul, Turkey 3 ITU, Agricultural and Environmental Informatics Research Center,34469, Maslak-Istanbul,Turkey alganci@itu.edu.tr, kayasina@itu.edu.tr, sertele@itu.edu.tr, berk@berk.tc ABSTRACT Cultivated land determination and their area estimation are important tasks for agricultural management. Derived information is mostly used in agricultural policies and precision agriculture, in specifically; yield estimation, irrigation and fertilization management and farmers declaration verification etc. The use of satellite image in crop type identification and area estimate is common for two decades due to its capability of monitoring large areas, rapid data acquisition and spectral response to crop properties. With launch of high and very high spatial resolution optical satellites in the last decade, such kind of analysis have gained importance as they provide information at big scale. With increasing spatial resolution of satellite images, image classification methods to derive the information form them have become important with increase of the spectral heterogeneity within land objects. In this research, pixel based classification with maximum likelihood algorithm and object based classification with nearest neighbor algorithm were applied to 2012 dated 2.5 m resolution SPOT 5 satellite images in order to investigate the accuracy of these methods in determination of cotton and corn planted lands and their area estimation. Study area was selected in Sanliurfa Province located on Southeastern Turkey that contributes to Turkey's agricultural production in a major way. Classification results were compared in terms of crop type identification using

  8. Evaluation method of lead measurement accuracy of gears using a wedge artefact

    NASA Astrophysics Data System (ADS)

    Komori, Masaharu; Takeoka, Fumi; Kubo, Aizoh; Okamoto, Kazuhiko; Osawa, Sonko; Sato, Osamu; Takatsuji, Toshiyuki

    2009-02-01

    The reduction of the vibration and noise of gears is an important issue in mechanical devices such as vehicles and wind turbines. The characteristics of the vibration and noise of gears are markedly affected by deviations of the tooth flank form of micrometre order; therefore, a strict quality control of the tooth flank form is required. The accuracy of the lead measurement for a gear-measuring instrument is usually evaluated using a master gear or a lead master. However, it is difficult to manufacture masters with high accuracy because the helix is a complicated geometrical form. In this paper, we propose a method of evaluating a gear-measuring instrument using a wedge artefact, which includes a highly precise plane surface. The concept of the wedge artefact is described and a mathematical model of the measuring condition of the wedge artefact is constructed. Theoretical measurement results for the wedge artefact are calculated. The wedge artefact is designed and produced on the basis of the theoretical measurement results. A measurement experiment using the wedge artefact is carried out and its effectiveness is verified.

  9. Using Generalizability Theory to Examine the Accuracy and Validity of Large-Scale ESL Writing Assessment

    ERIC Educational Resources Information Center

    Huang, Jinyan

    2012-01-01

    Using generalizability (G-) theory, this study examined the accuracy and validity of the writing scores assigned to secondary school ESL students in the provincial English examinations in Canada. The major research question that guided this study was: Are there any differences between the accuracy and construct validity of the analytic scores…

  10. Accuracies of southwell and force/stiffness methods in the prediction of buckling strength of hypersonic aircraft wing tubular panels

    NASA Technical Reports Server (NTRS)

    Ko, William L.

    1987-01-01

    Accuracies of the Southwell method and the force/stiffness (F/S) method are examined when the methods were used in the prediction of buckling loads of hypersonic aircraft wing tubular panels, based on nondestructive buckling test data. Various factors affecting the accuracies of the two methods were discussed. Effects of load cutoff point in the nondestructive buckling tests on the accuracies of the two methods were discussed in great detail. For the tubular panels under pure compression, the F/S method was found to give more accurate buckling load predictions than the Southwell method, which excessively overpredicts the buckling load. It was found that the Southwell method required a higher load cutoff point, as compared with the F/S method. In using the F/S method for predicting the buckling load of tubular panels under pure compression, the load cutoff point of approximately 50 percent of the critical load could give reasonably accurate predictions.

  11. Accuracy Assessment of Mobile Mapping Point Clouds Using the Existing Environment as Terrestrial Reference

    NASA Astrophysics Data System (ADS)

    Hofmann, S.; Brenner, C.

    2016-06-01

    Mobile mapping data is widely used in various applications, what makes it especially important for data users to get a statistically verified quality statement on the geometric accuracy of the acquired point clouds or its processed products. The accuracy of point clouds can be divided into an absolute and a relative quality, where the absolute quality describes the position of the point cloud in a world coordinate system such as WGS84 or UTM, whereas the relative accuracy describes the accuracy within the point cloud itself. Furthermore, the quality of processed products such as segmented features depends on the global accuracy of the point cloud but mainly on the quality of the processing steps. Several data sources with different characteristics and quality can be thought of as potential reference data, such as cadastral maps, orthophoto, artificial control objects or terrestrial surveys using a total station. In this work a test field in a selected residential area was acquired as reference data in a terrestrial survey using a total station. In order to reach high accuracy the stationing of the total station was based on a newly made geodetic network with a local accuracy of less than 3 mm. The global position of the network was determined using a long time GNSS survey reaching an accuracy of 8 mm. Based on this geodetic network a 3D test field with facades and street profiles was measured with a total station, each point with a two-dimensional position and altitude. In addition, the surface of poles of street lights, traffic signs and trees was acquired using the scanning mode of the total station. Comparing this reference data to the acquired mobile mapping point clouds of several measurement campaigns a detailed quality statement on the accuracy of the point cloud data is made. Additionally, the advantages and disadvantages of the described reference data source concerning availability, cost, accuracy and applicability are discussed.

  12. Development of a method for the determination of total bisphenol a at trace levels in human blood and urine and elucidation of factors influencing method accuracy and sensitivity.

    PubMed

    Markham, Dan; Waechter, John; Budinsky, Robert; Gries, Wolfgang; Beyer, Dieter; Snyder, Stephanie; Dimond, Stephen; Rajesh, V N; Rao, Narayana; Connolly, Paul; Neeley, Mark; Hentges, Steven

    2014-05-01

    This publication describes a method for the determination of total bisphenol A (BPA and conjugated BPA) following enzyme hydrolysis and is intended as a companion to our previously developed analytical method for the determination of free BPA (the aglycone) in human blood and urine using high-performance liquid chromatography-tandem mass spectrometry ( 1). That free BPA method provided a means to account for and/or eliminate background contamination and demonstrated accuracy and reproducibility in both matrices fortified with BPA or a surrogate analyte ((13)C BPA) at a low method quantitation limit (MQL) of 0.1-0.2 ng/mL. In contrast to the free BPA method results and based on stringent accuracy, precision and confirmation criteria set for the MQLs of the method developed for total BPA, the MQL achieved in blood was 1.020-2.550 and 0.510-1.020 ng/mL in urine. These data showed higher MQLs than the desired MQLs of 0.5 ng/mL (blood) and 0.2 ng/mL (urine) with increased variability between analyses which demonstrates the importance of generating method validation data with each analysis. In contrast, the MQL achieved for (13)C BPA-G (monoglucuronide as a surrogate analyte in blood was 0.2-0.5 and 0.2 ng/mL in urine illustrating that the method is capable of meeting lower MQL requirements if the contribution from exogenous BPA can be well controlled. This method for the determination total BPA in human blood and urine is intended to be used in conjunction with the free BPA method ( 1) to obtain accurate and complete BPA biomonitoring data to support human exposure assessments.

  13. A quantitative method for evaluating numerical simulation accuracy of time-transient Lamb wave propagation with its applications to selecting appropriate element size and time step.

    PubMed

    Wan, Xiang; Xu, Guanghua; Zhang, Qing; Tse, Peter W; Tan, Haihui

    2016-01-01

    Lamb wave technique has been widely used in non-destructive evaluation (NDE) and structural health monitoring (SHM). However, due to the multi-mode characteristics and dispersive nature, Lamb wave propagation behavior is much more complex than that of bulk waves. Numerous numerical simulations on Lamb wave propagation have been conducted to study its physical principles. However, few quantitative studies on evaluating the accuracy of these numerical simulations were reported. In this paper, a method based on cross correlation analysis for quantitatively evaluating the simulation accuracy of time-transient Lamb waves propagation is proposed. Two kinds of error, affecting the position and shape accuracies are firstly identified. Consequently, two quantitative indices, i.e., the GVE (group velocity error) and MACCC (maximum absolute value of cross correlation coefficient) derived from cross correlation analysis between a simulated signal and a reference waveform, are proposed to assess the position and shape errors of the simulated signal. In this way, the simulation accuracy on the position and shape is quantitatively evaluated. In order to apply this proposed method to select appropriate element size and time step, a specialized 2D-FEM program combined with the proposed method is developed. Then, the proper element size considering different element types and time step considering different time integration schemes are selected. These results proved that the proposed method is feasible and effective, and can be used as an efficient tool for quantitatively evaluating and verifying the simulation accuracy of time-transient Lamb wave propagation.

  14. Designing a Multi-Objective Multi-Support Accuracy Assessment of the 2001 National Land Cover Data (NLCD 2001) of the Conterminous United States

    EPA Science Inventory

    The database design and diverse application of NLCD 2001 pose significant challenges for accuracy assessment because numerous objectives are of interest, including accuracy of land cover, percent urban imperviousness, percent tree canopy, land-cover composition, and net change. ...

  15. Assessing posttraumatic stress in military service members: improving efficiency and accuracy.

    PubMed

    Fissette, Caitlin L; Snyder, Douglas K; Balderrama-Durbin, Christina; Balsis, Steve; Cigrang, Jeffrey; Talcott, G Wayne; Tatum, JoLyn; Baker, Monty; Cassidy, Daniel; Sonnek, Scott; Heyman, Richard E; Smith Slep, Amy M

    2014-03-01

    Posttraumatic stress disorder (PTSD) is assessed across many different populations and assessment contexts. However, measures of PTSD symptomatology often are not tailored to meet the needs and demands of these different populations and settings. In order to develop population- and context-specific measures of PTSD it is useful first to examine the item-level functioning of existing assessment methods. One such assessment measure is the 17-item PTSD Checklist-Military version (PCL-M; Weathers, Litz, Herman, Huska, & Keane, 1993). Although the PCL-M is widely used in both military and veteran health-care settings, it is limited by interpretations based on aggregate scores that ignore variability in item endorsement rates and relatedness to PTSD. Based on item response theory, this study conducted 2-parameter logistic analyses of the PCL-M in a sample of 196 service members returning from a yearlong, high-risk deployment to Iraq. Results confirmed substantial variability across items both in terms of their relatedness to PTSD and their likelihood of endorsement at any given level of PTSD. The test information curve for the full 17-item PCL-M peaked sharply at a value of θ = 0.71, reflecting greatest information at approximately the 76th percentile level of underlying PTSD symptom levels in this sample. Implications of findings are discussed as they relate to identifying more efficient, accurate subsets of items tailored to military service members as well as other specific populations and evaluation contexts.

  16. The modified equation approach to the stability and accuracy analysis of finite-difference methods

    NASA Technical Reports Server (NTRS)

    Warming, R. F.; Hyett, B. J.

    1974-01-01

    The stability and accuracy of finite-difference approximations to simple linear partial differential equations are analyzed by studying the modified partial differential equation. Aside from round-off error, the modified equation represents the actual partial differential equation solved when a numerical solution is computed using a finite-difference equation. The modified equation is derived by first expanding each term of a difference scheme in a Taylor series and then eliminating time derivatives higher than first order by certain algebraic manipulations. The connection between 'heuristic' stability theory based on the modified equation approach and the von Neumann (Fourier) method is established. In addition to the determination of necessary and sufficient conditions for computational stability, a truncated version of the modified equation can be used to gain insight into the nature of both dissipative and dispersive errors.

  17. Methods to improve pressure, temperature and velocity accuracies of filtered Rayleigh scattering measurements in gaseous flows

    NASA Astrophysics Data System (ADS)

    Doll, Ulrich; Burow, Eike; Stockhausen, Guido; Willert, Christian

    2016-12-01

    Frequency scanning filtered Rayleigh scattering is able to simultaneously provide time-averaged measurements of pressure, temperature and velocity in gaseous flows. By extending the underlying mathematical model, a robust alternative to existing approaches is introduced. Present and proposed model functions are then characterized during a detailed uncertainty analysis. Deviations between the analytical solution of a jet flow experiment and measured results could be related to laser-induced background radiation as well as the Rayleigh scattering’s spectral distribution. In applying a background correction method and by replacing the standard lineshape model by an empirical formulation, detrimental effects on pressure, temperature and velocity accuracies could be reduced below 15 hPa, 2.5 K and 2.7 m s-1.

  18. On the use of polymer gels for assessing the total geometrical accuracy in clinical Gamma Knife radiosurgery applications

    NASA Astrophysics Data System (ADS)

    Moutsatsos, A.; Karaiskos, P.; Petrokokkinos, L.; Zourari, K.; Pantelis, E.; Sakelliou, L.; Seimenis, I.; Constantinou, C.; Peraticou, A.; Georgiou, E.

    2010-11-01

    The nearly tissue equivalent MRI properties and the unique ability of registering 3D dose distributions of polymer gels were exploited to assess the total geometrical accuracy in clinical Gamma Knife applications, taking into account the combined effect of the unit's mechanical accuracy, dose delivery precision and the geometrical distortions inherent in MR images used for irradiation planning. Comparison between planned and experimental data suggests that the MR-related distortions due to susceptibility effects dominate the total clinical geometrical accuracy which was found within 1 mm. The dosimetric effect of the observed sub-millimetre uncertainties on single shot GK irradiation plans was assessed using the target percentage coverage criterion, and a considerable target dose underestimation was found.

  19. [Research on Accuracy and Stability of Inversing Vegetation Chlorophyll Content by Spectral Index Method].

    PubMed

    Jiang, Hai-ling; Yang, Hang; Chen, Xiao-ping; Wang, Shu-dong; Li, Xue-ke; Liu, Kai; Cen, Yi

    2015-04-01

    Spectral index method was widely applied to the inversion of crop chlorophyll content. In the present study, PSR3500 spectrometer and SPAD-502 chlorophyll fluorometer were used to acquire the spectrum and relative chlorophyll content (SPAD value) of winter wheat leaves on May 2nd 2013 when it was at the jointing stage of winter wheat. Then the measured spectra were resampled to simulate TM multispectral data and Hyperion hyperspectral data respectively, using the Gaussian spectral response function. We chose four typical spectral indices including normalized difference vegetation index (NDVD, triangle vegetation index (TVI), the ratio of modified transformed chlorophyll absorption ratio index (MCARI) to optimized soil adjusted vegetation index (OSAVI) (MCARI/OSAVI) and vegetation index based on universal pattern decomposition (VIUPD), which were constructed with the feature bands sensitive to the vegetation chlorophyll. After calculating these spectral indices based on the resampling TM and Hyperion data, the regression equation between spectral indices and chlorophyll content was established. For TM, the result indicates that VIUPD has the best correlation with chlorophyll (R2 = 0.819 7) followed by NDVI (R2 = 0.791 8), while MCARI/OSAVI and TVI also show a good correlation with R2 higher than 0.5. For the simulated Hyperion data, VIUPD again ranks first with R2 = 0.817 1, followed by MCARI/OSAVI (R2 = 0.658 6), while NDVI and TVI show very low values with R2 less than 0.2. It was demonstrated that VIUPD has the best accuracy and stability to estimate chlorophyll of winter wheat whether using simulated TM data or Hyperion data, which reaffirms that VIUPD is comparatively sensor independent. The chlorophyll estimation accuracy and stability of MCARI/OSAVI also works well, partly because OSAVI could reduce the influence of backgrounds. Two broadband spectral indices NDVI and TVI are weak for the chlorophyll estimation of simulated Hyperion data mainly because of

  20. Accuracy of the third molar index for assessing the legal majority of 18 years in Turkish population.

    PubMed

    Gulsahi, Ayse; De Luca, Stefano; Cehreli, S Burcak; Tirali, R Ebru; Cameriere, Roberto

    2016-09-01

    In the last few years, forced and unregistered child marriage has widely increased into Turkey. The aim of this study was to test the accuracy of cut-off value of 0.08 by measurement of third molar index (I3M) in assessing legal adult age of 18 years. Digital panoramic images of 293 Turkish children and young adults (165 girls and 128 boys), aged between 14 and 22 years, were analysed. Age distribution gradually decreases as I3M increases in both girls and boys. For girls, the sensitivity was 85.9% (95% CI 77.1-92.8%) and specificity was 100%. The proportion of correctly classified individuals was 92.7%. For boys, the sensitivity was 94.6% (95% CI 88.1-99.8%) and specificity was 100%. The proportion of correctly classified individuals was 97.6%. The cut-off value of 0.08 is a useful method to assess if a subject is older than 18 years of age or not.

  1. [Accuracy of the oscillometric method to measure blood pressure in children

    PubMed

    Rego Filho, E A; Mello, S F; Silva, C R; Vituri, D W; Bazoni, E; Gordan, L N

    1999-01-01

    OBJECTIVE: The aim of this study is to analyze the substitution of the standard auscultatory method by the oscillometric blood pressure monitor, independently of the validity of the intraarterial blood pressure measurement. The accuracy of the automatic oscillometric monitor was compared to the auscultatory mercury manometer blood pressure measurement in apparently healthy school age children. METHODS: A device able to perform 3 simultaneous readings are used: one reading by the monitor and the others by two "blind" observers. We studied 72 school age children with the following characteristics: mean age 9.5 (6.1-16.1) and 39 males (54.2%). RESULTS: The difference for the systolic and diastolic blood pressure obtained by the monitor was in average + 6.2 mmHg and + 10.0 mmHg, respectively, when compared to the observer's readings. There was neither a good correlation nor a good agreement between the two observers and the monitor in the blood pressure determination. CONCLUSIONS: We concluded that the substitution of the standard auscultatory method for the non-invasive oscillometric method to measure blood pressure in school age children can not be generally recommended.

  2. Computer-aided alignment method of optical lens with high accuracy

    NASA Astrophysics Data System (ADS)

    Xing, Song; Hou, Xiao-hua; Zhang, Xue-min; Ji, Bin-dong

    2016-09-01

    With the development of space and aviation industry, the optical systems with high resolution and better imaging quality are required. According to the alignment technical process, the factors of every step which have big influence to the imaging quality are analyzed. It is detected that the micro-stress assembly of the optical unit and the high co-axial precision of the entire optical system are the two important factors which are supposed to determine how well the imaging quality of the optical system is; also the technical methods are discussed to ensure these two factors from the engineering view. The reflective interference testing method to measure the surface figure and the transitive interference testing method to measure the wave aberration of the optical unit are combined to ensure the micro-stress assembly of the optical unit, so it will not bring astigmatism to the whole system imaging quality. Optical alignment machining and precision alignment are combined to ensure the high co-axial precision of the optical system. An optical lens of high accuracy is assembled by using these methods; the final wave aberration of optical lens is 0.022λ.

  3. Comparison of Explicitly Correlated Methods for Computing High-Accuracy Benchmark Energies for Noncovalent Interactions.

    PubMed

    Sirianni, Dominic A; Burns, Lori A; Sherrill, C David

    2017-01-10

    The reliability of explicitly correlated methods for providing benchmark-quality noncovalent interaction energies was tested at various levels of theory and compared to estimates of the complete basis set (CBS) limit. For all systems of the A24 test set, computations were performed using both aug-cc-pVXZ (aXZ; X = D, T, Q, 5) basis sets and specialized cc-pVXZ-F12 (XZ-F12; X = D, T, Q, 5) basis sets paired with explicitly correlated coupled cluster singles and doubles [CCSD-F12n (n = a, b, c)] with triple excitations treated by the canonical perturbative method and scaled to compensate for their lack of explicit correlation [(T**)]. Results show that aXZ basis sets produce smaller errors versus the CBS limit than XZ-F12 basis sets. The F12b ansatz results in the lowest average errors for aTZ and larger basis sets, while F12a is best for double-ζ basis sets. When using aXZ basis sets (X ≥ 3), convergence is achieved from above for F12b and F12c ansatzë and from below for F12a. The CCSD(T**)-F12b/aXZ approach converges quicker with respect to basis than any other combination, although the performance of CCSD(T**)-F12c/aXZ is very similar. Both CCSD(T**)-F12b/aTZ and focal point schemes employing density-fitted, frozen natural orbital [DF-FNO] CCSD(T)/aTZ exhibit similar accuracy and computational cost, and both are much more computationally efficient than large-basis conventional CCSD(T) computations of similar accuracy.

  4. Diagnostic accuracy assessment of Sensititre and agar disk diffusion for determining antimicrobial resistance profiles of bovine clinical mastitis pathogens.

    PubMed

    Saini, V; Riekerink, R G M Olde; McClure, J T; Barkema, H W

    2011-04-01

    Determining the accuracy and precision of a measuring instrument is pertinent in antimicrobial susceptibility testing. This study was conducted to predict the diagnostic accuracy of the Sensititre MIC mastitis panel (Sensititre) and agar disk diffusion (ADD) method with reference to the manual broth microdilution test method for antimicrobial resistance profiling of Escherichia coli (n = 156), Staphylococcus aureus (n = 154), streptococcal (n = 116), and enterococcal (n = 31) bovine clinical mastitis isolates. The activities of ampicillin, ceftiofur, cephalothin, erythromycin, oxacillin, penicillin, the penicillin-novobiocin combination, pirlimycin, and tetracycline were tested against the isolates. Diagnostic accuracy was determined by estimating the area under the receiver operating characteristic curve; intertest essential and categorical agreements were determined as well. Sensititre and the ADD method demonstrated moderate to highly accurate (71 to 99%) and moderate to perfect (71 to 100%) predictive accuracies for 74 and 76% of the isolate-antimicrobial MIC combinations, respectively. However, the diagnostic accuracy was low for S. aureus-ceftiofur/oxacillin combinations and other streptococcus-ampicillin combinations by either testing method. Essential agreement between Sensititre automatic MIC readings and MIC readings obtained by the broth microdilution test method was 87%. Essential agreement between Sensititre automatic and manual MIC reading methods was 97%. Furthermore, the ADD test method and Sensititre MIC method exhibited 92 and 91% categorical agreement (sensitive, intermediate, resistant) of results, respectively, compared with the reference method. However, both methods demonstrated lower agreement for E. coli-ampicillin/cephalothin combinations than for Gram-positive isolates. In conclusion, the Sensititre and ADD methods had moderate to high diagnostic accuracy and very good essential and categorical agreement for most udder pathogen

  5. Assessment of Completeness and Positional Accuracy of Linear Features in Volunteered Geographic Information (vgi)

    NASA Astrophysics Data System (ADS)

    Eshghi, M.; Alesheikh, A. A.

    2015-12-01

    Recent advances in spatial data collection technologies and online services dramatically increase the contribution of ordinary people to produce, share, and use geographic information. Collecting spatial data as well as disseminating them on the internet by citizens has led to a huge source of spatial data termed as Volunteered Geographic Information (VGI) by Mike Goodchild. Although, VGI has produced previously unavailable data assets, and enriched existing ones. But its quality can be highly variable and challengeable. This presents several challenges to potential end users who are concerned about the validation and the quality assurance of the data which are collected. Almost, all the existing researches are based on how to find accurate VGI data from existing VGI data which consist of a) comparing the VGI data with the accurate official data, or b) in cases that there is no access to correct data; therefore, looking for an alternative way to determine the quality of VGI data is essential, and so forth. In this paper it has been attempt to develop a useful method to reach this goal. In this process, the positional accuracy of linear feature of Iran, Tehran OSM data have been analyzed.

  6. Accuracy assessment of building point clouds automatically generated from iphone images

    NASA Astrophysics Data System (ADS)

    Sirmacek, B.; Lindenbergh, R.

    2014-06-01

    Low-cost sensor generated 3D models can be useful for quick 3D urban model updating, yet the quality of the models is questionable. In this article, we evaluate the reliability of an automatic point cloud generation method using multi-view iPhone images or an iPhone video file as an input. We register such automatically generated point cloud on a TLS point cloud of the same object to discuss accuracy, advantages and limitations of the iPhone generated point clouds. For the chosen example showcase, we have classified 1.23% of the iPhone point cloud points as outliers, and calculated the mean of the point to point distances to the TLS point cloud as 0.11 m. Since a TLS point cloud might also include measurement errors and noise, we computed local noise values for the point clouds from both sources. Mean (μ) and standard deviation (σ) of roughness histograms are calculated as (μ1 = 0.44 m., σ1 = 0.071 m.) and (μ2 = 0.025 m., σ2 = 0.037 m.) for the iPhone and TLS point clouds respectively. Our experimental results indicate possible usage of the proposed automatic 3D model generation framework for 3D urban map updating, fusion and detail enhancing, quick and real-time change detection purposes. However, further insights should be obtained first on the circumstances that are needed to guarantee a successful point cloud generation from smartphone images.

  7. Improving accuracy in the MPM method using a null space filter

    NASA Astrophysics Data System (ADS)

    Gritton, Chris; Berzins, Martin

    2017-01-01

    The material point method (MPM) has been very successful in providing solutions to many challenging problems involving large deformations. Nevertheless there are some important issues that remain to be resolved with regard to its analysis. One key challenge applies to both MPM and particle-in-cell (PIC) methods and arises from the difference between the number of particles and the number of the nodal grid points to which the particles are mapped. This difference between the number of particles and the number of grid points gives rise to a non-trivial null space of the linear operator that maps particle values onto nodal grid point values. In other words, there are non-zero particle values that when mapped to the grid point nodes result in a zero value there. Moreover, when the nodal values at the grid points are mapped back to particles, part of those particle values may be in that same null space. Given positive mapping weights from particles to nodes such null space values are oscillatory in nature. While this problem has been observed almost since the beginning of PIC methods there are still elements of it that are problematical today as well as methods that transcend it. The null space may be viewed as being connected to the ringing instability identified by Brackbill for PIC methods. It will be shown that it is possible to remove these null space values from the solution using a null space filter. This filter improves the accuracy of the MPM methods using an approach that is based upon a local singular value decomposition (SVD) calculation. This local SVD approach is compared against the global SVD approach previously considered by the authors and to a recent MPM method by Zhang and colleagues.

  8. Springback Control in Industrial Bending Operations: Assessing the Accuracy of Three Commercial FEA Codes

    NASA Astrophysics Data System (ADS)

    Welo, Torgeir; Granly, Bjørg M.; Elverum, Christer; Søvik, Odd P.; Sørbø, Steinar

    2011-05-01

    Over the past two decades, a quantum leap has been made in FE technology for metal forming applications, including methods, algorithms, models and hardware capabilities. A myriad of research articles reports on methodologies that provide excellent capabilities in reproducing springback obtained from physical experiments. However, it is felt that we are not yet to the point where current modeling practice provides satisfactory value to tool designers and manufacturing engineers, particularly when the results have to be available before the first piece of tool steel has been cut; the main reasons being lack of accuracy in predicting elastic springback. The main objective of the present work is to validate springback capabilities using a strategy that integrates industrial tool simulation practice with carefully controlled physical experiments conducted in an academic setting. An industry-like (rotary) draw bending machine has been built and equipped with advanced measurement capabilities. Extruded rectangular, hollow aluminum alloy AA6060 sections were heat treated to two different tempers to produce a range of material properties prior to forming into two different bending angles. The selected set-up represents a challenging benchmark due to tight-radius bending and complex contact conditions, meaning that elastic springback is resulting from interaction effects between excessive local cross-sectional distortions and global bending mechanisms. The material properties were obtained by tensile testing, curve-fitting data to a conventional isotropic Ludwik-type material model. The bending process was modeled in three different commercial FE codes following best practice, including LS-Dyna, Stampack and Abaqus (explicit). The springback analyses were done prior to bending tests as would be done in an industrial tool design process. After having completed the bending tests and carefully measured the released bend angle for the different combinations, the results were

  9. A TECHNIQUE FOR ASSESSING THE ACCURACY OF SUB-PIXEL IMPERVIOUS SURFACE ESTIMATES DERIVED FROM LANDSAT TM IMAGERY

    EPA Science Inventory

    We developed a technique for assessing the accuracy of sub-pixel derived estimates of impervious surface extracted from LANDSAT TM imagery. We utilized spatially coincident
    sub-pixel derived impervious surface estimates, high-resolution planimetric GIS data, vector--to-
    r...

  10. Classification Accuracy of Oral Reading Fluency and Maze in Predicting Performance on Large-Scale Reading Assessments

    ERIC Educational Resources Information Center

    Decker, Dawn M.; Hixson, Michael D.; Shaw, Amber; Johnson, Gloria

    2014-01-01

    The purpose of this study was to examine whether using a multiple-measure framework yielded better classification accuracy than oral reading fluency (ORF) or maze alone in predicting pass/fail rates for middle-school students on a large-scale reading assessment. Participants were 178 students in Grades 7 and 8 from a Midwestern school district.…

  11. How Nonrecidivism Affects Predictive Accuracy: Evidence from a Cross-Validation of the Ontario Domestic Assault Risk Assessment (ODARA)

    ERIC Educational Resources Information Center

    Hilton, N. Zoe; Harris, Grant T.

    2009-01-01

    Prediction effect sizes such as ROC area are important for demonstrating a risk assessment's generalizability and utility. How a study defines recidivism might affect predictive accuracy. Nonrecidivism is problematic when predicting specialized violence (e.g., domestic violence). The present study cross-validates the ability of the Ontario…

  12. Disease severity assessment in epidemiological studies: accuracy and reliability of visual estimates of Septoria leaf blotch (SLB) in winter wheat

    Technology Transfer Automated Retrieval System (TEKTRAN)

    The accuracy and reliability of visual assessments of SLB severity by raters (i.e. one plant pathologist with extensive experience and three other raters trained prior to field observations using standard area diagrams and DISTRAIN) was determined by comparison with assumed actual values obtained by...

  13. Diagnostic Accuracy of Computer-Aided Assessment of Intranodal Vascularity in Distinguishing Different Causes of Cervical Lymphadenopathy.

    PubMed

    Ying, Michael; Cheng, Sammy C H; Ahuja, Anil T

    2016-08-01

    Ultrasound is useful in assessing cervical lymphadenopathy. Advancement of computer science technology allows accurate and reliable assessment of medical images. The aim of the study described here was to evaluate the diagnostic accuracy of computer-aided assessment of the intranodal vascularity index (VI) in differentiating the various common causes of cervical lymphadenopathy. Power Doppler sonograms of 347 patients (155 with metastasis, 23 with lymphoma, 44 with tuberculous lymphadenitis, 125 reactive) with palpable cervical lymph nodes were reviewed. Ultrasound images of cervical nodes were evaluated, and the intranodal VI was quantified using a customized computer program. The diagnostic accuracy of using the intranodal VI to distinguish different disease groups was evaluated and compared. Metastatic and lymphomatous lymph nodes tend to be more vascular than tuberculous and reactive lymph nodes. The intranodal VI had the highest diagnostic accuracy in distinguishing metastatic and tuberculous nodes with a sensitivity of 80%, specificity of 73%, positive predictive value of 91%, negative predictive value of 51% and overall accuracy of 68% when a cutoff VI of 22% was used. Computer-aided assessment provides an objective and quantitative way to evaluate intranodal vascularity. The intranodal VI is a useful parameter in distinguishing certain causes of cervical lymphadenopathy and is particularly useful in differentiating metastatic and tuberculous lymph nodes. However, it has limited value in distinguishing lymphomatous nodes from metastatic and reactive nodes.

  14. On the accuracy of indirect methods for estimating the sizes of asteroids

    NASA Astrophysics Data System (ADS)

    Morozhenko, A. V.; Vidmachenko, A. P.

    2013-09-01

    Using data on the maximum value of negative polarization degree is shown to be the least reliable in estimating asteroid albedos. To improve the reliability of the method based on the information relative to the positive polarization branch slope, the approximation coefficient values are bound to be chosen for a particular type of asteroids. The same situation is in the case of the thermal method where each type needs a corresponding value of the phase integral q. In addition, the accuracy of both methods increases if a corresponding type of characteristic phase dependence of brightness is used for the transition from A(0) to A(alpha), including the opposition effect. Our simulation within the frame work of the Irwin-Yanovitskij modification of the Hapke shadow model shows that the values of the phase coefficient beta for (10°

  15. Structural health monitoring ultrasonic thickness measurement accuracy and reliability of various time-of-flight calculation methods

    NASA Astrophysics Data System (ADS)

    Eason, Thomas J.; Bond, Leonard J.; Lozev, Mark G.

    2016-02-01

    The accuracy, precision, and reliability of ultrasonic thickness structural health monitoring systems are discussed in-cluding the influence of systematic and environmental factors. To quantify some of these factors, a compression wave ultrasonic thickness structural health monitoring experiment is conducted on a flat calibration block at ambient temperature with forty four thin-film sol-gel transducers and various time-of-flight thickness calculation methods. As an initial calibration, the voltage response signals from each sensor are used to determine the common material velocity as well as the signal offset unique to each calculation method. Next, the measurement precision of the thickness error of each method is determined with a proposed weighted censored relative maximum likelihood analysis technique incorporating the propagation of asymmetric measurement uncertainty. The results are presented as upper and lower confidence limits analogous to the a90/95 terminology used in industry recognized Probability-of-Detection assessments. Future work is proposed to apply the statistical analysis technique to quantify measurement precision of various thickness calculation methods under different environmental conditions such as high temperature, rough back-wall surface, and system degradation with an intended application to monitor naphthenic acid corrosion in oil refineries.

  16. Multinomial tree models for assessing the status of the reference in studies of the accuracy of tools for binary classification

    PubMed Central

    Botella, Juan; Huang, Huiling; Suero, Manuel

    2013-01-01

    Studies that evaluate the accuracy of binary classification tools are needed. Such studies provide 2 × 2 cross-classifications of test outcomes and the categories according to an unquestionable reference (or gold standard). However, sometimes a suboptimal reliability reference is employed. Several methods have been proposed to deal with studies where the observations are cross-classified with an imperfect reference. These methods require that the status of the reference, as a gold standard or as an imperfect reference, is known. In this paper a procedure for determining whether it is appropriate to maintain the assumption that the reference is a gold standard or an imperfect reference, is proposed. This procedure fits two nested multinomial tree models, and assesses and compares their absolute and incremental fit. Its implementation requires the availability of the results of several independent studies. These should be carried out using similar designs to provide frequencies of cross-classification between a test and the reference under investigation. The procedure is applied in two examples with real data. PMID:24106484

  17. Problems by Assessment of Accuracy at Processing Joint Rows of Digitized Astronomical Images Obtained with Different Instruments

    NASA Astrophysics Data System (ADS)

    Kazantseva, L.

    2016-09-01

    Current methods for determining the astrometry positions of digitized photographic plates have difficulty for correctly using data from different telescopes. Factors such as scale of a plate, exposure duration, size of the plate and its quality, affect the accuracy of the results. Scanner options add their mistakes. Often, it is important to get results in prolonged time period. In such cases, we use observational data that were obtained using different telescopes. Then there is the need to balance the various data accuracy. The report provides a methodical approach. The method has been tested for some standard star fields.

  18. Accuracy of Conventional Diagnostic Methods for Identifying Structural Changes in Patients with Focal Epilepsy

    PubMed Central

    Dakaj, Nazim; Kruja, Jera; Jashari, Fisnik; Boshnjaku, Dren; Shatri, Nexhat; Zeqiraj, Kamber

    2016-01-01

    Background: Epilepsy is a neurological disorder characterized by abnormal firing of nerve impulses in the brain. Aim: This study aims to investigate the frequency of appearance of pathological changes in conventional examination methods (electroencephalography–EEG, brain computerized tomography -CT or brain magnetic resonance imaging – MRI) in patients with epilepsy, and relationship between clinical manifestations and localization of changes in CT or MRI. Methods: In this study we have included 110 patients with focal epilepsy who fulfilled the inclusion criteria out of 557 initially diagnosed patients. Detailed clinical examination together with brain imaging (CT and MRI) and electroencephalography examination was performed. We have evaluated the accuracy of each diagnostic method to localize the epileptic focus. Diagnosis of epilepsy was determined by the ILAE (International League Against Epilepsy) criteria of the year 1989, and classification of epileptic seizures was made according to the ILAE classification 2010. Results: Electroencephalography presented changes in 60.9% of patients; brain CT in 42.1%, and MRI in 78% of the patients. The results of our study showed that clinical manifestations were not always conveyed with pathological changes in conventional examining methods performed. Of the total of 79 patients with changes in imaging (8 with changes in CT and 71 in MRI), 79.7% presented a clinical picture compatible with the region in which morphological changes were found, while in 20.3% of patients the presented morphological changes were not aligned with the clinical picture. Conclusion: In patients with epilepsy, conventional examination methods do not always find pathological changes, while clinical manifestations of epilepsy did not always coincide with the location of changes in imaging. Further studies are needed to see if there is clear border between focal and generalized epilepsy. PMID:28077892

  19. Accuracy assessment of single and double difference models for the single epoch GPS compass

    NASA Astrophysics Data System (ADS)

    Chen, Wantong; Qin, Honglei; Zhang, Yanzhong; Jin, Tian

    2012-02-01

    The single epoch GPS compass is an important field of study, since it is a valuable technique for the orientation estimation of vehicles and it can guarantee a total independence from carrier phase slips in practical applications. To achieve highly accurate angular estimates, the unknown integer ambiguities of the carrier phase observables need to be resolved. Past researches focus on the ambiguity resolution for single epoch; however, accuracy is another significant problem for many challenging applications. In this contribution, the accuracy is evaluated for the non-common clock scheme of the receivers and the common clock scheme of the receivers, respectively. We focus on three scenarios for either scheme: single difference model vs. double difference model, single frequency model vs. multiple frequency model and optimal linear combinations vs. traditional triple-frequency least squares. We deduce the short baseline precision for a number of different available models and analyze the difference in accuracy for those models. Compared with the single or double difference model of the non-common clock scheme, the single difference model of the common clock scheme can greatly reduce the vertical component error of baseline vector, which results in higher elevation accuracy. The least squares estimator can also reduce the error of fixed baseline vector with the aid of the multi-frequency observation, thereby improving the attitude accuracy. In essence, the "accuracy improvement" is attributed to the difference in accuracy for different models, not a real improvement for any specific model. If all noise levels of GPS triple frequency carrier phase are assumed the same in unit of cycles, it can be proved that the optimal linear combination approach is equivalent to the traditional triple-frequency least squares, no matter which scheme is utilized. Both simulations and actual experiments have been performed to verify the correctness of theoretical analysis.

  20. New mobile methods for dietary assessment: review of image-assisted and image-based dietary assessment methods.

    PubMed

    Boushey, C J; Spoden, M; Zhu, F M; Delp, E J; Kerr, D A

    2016-12-12

    For nutrition practitioners and researchers, assessing dietary intake of children and adults with a high level of accuracy continues to be a challenge. Developments in mobile technologies have created a role for images in the assessment of dietary intake. The objective of this review was to examine peer-reviewed published papers covering development, evaluation and/or validation of image-assisted or image-based dietary assessment methods from December 2013 to January 2016. Images taken with handheld devices or wearable cameras have been used to assist traditional dietary assessment methods for portion size estimations made by dietitians (image-assisted methods). Image-assisted approaches can supplement either dietary records or 24-h dietary recalls. In recent years, image-based approaches integrating application technology for mobile devices have been developed (image-based methods). Image-based approaches aim at capturing all eating occasions by images as the primary record of dietary intake, and therefore follow the methodology of food records. The present paper reviews several image-assisted and image-based methods, their benefits and challenges; followed by details on an image-based mobile food record. Mobile technology offers a wide range of feasible options for dietary assessment, which are easier to incorporate into daily routines. The presented studies illustrate that image-assisted methods can improve the accuracy of conventional dietary assessment methods by adding eating occasion detail via pictures captured by an individual (dynamic images). All of the studies reduced underreporting with the help of images compared with results with traditional assessment methods. Studies with larger sample sizes are needed to better delineate attributes with regards to age of user, degree of error and cost.

  1. A priori evaluation of two-stage cluster sampling for accuracy assessment of large-area land-cover maps

    USGS Publications Warehouse

    Wickham, J.D.; Stehman, S.V.; Smith, J.H.; Wade, T.G.; Yang, L.

    2004-01-01

    Two-stage cluster sampling reduces the cost of collecting accuracy assessment reference data by constraining sample elements to fall within a limited number of geographic domains (clusters). However, because classification error is typically positively spatially correlated, within-cluster correlation may reduce the precision of the accuracy estimates. The detailed population information to quantify a priori the effect of within-cluster correlation on precision is typically unavailable. Consequently, a convenient, practical approach to evaluate the likely performance of a two-stage cluster sample is needed. We describe such an a priori evaluation protocol focusing on the spatial distribution of the sample by land-cover class across different cluster sizes and costs of different sampling options, including options not imposing clustering. This protocol also assesses the two-stage design's adequacy for estimating the precision of accuracy estimates for rare land-cover classes. We illustrate the approach using two large-area, regional accuracy assessments from the National Land-Cover Data (NLCD), and describe how the a priorievaluation was used as a decision-making tool when implementing the NLCD design.

  2. Multispectral image compression methods for improvement of both colorimetric and spectral accuracy

    NASA Astrophysics Data System (ADS)

    Liang, Wei; Zeng, Ping; Xiao, Zhaolin; Xie, Kun

    2016-07-01

    We propose that both colorimetric and spectral distortion in compressed multispectral images can be reduced by a composite model, named OLCP(W)-X (OptimalLeaders_Color clustering-PCA-W weighted-X coding). In the model, first the spectral-colorimetric clustering is designed for sparse equivalent representation by generating spatial basis. Principal component analysis (PCA) is subsequently used in the manipulation of spatial basis for spectral redundancy removal. Then error compensation mechanism is presented to produce predicted difference image, and finally combined with visual characteristic matrix W, and the created image is compressed by traditional multispectral image coding schemes. We introduce four model-based algorithms to explain their validity. The first two algorithms are OLCPWKWS (OLC-PCA-W-KLT-WT-SPIHT) and OLCPKWS, in which Karhunen-Loeve transform, wavelet transform, and set partitioning in hierarchical trees coding are applied for the created image compression. And the latter two methods are OLCPW-JPEG2000-MCT and OLCP-JPEG2000-MCT. Experimental results show that, compared with the corresponding traditional coding, the proposed OLCPW-X schemes can significantly improve the colorimetric accuracy of rebuilding images under various illumination conditions and generally achieve satisfactory peak signal-to-noise ratio under the same compression ratio. And OLCP-X methods could always ensure superior spectrum reconstruction. Furthermore, our model has excellent performance on user interaction.

  3. Improved Accuracy of the Inherent Shrinkage Method for Fast and More Reliable Welding Distortion Calculations

    NASA Astrophysics Data System (ADS)

    Mendizabal, A.; González-Díaz, J. B.; San Sebastián, M.; Echeverría, A.

    2016-07-01

    This paper describes the implementation of a simple strategy adopted for the inherent shrinkage method (ISM) to predict welding-induced distortion. This strategy not only makes it possible for the ISM to reach accuracy levels similar to the detailed transient analysis method (considered the most reliable technique for calculating welding distortion) but also significantly reduces the time required for these types of calculations. This strategy is based on the sequential activation of welding blocks to account for welding direction and transient movement of the heat source. As a result, a significant improvement in distortion prediction is achieved. This is demonstrated by experimentally measuring and numerically analyzing distortions in two case studies: a vane segment subassembly of an aero-engine, represented with 3D-solid elements, and a car body component, represented with 3D-shell elements. The proposed strategy proves to be a good alternative for quickly estimating the correct behaviors of large welded components and may have important practical applications in the manufacturing industry.

  4. A Method to Improve the Accuracy of Particle Diameter Measurements from Shadowgraph Images

    NASA Astrophysics Data System (ADS)

    Erinin, Martin A.; Wang, Dan; Liu, Xinan; Duncan, James H.

    2015-11-01

    A method to improve the accuracy of the measurement of the diameter of particles using shadowgraph images is discussed. To obtain data for analysis, a transparent glass calibration reticle, marked with black circular dots of known diameters, is imaged with a high-resolution digital camera using backlighting separately from both a collimated laser beam and diffuse white light. The diameter and intensity of each dot is measured by fitting an inverse hyperbolic tangent function to the particle image intensity map. Using these calibration measurements, a relationship between the apparent diameter and intensity of the dot and its actual diameter and position relative to the focal plane of the lens is determined. It is found that the intensity decreases and apparent diameter increases/decreases (for collimated/diffuse light) with increasing distance from the focal plane. Using the relationships between the measured properties of each dot and its actual size and position, an experimental calibration method has been developed to increase the particle-diameter-dependent range of distances from the focal plane for which accurate particle diameter measurements can be made. The support of the National Science Foundation under grant OCE0751853 from the Division of Ocean Sciences is gratefully acknowledged.

  5. High-accuracy measurement of low-water-content in liquid using NIR spectral absorption method

    NASA Astrophysics Data System (ADS)

    Peng, Bao-Jin; Wan, Xu; Jin, Hong-Zhen; Zhao, Yong; Mao, He-Fa

    2005-01-01

    Water content measurement technologies are very important for quality inspection of food, medicine products, chemical products and many other industry fields. In recent years, requests for accurate low-water-content measurement in liquid are more and more exigent, and great interests have been shown from the research and experimental work. With the development and advancement of modern production and control technologies, more accurate water content technology is needed. In this paper, a novel experimental setup based on near-infrared (NIR) spectral technology and fiber-optic sensor (OFS) is presented. It has a good measurement accuracy about -/+ 0.01%, which is better, to our knowledge, than most other methods published until now. It has a high measurement resolution of 0.001% in the measurement range from zero to 0.05% for water-in-alcohol measurement, and the water-in-oil measurement is carried out as well. In addition, the advantages of this method also include pollution-free to the measured liquid, fast measurement and so on.

  6. Accuracy of various methods of localization of the orifice of the coronary sinus at electrophysiologic study.

    PubMed

    Davis, L M; Byth, K; Lau, K C; Uther, J B; Richards, D A; Ross, D L

    1992-08-01

    The coronary sinus (CS) orifice is an important reference point for determining electrode and, thereby, accessory pathway location at electrophysiologic study. The reliability of fluoroscopic landmarks used to identify the CS orifice is not known. This study compared the accuracy of several fluoroscopic landmarks for identifying the CS orifice with the location defined by radiopaque contrast injection of the CS. Forty patients were studied. Radiographic markers of the CS orifice that were examined included: (1) the point at which the CS catheter prolapsed during advancement, (2) the point of maximum convexity of the CS catheter when a superior vena caval approach was used, (3) the right side of the ventricular septum, and (4) the relation to the underlying vertebrae. The least-significant difference method of multiple comparisons was used for statistical analysis. The point at which the CS catheter prolapsed was the most accurate noncontrast method for determining the location of the CS orifice (p less than 0.05), but was possible without the use of excessive force in only 48% of patients. The point of catheter prolapse was a median of 1 mm (range 0 to 11) from the true location of the os. Errors with other examined landmarks ranged up to 3 cm. Identification of the CS orifice is best performed by radiopaque contrast injection. The point of prolapse during catheter advancement in the CS is an accurate alternative when contrast injection is not feasible. Other noncontrast fluoroscopic landmarks are less reliable and are best avoided.(ABSTRACT TRUNCATED AT 250 WORDS)

  7. Multinational assessment of accuracy of equations for predicting risk of kidney failure: a meta-analysis

    PubMed Central

    Tangri, Navdeep; Grams, Morgan E.; Levey, Andrew S.; Coresh, Josef; Appel, Lawrence; Astor, Brad C.; Chodick, Gabriel; Collins, Allan J.; Djurdjev, Ognjenka; Elley, C. Raina; Evans, Marie; Garg, Amit X.; Hallan, Stein I.; Inker, Lesley; Ito, Sadayoshi; Jee, Sun Ha; Kovesdy, Csaba P.; Kronenberg, Florian; Lambers Heerspink, Hiddo J.; Marks, Angharad; Nadkarni, Girish N.; Navaneethan, Sankar D.; Nelson, Robert G.; Titze, Stephanie; Sarnak, Mark J.; Stengel, Benedicte; Woodward, Mark; Iseki, Kunitoshi

    2016-01-01

    Importance Identifying patients at risk of chronic kidney disease (CKD) progression may facilitate more optimal nephrology care. Kidney failure risk equations (KFREs) were previously developed and validated in two Canadian cohorts. Validation in other regions and in CKD populations not under the care of a nephrologist is needed. Objective To evaluate the accuracy of the KFREs across different geographic regions and patient populations through individual-participant data meta-analysis. Data Sources Thirty-one cohorts, including 721,357 participants with CKD Stages 3–5 in over 30 countries spanning 4 continents, were studied. These cohorts collected data from 1982 through 2014. Study Selection Cohorts participating in the CKD Prognosis Consortium with data on end-stage renal disease. Data Extraction and Synthesis Data were obtained and statistical analyses were performed between July 2012 and June 2015. Using the risk factors from the original KFREs, cohort-specific hazard ratios were estimated, and combined in meta-analysis to form new “pooled” KFREs. Original and pooled equation performance was compared, and the need for regional calibration factors was assessed. Main Outcome and Measure Kidney failure (treatment by dialysis or kidney transplantation). Results During a median follow-up of 4 years, 23,829 cases of kidney failure were observed. The original KFREs achieved excellent discrimination (ability to differentiate those who developed kidney failure from those who did not) across all cohorts (overall C statistic, 0.90 (95% CI 0.89–0.92) at 2 years and 0.88 (95% CI 0.86–0.90) at 5 years); discrimination in subgroups by age, race, and diabetes status was similar. There was no improvement with the pooled equations. Calibration (the difference between observed and predicted risk) was adequate in North American cohorts, but the original KFREs overestimated risk in some non-North American cohorts. Addition of a calibration factor that lowered the baseline

  8. Accuracy Assessment of Underwater Photogrammetric Three Dimensional Modelling for Coral Reefs

    NASA Astrophysics Data System (ADS)

    Guo, T.; Capra, A.; Troyer, M.; Gruen, A.; Brooks, A. J.; Hench, J. L.; Schmitt, R. J.; Holbrook, S. J.; Dubbini, M.

    2016-06-01

    Recent advances in automation of photogrammetric 3D modelling software packages have stimulated interest in reconstructing highly accurate 3D object geometry in unconventional environments such as underwater utilizing simple and low-cost camera systems. The accuracy of underwater 3D modelling is affected by more parameters than in single media cases. This study is part of a larger project on 3D measurements of temporal change of coral cover in tropical waters. It compares the accuracies of 3D point clouds generated by using images acquired from a system camera mounted in an underwater housing and the popular GoPro cameras respectively. A precisely measured calibration frame was placed in the target scene in order to provide accurate control information and also quantify the errors of the modelling procedure. In addition, several objects (cinder blocks) with various shapes were arranged in the air and underwater and 3D point clouds were generated by automated image matching. These were further used to examine the relative accuracy of the point cloud generation by comparing the point clouds of the individual objects with the objects measured by the system camera in air (the best possible values). Given a working distance of about 1.5 m, the GoPro camera can achieve a relative accuracy of 1.3 mm in air and 2.0 mm in water. The system camera achieved an accuracy of 1.8 mm in water, which meets our requirements for coral measurement in this system.

  9. Effects of tangential-type boundary condition discontinuities on the accuracy of the lattice Boltzmann method for heat and mass transfer

    NASA Astrophysics Data System (ADS)

    Li, Like; AuYeung, Nick; Mei, Renwei; Klausner, James F.

    2016-08-01

    We present a systematic study on the effects of tangential-type boundary condition discontinuities on the accuracy of the lattice Boltzmann equation (LBE) method for Dirichlet and Neumann problems in heat and mass transfer modeling. The second-order accurate boundary condition treatments for continuous Dirichlet and Neumann problems are directly implemented for the corresponding discontinuous boundary conditions. Results from three numerical tests, including both straight and curved boundaries, are presented to show the accuracy and order of convergence of the LBE computations. Detailed error assessments are conducted for the interior temperature or concentration (denoted as a scalar ϕ) and the interior derivatives of ϕ for both types of boundary conditions, for the boundary flux in the Dirichlet problem and for the boundary ϕ values in the Neumann problem. When the discontinuity point on the straight boundary is placed at the center of the unit lattice in the Dirichlet problem, it yields only first-order accuracy for the interior distribution of ϕ, first-order accuracy for the boundary flux, and zeroth-order accuracy for the interior derivatives compared with the second-order accuracy of all quantities of interest for continuous boundary conditions. On the lattice scale, the LBE solution for the interior derivatives near the singularity is largely independent of the resolution and correspondingly the local distribution of the absolute errors is almost invariant with the changing resolution. For Neumann problems, when the discontinuity is placed at the lattice center, second-order accuracy is preserved for the interior distribution of ϕ; and a "superlinear" convergence order of 1.5 for the boundary ϕ values and first-order accuracy for the interior derivatives are obtained. For straight boundaries with the discontinuity point arbitrarily placed within the lattice and curved boundaries, the boundary flux becomes zeroth-order accurate for Dirichlet problems

  10. Accuracy of qualitative analysis for assessment of skilled baseball pitching technique.

    PubMed

    Nicholls, Rochelle; Fleisig, Glenn; Elliott, Bruce; Lyman, Stephen; Osinski, Edmund

    2003-07-01

    Baseball pitching must be performed with correct technique if injuries are to be avoided and performance maximized. High-speed video analysis is accepted as the most accurate and objective method for evaluation of baseball pitching mechanics. The aim of this research was to develop an equivalent qualitative analysis method for use with standard video equipment. A qualitative analysis protocol (QAP) was developed for 24 kinematic variables identified as important to pitching performance. Twenty male baseball pitchers were videotaped using 60 Hz camcorders, and their technique evaluated using the QAP, by two independent raters. Each pitcher was also assessed using a 6-camera 200 Hz Motion Analysis system (MAS). Four QAP variables (22%) showed significant similarity with MAS results. Inter-rater reliability showed agreement on 33% of QAP variables. It was concluded that a complete and accurate profile of an athlete's pitching mechanics cannot be made using the QAP in its current form, but it is possible such simple forms of biomechanical analysis could yield accurate results before 3-D methods become obligatory.

  11. ChronRater: A simple approach to assessing the accuracy of age models from Holocene sediment cores

    NASA Astrophysics Data System (ADS)

    Kaufman, D. S.; Balascio, N. L.; McKay, N. P.; Sundqvist, H. S.

    2013-12-01

    We have assembled a database of previously published Holocene proxy climate records from the Arctic, with the goal of reconstructing the spatial-temporal pattern of past climate changes. The focus is on well-dated, highly resolved, continuous records that extend to at least 6 ka, most of which (90%) are from sedimentary sequences sampled in cores from lakes and oceans. The database includes the original geochronological data (radiocarbon ages) for each record so that the accuracy of the underlying age models can be assessed uniformly. Determining the accuracy of age control for sedimentary sequences is difficult because it depends on many factors, some of which are difficult to quantify. Nevertheless, the geochronological accuracy of each time series in the database must be assessed systematically to objectively identify those that are appropriate to address a particular level of temporal inquiry. We have therefore devised a scoring scheme to rate the accuracy of age models that focuses on the most important factors and uses just the most commonly published information to determine the overall geochronological accuracy. The algorithm, "ChronRater" is written in the open-source statistical package, R. It relies on three characteristics of dated materials and their downcore trends: (1) The delineation of the downcore trend, which is quantified based on three attributes, namely: (a) the frequency of ages, (b) the regularity of their spacing, and (c) the uniformity of the sedimentation rate. (2) The quality of the dated materials, as determined by: (a) the proportion of outliers and downcore reversals, and (b) the type of materials analyzed and the extent to which their ages are verified by independent information as judged by a five-point scale for the entire sequence of ages. And (3) the overall uncertainty in the calibrated ages, which includes the analytical precision and the associated calibrated age ranges. Although our geochronological accuracy score is

  12. Accuracy of Two Motor Assessments during the First Year of Life in Preterm Infants for Predicting Motor Outcome at Preschool Age

    PubMed Central

    Spittle, Alicia J.; Lee, Katherine J.; Spencer-Smith, Megan; Lorefice, Lucy E.; Anderson, Peter J.; Doyle, Lex W.

    2015-01-01

    Aim The primary aim of this study was to investigate the accuracy of the Alberta Infant Motor Scale (AIMS) and Neuro-Sensory Motor Developmental Assessment (NSMDA) over the first year of life for predicting motor impairment at 4 years in preterm children. The secondary aims were to assess the predictive value of serial assessments over the first year and when using a combination of these two assessment tools in follow-up. Method Children born <30 weeks’ gestation were prospectively recruited and assessed at 4, 8 and 12 months’ corrected age using the AIMS and NSMDA. At 4 years’ corrected age children were assessed for cerebral palsy (CP) and motor impairment using the Movement Assessment Battery for Children 2nd-edition (MABC-2). We calculated accuracy of the AIMS and NSMDA for predicting CP and MABC-2 scores ≤15th (at-risk of motor difficulty) and ≤5th centile (significant motor difficulty) for each test (AIMS and NSMDA) at 4, 8 and 12 months, for delay on one, two or all three of the time points over the first year, and finally for delay on both tests at each time point. Results Accuracy for predicting motor impairment was good for each test at each age, although false positives were common. Motor impairment on the MABC-2 (scores ≤5th and ≤15th) was most accurately predicted by the AIMS at 4 months, whereas CP was most accurately predicted by the NSMDA at 12 months. In regards to serial assessments, the likelihood ratio for motor impairment increased with the number of delayed assessments. When combining both the NSMDA and AIMS the best accuracy was achieved at 4 months, although results were similar at 8 and 12 months. Interpretation Motor development during the first year of life in preterm infants assessed with the AIMS and NSMDA is predictive of later motor impairment at preschool age. However, false positives are common and therefore it is beneficial to follow-up children at high risk of motor impairment at more than one time point, or to use a

  13. A new method for the accuracy evaluation of a manufactured piece

    NASA Astrophysics Data System (ADS)

    Oniga, E. V.; Cardei, M.

    2015-11-01

    To evaluate the accuracy of a manufactured piece, it must be measured and compared with a reference model, namely the designed 3D model, based on geometrical elements. In this paper a new method for the precision evaluation of a manufactured piece is proposed, which implies the creation of the piece digital 3D model based on digital images and its transformation into a 3D mesh surface. The differences between the two models, the designed model and the new created one, are calculated using the Hausdorff distance. The aim of this research is to determine the differences between two 3D models, especially CAD models, with high precision, in a completely automated way. To obtain the results, a small piece has been photographed with a digital camera, that was calibrated using a 3D calibration object, a target consisting of a number of 42 points, 36 placed in the corners of 9 wood cubes with different heights and 6 of them placed at the middle of the distance between the cubes, on a board. This target was previously tested, the tests showing that using this calibration target instead of a 2D calibration grid, the precision of the final 3D model is improved with approximatly 50%. The 3D model of the manufactured piece was created using two methods. First, based on digital images, a point cloud was automatically generated and after the filtering process, the remaining points were interpolated, obtaining the piece 3D model as a mesh surface. Second, the piece 3D model was created using also the digital images, based on its characteristic points, resulting a CAD model, that was transformed into a mesh surface. Finally, the two 3D models were compared with the designed model, using the CloudCompare software, thus resulting the imperfections of the manufactured piece. The proposed method highlights the differences between the two models using a color palette, offering at the same time a global comparison.

  14. First-principle modelling of forsterite surface properties: Accuracy of methods and basis sets.

    PubMed

    Demichelis, Raffaella; Bruno, Marco; Massaro, Francesco R; Prencipe, Mauro; De La Pierre, Marco; Nestola, Fabrizio

    2015-07-15

    The seven main crystal surfaces of forsterite (Mg2 SiO4 ) were modeled using various Gaussian-type basis sets, and several formulations for the exchange-correlation functional within the density functional theory (DFT). The recently developed pob-TZVP basis set provides the best results for all properties that are strongly dependent on the accuracy of the wavefunction. Convergence on the structure and on the basis set superposition error-corrected surface energy can be reached also with poorer basis sets. The effect of adopting different DFT functionals was assessed. All functionals give the same stability order for the various surfaces. Surfaces do not exhibit any major structural differences when optimized with different functionals, except for higher energy orientations where major rearrangements occur around the Mg sites at the surface or subsurface. When dispersions are not accounted for, all functionals provide similar surface energies. The inclusion of empirical dispersions raises the energy of all surfaces by a nearly systematic value proportional to the scaling factor s of the dispersion formulation. An estimation for the surface energy is provided through adopting C6 coefficients that are more suitable than the standard ones to describe O-O interactions in minerals. A 2 × 2 supercell of the most stable surface (010) was optimized. No surface reconstruction was observed. The resulting structure and surface energy show no difference with respect to those obtained when using the primitive cell. This result validates the (010) surface model here adopted, that will serve as a reference for future studies on adsorption and reactivity of water and carbon dioxide at this interface.

  15. Extended canonical Monte Carlo methods: Improving accuracy of microcanonical calculations using a reweighting technique

    NASA Astrophysics Data System (ADS)

    Velazquez, L.; Castro-Palacio, J. C.

    2015-03-01

    Velazquez and Curilef [J. Stat. Mech. (2010) P02002, 10.1088/1742-5468/2010/02/P02002; J. Stat. Mech. (2010) P04026, 10.1088/1742-5468/2010/04/P04026] have proposed a methodology to extend Monte Carlo algorithms that are based on canonical ensemble. According to our previous study, their proposal allows us to overcome slow sampling problems in systems that undergo any type of temperature-driven phase transition. After a comprehensive review about ideas and connections of this framework, we discuss the application of a reweighting technique to improve the accuracy of microcanonical calculations, specifically, the well-known multihistograms method of Ferrenberg and Swendsen [Phys. Rev. Lett. 63, 1195 (1989), 10.1103/PhysRevLett.63.1195]. As an example of application, we reconsider the study of the four-state Potts model on the square lattice L ×L with periodic boundary conditions. This analysis allows us to detect the existence of a very small latent heat per site qL during the occurrence of temperature-driven phase transition of this model, whose size dependence seems to follow a power law qL(L ) ∝(1/L ) z with exponent z ≃0 .26 ±0 .02. Discussed is the compatibility of these results with the continuous character of temperature-driven phase transition when L →+∞ .

  16. Accuracy analysis of the Null-Screen method for the evaluation of flat heliostats

    NASA Astrophysics Data System (ADS)

    Cebrian-Xochihuila, P.; Huerta-Carranza, O.; Díaz-Uribe, R.

    2016-04-01

    In this work we develop an algorithm to determinate the accuracy of the Null-Screen Method, used for the testing of flat heliostats used as solar concentrators in a central tower configuration. We simulate the image obtained on a CCD camera when an orderly distribution of points are displayed on a Null-Screen perpendicular to the heliostat under test. The deformations present in the heliostat are represented as a cosine function of the position with different periods and amplitudes. As a resolution criterion, a deformation on the mirror can be detected when the differences in position between the spots on the image plane for the deformed surface as compared with those obtained for an ideally flat heliostat are equal to one pixel. For 6.4μm pixel size and 18mm focal length, the minimum deformation we can measure in the heliostat, correspond to amplitude equal a 122μm for a period equal to 1m; this is equivalent to 0.8mrad in slope. This result depends on the particular configuration used during the test and the size of the heliostat.

  17. Understanding the accuracy of parental perceptions of child physical activity: a mixed methods analysis

    PubMed Central

    Kesten, Joanna M.; Jago, Russell; Sebire, Simon J.; Edwards, Mark J.; Pool, Laura; Zahra, Jesmond; Thompson, Janice L.

    2016-01-01

    Background Interventions to increase children’s physical activity (PA) have achieved limited success. This may be attributed to inaccurate parental perceptions of their children’s PA and a lack of recognition of a need to change activity levels. Methods Fifty-three parents participated in semi-structured interviews to determine perceptions of child PA. Perceptions were compared to children’s measured MVPA (classified as meeting or not meeting UK guidelines) to produce three categories: “accurate”, “over-estimate”, “under-estimate”. Deductive content analysis was performed to understand the accuracy of parental perceptions. Results All parents of children meeting the PA guidelines accurately perceived their child’s PA; whilst the majority of parents whose child did not meet the guidelines overestimated their PA. Most parents were unconcerned about their child’s PA level, viewing them as naturally active and willing to be active. Qualitative explanations for perceptions of insufficient activity included children having health problems and preferences for inactive pursuits, and parents having difficulty facilitating PA in poor weather and not always observing their child’s PA level. Social comparisons also influenced parental perceptions. Conclusions Strategies to improve parental awareness of child PA are needed. Perceptions of child PA may be informed by child “busyness”, being unaware of activity levels, and social comparisons. PMID:25872227

  18. Accuracy of forced oscillation technique to assess lung function in geriatric COPD population

    PubMed Central

    Tse, Hoi Nam; Tseng, Cee Zhung Steven; Wong, King Ying; Yee, Kwok Sang; Ng, Lai Yun

    2016-01-01

    Introduction Performing lung function test in geriatric patients has never been an easy task. With well-established evidence indicating impaired small airway function and air trapping in patients with geriatric COPD, utilizing forced oscillation technique (FOT) as a supplementary tool may aid in the assessment of lung function in this population. Aims To study the use of FOT in the assessment of airflow limitation and air trapping in geriatric COPD patients. Study design A cross-sectional study in a public hospital in Hong Kong. ClinicalTrials.gov ID: NCT01553812. Methods Geriatric patients who had spirometry-diagnosed COPD were recruited, with both FOT and plethysmography performed. “Resistance” and “reactance” FOT parameters were compared to plethysmography for the assessment of air trapping and airflow limitation. Results In total, 158 COPD subjects with a mean age of 71.9±0.7 years and percentage of forced expiratory volume in 1 second of 53.4±1.7 L were recruited. FOT values had a good correlation (r=0.4–0.7) to spirometric data. In general, X values (reactance) were better than R values (resistance), showing a higher correlation with spirometric data in airflow limitation (r=0.07–0.49 vs 0.61–0.67), small airway (r=0.05–0.48 vs 0.56–0.65), and lung volume (r=0.12–0.29 vs 0.43–0.49). In addition, resonance frequency (Fres) and frequency dependence (FDep) could well identify the severe type (percentage of forced expiratory volume in 1 second <50%) of COPD with high sensitivity (0.76, 0.71) and specificity (0.72, 0.64) (area under the curve: 0.8 and 0.77, respectively). Moreover, X values could stratify different severities of air trapping, while R values could not. Conclusion FOT may act as a simple and accurate tool in the assessment of severity of airflow limitation, small and central airway function, and air trapping in patients with geriatric COPD who have difficulties performing conventional lung function test. Moreover, reactance

  19. Pharmacokinetic digital phantoms for accuracy assessment of image-based dosimetry in (177)Lu-DOTATATE peptide receptor radionuclide therapy.

    PubMed

    Brolin, Gustav; Gustafsson, Johan; Ljungberg, Michael; Gleisner, Katarina Sjögreen

    2015-08-07

    Patient-specific image-based dosimetry is considered to be a useful tool to limit toxicity associated with peptide receptor radionuclide therapy (PRRT). To facilitate the establishment and reliability of absorbed-dose response relationships, it is essential to assess the accuracy of dosimetry in clinically realistic scenarios. To this end, we developed pharmacokinetic digital phantoms corresponding to patients treated with (177)Lu-DOTATATE. Three individual voxel phantoms from the XCAT population were generated and assigned a dynamic activity distribution based on a compartment model for (177)Lu-DOTATATE, designed specifically for this purpose. The compartment model was fitted to time-activity data from 10 patients, primarily acquired using quantitative scintillation camera imaging. S values for all phantom source-target combinations were calculated based on Monte-Carlo simulations. Combining the S values and time-activity curves, reference values of the absorbed dose to the phantom kidneys, liver, spleen, tumours and whole-body were calculated. The phantoms were used in a virtual dosimetry study, using Monte-Carlo simulated gamma-camera images and conventional methods for absorbed-dose calculations. The characteristics of the SPECT and WB planar images were found to well represent those of real patient images, capturing the difficulties present in image-based dosimetry. The phantoms are expected to be useful for further studies and optimisation of clinical dosimetry in (177)Lu PRRT.

  20. Accuracy of a Low-Cost Novel Computer-Vision Dynamic Movement Assessment: Potential Limitations and Future Directions

    NASA Astrophysics Data System (ADS)

    McGroarty, M.; Giblin, S.; Meldrum, D.; Wetterling, F.

    2016-04-01

    The aim of the study was to perform a preliminary validation of a low cost markerless motion capture system (CAPTURE) against an industry gold standard (Vicon). Measurements of knee valgus and flexion during the performance of a countermovement jump (CMJ) between CAPTURE and Vicon were compared. After correction algorithms were applied to the raw CAPTURE data acceptable levels of accuracy and precision were achieved. The knee flexion angle measured for three trials using Capture deviated by -3.8° ± 3° (left) and 1.7° ± 2.8° (right) compared to Vicon. The findings suggest that low-cost markerless motion capture has potential to provide an objective method for assessing lower limb jump and landing mechanics in an applied sports setting. Furthermore, the outcome of the study warrants the need for future research to examine more fully the potential implications of the use of low-cost markerless motion capture in the evaluation of dynamic movement for injury prevention.

  1. Pharmacokinetic digital phantoms for accuracy assessment of image-based dosimetry in 177Lu-DOTATATE peptide receptor radionuclide therapy

    NASA Astrophysics Data System (ADS)

    Brolin, Gustav; Gustafsson, Johan; Ljungberg, Michael; Sjögreen Gleisner, Katarina

    2015-08-01

    Patient-specific image-based dosimetry is considered to be a useful tool to limit toxicity associated with peptide receptor radionuclide therapy (PRRT). To facilitate the establishment and reliability of absorbed-dose response relationships, it is essential to assess the accuracy of dosimetry in clinically realistic scenarios. To this end, we developed pharmacokinetic digital phantoms corresponding to patients treated with 177Lu-DOTATATE. Three individual voxel phantoms from the XCAT population were generated and assigned a dynamic activity distribution based on a compartment model for 177Lu-DOTATATE, designed specifically for this purpose. The compartment model was fitted to time-activity data from 10 patients, primarily acquired using quantitative scintillation camera imaging. S values for all phantom source-target combinations were calculated based on Monte-Carlo simulations. Combining the S values and time-activity curves, reference values of the absorbed dose to the phantom kidneys, liver, spleen, tumours and whole-body were calculated. The phantoms were used in a virtual dosimetry study, using Monte-Carlo simulated gamma-camera images and conventional methods for absorbed-dose calculations. The characteristics of the SPECT and WB planar images were found to well represent those of real patient images, capturing the difficulties present in image-based dosimetry. The phantoms are expected to be useful for further studies and optimisation of clinical dosimetry in 177Lu PRRT.

  2. Applying Signal-Detection Theory to the Study of Observer Accuracy and Bias in Behavioral Assessment

    ERIC Educational Resources Information Center

    Lerman, Dorothea C.; Tetreault, Allison; Hovanetz, Alyson; Bellaci, Emily; Miller, Jonathan; Karp, Hilary; Mahmood, Angela; Strobel, Maggie; Mullen, Shelley; Keyl, Alice; Toupard, Alexis

    2010-01-01

    We evaluated the feasibility and utility of a laboratory model for examining observer accuracy within the framework of signal-detection theory (SDT). Sixty-one individuals collected data on aggression while viewing videotaped segments of simulated teacher-child interactions. The purpose of Experiment 1 was to determine if brief feedback and…

  3. Interrater Reliability Estimators Commonly Used in Scoring Language Assessments: A Monte Carlo Investigation of Estimator Accuracy

    ERIC Educational Resources Information Center

    Morgan, Grant B.; Zhu, Min; Johnson, Robert L.; Hodge, Kari J.

    2014-01-01

    Common estimators of interrater reliability include Pearson product-moment correlation coefficients, Spearman rank-order correlations, and the generalizability coefficient. The purpose of this study was to examine the accuracy of estimators of interrater reliability when varying the true reliability, number of scale categories, and number of…

  4. An Accuracy--Response Time Capacity Assessment Function that Measures Performance against Standard Parallel Predictions

    ERIC Educational Resources Information Center

    Townsend, James T.; Altieri, Nicholas

    2012-01-01

    Measures of human efficiency under increases in mental workload or attentional limitations are vital in studying human perception, cognition, and action. Assays of efficiency as workload changes have typically been confined to either reaction times (RTs) or accuracy alone. Within the realm of RTs, a nonparametric measure called the "workload…

  5. Development of a Mathematical Model to Assess the Accuracy of Difference between Geodetic Heights

    ERIC Educational Resources Information Center

    Gairabekov, Ibragim; Kliushin, Evgenii; Gayrabekov, Magomed-Bashir; Ibragimova, Elina; Gayrabekova, Amina

    2016-01-01

    The article includes the results of theoretical studies of the accuracy of geodetic height survey and marks points on the Earth's surface using satellite technology. The dependence of the average square error of geodetic heights difference survey from the distance to the base point was detected. It is being proved that by using satellite…

  6. Portable device to assess dynamic accuracy of global positioning systems (GPS) receivers used in agricultural aircraft

    Technology Transfer Automated Retrieval System (TEKTRAN)

    A device was designed to test the dynamic accuracy of Global Positioning System (GPS) receivers used in aerial vehicles. The system works by directing a sun-reflected light beam from the ground to the aircraft using mirrors. A photodetector is placed pointing downward from the aircraft and circuitry...

  7. Comparative analysis of Worldview-2 and Landsat 8 for coastal saltmarsh mapping accuracy assessment

    NASA Astrophysics Data System (ADS)

    Rasel, Sikdar M. M.; Chang, Hsing-Chung; Diti, Israt Jahan; Ralph, Tim; Saintilan, Neil

    2016-05-01

    Coastal saltmarsh and their constituent components and processes are of an interest scientifically due to their ecological function and services. However, heterogeneity and seasonal dynamic of the coastal wetland system makes it challenging to map saltmarshes with remotely sensed data. This study selected four important saltmarsh species Pragmitis australis, Sporobolus virginicus, Ficiona nodosa and Schoeloplectus sp. as well as a Mangrove and Pine tree species, Avecinia and Casuarina sp respectively. High Spatial Resolution Worldview-2 data and Coarse Spatial resolution Landsat 8 imagery were selected in this study. Among the selected vegetation types some patches ware fragmented and close to the spatial resolution of Worldview-2 data while and some patch were larger than the 30 meter resolution of Landsat 8 data. This study aims to test the effectiveness of different classifier for the imagery with various spatial and spectral resolutions. Three different classification algorithm, Maximum Likelihood Classifier (MLC), Support Vector Machine (SVM) and Artificial Neural Network (ANN) were tested and compared with their mapping accuracy of the results derived from both satellite imagery. For Worldview-2 data SVM was giving the higher overall accuracy (92.12%, kappa =0.90) followed by ANN (90.82%, Kappa 0.89) and MLC (90.55%, kappa = 0.88). For Landsat 8 data, MLC (82.04%) showed the highest classification accuracy comparing to SVM (77.31%) and ANN (75.23%). The producer accuracy of the classification results were also presented in the paper.

  8. An automated method for the evaluation of the pointing accuracy of Sun-tracking devices

    NASA Astrophysics Data System (ADS)

    Baumgartner, Dietmar J.; Pötzi, Werner; Freislich, Heinrich; Strutzmann, Heinz; Veronig, Astrid M.; Rieder, Harald E.

    2017-03-01

    The accuracy of solar radiation measurements, for direct (DIR) and diffuse (DIF) radiation, depends significantly on the precision of the operational Sun-tracking device. Thus, rigid targets for instrument performance and operation have been specified for international monitoring networks, e.g., the Baseline Surface Radiation Network (BSRN) operating under the auspices of the World Climate Research Program (WCRP). Sun-tracking devices that fulfill these accuracy requirements are available from various instrument manufacturers; however, none of the commercially available systems comprise an automatic accuracy control system allowing platform operators to independently validate the pointing accuracy of Sun-tracking sensors during operation. Here we present KSO-STREAMS (KSO-SunTRackEr Accuracy Monitoring System), a fully automated, system-independent, and cost-effective system for evaluating the pointing accuracy of Sun-tracking devices. We detail the monitoring system setup, its design and specifications, and the results from its application to the Sun-tracking system operated at the Kanzelhöhe Observatory (KSO) Austrian radiation monitoring network (ARAD) site. The results from an evaluation campaign from March to June 2015 show that the tracking accuracy of the device operated at KSO lies within BSRN specifications (i.e., 0.1° tracking accuracy) for the vast majority of observations (99.8 %). The evaluation of manufacturer-specified active-tracking accuracies (0.02°), during periods with direct solar radiation exceeding 300 W m-2, shows that these are satisfied in 72.9 % of observations. Tracking accuracies are highest during clear-sky conditions and on days where prevailing clear-sky conditions are interrupted by frontal movement; in these cases, we obtain the complete fulfillment of BSRN requirements and 76.4 % of observations within manufacturer-specified active-tracking accuracies. Limitations to tracking surveillance arise during overcast conditions and

  9. Comparison of reconstruction methods and quantitative accuracy in Siemens Inveon PET scanner

    NASA Astrophysics Data System (ADS)

    Ram Yu, A.; Kim, Jin Su; Kang, Joo Hyun; Moo Lim, Sang

    2015-04-01

    PET reconstruction is key to the quantification of PET data. To our knowledge, no comparative study of reconstruction methods has been performed to date. In this study, we compared reconstruction methods with various filters in terms of their spatial resolution, non-uniformities (NU), recovery coefficients (RCs), and spillover ratios (SORs). In addition, the linearity of reconstructed radioactivity between linearity of measured and true concentrations were also assessed. A Siemens Inveon PET scanner was used in this study. Spatial resolution was measured with NEMA standard by using a 1 mm3 sized 18F point source. Image quality was assessed in terms of NU, RC and SOR. To measure the effect of reconstruction algorithms and filters, data was reconstructed using FBP, 3D reprojection algorithm (3DRP), ordered subset expectation maximization 2D (OSEM 2D), and maximum a posteriori (MAP) with various filters or smoothing factors (β). To assess the linearity of reconstructed radioactivity, image quality phantom filled with 18F was used using FBP, OSEM and MAP (β =1.5 & 5 × 10-5). The highest achievable volumetric resolution was 2.31 mm3 and the highest RCs were obtained when OSEM 2D was used. SOR was 4.87% for air and 3.97% for water, obtained OSEM 2D reconstruction was used. The measured radioactivity of reconstruction image was proportional to the injected one for radioactivity below 16 MBq/ml when FBP or OSEM 2D reconstruction methods were used. By contrast, when the MAP reconstruction method was used, activity of reconstruction image increased proportionally, regardless of the amount of injected radioactivity. When OSEM 2D or FBP were used, the measured radioactivity concentration was reduced by 53% compared with true injected radioactivity for radioactivity <16 MBq/ml. The OSEM 2D reconstruction method provides the highest achievable volumetric resolution and highest RC among all the tested methods and yields a linear relation between the measured and true

  10. Accuracy Assessment of Direct Georeferencing for Photogrammetric Applications on Small Unmanned Aerial Platforms

    NASA Astrophysics Data System (ADS)

    Mian, O.; Lutes, J.; Lipa, G.; Hutton, J. J.; Gavelle, E.; Borghini, S.

    2016-03-01

    Microdrones md4-1000 quad-rotor VTOL UAV. The Sony A7R and each lens combination were focused and calibrated terrestrially using the Applanix camera calibration facility, and then integrated with the APX-15 GNSS-Inertial system using a custom mount specifically designed for UAV applications. The mount is constructed in such a way as to maintain the stability of both the interior orientation and IMU boresight calibration over shock and vibration, thus turning the Sony A7R into a metric imaging solution. In July and August 2015, Applanix and Avyon carried out a series of test flights of this system. The goal of these test flights was to assess the performance of DMS APX-15 direct georeferencing system under various scenarios. Furthermore, an examination of how DMS APX-15 can be used to produce accurate map products without the use of ground control points and with reduced sidelap was also carried out. Reducing the side lap for survey missions performed by small UAVs can significantly increase the mapping productivity of these platforms. The area mapped during the first flight campaign was a 250m x 300m block and a 775m long railway corridor in a rural setting in Ontario, Canada. The second area mapped was a 450m long corridor over a dam known as Fryer Dam (over Richelieu River in Quebec, Canada). Several ground control points were distributed within both test areas. The flight over the block area included 8 North-South lines and 1 cross strip flown at 80m AGL, resulting in a ~1cm GSD. The flight over the railway corridor included 2 North-South lines also flown at 80m AGL. Similarly, the flight over the dam corridor included 2 North-South lines flown at 50m AGL. The focus of this paper was to analyse the results obtained from the two corridors. Test results from both areas were processed using Direct Georeferencing techniques, and then compared for accuracy against the known positions of ground control points in each test area. The GNSS-Inertial data collected by the APX-15 was

  11. Accuracy of Cameriere, Haavikko, and Willems radiographic methods on age estimation on Bosnian-Herzegovian children age groups 6-13.

    PubMed

    Galić, Ivan; Vodanović, Marin; Cameriere, Roberto; Nakaš, Enita; Galić, Elizabeta; Selimović, Edin; Brkić, Hrvoje

    2011-03-01

    The aim of this cross-sectional study was to compare the accuracy of the Cameriere European formula (Cameriere), adopted Haavikko method from 1974 (Haavikko), and revisited Demirjian method by Willems (Willems) for age estimation on orthopantomograms (OPGs) of Bosnian-Herzegovian (BH) children age groups 6-13 years. The accuracy was determined as difference between estimated dental age (DA) and chronological age (CA) and the absolute accuracy (absolute difference) was assessed by analyzing OPGs of 591 girls and 498 boys. The Cameriere method overestimated the mean age by 0.09 year for girls and underestimated by -0.02 year for boys. The Haavikko method underestimated the mean age by -0.29 year for girls and -0.09 year for boys. The Willems method overestimated the mean age by 0.24 year in girls and by 0.42 year in boys. The absolute accuracies were 0.53 year for girls and 0.55 year for boys for Cameriere method; for Haavikko method, 0.59 year for girls and 0.62 year for boys; and for Willems method 0.69 year for girls and 0.67 year for boys. In conclusion, Cameriere method is the most accurate for estimating the age of BH children age groups 6-13 years using OPGs, following adopted Haavikko method and Willems method.

  12. Three-dimensional accuracy of different correction methods for cast implant bars

    PubMed Central

    Kwon, Ji-Yung; Kim, Chang-Whe; Lim, Young-Jun; Kwon, Ho-Beom

    2014-01-01

    PURPOSE The aim of the present study was to evaluate the accuracy of three techniques for correction of cast implant bars. MATERIALS AND METHODS Thirty cast implant bars were fabricated on a metal master model. All cast implant bars were sectioned at 5 mm from the left gold cylinder using a disk of 0.3 mm thickness, and then each group of ten specimens was corrected by gas-air torch soldering, laser welding, and additional casting technique. Three dimensional evaluation including horizontal, vertical, and twisting measurements was based on measurement and comparison of (1) gap distances of the right abutment replica-gold cylinder interface at buccal, distal, lingual side, (2) changes of bar length, and (3) axis angle changes of the right gold cylinders at the step of the post-correction measurements on the three groups with a contact and non-contact coordinate measuring machine. One-way analysis of variance (ANOVA) and paired t-test were performed at the significance level of 5%. RESULTS Gap distances of the cast implant bars after correction procedure showed no statistically significant difference among groups. Changes in bar length between pre-casting and post-correction measurement were statistically significance among groups. Axis angle changes of the right gold cylinders were not statistically significance among groups. CONCLUSION There was no statistical significance among three techniques in horizontal, vertical and axial errors. But, gas-air torch soldering technique showed the most consistent and accurate trend in the correction of implant bar error. However, Laser welding technique, showed a large mean and standard deviation in vertical and twisting measurement and might be technique-sensitive method. PMID:24605205

  13. Spatial and temporal analysis on the distribution of active radio-frequency identification (RFID) tracking accuracy with the Kriging method.

    PubMed

    Liu, Xin; Shannon, Jeremy; Voun, Howard; Truijens, Martijn; Chi, Hung-Lin; Wang, Xiangyu

    2014-10-29

    Radio frequency identification (RFID) technology has already been applied in a number of areas to facilitate the tracking process. However, the insufficient tracking accuracy of RFID is one of the problems that impedes its wider application. Previous studies focus on examining the accuracy of discrete points RFID, thereby leaving the tracking accuracy of the areas between the observed points unpredictable. In this study, spatial and temporal analysis is applied to interpolate the continuous distribution of RFID tracking accuracy based on the Kriging method. An implementation trial has been conducted in the loading and docking area in front of a warehouse to validate this approach. The results show that the weak signal area can be easily identified by the approach developed in the study. The optimum distance between two RFID readers and the effect of the sudden removal of readers are also presented by analysing the spatial and temporal variation of RFID tracking accuracy. This study reveals the correlation between the testing time and the stability of RFID tracking accuracy. Experimental results show that the proposed approach can be used to assist the RFID system setup process to increase tracking accuracy.

  14. Spatial and Temporal Analysis on the Distribution of Active Radio-Frequency Identification (RFID) Tracking Accuracy with the Kriging Method

    PubMed Central

    Liu, Xin; Shannon, Jeremy; Voun, Howard; Truijens, Martijn; Chi, Hung-Lin; Wang, Xiangyu

    2014-01-01

    Radio frequency identification (RFID) technology has already been applied in a number of areas to facilitate the tracking process. However, the insufficient tracking accuracy of RFID is one of the problems that impedes its wider application. Previous studies focus on examining the accuracy of discrete points RFID, thereby leaving the tracking accuracy of the areas between the observed points unpredictable. In this study, spatial and temporal analysis is applied to interpolate the continuous distribution of RFID tracking accuracy based on the Kriging method. An implementation trial has been conducted in the loading and docking area in front of a warehouse to validate this approach. The results show that the weak signal area can be easily identified by the approach developed in the study. The optimum distance between two RFID readers and the effect of the sudden removal of readers are also presented by analysing the spatial and temporal variation of RFID tracking accuracy. This study reveals the correlation between the testing time and the stability of RFID tracking accuracy. Experimental results show that the proposed approach can be used to assist the RFID system setup process to increase tracking accuracy. PMID:25356648

  15. Long-term deflections of reinforced concrete elements: accuracy analysis of predictions by different methods

    NASA Astrophysics Data System (ADS)

    Gribniak, Viktor; Bacinskas, Darius; Kacianauskas, Rimantas; Kaklauskas, Gintaris; Torres, Lluis

    2013-08-01

    Long-term deflection response of reinforced concrete flexural members is influenced by the interaction of complex physical phenomena, such as concrete creep, shrinkage and cracking, which makes their prediction difficult. A number of approaches are proposed by design codes with different degrees of simplification and accuracy. This paper statistically investigates accuracy of long-term deflection predictions made by some of the most widely used design codes ( Eurocode 2, ACI 318, ACI 435, and the new Russian code SP 52-101) and a numerical technique proposed by the authors. The accuracy is analyzed using test data of 322 reinforced concrete members from 27 test programs reported in the literature. The predictions of each technique are discussed, and a comparative analysis is made showing the influence of different parameters, such as sustained loading duration, compressive strength of concrete, loading intensity and reinforcement ratio, on the prediction accuracy.

  16. Accuracy assessment of airborne photogrammetrically derived high-resolution digital elevation models in a high mountain environment

    NASA Astrophysics Data System (ADS)

    Müller, Johann; Gärtner-Roer, Isabelle; Thee, Patrick; Ginzler, Christian

    2014-12-01

    High-resolution digital elevation models (DEMs) generated by airborne remote sensing are frequently used to analyze landform structures (monotemporal) and geomorphological processes (multitemporal) in remote areas or areas of extreme terrain. In order to assess and quantify such structures and processes it is necessary to know the absolute accuracy of the available DEMs. This study assesses the absolute vertical accuracy of DEMs generated by the High Resolution Stereo Camera-Airborne (HRSC-A), the Leica Airborne Digital Sensors 40/80 (ADS40 and ADS80) and the analogue camera system RC30. The study area is located in the Turtmann valley, Valais, Switzerland, a glacially and periglacially formed hanging valley stretching from 2400 m to 3300 m a.s.l. The photogrammetrically derived DEMs are evaluated against geodetic field measurements and an airborne laser scan (ALS). Traditional and robust global and local accuracy measurements are used to describe the vertical quality of the DEMs, which show a non Gaussian distribution of errors. The results show that all four sensor systems produce DEMs with similar accuracy despite their different setups and generations. The ADS40 and ADS80 (both with a ground sampling distance of 0.50 m) generate the most accurate DEMs in complex high mountain areas with a RMSE of 0.8 m and NMAD of 0.6 m They also show the highest accuracy relating to flying height (0.14‰). The pushbroom scanning system HRSC-A produces a RMSE of 1.03 m and a NMAD of 0.83 m (0.21‰ accuracy of the flying height and 10 times the ground sampling distance). The analogue camera system RC30 produces DEMs with a vertical accuracy of 1.30 m RMSE and 0.83 m NMAD (0.17‰ accuracy of the flying height and two times the ground sampling distance). It is also shown that the performance of the DEMs strongly depends on the inclination of the terrain. The RMSE of areas up to an inclination <40° is better than 1 m. In more inclined areas the error and outlier occurrence

  17. Predictive accuracy of three field methods for estimating relative body fatness of nonobese and obese women.

    PubMed

    Heyward, V H; Cook, K L; Hicks, V L; Jenkins, K A; Quatrochi, J A; Wilson, W L

    1992-03-01

    Three methods of body composition assessment were used to estimate percent body fat (%BF) in nonobese (n = 77) and obese (n = 71) women, 20-72 yrs of age. Skinfolds (SKF), bioelectrical impedance (BIA), and near-infrared interactance (NIR) methods were compared to criterion-derived %BF from hydrostatic weighing (%BFHW). Nonobese subjects had < 30% BFHW and obese subjects had > or = 30% BFHW. The Jackson, Pollock, and Ward SKF equation and the manufacturer's equations for BIA (Valhalla) and NIR (Futrex-5000) were used. For nonobese women there were no significant differences between mean %BFHW and %BFSKF, %BFBIA, and %BFNIR. The rs and SEEs were 0.65 and 3.4% BF for SKF, 0.61 and 3.6% BF for BIA, and 0.58 and 3.7% BF for NIR for nonobese subjects. For obese women, mean %BFHW was significantly underestimated by the SKF, BIA, and NIR methods. The rs and SEEs for the obese group were 0.59 and 3.4% BF for SKF, 0.56 and 3.5% BF for BIA, and 0.36 and 3.9% BF for NIR. The total errors of the equations ranged from 5.6 to 8.0% BF in the obese group. It is concluded that all three field methods accurately estimate %BF for nonobese women; however, none of the methods is suitable for estimating %BF for obese women.

  18. Assessing the accuracy of software predictions of mammalian and microbial metabolites

    EPA Science Inventory

    New chemical development and hazard assessments benefit from accurate predictions of mammalian and microbial metabolites. Fourteen biotransformation libraries encoded in eight software packages that predict metabolite structures were assessed for their sensitivity (proportion of ...

  19. Assessing the accuracy of protein structures by quantum mechanical computations of 13C(alpha) chemical shifts.

    PubMed

    Vila, Jorge A; Scheraga, Harold A

    2009-10-20

    Two major techniques have been used to determine the three-dimensional structures of proteins: X-ray diffraction and NMR spectroscopy. In particular, the validation of NMR-derived protein structures is one of the most challenging problems in NMR spectroscopy. Therefore, researchers have proposed a plethora of methods to determine the accuracy and reliability of protein structures. Despite these proposals, there is a growing need for more sophisticated, physics-based structure validation methods. This approach will enable us to (a) characterize the "quality" of the NMR-derived ensemble as a whole by a single parameter, (b) unambiguously identify flaws in the sequence at a residue level, and (c) provide precise information, such as sets of backbone and side-chain torsional angles, that we can use to detect local flaws. Rather than reviewing all of the existing validation methods, this Account describes the contributions of our research group toward a solution of the long-standing problem of both global and local structure validation of NMR-derived protein structures. We emphasize a recently introduced physics-based methodology that makes use of observed and computed (13)C(alpha) chemical shifts (at the density functional theory (DFT) level of theory) for an accurate validation of protein structures in solution and in crystals. By assessing the ability of computed (13)C(alpha) chemical shifts to reproduce observed (13)C(alpha) chemical shifts of a single structure or ensemble of structures in solution and in crystals, we accomplish a global validation by using the conformationally averaged root-mean-square deviation, ca-rmsd, as a scoring function. In addition, the method enables us to provide local validation by identifying a set of individual amino acid conformations for which the computed and observed (13)C(alpha) chemical shifts do not agree within a certain error range and may represent a nonreliable fold of the protein model. Although it is computationally

  20. Accuracy in Student Self-Assessment: Directions and Cautions for Research

    ERIC Educational Resources Information Center

    Brown, Gavin T. L.; Andrade, Heidi L.; Chen, Fei

    2015-01-01

    Student self-assessment is a central component of current conceptions of formative and classroom assessment. The research on self-assessment has focused on its efficacy in promoting both academic achievement and self-regulated learning, with little concern for issues of validity. Because reliability of testing is considered a sine qua non for the…

  1. How Does One Assess the Accuracy of Academic Success Predictors? ROC Analysis Applied to University Entrance Factors

    ERIC Educational Resources Information Center

    Vivo, Juana-Maria; Franco, Manuel

    2008-01-01

    This article attempts to present a novel application of a method of measuring accuracy for academic success predictors that could be used as a standard. This procedure is known as the receiver operating characteristic (ROC) curve, which comes from statistical decision techniques. The statistical prediction techniques provide predictor models and…

  2. Image-based in vivo assessment of targeting accuracy of stereotactic brain surgery in experimental rodent models

    NASA Astrophysics Data System (ADS)

    Rangarajan, Janaki Raman; Vande Velde, Greetje; van Gent, Friso; de Vloo, Philippe; Dresselaers, Tom; Depypere, Maarten; van Kuyck, Kris; Nuttin, Bart; Himmelreich, Uwe; Maes, Frederik

    2016-11-01

    Stereotactic neurosurgery is used in pre-clinical research of neurological and psychiatric disorders in experimental rat and mouse models to engraft a needle or electrode at a pre-defined location in the brain. However, inaccurate targeting may confound the results of such experiments. In contrast to the clinical practice, inaccurate targeting in rodents remains usually unnoticed until assessed by ex vivo end-point histology. We here propose a workflow for in vivo assessment of stereotactic targeting accuracy in small animal studies based on multi-modal post-operative imaging. The surgical trajectory in each individual animal is reconstructed in 3D from the physical implant imaged in post-operative CT and/or its trace as visible in post-operative MRI. By co-registering post-operative images of individual animals to a common stereotaxic template, targeting accuracy is quantified. Two commonly used neuromodulation regions were used as targets. Target localization errors showed not only variability, but also inaccuracy in targeting. Only about 30% of electrodes were within the subnucleus structure that was targeted and a-specific adverse effects were also noted. Shifting from invasive/subjective 2D histology towards objective in vivo 3D imaging-based assessment of targeting accuracy may benefit a more effective use of the experimental data by excluding off-target cases early in the study.

  3. Image-based in vivo assessment of targeting accuracy of stereotactic brain surgery in experimental rodent models

    PubMed Central

    Rangarajan, Janaki Raman; Vande Velde, Greetje; van Gent, Friso; De Vloo, Philippe; Dresselaers, Tom; Depypere, Maarten; van Kuyck, Kris; Nuttin, Bart; Himmelreich, Uwe; Maes, Frederik

    2016-01-01

    Stereotactic neurosurgery is used in pre-clinical research of neurological and psychiatric disorders in experimental rat and mouse models to engraft a needle or electrode at a pre-defined location in the brain. However, inaccurate targeting may confound the results of such experiments. In contrast to the clinical practice, inaccurate targeting in rodents remains usually unnoticed until assessed by ex vivo end-point histology. We here propose a workflow for in vivo assessment of stereotactic targeting accuracy in small animal studies based on multi-modal post-operative imaging. The surgical trajectory in each individual animal is reconstructed in 3D from the physical implant imaged in post-operative CT and/or its trace as visible in post-operative MRI. By co-registering post-operative images of individual animals to a common stereotaxic template, targeting accuracy is quantified. Two commonly used neuromodulation regions were used as targets. Target localization errors showed not only variability, but also inaccuracy in targeting. Only about 30% of electrodes were within the subnucleus structure that was targeted and a-specific adverse effects were also noted. Shifting from invasive/subjective 2D histology towards objective in vivo 3D imaging-based assessment of targeting accuracy may benefit a more effective use of the experimental data by excluding off-target cases early in the study. PMID:27901096

  4. Accuracy assessment of the global ionospheric model over the Southern Ocean based on dynamic observation

    NASA Astrophysics Data System (ADS)

    Luo, Xiaowen; Xu, Huajun; Li, Zishen; Zhang, Tao; Gao, Jinyao; Shen, Zhongyan; Yang, Chunguo; Wu, Ziyin

    2017-02-01

    The global ionospheric model based on the reference stations of the Global Navigation Satellite System (GNSS) of the International GNSS Services is presently the most commonly used products of the global ionosphere. It is very important to comprehensively analyze and evaluate the accuracy and reliability of the model for the reasonable use of this kind of ionospheric product. In terms of receiver station deployment, this work is different from the traditional performance evaluation of the global ionosphere model based on observation data of ground-based static reference stations. The preliminary evaluation and analysis of the the global ionospheric model was conducted with the dynamic observation data across different latitudes over the southern oceans. The validation results showed that the accuracy of the global ionospheric model over the southern oceans is about 5 TECu, which deviates from the measured ionospheric TEC by about -0.6 TECu.

  5. Aneroid sphygmomanometers. An assessment of accuracy at a university hospital and clinics.

    PubMed

    Bailey, R H; Knaus, V L; Bauer, J H

    1991-07-01

    Defects of aneroid sphygmomanometers are a source of error in blood pressure measurement. We inspected 230 aneroid sphygmomanometers for physical defects and compared their accuracy against a standard mercury manometer at five different pressure points. An aneroid sphygmomanometer was defined as intolerant if it deviated from the mercury manometer by greater than +/- 3 mm Hg at two or more of the test points. The three most common physical defects were indicator needles not pointing to the "zero box," cracked face plates, and defective tubing. Eighty (34.8 of the 230 aneroid sphygmomanometers were determined to be intolerant with the greatest frequency of deviation seen at pressure levels of 150 mm Hg or greater. We recommend that aneroid manometers be inspected for physical defects and calibrated for accuracy against a standard mercury manometer at 6-month intervals to prevent inaccurate blood pressure measurements.

  6. The Effects of Various Item Selection Methods on the Classification Accuracy and Classification Consistency of Criterion-Referenced Instruments.

    ERIC Educational Resources Information Center

    Smith, Douglas U.

    This study examined the effects of certain item selection methods on the classification accuracy and classification consistency of criterion-referenced instruments. Three item response data sets, representing varying situations of instructional effectiveness, were simulated. Five methods of item selection were then applied to each data set for the…

  7. Task and Observer Skill Factors in Accuracy of Assessment of Performance

    DTIC Science & Technology

    1977-04-01

    BEHAVIORAL and SOCIAL SCIENCES r -j 1300 Wilson Boulevard ^ Arlingfon, Virginia 22209 Approved for public release; distributed unlimitacl U. S...Unclassified /i Unclafislfli’L BCCUKtTY CLASSIFICATION OF THIS PAGEfIVh«n Data Entat-d) ( r .’O continued) H^wQn the basis of these data, a...34 approach has been tak»in to questions of accuracy of clinical judge- ments (Sarbin, Taft, & Bailey, I960; Bierl , Atkins, Briar, Leaman

  8. Quantitative Assessment of the Accuracy of Constitutive Laws for Plasticity with an Emphasis on Cyclic Deformation

    DTIC Science & Technology

    1993-04-01

    new law, the B-L law. The experimental database is constructed from a5 series of constant amplitude and random amplitude strain controlled cyclic...description of the experimental instrumentation is given in Appendix I. The cyclic plasticity experiments were performed under strain control at room5...instrumentation is present and control accuracy is not as good, the increments or difference of strain at two adjacent sampling intervals should be

  9. Accuracy assessment of high frequency 3D ultrasound for digital impression-taking of prepared teeth

    NASA Astrophysics Data System (ADS)

    Heger, Stefan; Vollborn, Thorsten; Tinschert, Joachim; Wolfart, Stefan; Radermacher, Klaus

    2013-03-01

    Silicone based impression-taking of prepared teeth followed by plaster casting is well-established but potentially less reliable, error-prone and inefficient, particularly in combination with emerging techniques like computer aided design and manufacturing (CAD/CAM) of dental prosthesis. Intra-oral optical scanners for digital impression-taking have been introduced but until now some drawbacks still exist. Because optical waves can hardly penetrate liquids or soft-tissues, sub-gingival preparations still need to be uncovered invasively prior to scanning. High frequency ultrasound (HFUS) based micro-scanning has been recently investigated as an alternative to optical intra-oral scanning. Ultrasound is less sensitive against oral fluids and in principal able to penetrate gingiva without invasively exposing of sub-gingival preparations. Nevertheless, spatial resolution as well as digitization accuracy of an ultrasound based micro-scanning system remains a critical parameter because the ultrasound wavelength in water-like media such as gingiva is typically smaller than that of optical waves. In this contribution, the in-vitro accuracy of ultrasound based micro-scanning for tooth geometry reconstruction is being investigated and compared to its extra-oral optical counterpart. In order to increase the spatial resolution of the system, 2nd harmonic frequencies from a mechanically driven focused single element transducer were separated and corresponding 3D surface models were calculated for both fundamentals and 2nd harmonics. Measurements on phantoms, model teeth and human teeth were carried out for evaluation of spatial resolution and surface detection accuracy. Comparison of optical and ultrasound digital impression taking indicate that, in terms of accuracy, ultrasound based tooth digitization can be an alternative for optical impression-taking.

  10. Power Series Approximation for the Correlation Kernel Leading to Kohn-Sham Methods Combining Accuracy, Computational Efficiency, and General Applicability

    NASA Astrophysics Data System (ADS)

    Erhard, Jannis; Bleiziffer, Patrick; Görling, Andreas

    2016-09-01

    A power series approximation for the correlation kernel of time-dependent density-functional theory is presented. Using this approximation in the adiabatic-connection fluctuation-dissipation (ACFD) theorem leads to a new family of Kohn-Sham methods. The new methods yield reaction energies and barriers of unprecedented accuracy and enable a treatment of static (strong) correlation with an accuracy of high-level multireference configuration interaction methods but are single-reference methods allowing for a black-box-like handling of static correlation. The new methods exhibit a better scaling of the computational effort with the system size than rivaling wave-function-based electronic structure methods. Moreover, the new methods do not suffer from the problem of singularities in response functions plaguing previous ACFD methods and therefore are applicable to any type of electronic system.

  11. Future dedicated Venus-SGG flight mission: Accuracy assessment and performance analysis

    NASA Astrophysics Data System (ADS)

    Zheng, Wei; Hsu, Houtse; Zhong, Min; Yun, Meijuan

    2016-01-01

    This study concentrates principally on the systematic requirements analysis for the future dedicated Venus-SGG (spacecraft gravity gradiometry) flight mission in China in respect of the matching measurement accuracies of the spacecraft-based scientific instruments and the orbital parameters of the spacecraft. Firstly, we created and proved the single and combined analytical error models of the cumulative Venusian geoid height influenced by the gravity gradient error of the spacecraft-borne atom-interferometer gravity gradiometer (AIGG) and the orbital position error and orbital velocity error tracked by the deep space network (DSN) on the Earth station. Secondly, the ultra-high-precision spacecraft-borne AIGG is propitious to making a significant contribution to globally mapping the Venusian gravitational field and modeling the geoid with unprecedented accuracy and spatial resolution through weighing the advantages and disadvantages among the electrostatically suspended gravity gradiometer, the superconducting gravity gradiometer and the AIGG. Finally, the future dedicated Venus-SGG spacecraft had better adopt the optimal matching accuracy indices consisting of 3 × 10-13/s2 in gravity gradient, 10 m in orbital position and 8 × 10-4 m/s in orbital velocity and the preferred orbital parameters comprising an orbital altitude of 300 ± 50 km, an observation time of 60 months and a sampling interval of 1 s.

  12. Enhancing Institutional Assessment Efforts through Qualitative Methods

    ERIC Educational Resources Information Center

    Van Note Chism, Nancy; Banta, Trudy W.

    2007-01-01

    Qualitative methods can do much to describe context and illuminate the why behind patterns encountered in institutional assessment. Alone, or in combination with quantitative methods, they should be the approach of choice for many of the most important assessment questions. (Contains 1 table.)

  13. Accuracy Assessment of Geometrical Elements for Setting-Out in Horizontal Plane of Conveying Chambers at the Bauxite Mine "KOSTURI" Srebrenica

    NASA Astrophysics Data System (ADS)

    Milutinović, Aleksandar; Ganić, Aleksandar; Tokalić, Rade

    2014-03-01

    Setting-out of objects on the exploitation field of the mine, both in surface mining and in the underground mines, is determined by the specified setting-out accuracy of reference points, which are best to define spatial position of the object projected. For the purpose of achieving of the specified accuracy, it is necessary to perform a priori accuracy assessment of parameters, which are to be used when performing setting-out. Based on the a priori accuracy assessment, verification of the quality of geometrical setting- -out elements specified in the layout; definition of the accuracy for setting-out of geometrical elements; selection of setting-out method; selection at the type and class of instruments and tools that need to be applied in order to achieve predefined accuracy. The paper displays the accuracy assessment of geometrical elements for setting-out of the main haul gallery, haul downcast and helical conveying downcasts in shape of an inclined helix in horizontal plane, using the example of the underground bauxite mine »Kosturi«, Srebrenica. Wytyczanie obiektów na polu wydobywczym w kopalniach, zarówno podziemnych jak i odkrywkowych, zależy w dużej mierze od określonej dokładności wytyczania punktów referencyjnych, przy pomocy których określane jest następnie położenie przestrzenne pozostałych obiektów. W celu uzyskania założonej dokładności, należy przeprowadzić wstępną analizę dokładności oszacowania parametrów które następnie wykorzystane będą w procesie wytyczania. W oparciu o wyniki wstępnej analizy dokładności dokonuje się weryfikacji jakości geometrycznego wytyczenia elementów zaznaczonych na szkicu, uwzględniając te wyniki dobrać należy odpowiednią metodę wytyczania i rodzaj oraz klasę wykorzystywanych narzędzi i instrumentów, tak by osiągnąć założony poziom dokładności. W pracy przedstawiono oszacowanie dokładności wytyczania elementów geometrycznych dla głównego chodnika transportowego

  14. Screeners and brief assessment methods.

    PubMed

    Pérez Rodrigo, Carmen; Morán Fagúndez, Luis Juan; Riobó Serván, Pilar; Aranceta Bartrina, Javier

    2015-02-26

    In the last two decades easy-to-use simple instruments have been developed and validated to assess specific aspects of the diet or a general profile that can be compared with a reference dietary pattern as the Mediterranean Diet or with the recommendations of the Dietary Guidelines. Brief instruments are rapid, simple and easy to use tools that can be implemented by unskilled personnel without specific training. These tools are useful both in clinical settings and in Primary Health Care or in the community as a tool for triage, as a screening tool to identify individuals or groups of people at risk who require further care or even they have been used in studies to investigate associations between specific aspects of the diet and health outcomes. They are also used in interventions focused on changing eating behaviors as a diagnostic tool, for self-evaluation purposes, or to provide tailored advice in web based interventions or mobile apps. There are some specific instruments for use in children, adults, elderly or specific population groups.

  15. Constraining OCT with Knowledge of Device Design Enables High Accuracy Hemodynamic Assessment of Endovascular Implants

    PubMed Central

    Brown, Jonathan; Lopes, Augusto C.; Kunio, Mie; Kolachalama, Vijaya B.; Edelman, Elazer R.

    2016-01-01

    Background Stacking cross-sectional intravascular images permits three-dimensional rendering of endovascular implants, yet introduces between-frame uncertainties that limit characterization of device placement and the hemodynamic microenvironment. In a porcine coronary stent model, we demonstrate enhanced OCT reconstruction with preservation of between-frame features through fusion with angiography and a priori knowledge of stent design. Methods and Results Strut positions were extracted from sequential OCT frames. Reconstruction with standard interpolation generated discontinuous stent structures. By computationally constraining interpolation to known stent skeletons fitted to 3D ‘clouds’ of OCT-Angio-derived struts, implant anatomy was resolved, accurately rendering features from implant diameter and curvature (n = 1 vessels, r2 = 0.91, 0.90, respectively) to individual strut-wall configurations (average displacement error ~15 μm). This framework facilitated hemodynamic simulation (n = 1 vessel), showing the critical importance of accurate anatomic rendering in characterizing both quantitative and basic qualitative flow patterns. Discontinuities with standard approaches systematically introduced noise and bias, poorly capturing regional flow effects. In contrast, the enhanced method preserved multi-scale (local strut to regional stent) flow interactions, demonstrating the impact of regional contexts in defining the hemodynamic consequence of local deployment errors. Conclusion Fusion of planar angiography and knowledge of device design permits enhanced OCT image analysis of in situ tissue-device interactions. Given emerging interests in simulation-derived hemodynamic assessment as surrogate measures of biological risk, such fused modalities offer a new window into patient-specific implant environments. PMID:26906566

  16. Reproducibility and accuracy of body composition assessments in mice by dual energy x-ray absorptiometry and time domain nuclear magnetic resonance.

    PubMed

    Halldorsdottir, Solveig; Carmody, Jill; Boozer, Carol N; Leduc, Charles A; Leibel, Rudolph L

    2009-01-01

    OBJECTIVE: To assess the accuracy and reproducibility of dual-energy absorptiometry (DXA; PIXImus(™)) and time domain nuclear magnetic resonance (TD-NMR; Bruker Optics) for the measurement of body composition of lean and obese mice. SUBJECTS AND MEASUREMENTS: Thirty lean and obese mice (body weight range 19-67 g) were studied. Coefficients of variation for repeated (x 4) DXA and NMR scans of mice were calculated to assess reproducibility. Accuracy was assessed by comparing DXA and NMR results of ten mice to chemical carcass analyses. Accuracy of the respective techniques was also assessed by comparing DXA and NMR results obtained with ground meat samples to chemical analyses. Repeated scans of 10-25 gram samples were performed to test the sensitivity of the DXA and NMR methods to variation in sample mass. RESULTS: In mice, DXA and NMR reproducibility measures were similar for fat tissue mass (FTM) (DXA coefficient of variation [CV]=2.3%; and NMR CV=2.8%) (P=0.47), while reproducibility of lean tissue mass (LTM) estimates were better for DXA (1.0%) than NMR (2.2%) (

    accuracy, in mice, DXA overestimated (vs chemical composition) LTM (+1.7 ± 1.3 g [SD], ~ 8%, P <0.001) as well as FTM (+2.0 ± 1.2 g, ~ 46%, P <0.001). NMR estimated LTM and FTM virtually identical to chemical composition analysis (LTM: -0.05 ± 0.5 g, ~0.2%, P =0.79) (FTM: +0.02 ± 0.7 g, ~15%, P =0.93). DXA and NMR-determined LTM and FTM measurements were highly correlated with the corresponding chemical analyses (r(2)=0.92 and r(2)=0.99 for DXA LTM and FTM, respectively; r(2)=0.99 and r(2)=0.99 for NMR LTM and FTM, respectively.) Sample mass did not affect accuracy in assessing chemical composition of small ground meat samples by either DXA or NMR. CONCLUSION: DXA and NMR provide comparable levels of reproducibility in measurements of body composition lean and obese mice. While DXA and NMR measures are highly correlated with chemical analysis measures, DXA consistently

  17. Reproducibility and accuracy of body composition assessments in mice by dual energy x-ray absorptiometry and time domain nuclear magnetic resonance

    PubMed Central

    Halldorsdottir, Solveig; Carmody, Jill; Boozer, Carol N.; Leduc, Charles A.; Leibel, Rudolph L.

    2011-01-01

    Objective To assess the accuracy and reproducibility of dual-energy absorptiometry (DXA; PIXImus™) and time domain nuclear magnetic resonance (TD-NMR; Bruker Optics) for the measurement of body composition of lean and obese mice. Subjects and measurements Thirty lean and obese mice (body weight range 19–67 g) were studied. Coefficients of variation for repeated (x 4) DXA and NMR scans of mice were calculated to assess reproducibility. Accuracy was assessed by comparing DXA and NMR results of ten mice to chemical carcass analyses. Accuracy of the respective techniques was also assessed by comparing DXA and NMR results obtained with ground meat samples to chemical analyses. Repeated scans of 10–25 gram samples were performed to test the sensitivity of the DXA and NMR methods to variation in sample mass. Results In mice, DXA and NMR reproducibility measures were similar for fat tissue mass (FTM) (DXA coefficient of variation [CV]=2.3%; and NMR CV=2.8%) (P=0.47), while reproducibility of lean tissue mass (LTM) estimates were better for DXA (1.0%) than NMR (2.2%) (

    accuracy, in mice, DXA overestimated (vs chemical composition) LTM (+1.7 ± 1.3 g [SD], ~ 8%, P <0.001) as well as FTM (+2.0 ± 1.2 g, ~ 46%, P <0.001). NMR estimated LTM and FTM virtually identical to chemical composition analysis (LTM: −0.05 ± 0.5 g, ~0.2%, P =0.79) (FTM: +0.02 ± 0.7 g, ~15%, P =0.93). DXA and NMR-determined LTM and FTM measurements were highly correlated with the corresponding chemical analyses (r2=0.92 and r2=0.99 for DXA LTM and FTM, respectively; r2=0.99 and r2=0.99 for NMR LTM and FTM, respectively.) Sample mass did not affect accuracy in assessing chemical composition of small ground meat samples by either DXA or NMR. Conclusion DXA and NMR provide comparable levels of reproducibility in measurements of body composition lean and obese mice. While DXA and NMR measures are highly correlated with chemical analysis measures, DXA consistently overestimates LTM

  18. A Comparative Analysis of Diagnostic Accuracy of Focused Assessment With Sonography for Trauma Performed by Emergency Medicine and Radiology Residents

    PubMed Central

    Zamani, Majid; Masoumi, Babak; Esmailian, Mehrdad; Habibi, Amin; Khazaei, Mehdi; Mohammadi Esfahani, Mohammad

    2015-01-01

    Background: Focused assessment with sonography in trauma (FAST) is a method for prompt detection of the abdominal free fluid in patients with abdominal trauma. Objectives: This study was conducted to compare the diagnostic accuracy of FAST performed by emergency medicine residents (EMR) and radiology residents (RRs) in detecting peritoneal free fluids. Patients and Methods: Patients triaged in the emergency department with blunt abdominal trauma, high energy trauma, and multiple traumas underwent a FAST examination by EMRs and RRs with the same techniques to obtain the standard views. Ultrasound findings for free fluid in peritoneal cavity for each patient (positive/negative) were compared with the results of computed tomography, operative exploration, or observation as the final outcome. Results: A total of 138 patients were included in the final analysis. Good diagnostic agreement was noted between the results of FAST scans performed by EMRs and RRs (κ = 0.701, P < 0.001), also between the results of EMRs-performed FAST and the final outcome (κ = 0.830, P < 0.0010), and finally between the results of RRs-performed FAST and final outcome (κ = 0.795, P < 0.001). No significant differences were noted between EMRs- and RRs-performed FASTs regarding sensitivity (84.6% vs 84.6%), specificity (98.4% vs 97.6%), positive predictive value (84.6% vs 84.6%), and negative predictive value (98.4% vs 98.4%). Conclusions: Trained EMRs like their fellow RRs have the ability to perform FAST scan with high diagnostic value in patients with blunt abdominal trauma. PMID:26756009

  19. Assessment of User Home Location Geoinference Methods

    SciTech Connect

    Harrison, Joshua J.; Bell, Eric B.; Corley, Courtney D.; Dowling, Chase P.; Cowell, Andrew J.

    2015-05-29

    This study presents an assessment of multiple approaches to determine the home and/or other important locations to a Twitter user. In this study, we present a unique approach to the problem of geotagged data sparsity in social media when performing geoinferencing tasks. Given the sparsity of explicitly geotagged Twitter data, the ability to perform accurate and reliable user geolocation from a limited number of geotagged posts has proven to be quite useful. In our survey, we have achieved accuracy rates of over 86% in matching Twitter user profile locations with their inferred home locations derived from geotagged posts.

  20. Increased Throwing Accuracy Improves Children's Catching Performance in a Ball-Catching Task from the Movement Assessment Battery (MABC-2).

    PubMed

    Dirksen, Tim; De Lussanet, Marc H E; Zentgraf, Karen; Slupinski, Lena; Wagner, Heiko

    2016-01-01

    The Movement Assessment Battery for Children (MABC-2) is a functional test for identifying deficits in the motor performance of children. The test contains a ball-catching task that requires the children to catch a self-thrown ball with one hand. As the task can be executed with a variety of different catching strategies, it is assumed that the task success can also vary considerably. Even though it is not clear, whether the performance merely depends on the catching skills or also to some extent on the throwing skills, the MABC-2 takes into account only the movement outcome. Therefore, the purpose of the current study was to examine (1) to what extent the throwing accuracy has an effect on the children's catching performance and (2) to what extent the throwing accuracy influences their choice of catching strategy. In line with the test manual, the children's catching performance was quantified on basis of the number of correctly caught balls. The throwing accuracy and the catching strategy were quantified by applying a kinematic analysis on the ball's trajectory and the hand movements. Based on linear regression analyses, we then investigated the relation between throwing accuracy, catching performance and catching strategy. The results show that an increased throwing accuracy is significantly correlated with an increased catching performance. Moreover, a higher throwing accuracy is significantly correlated with a longer duration of the hand on the ball's parabola, which indicates that throwing the ball more accurately could enable the children to effectively reduce the requirements on temporal precision. As the children's catching performance and their choice of catching strategy in the ball-catching task of the MABC-2 are substantially determined by their throwing accuracy, the test evaluation should not be based on the movement outcome alone, but should also take into account the children's throwing performance. Our findings could be of particular value for the

  1. Increased Throwing Accuracy Improves Children's Catching Performance in a Ball-Catching Task from the Movement Assessment Battery (MABC-2)

    PubMed Central

    Dirksen, Tim; De Lussanet, Marc H. E.; Zentgraf, Karen; Slupinski, Lena; Wagner, Heiko

    2016-01-01

    The Movement Assessment Battery for Children (MABC-2) is a functional test for identifying deficits in the motor performance of children. The test contains a ball-catching task that requires the children to catch a self-thrown ball with one hand. As the task can be executed with a variety of different catching strategies, it is assumed that the task success can also vary considerably. Even though it is not clear, whether the performance merely depends on the catching skills or also to some extent on the throwing skills, the MABC-2 takes into account only the movement outcome. Therefore, the purpose of the current study was to examine (1) to what extent the throwing accuracy has an effect on the children's catching performance and (2) to what extent the throwing accuracy influences their choice of catching strategy. In line with the test manual, the children's catching performance was quantified on basis of the number of correctly caught balls. The throwing accuracy and the catching strategy were quantified by applying a kinematic analysis on the ball's trajectory and the hand movements. Based on linear regression analyses, we then investigated the relation between throwing accuracy, catching performance and catching strategy. The results show that an increased throwing accuracy is significantly correlated with an increased catching performance. Moreover, a higher throwing accuracy is significantly correlated with a longer duration of the hand on the ball's parabola, which indicates that throwing the ball more accurately could enable the children to effectively reduce the requirements on temporal precision. As the children's catching performance and their choice of catching strategy in the ball-catching task of the MABC-2 are substantially determined by their throwing accuracy, the test evaluation should not be based on the movement outcome alone, but should also take into account the children's throwing performance. Our findings could be of particular value for the

  2. PLÉIADES Project: Assessment of Georeferencing Accuracy, Image Quality, Pansharpening Performence and Dsm/dtm Quality

    NASA Astrophysics Data System (ADS)

    Topan, Hüseyin; Cam, Ali; Özendi, Mustafa; Oruç, Murat; Jacobsen, Karsten; Taşkanat, Talha

    2016-06-01

    Pléiades 1A and 1B are twin optical satellites of Optical and Radar Federated Earth Observation (ORFEO) program jointly running by France and Italy. They are the first satellites of Europe with sub-meter resolution. Airbus DS (formerly Astrium Geo) runs a MyGIC (formerly Pléiades Users Group) program to validate Pléiades images worldwide for various application purposes. The authors conduct three projects, one is within this program, the second is supported by BEU Scientific Research Project Program, and the third is supported by TÜBİTAK. Assessment of georeferencing accuracy, image quality, pansharpening performance and Digital Surface Model/Digital Terrain Model (DSM/DTM) quality subjects are investigated in these projects. For these purposes, triplet panchromatic (50 cm Ground Sampling Distance (GSD)) and VNIR (2 m GSD) Pléiades 1A images were investigated over Zonguldak test site (Turkey) which is urbanised, mountainous and covered by dense forest. The georeferencing accuracy was estimated with a standard deviation in X and Y (SX, SY) in the range of 0.45m by bias corrected Rational Polynomial Coefficient (RPC) orientation, using ~170 Ground Control Points (GCPs). 3D standard deviation of ±0.44m in X, ±0.51m in Y, and ±1.82m in Z directions have been reached in spite of the very narrow angle of convergence by bias corrected RPC orientation. The image quality was also investigated with respect to effective resolution, Signal to Noise Ratio (SNR) and blur coefficient. The effective resolution was estimated with factor slightly below 1.0, meaning that the image quality corresponds to the nominal resolution of 50cm. The blur coefficients were achieved between 0.39-0.46 for triplet panchromatic images, indicating a satisfying image quality. SNR is in the range of other comparable space borne images which may be caused by de-noising of Pléiades images. The pansharpened images were generated by various methods, and are validated by most common statistical

  3. On the Spatial and Temporal Accuracy of Overset Grid Methods for Moving Body Problems

    NASA Technical Reports Server (NTRS)

    Meakin, Robert L.

    1996-01-01

    A study of numerical attributes peculiar to an overset grid approach to unsteady aerodynamics prediction is presented. Attention is focused on the effect of spatial error associated with interpolation of intergrid boundary conditions and temporal error associated with explicit update of intergrid boundary points on overall solution accuracy. A set of numerical experiments are used to verify whether or not the use of simple interpolation for intergrid boundary conditions degrades the formal accuracy of a conventional second-order flow solver, and to quantify the error associated with explicit updating of intergrid boundary points. Test conditions correspond to the transonic regime. The validity of the numerical results presented here are established by comparison with existing numerical results of documented accuracy, and by direct comparison with experimental results.

  4. An Accuracy Evaluation of Unstructured Node-Centred Finite Volume Methods

    NASA Technical Reports Server (NTRS)

    Svard, Magnus; Gong, Jing; Nordstrom, Jan

    2006-01-01

    Node-centred edge-based finite volume approximations are very common in computational fluid dynamics since they are assumed to run on structured, unstructured and even on mixed grids. We analyse the accuracy properties of both first and second derivative approximations and conclude that these schemes can not be used on arbitrary grids as is often assumed. For the Euler equations first-order accuracy can be obtained if care is taken when constructing the grid. For the Navier-Stokes equations, the grid restrictions are so severe that these finite volume schemes have little advantage over structured finite difference schemes. Our theoretical results are verified through extensive computations.

  5. Analysis of accuracy of digital elevation models created from captured data by digital photogrammetry method

    NASA Astrophysics Data System (ADS)

    Hudec, P.

    2011-12-01

    A digital elevation model (DEM) is an important part of many geoinformatic applications. For the creation of DEM, spatial data collected by geodetic measurements in the field, photogrammetric processing of aerial survey photographs, laser scanning and secondary sources (analogue maps) are used. It is very important from a user's point of view to know the vertical accuracy of a DEM. The article describes the verification of the vertical accuracy of a DEM for the region of Medzibodrožie, which was created using digital photogrammetry for the purposes of water resources management and modeling and resolving flood cases based on geodetic measurements in the field.

  6. Assessing Accuracy of Exchange-Correlation Functionals for the Description of Atomic Excited States

    NASA Astrophysics Data System (ADS)

    Makowski, Marcin; Hanas, Martyna

    2016-09-01

    The performance of exchange-correlation functionals for the description of atomic excitations is investigated. A benchmark set of excited states is constructed and experimental data is compared to Time-Dependent Density Functional Theory (TDDFT) calculations. The benchmark results show that for the selected group of functionals good accuracy may be achieved and the quality of predictions provided is competitive to computationally more demanding coupled-cluster approaches. Apart from testing the standard TDDFT approaches, also the role of self-interaction error plaguing DFT calculations and the adiabatic approximation to the exchange-correlation kernels is given some insight.

  7. Assessment methods for the evaluation of vitiligo.

    PubMed

    Alghamdi, K M; Kumar, A; Taïeb, A; Ezzedine, K

    2012-12-01

    There is no standardized method for assessing vitiligo. In this article, we review the literature from 1981 to 2011 on different vitiligo assessment methods. We aim to classify the techniques available for vitiligo assessment as subjective, semi-objective or objective; microscopic or macroscopic; and as based on morphometry or colorimetry. Macroscopic morphological measurements include visual assessment, photography in natural or ultraviolet light, photography with computerized image analysis and tristimulus colorimetry or spectrophotometry. Non-invasive micromorphological methods include confocal laser microscopy (CLM). Subjective methods include clinical evaluation by a dermatologist and a vitiligo disease activity score. Semi-objective methods include the Vitiligo Area Scoring Index (VASI) and point-counting methods. Objective methods include software-based image analysis, tristimulus colorimetry, spectrophotometry and CLM. Morphometry is the measurement of the vitiliginous surface area, whereas colorimetry quantitatively analyses skin colour changes caused by erythema or pigment. Most methods involve morphometry, except for the chromameter method, which assesses colorimetry. Some image analysis software programs can assess both morphometry and colorimetry. The details of these programs (Corel Draw, Image Pro Plus, AutoCad and Photoshop) are discussed in the review. Reflectance confocal microscopy provides real-time images and has great potential for the non-invasive assessment of pigmentary lesions. In conclusion, there is no single best method for assessing vitiligo. This review revealed that VASI, the rule of nine and Wood's lamp are likely to be the best techniques available for assessing the degree of pigmentary lesions and measuring the extent and progression of vitiligo in the clinic and in clinical trials.

  8. An Assessment of the Accuracy of Admittance and Coherence Estimates Using Synthetic Data

    NASA Astrophysics Data System (ADS)

    Crosby, A.

    2006-12-01

    The estimation of the effective elastic thickness of the lithosphere (T_e) using spectral relationships between gravity and topography has become a controversial topic in recent years. However, one area which has received relatively little attention is the bias in estimates of T_e and the internal loading fraction (F_2) which results from spectral leakage and noise when using the multi-tapered free-air admittance method. In this study, I use grids of synthetic data to assess the magnitude of that bias. I also assess the bias which occurs when T_e within other planets is estimated using the admittance between observed and topographic line-of-sight accelerations of orbiting satellites. I find that leakage can cause the estimated admittance and coherence to be significantly in error, but only if the box in which they are estimated is too small. The definition of `small' depends on the redness of the gravity spectrum. On the Earth, there is minimal error in the estimate of T_e if the admittance between surface gravity and topography is estimated within a box at least 3000-km-wide. When the true T_e is less than 20~km and the true coherence is high, the errors in the estimate of T_e are mostly less than 5~km for all box sizes greater than 1000~km. On the other hand, when the true T_e is greater than 20~km and the box size is 1000~km, the best-fit T_e is likely to be at least 5-10~km less than the true T_e. Even when the true coherence is high, it is not possible to use the free-air admittance to distinguish between real and spurious small fractions of internal loading when the boxes are smaller than 2000~km in size. Furthermore, the trade-off between T_e and F_2 means that even small amounts of leakage can shift the best-fit values of T_e and F_2 by an appreciable amount when the true F_2 is greater than zero. Geological noise in the gravity is caused by subsurface loads, the flexural surface expression of which has been erased by erosion and deposition. I find that

  9. Assessment of mesoscopic particle-based methods in microfluidic geometries

    NASA Astrophysics Data System (ADS)

    Zhao, Tongyang; Wang, Xiaogong; Jiang, Lei; Larson, Ronald G.

    2013-08-01

    We assess the accuracy and efficiency of two particle-based mesoscopic simulation methods, namely, Dissipative Particle Dynamics (DPD) and Stochastic Rotation Dynamics (SRD) for predicting a complex flow in a microfluidic geometry. Since both DPD and SRD use soft or weakly interacting particles to carry momentum, both methods contain unavoidable inertial effects and unphysically high fluid compressibility. To assess these effects, we compare the predictions of DPD and SRD for both an exact Stokes-flow solution and nearly exact solutions at finite Reynolds numbers from the finite element method for flow in a straight channel with periodic slip boundary conditions. This flow represents a periodic electro-osmotic flow, which is a complex flow with an analytical solution for zero Reynolds number. We find that SRD is roughly ten-fold faster than DPD in predicting the flow field, with better accuracy at low Reynolds numbers. However, SRD has more severe problems with compressibility effects than does DPD, which limits the Reynolds numbers attainable in SRD to around 25-50, while DPD can achieve Re higher than this before compressibility effects become too large. However, since the SRD method runs much faster than DPD does, we can afford to enlarge the number of grid cells in SRD to reduce the fluid compressibility at high Reynolds number. Our simulations provide a method to estimate the range of conditions for which SRD or DPD is preferable for mesoscopic simulations.

  10. Accuracy assessment of NOAA gridded daily reference evapotranspiration for the Texas High Plains

    USGS Publications Warehouse

    Moorhead, Jerry; Gowda, Prasanna H.; Hobbins, Michael; Senay, Gabriel; Paul, George; Marek, Thomas; Porter, Dana

    2015-01-01

    The National Oceanic and Atmospheric Administration (NOAA) provides daily reference evapotranspiration (ETref) maps for the contiguous United States using climatic data from North American Land Data Assimilation System (NLDAS). This data provides large-scale spatial representation of ETref, which is essential for regional scale water resources management. Data used in the development of NOAA daily ETref maps are derived from observations over surfaces that are different from short (grass — ETos) or tall (alfalfa — ETrs) reference crops, often in nonagricultural settings, which carries an unknown discrepancy between assumed and actual conditions. In this study, NOAA daily ETos and ETrs maps were evaluated for accuracy, using observed data from the Texas High Plains Evapotranspiration (TXHPET) network. Daily ETos, ETrs and the climatic data (air temperature, wind speed, and solar radiation) used for calculating ETref were extracted from the NOAA maps for TXHPET locations and compared against ground measurements on reference grass surfaces. NOAA ETrefmaps generally overestimated the TXHPET observations (1.4 and 2.2 mm/day ETos and ETrs, respectively), which may be attributed to errors in the NLDAS modeled air temperature and wind speed, to which reference ETref is most sensitive. Therefore, a bias correction to NLDAS modeled air temperature and wind speed data, or adjustment to the resulting NOAA ETref, may be needed to improve the accuracy of NOAA ETref maps.

  11. Assessing the speed--accuracy trade-off effect on the capacity of information processing.

    PubMed

    Donkin, Chris; Little, Daniel R; Houpt, Joseph W

    2014-06-01

    The ability to trade accuracy for speed is fundamental to human decision making. The speed-accuracy trade-off (SAT) effect has received decades of study, and is well understood in relatively simple decisions: collecting more evidence before making a decision allows one to be more accurate but also slower. The SAT in more complex paradigms has been given less attention, largely due to limits in the models and statistics that can be applied to such tasks. Here, we have conducted the first analysis of the SAT in multiple signal processing, using recently developed technologies for measuring capacity that take into account both response time and choice probability. We show that the primary influence of caution in our redundant-target experiments is on the threshold amount of evidence required to trigger a response. However, in a departure from the usual SAT effect, we found that participants strategically ignored redundant information when they were forced to respond quickly, but only when the additional stimulus was reliably redundant. Interestingly, because the capacity of the system was severely limited on redundant-target trials, ignoring additional targets meant that processing was more efficient when making fast decisions than when making slow and accurate decisions, where participants' limited resources had to be divided between the 2 stimuli.

  12. Development of a liquid chromatography-mass spectrometry method for the high-accuracy determination of creatinine in serum.

    PubMed

    Stokes, Peter; O'Connor, Gavin

    2003-08-25

    An LC-MS method for the high-accuracy determination of creatinine in serum has been developed and used to provide results for an international measurement evaluation programme (IMEP) and the Comité Consultatif pour la Quantité de Matière (CCQM) international inter-laboratory studies. An assessment of different sample preparation methods, including ion-exchange chromatography, solid-phase extraction, plasma ultrafiltration and ethanol protein precipitation, revealed that no bias or reduced precision was associated with the quicker less extensive clean-up methods, when using liquid chromatography-isotope dilution mass spectrometry (LC-IDMS) for quantitation. A number of different calibration regimes were also investigated. External calibration was shown to provide adequate calibration for most routine analysis with a relative associated expanded uncertainty (k=2) of 6% at the 95% confidence level. The use of a non-isotopically labelled internal standard was shown to improve the relative expanded uncertainty (k=2) to 4%. However, the difference in retention time between the internal standard and the creatinine was such that a matrix interferent produced an observed bias of over 16%. The use of an isotopically labelled internal standard was shown to reduce any bias to less than 0.2% with an expanded uncertainty (k=2) of less than 0.3%. The developed method was then used, in a blind trial organised jointly by IMEP and CCQM, to determine the amount of creatinine in human serum. The method performed well against the established reference method of ion-exchange chromatography followed by derivatisation gas chromatography (GC)-IDMS. The observed difference between the values determined by LC-IDMS and the key comparison reference value (average of all the submitted results) was less than 0.3%. The biggest advantage of the described method is in the speed of analysis. With a chromatographic run time of less than 10 min and sample preparation consisting of a simple protein

  13. Effect of the revisit interval and temporal upscaling methods on the accuracy of remotely sensed evapotranspiration estimates

    NASA Astrophysics Data System (ADS)

    Alfieri, Joseph G.; Anderson, Martha C.; Kustas, William P.; Cammalleri, Carmelo

    2017-01-01

    Accurate spatially distributed estimates of actual evapotranspiration (ET) derived from remotely sensed data are critical to a broad range of practical and operational applications. However, due to lengthy return intervals and cloud cover, data acquisition is not continuous over time, particularly for satellite sensors operating at medium ( ˜ 100 m) or finer resolutions. To fill the data gaps between clear-sky data acquisitions, interpolation methods that take advantage of the relationship between ET and other environmental properties that can be continuously monitored are often used. This study sought to evaluate the accuracy of this approach, which is commonly referred to as temporal upscaling, as a function of satellite revisit interval. Using data collected at 20 Ameriflux sites distributed throughout the contiguous United States and representing four distinct land cover types (cropland, grassland, forest, and open-canopy) as a proxy for perfect retrievals on satellite overpass dates, this study assesses daily ET estimates derived using five different reference quantities (incident solar radiation, net radiation, available energy, reference ET, and equilibrium latent heat flux) and three different interpolation methods (linear, cubic spline, and Hermite spline). Not only did the analyses find that the temporal autocorrelation, i.e., persistence, of all of the reference quantities was short, it also found that those land cover types with the greatest ET exhibited the least persistence. This carries over to the error associated with both the various scaled quantities and flux estimates. In terms of both the root mean square error (RMSE) and mean absolute error (MAE), the errors increased rapidly with increasing return interval following a logarithmic relationship. Again, those land cover types with the greatest ET showed the largest errors. Moreover, using a threshold of 20 % relative error, this study indicates that a return interval of no more than 5 days is

  14. Accuracy of a hybrid finite-element method for solving a scattering Schrödinger equation

    NASA Astrophysics Data System (ADS)

    Power, Joseph; Rawitscher, George

    2012-12-01

    This hybrid method [finite-element discrete variable representation (FE-DVR)], introduced by Resigno and McCurdy [Phys. Rev. APLRAAN1050-294710.1103/PhysRevA.62.032706 62, 032706 (2000)], uses Lagrange polynomials in each partition, rather than “hat” functions or Gaussian functions. These polynomials are discrete variable representation functions, and they are orthogonal under the Gauss-Lobatto quadrature discretization approximation. Accuracy analyses of this method are performed for the case of a one-dimensional Schrödinger equation with various types of local and nonlocal potentials for scattering boundary conditions. The accuracy is ascertained by a comparison with a spectral Chebyshev integral equation method, accurate to 1:10-11. For an accuracy of the phase shift of 1:10-8, the FE-DVR method is found to be 100 times faster than a sixth-order finite-difference method (Numerov), it is easy to program, and it can routinely achieve an accuracy of better than 1:10-8 for the numerical examples studied.

  15. Accuracy of a hybrid finite-element method for solving a scattering Schrödinger equation.

    PubMed

    Power, Joseph; Rawitscher, George

    2012-12-01

    This hybrid method [finite-element discrete variable representation (FE-DVR)], introduced by Resigno and McCurdy [Phys. Rev. A 62, 032706 (2000)], uses Lagrange polynomials in each partition, rather than "hat" functions or Gaussian functions. These polynomials are discrete variable representation functions, and they are orthogonal under the Gauss-Lobatto quadrature discretization approximation. Accuracy analyses of this method are performed for the case of a one-dimensional Schrödinger equation with various types of local and nonlocal potentials for scattering boundary conditions. The accuracy is ascertained by a comparison with a spectral Chebyshev integral equation method, accurate to 1:10⁻¹¹. For an accuracy of the phase shift of 1:10⁻⁸, the FE-DVR method is found to be 100 times faster than a sixth-order finite-difference method (Numerov), it is easy to program, and it can routinely achieve an accuracy of better than 1:10⁻⁸ for the numerical examples studied.

  16. Accuracy of plant specimen disease severity estimates: concepts, history, methods, ramifications and challenges for the future

    Technology Transfer Automated Retrieval System (TEKTRAN)

    Knowledge of the extent of the symptoms of a plant disease, generally referred to as severity, is key to both fundamental and applied aspects of plant pathology. Most commonly, severity is obtained visually and the accuracy of each estimate (closeness to the actual value) by individual raters is par...

  17. Presentation accuracy of the web revisited: animation methods in the HTML5 era.

    PubMed

    Garaizar, Pablo; Vadillo, Miguel A; López-de-Ipiña, Diego

    2014-01-01

    Using the Web to run behavioural and social experiments quickly and efficiently has become increasingly popular in recent years, but there is some controversy about the suitability of using the Web for these objectives. Several studies have analysed the accuracy and precision of different web technologies in order to determine their limitations. This paper updates the extant evidence about presentation accuracy and precision of the Web and extends the study of the accuracy and precision in the presentation of multimedia stimuli to HTML5-based solutions, which were previously untested. The accuracy and precision in the presentation of visual content in classic web technologies is acceptable for use in online experiments, although some results suggest that these technologies should be used with caution in certain circumstances. Declarative animations based on CSS are the best alternative when animation intervals are above 50 milliseconds. The performance of procedural web technologies based on the HTML5 standard is similar to that of previous web technologies. These technologies are being progressively adopted by the scientific community and have promising futures, which makes their use advisable to utilizing more obsolete technologies.

  18. Placement accuracy gauge for electrical components and method of using same

    SciTech Connect

    Biggs, P.M.; Dancer, L.K.; Yerganian, S.S.

    1988-10-11

    This patent describes a placement accuracy gauge for checking the accuracy of a machine which positions discrete electrical components on printed circuit boards wherein the machine utilizes a positioning head for releasably gripping the components, the machine effecting movement of the components and printed circuit boards relative to one another both orthogonally and rotationally; the placement accuracy gauge comprising: a substrate of transparent material having grid lines thereon and being dimensioned substantially the same as one of the printed circuit boards processed by the machine, the substrate having a surface divided into first, second and third regions; the grid lines in the first region being positioned in circle divided into segments with each segment having a line therethrough corresponding to a non-standard angle of machine rotation, the grid lines in each segment being parallel and perpendicular to the selected angle line; the second region including an array of grid lines arranged in target areas with each target area including two orthogonal grid loines and two diagonal grid lines with all grid lines intersecting at a common point, the second region further including an array of fiducial points recognizable by the machine; the third region having a plurality of grid lines arranged orthogonally with respect to one another in a rectangular area; whereby when the placement accuracy gauge is placed in the machine and components are placed within the grid areas and held thereto by a layer of adhesive.

  19. Presentation Accuracy of the Web Revisited: Animation Methods in the HTML5 Era

    PubMed Central

    Garaizar, Pablo; Vadillo, Miguel A.; López-de-Ipiña, Diego

    2014-01-01

    Using the Web to run behavioural and social experiments quickly and efficiently has become increasingly popular in recent years, but there is some controversy about the suitability of using the Web for these objectives. Several studies have analysed the accuracy and precision of different web technologies in order to determine their limitations. This paper updates the extant evidence about presentation accuracy and precision of the Web and extends the study of the accuracy and precision in the presentation of multimedia stimuli to HTML5-based solutions, which were previously untested. The accuracy and precision in the presentation of visual content in classic web technologies is acceptable for use in online experiments, although some results suggest that these technologies should be used with caution in certain circumstances. Declarative animations based on CSS are the best alternative when animation intervals are above 50 milliseconds. The performance of procedural web technologies based on the HTML5 standard is similar to that of previous web technologies. These technologies are being progressively adopted by the scientific community and have promising futures, which makes their use advisable to utilizing more obsolete technologies. PMID:25302791

  20. Methods of assessment of antiepileptic drugs.

    PubMed Central

    Milligan, N; Richens, A

    1981-01-01

    Epilepsy is a symptom with protean manifestations and as such it is a difficult disease in which to carry out a therapeutic trial. The methods available to research workers for the assessment of new antiepileptic drugs are hampered by the fact that epilepsy is a fluctuant condition. Although it is a chronic disorder open to study using cross-over trials and within-patient comparisons, accurate assessment cannot be easily made at any one point in time. Research workers are therefore automatically placed at a time factor disadvantage and this is especially so for those searching for quick methods of evaluating new compounds. The need for a quick and reliable method of assessing a new antiepileptic drug has long been appreciated. This article will discuss the methods currently available and we will begin by considering the most commonly used method of assessment with particular reference to some of the problems involved in conducting a controlled clinical trial in epilepsy. PMID:7272157

  1. Validity and accuracy of maternal tactile assessment for fever in under-five children in North Central Nigeria: a cross-sectional study

    PubMed Central

    Abdulkadir, Mohammed Baba; Johnson, Wahab Babatunde Rotimi; Ibraheem, Rasheedah Mobolaji

    2014-01-01

    Objectives This study seeks to determine not only the reliability of parental touch in detecting fever as compared to rectal thermometry in under-five children, but also the sociodemographic factors that may predict its reliability. Setting The study was carried out in the Emergency Paediatric Unit of a tertiary hospital in North Central Nigeria. Participants 409 children aged less than 5 years with a history of fever in the 48 h prior to presentation and their mothers were recruited consecutively. All the children recruited completed the study. Children with clinical parameters suggestive of shock, and those who were too ill, were excluded from the study. Primary and secondary outcome measures The primary outcome was the proportion of mothers who could accurately predict if their child was febrile or not (defined by rectal temperature) using tactile assessment only. Secondary outcomes were the validity and accuracy of touch in detecting fever and factors related to its accuracy. Results About 85% of the children were febrile using rectal thermometry. The sensitivity, specificity, positive predictive and negative predictive values for touch as a screening tool were 63%, 54%, 88.3% and 21%, respectively. High maternal socioeconomic status and low maternal age influenced positively the accuracy of touch in correctly determining the presence or absence of fever. Conclusions This study has shown that tactile assessment of temperature is not reliable and that absence of fever in a previously febrile child should be confirmed by objective methods of temperature measurement. PMID:25304190

  2. Accuracy of maximum likelihood and least-squares estimates in the lidar slope method with noisy data.

    PubMed

    Eberhard, Wynn L

    2017-04-01

    The maximum likelihood estimator (MLE) is derived for retrieving the extinction coefficient and zero-range intercept in the lidar slope method in the presence of random and independent Gaussian noise. Least-squares fitting, weighted by the inverse of the noise variance, is equivalent to the MLE. Monte Carlo simulations demonstrate that two traditional least-squares fitting schemes, which use different weights, are less accurate. Alternative fitting schemes that have some positive attributes are introduced and evaluated. The principal factors governing accuracy of all these schemes are elucidated. Applying these schemes to data with Poisson rather than Gaussian noise alters accuracy little, even when the signal-to-noise ratio is low. Methods to estimate optimum weighting factors in actual data are presented. Even when the weighting estimates are coarse, retrieval accuracy declines only modestly. Mathematical tools are described for predicting retrieval accuracy. Least-squares fitting with inverse variance weighting has optimum accuracy for retrieval of parameters from single-wavelength lidar measurements when noise, errors, and uncertainties are Gaussian distributed, or close to optimum when only approximately Gaussian.

  3. Assessment of accuracy of adopted centre of mass corrections for the Etalon geodetic satellites

    NASA Astrophysics Data System (ADS)

    Appleby, Graham; Dunn, Peter; Otsubo, Toshimichi; Rodriguez, Jose

    2016-04-01

    Accurate centre-of-mass corrections are key parameters in the analysis of satellite laser ranging observations. In order to meet current accuracy requirements, the vector from the reflection point of a laser retroreflector array to the centre of mass of the orbiting spacecraft must be known with mm-level accuracy. In general, the centre-of-mass correction will be dependent on the characteristics of the target (geometry, constr