Science.gov

Sample records for acceptable error range

  1. Tropospheric range error parameters: Further studies

    NASA Technical Reports Server (NTRS)

    Hopfield, H. S.

    1972-01-01

    Improved parameters are presented for predicting the tropospheric effect on electromagnetic range measurements from surface meteorological data. Parameters are given for computing the dry component of the zenith radio range effect from surface pressure alone with an rms error of 1 to 2 mm, or the total range effect from the dry and wet components of the surface refractivity, N, and a two-part quartic profile model. The parameters were obtained from meteorological balloon data with improved procedures, including the conversion of the geopotential heights of the balloon data to actual or geometric heights before using the data. The revised values of the parameter k show more latitude variation than is accounted for by the variation of g. This excess variation of k indicates a small latitude variation in the mean molecular weight of air and yields information about the latitude-varying water vapor content of air.

  2. Preliminary error budget for an optical ranging system: Range, range rate, and differenced range observables

    NASA Technical Reports Server (NTRS)

    Folkner, W. M.; Finger, M. H.

    1990-01-01

    Future missions to the outer solar system or human exploration of Mars may use telemetry systems based on optical rather than radio transmitters. Pulsed laser transmission can be used to deliver telemetry rates of about 100 kbits/sec with an efficiency of several bits for each detected photon. Navigational observables that can be derived from timing pulsed laser signals are discussed. Error budgets are presented based on nominal ground stations and spacecraft-transceiver designs. Assuming a pulsed optical uplink signal, two-way range accuracy may approach the few centimeter level imposed by the troposphere uncertainty. Angular information can be achieved from differenced one-way range using two ground stations with the accuracy limited by the length of the available baseline and by clock synchronization and troposphere errors. A method of synchronizing the ground station clocks using optical ranging measurements is presented. This could allow differenced range accuracy to reach the few centimeter troposphere limit.

  3. Atmospheric refraction errors in laser ranging systems

    NASA Technical Reports Server (NTRS)

    Gardner, C. S.; Rowlett, J. R.

    1976-01-01

    The effects of horizontal refractivity gradients on the accuracy of laser ranging systems were investigated by ray tracing through three dimensional refractivity profiles. The profiles were generated by performing a multiple regression on measurements from seven or eight radiosondes, using a refractivity model which provided for both linear and quadratic variations in the horizontal direction. The range correction due to horizontal gradients was found to be an approximately sinusoidal function of azimuth having a minimum near 0 deg azimuth and a maximum near 180 deg azimuth. The peak to peak variation was approximately 5 centimeters at 10 deg elevation and decreased to less than 1 millimeter at 80 deg elevation.

  4. Statistics of the residual refraction errors in laser ranging data

    NASA Technical Reports Server (NTRS)

    Gardner, C. S.

    1977-01-01

    A theoretical model for the range error covariance was derived by assuming that the residual refraction errors are due entirely to errors in the meteorological data which are used to calculate the atmospheric correction. The properties of the covariance function are illustrated by evaluating the theoretical model for the special case of a dense network of weather stations uniformly distributed within a circle.

  5. Poster error probability in the Mu-11 Sequential Ranging System

    NASA Technical Reports Server (NTRS)

    Coyle, C. W.

    1981-01-01

    An expression is derived for the posterior error probability in the Mu-2 Sequential Ranging System. An algorithm is developed which closely bounds the exact answer and can be implemented in the machine software. A computer simulation is provided to illustrate the improved level of confidence in a ranging acquisition using this figure of merit as compared to that using only the prior probabilities. In a simulation of 20,000 acquisitions with an experimentally determined threshold setting, the algorithm detected 90% of the actual errors and made false indication of errors on 0.2% of the acquisitions.

  6. Laser ranging error budget for the Topex/Poseidon satellite

    NASA Technical Reports Server (NTRS)

    Schwartz, Jon A.

    1990-01-01

    A laser ranging error budget is detailed, and a specific error budget is derived for the Topex/Poseidon satellite. A ranging uncertainty of 0.76 cm is predicted for Topex/Poseidon at 20 deg elevation using the presently designed laser retroreflector array and only modest improvements in present system operations. Atmospheric refraction and satellite attitude effects cause the predicted range error to vary with satellite elevation angle from 0.71 cm at zenith to 0.76 cm at 20 deg elevation. This a priori error budget compares well with the about 1.2-cm rms a posteriori polynomial orbital fit using existing data taken for an extant satellite of similar size and orbit.

  7. Error analysis of combined stereo/optical-flow passive ranging

    NASA Technical Reports Server (NTRS)

    Barniv, Yair

    1991-01-01

    The motion of an imaging sensor causes each imaged point of the scene to correspondingly describe a time trajectory on the image plane. The trajectories of all imaged points are reminiscent of a flow (e.g., of liquid) which is the source of the term 'optical flow'. Optical-flow ranging is a method by which the stream of two-dimensional images obtained from a forward-looking forward-moving passive sensor is used to compute depth (or range) to points in the field of view. Another well-known ranging method consists of triangulation based on stereo images obtained from at least two stationary sensors. In this paper we analyze the potential accuracies of a combined optical flow and stereo passive-ranging system in the context of helicopter nap-of-the-earth obstacle avoidance. The Cramer-Rao lower bound is developed for the combined system under the assumption of an unknown angular bias error common to both cameras of a stereo pair. It is shown that the depth accuracy degradations caused by a bias error is negligible for a combined optical-flow and stereo system as compared to a monocular optical-flow system.

  8. Entropy-Based TOA Estimation and SVM-Based Ranging Error Mitigation in UWB Ranging Systems.

    PubMed

    Yin, Zhendong; Cui, Kai; Wu, Zhilu; Yin, Liang

    2015-01-01

    The major challenges for Ultra-wide Band (UWB) indoor ranging systems are the dense multipath and non-line-of-sight (NLOS) problems of the indoor environment. To precisely estimate the time of arrival (TOA) of the first path (FP) in such a poor environment, a novel approach of entropy-based TOA estimation and support vector machine (SVM) regression-based ranging error mitigation is proposed in this paper. The proposed method can estimate the TOA precisely by measuring the randomness of the received signals and mitigate the ranging error without the recognition of the channel conditions. The entropy is used to measure the randomness of the received signals and the FP can be determined by the decision of the sample which is followed by a great entropy decrease. The SVM regression is employed to perform the ranging-error mitigation by the modeling of the regressor between the characteristics of received signals and the ranging error. The presented numerical simulation results show that the proposed approach achieves significant performance improvements in the CM1 to CM4 channels of the IEEE 802.15.4a standard, as compared to conventional approaches. PMID:26007726

  9. Entropy-Based TOA Estimation and SVM-Based Ranging Error Mitigation in UWB Ranging Systems

    PubMed Central

    Yin, Zhendong; Cui, Kai; Wu, Zhilu; Yin, Liang

    2015-01-01

    The major challenges for Ultra-wide Band (UWB) indoor ranging systems are the dense multipath and non-line-of-sight (NLOS) problems of the indoor environment. To precisely estimate the time of arrival (TOA) of the first path (FP) in such a poor environment, a novel approach of entropy-based TOA estimation and support vector machine (SVM) regression-based ranging error mitigation is proposed in this paper. The proposed method can estimate the TOA precisely by measuring the randomness of the received signals and mitigate the ranging error without the recognition of the channel conditions. The entropy is used to measure the randomness of the received signals and the FP can be determined by the decision of the sample which is followed by a great entropy decrease. The SVM regression is employed to perform the ranging-error mitigation by the modeling of the regressor between the characteristics of received signals and the ranging error. The presented numerical simulation results show that the proposed approach achieves significant performance improvements in the CM1 to CM4 channels of the IEEE 802.15.4a standard, as compared to conventional approaches. PMID:26007726

  10. Range of acceptable stimulus intensities: an estimator of dynamic range for intensive perceptual continua.

    PubMed

    Teghtsoonian, R; Teghtsoonian, M

    1997-07-01

    The dynamic range (DR) of a sensory system is the span (usually given in log units) from the lowest to highest intensities over which a continuously graded response is evoked, and may be a distinctive feature of each such system. Teghtsoonian (1971) proposed that, although DR varies widely over sensory systems, its subjective size (SDR) is invariant. Assuming the psychophysical power law, the exponent for any continuum is given by the ratio of subjective span to DR, both quantities expressed logarithmically. Thus, exponents are inversely related to the DR and many be interpreted as indexes of it. Because DR can be difficult or even dangerous to measure directly, we sought to define a smaller range representing some fixed proportion of DR that could be used in its place to test the hypothesis of an invariant subjective range. Observations manipulated the intensities of five target continua to produce the broadest range they found acceptable and reasonably comfortable, a range of acceptable stimulus intensities (RASIN). Combined with an assumed constant SDR (derived from previous research), RASINs accurately predicted exponents obtained by magnitude production from the same observers on the five continua, as well as exponents reported in the literature. PMID:9259639

  11. Atmospheric refraction effects on baseline error in satellite laser ranging systems

    NASA Technical Reports Server (NTRS)

    Im, K. E.; Gardner, C. S.

    1982-01-01

    Because of the mathematical complexities involved in exact analyses of baseline errors, it is not easy to isolate atmospheric refraction effects; however, by making certain simplifying assumptions about the ranging system geometry, relatively simple expressions can be derived which relate the baseline errors directly to the refraction errors. The results indicate that even in the absence of other errors, the baseline error for intercontinental baselines can be more than an order of magnitude larger than the refraction error.

  12. Decreasing range resolution of a SAR image to permit correction of motion measurement errors beyond the SAR range resolution

    SciTech Connect

    Doerry, Armin W.; Heard, Freddie E.; Cordaro, J. Thomas

    2010-07-20

    Motion measurement errors that extend beyond the range resolution of a synthetic aperture radar (SAR) can be corrected by effectively decreasing the range resolution of the SAR in order to permit measurement of the error. Range profiles can be compared across the slow-time dimension of the input data in order to estimate the error. Once the error has been determined, appropriate frequency and phase correction can be applied to the uncompressed input data, after which range and azimuth compression can be performed to produce a desired SAR image.

  13. 76 FR 37793 - Viking Range Corporation, Provisional Acceptance of a Settlement Agreement and Order

    Federal Register 2010, 2011, 2012, 2013, 2014

    2011-06-28

    ... COMMISSION Viking Range Corporation, Provisional Acceptance of a Settlement Agreement and Order AGENCY.... 16. Upon provisional acceptance of the Agreement by the Commission, the Agreement shall be placed on.... Upon the Commission's final acceptance of the Agreement and issuance of the final Order,...

  14. Close-range radar rainfall estimation and error analysis

    NASA Astrophysics Data System (ADS)

    van de Beek, C. Z.; Leijnse, H.; Hazenberg, P.; Uijlenhoet, R.

    2012-04-01

    It is well-known that quantitative precipitation estimation (QPE) is affected by many sources of error. The most important of these are 1) radar calibration, 2) wet radome attenuation, 3) rain attenuation, 4) vertical profile of reflectivity, 5) variations in drop size distribution, and 6) sampling effects. The study presented here is an attempt to separate and quantify these sources of error. For this purpose, QPE is performed very close to the radar (~1-2 km) so that 3), 4), and 6) will only play a minor role. Error source 5) can be corrected for because of the availability of two disdrometers (instruments that measure the drop size distribution). A 3-day rainfall event (25-27 August 2010) that produced more than 50 mm in De Bilt, The Netherlands is analyzed. Radar, rain gauge, and disdrometer data from De Bilt are used for this. It is clear from the analyses that without any corrections, the radar severely underestimates the total rain amount (only 25 mm). To investigate the effect of wet radome attenuation, stable returns from buildings close to the radar are analyzed. It is shown that this may have caused an underestimation up to ~4 dB. The calibration of the radar is checked by looking at received power from the sun. This turns out to cause another 1 dB of underestimation. The effect of variability of drop size distributions is shown to cause further underestimation. Correcting for all of these effects yields a good match between radar QPE and gauge measurements.

  15. Correction of motion measurement errors beyond the range resolution of a synthetic aperture radar

    SciTech Connect

    Doerry, Armin W.; Heard, Freddie E.; Cordaro, J. Thomas

    2008-06-24

    Motion measurement errors that extend beyond the range resolution of a synthetic aperture radar (SAR) can be corrected by effectively decreasing the range resolution of the SAR in order to permit measurement of the error. Range profiles can be compared across the slow-time dimension of the input data in order to estimate the error. Once the error has been determined, appropriate frequency and phase correction can be applied to the uncompressed input data, after which range and azimuth compression can be performed to produce a desired SAR image.

  16. Error analysis for a spaceborne laser ranging system

    NASA Technical Reports Server (NTRS)

    Pavlis, E. C.

    1979-01-01

    The dependence (or independence) of baseline accuracies, obtained from a typical mission of a spaceborne ranging system, on several factors is investigated. The emphasis is placed on a priori station information, but factors such as the elevation cut-off angle, the geometry of the network, the mean orbital height, and to a limited extent geopotential modeling are also examined. The results are obtained through simulations, but some theoretical justification is also given. Guidelines for freeing the results from these dependencies are suggested for most of the factors.

  17. Example Procedures for Developing Acceptance-Range Criteria for BESTEST-EX

    SciTech Connect

    Judkoff, R.; Polly, B.; Bianchi, M.; Neymark, J.

    2010-08-01

    This document provides an example procedure for establishing acceptance-range criteria to assess results from software undergoing BESTEST-EX. This example method for BESTEST-EX is a modified version of the method described in HERS BESTEST.

  18. Improved estimates of the range of errors on photomasks using measured values of skewness and kurtosis

    NASA Astrophysics Data System (ADS)

    Hamaker, Henry Chris

    1995-12-01

    Statistical process control (SPC) techniques often use six times the standard deviation sigma to estimate the range of errors within a process. Two assumptions are inherent in this choice of metric for the range: (1) the normal distribution adequately describes the errors, and (2) the fraction of errors falling within plus or minus 3 sigma, about 99.73%, is sufficiently large that we may consider the fraction occurring outside this range to be negligible. In state-of-the-art photomasks, however, the assumption of normality frequently breaks down, and consequently plus or minus 3 sigma is not a good estimate of the range of errors. In this study, we show that improved estimates for the effective maximum error Em, which is defined as the value for which 99.73% of all errors fall within plus or minus Em of the mean mu, may be obtained by quantifying the deviation from normality of the error distributions using the skewness and kurtosis of the error sampling. Data are presented indicating that in laser reticle- writing tools, Em less than or equal to 3 sigma. We also extend this technique for estimating the range of errors to specifications that are usually described by mu plus 3 sigma. The implications for SPC are examined.

  19. Modeling methodology for MLS range navigation system errors using flight test data

    NASA Technical Reports Server (NTRS)

    Karmali, M. S.; Phatak, A. V.

    1982-01-01

    Flight test data was used to develop a methodology for modeling MLS range navigation system errors. The data used corresponded to the constant velocity and glideslope approach segment of a helicopter landing trajectory. The MLS range measurement was assumed to consist of low frequency and random high frequency components. The random high frequency component was extracted from the MLS range measurements. This was done by appropriate filtering of the range residual generated from a linearization of the range profile for the final approach segment. This range navigation system error was then modeled as an autoregressive moving average (ARMA) process. Maximum likelihood techniques were used to identify the parameters of the ARMA process.

  20. Error analysis for Mariner Venus/Mercury 1973 conducted at the JPL Mesa west antenna range

    NASA Technical Reports Server (NTRS)

    Vincent, N. L.; Smith, C. A.; Brejcha, A. J.; Curtis, H. A.

    1973-01-01

    Theoretical analysis and experimental data are combined to yield the errors to be used with antenna gain, antenna patterns, and RF cable insertion loss measurements for the Mariner Venus-Mercury 1973 Flight Project. These errors apply to measurements conducted at the JPL Mesa, West Antenna Range, on the high gain antenna, low gain antenna, and RF coaxial cables.

  1. Assessment of an adjustment factor to model radar range dependent error

    NASA Astrophysics Data System (ADS)

    Sebastianelli, S.; Russo, F.; Napolitano, F.; Baldini, L.

    2012-09-01

    Quantitative radar precipitation estimates are affected by errors determined by many causes such as radar miscalibration, range degradation, attenuation, ground clutter, variability of Z-R relation, variability of drop size distribution, vertical air motion, anomalous propagation and beam-blocking. Range degradation (including beam broadening and sampling of precipitation at an increasing altitude)and signal attenuation, determine a range dependent behavior of error. The aim of this work is to model the range-dependent error through an adjustment factor derived from the G/R ratio trend against the range, where G and R are the corresponding rain gauge and radar rainfall amounts computed at each rain gauge location. Since range degradation and signal attenuation effects are negligible close to the radar, resultsshowthatwithin 40 km from radar the overall range error is independent of the distance from Polar 55C and no range-correction is needed. Nevertheless, up to this distance, the G/R ratiocan showa concave trend with the range, which is due to the melting layer interception by the radar beam during stratiform events.

  2. STUDY TO ESTABLISH THE ACCEPTANCE RANGE FOR PEROXYL RADICALS SCAVENGER CAPACITY OF NATURAL SOD.

    PubMed

    Lupu, Andreea-Roxana; Cremer, Lidia

    2015-01-01

    In the context of an emerging market of food supplements, the proven quality of the antioxidant products should be the main criteria for using them. The production process has to be carefully controlled and complementary tests are needed to demonstrate the correspondence between real and declared properties of final product. Using well characterized compounds with proven antioxidant activity in biological systems as reference brings a plus of rigorously to the testing protocol. The aim of this study was to determine the acceptance range for the antioxidant (peroxyl radicals scavenger) capacity of "Natural SOD" by using for comparison ascorbic acid (vitamin C). The established acceptance range complete our previous results concerning the antioxidant capacity of Natural SOD using validated ORAC method and creates premises for supplementary checking of the batches in the current production and improving the product quality. PMID:27328523

  3. Stray signal requirements for compact range reflectors based on RCS measurement errors

    NASA Technical Reports Server (NTRS)

    Lee, Teh-Hong; Burnside, Walter D.

    1991-01-01

    The authors present a performance criterion for compact range reflectors such that their edge diffracted stray signal levels meet a reasonable radar cross section (RCS) measurement error requirement. It is shown by example that one of the significant error sources is the diffracted fields emanating from the edges or junctions of the reflector. This measurement error is demonstrated by placing a diagonal square flat plate in the target zone and rotating it to appropriate angles. These angles are determined by bisecting the plane wave and stray signal directions. This results in a peak bistatic measurement of the edge diffracted stray signal. It is proposed that the diagonal flat plate be used to evaluate new reflector designs as well as existing systems. A reasonable stray signal performance level has been developed so that new reflector systems can be characterized in terms of an RCS measurement error requirement.

  4. Comparing range data across the slow-time dimension to correct motion measurement errors beyond the range resolution of a synthetic aperture radar

    SciTech Connect

    Doerry, Armin W.; Heard, Freddie E.; Cordaro, J. Thomas

    2010-08-17

    Motion measurement errors that extend beyond the range resolution of a synthetic aperture radar (SAR) can be corrected by effectively decreasing the range resolution of the SAR in order to permit measurement of the error. Range profiles can be compared across the slow-time dimension of the input data in order to estimate the error. Once the error has been determined, appropriate frequency and phase correction can be applied to the uncompressed input data, after which range and azimuth compression can be performed to produce a desired SAR image.

  5. Using kriging to bound satellite ranging errors due to the ionosphere

    NASA Astrophysics Data System (ADS)

    Blanch, Juan

    The Global Positioning System (GPS) has the potential to become the primary navigational aid for civilian aircraft, thanks to satellite based augmentation systems (SBAS). SBAS systems, including the United State's Wide Area Augmentation System (WAAS), provide corrections and hard bounds on the user errors. The ionosphere is the largest and least predictable source of error. The only ionospheric information available to WAAS is a set of range delay measurements taken at reference stations. From this data, the master station must compute a real time estimate of the ionospheric delay and a hard error bound valid for any user. The variability of the ionospheric behavior has caused the confidence bounds corresponding to the ionosphere to be very large in WAAS. These ranging bounds translate into conservative bounds on user position error. These position error bounds (called protection levels) have values of 30 to 50 meters. Since these values fluctuate near the maximum tolerable limit, WAAS is not always available. In order to increase the availability of WAAS, we must decrease the confidence bounds corresponding to ionospheric uncertainty while maintaining integrity. In this work, I present an ionospheric estimation algorithm based on kriging. I first introduce a simple model of the Vertical Ionospheric Delay that captures both the deterministic behavior and the random behavior of the ionosphere. Under this model, the kriging method is optimal. More importantly kriging provides an estimation variance that can be translated into an error bound. However, this method must be modified for three reasons; first, the state of the ionosphere is unknown and can only be estimated through real time measurements; second, because of bandwidth constraints, the user cannot receive all the measurements and third there is noise in the measurements. I will show how these three obstacles can be overcome. The algorithm presented here provides a reduction in the error bound corresponding

  6. Sensitivity analysis of short-arc station coordinate determinations from range data. [geocentric coordinate range errors in satellite tracking

    NASA Technical Reports Server (NTRS)

    Schutz, B. E.; Tapley, B. D.

    1976-01-01

    The accurate determination of the geocentric coordinates of a tracking station is essential for most geodetic and geophysical satellite applications. Since most of these satellites are close to the earth, the geopotential model is a dominant source of error which significantly influences station coordinate determinations. Other sources, such as GM error and drag, also influence the accuracy of the station coordinate determination. One technique for reducing the effect of these errors is to use short-arcs consisting of a few passes of the satellite over the tracking station. This paper analyzes the sensitivity of short-arc station coordinate estimates to various errors in the physical model, to the number of observations, and to the station-satellite geometry using simulated as well as real data.

  7. Dose Uncertainties in IMPT for Oropharyngeal Cancer in the Presence of Anatomical, Range, and Setup Errors

    SciTech Connect

    Kraan, Aafke C.; Water, Steven van de; Teguh, David N.; Al-Mamgani, Abrahim; Madden, Tom; Kooy, Hanne M.; Heijmen, Ben J.M.; Hoogeman, Mischa S.

    2013-12-01

    Purpose: Setup, range, and anatomical uncertainties influence the dose delivered with intensity modulated proton therapy (IMPT), but clinical quantification of these errors for oropharyngeal cancer is lacking. We quantified these factors and investigated treatment fidelity, that is, robustness, as influenced by adaptive planning and by applying more beam directions. Methods and Materials: We used an in-house treatment planning system with multicriteria optimization of pencil beam energies, directions, and weights to create treatment plans for 3-, 5-, and 7-beam directions for 10 oropharyngeal cancer patients. The dose prescription was a simultaneously integrated boost scheme, prescribing 66 Gy to primary tumor and positive neck levels (clinical target volume-66 Gy; CTV-66 Gy) and 54 Gy to elective neck levels (CTV-54 Gy). Doses were recalculated in 3700 simulations of setup, range, and anatomical uncertainties. Repeat computed tomography (CT) scans were used to evaluate an adaptive planning strategy using nonrigid registration for dose accumulation. Results: For the recalculated 3-beam plans including all treatment uncertainty sources, only 69% (CTV-66 Gy) and 88% (CTV-54 Gy) of the simulations had a dose received by 98% of the target volume (D98%) >95% of the prescription dose. Doses to organs at risk (OARs) showed considerable spread around planned values. Causes for major deviations were mixed. Adaptive planning based on repeat imaging positively affected dose delivery accuracy: in the presence of the other errors, percentages of treatments with D98% >95% increased to 96% (CTV-66 Gy) and 100% (CTV-54 Gy). Plans with more beam directions were not more robust. Conclusions: For oropharyngeal cancer patients, treatment uncertainties can result in significant differences between planned and delivered IMPT doses. Given the mixed causes for major deviations, we advise repeat diagnostic CT scans during treatment, recalculation of the dose, and if required, adaptive

  8. Development of Algorithms and Error Analyses for the Short Baseline Lightning Detection and Ranging System

    NASA Technical Reports Server (NTRS)

    Starr, Stanley O.

    1998-01-01

    NASA, at the John F. Kennedy Space Center (KSC), developed and operates a unique high-precision lightning location system to provide lightning-related weather warnings. These warnings are used to stop lightning- sensitive operations such as space vehicle launches and ground operations where equipment and personnel are at risk. The data is provided to the Range Weather Operations (45th Weather Squadron, U.S. Air Force) where it is used with other meteorological data to issue weather advisories and warnings for Cape Canaveral Air Station and KSC operations. This system, called Lightning Detection and Ranging (LDAR), provides users with a graphical display in three dimensions of 66 megahertz radio frequency events generated by lightning processes. The locations of these events provide a sound basis for the prediction of lightning hazards. This document provides the basis for the design approach and data analysis for a system of radio frequency receivers to provide azimuth and elevation data for lightning pulses detected simultaneously by the LDAR system. The intent is for this direction-finding system to correct and augment the data provided by LDAR and, thereby, increase the rate of valid data and to correct or discard any invalid data. This document develops the necessary equations and algorithms, identifies sources of systematic errors and means to correct them, and analyzes the algorithms for random error. This data analysis approach is not found in the existing literature and was developed to facilitate the operation of this Short Baseline LDAR (SBLDAR). These algorithms may also be useful for other direction-finding systems using radio pulses or ultrasonic pulse data.

  9. Single-plane versus three-plane methods for relative range error evaluation of medium-range 3D imaging systems

    NASA Astrophysics Data System (ADS)

    MacKinnon, David K.; Cournoyer, Luc; Beraldin, J.-Angelo

    2015-05-01

    Within the context of the ASTM E57 working group WK12373, we compare the two methods that had been initially proposed for calculating the relative range error of medium-range (2 m to 150 m) optical non-contact 3D imaging systems: the first is based on a single plane (single-plane assembly) and the second on an assembly of three mutually non-orthogonal planes (three-plane assembly). Both methods are evaluated for their utility in generating a metric to quantify the relative range error of medium-range optical non-contact 3D imaging systems. We conclude that the three-plane assembly is comparable to the single-plane assembly with regard to quantification of relative range error while eliminating the requirement to isolate the edges of the target plate face.

  10. Target error for image-to-physical space registration: preliminary clinical results using laser range scanning

    NASA Astrophysics Data System (ADS)

    Cao, Aize; Miga, Michael I.; Dumpuri, P.; Ding, S.; Dawant, B. M.; Thompson, R. C.

    2007-03-01

    In this paper, preliminary results from an image-to-physical space registration platform are presented. The current platform employs traditional and novel methods of registration which use a variety of data sources to include: traditional synthetic skin-fiducial point-based registration, surface registration based on facial contours, brain feature point-based registration, brain vessel-to-vessel registration, and a more comprehensive cortical surface registration method that utilizes both geometric and intensity information from both the image volume and physical patient. The intraoperative face and cortical surfaces were digitized using a laser range scanner (LRS) capable of producing highly resolved textured point clouds. In two in vivo cases, a series of registrations were performed using these techniques and compared within the context of a true target error. One of the advantages of using a textured point cloud data stream is that true targets among the physical cortical surface and the preoperative image volume can be identified and used to assess image-to-physical registration methods. The results suggest that iterative closest point (ICP) method for intraoperative face surface registration is equivalent to point-based registration (PBR) method of skin fiducial markers. With regard to the initial image and physical space registration, for patient 1, mean target registration error (TRE) were 3.1+/-0.4 mm and 3.6 +/-0.9 mm for face ICP and skin fiducial PBR, respectively. For patient 2, the mean TRE were 5.7 +/-1.3 mm, and 6.6 +/-0.9 mm for face ICP and skin fiducial PBR, respectively. With regard to intraoperative cortical surface registration, SurfaceMI outperformed feature based PBR and vessel ICP with 1.7+/-1.8 mm for patient 1. For patient 2, the best result was achieved by using vessel ICP with 1.9+/-0.5 mm.

  11. Diagnosis of the source of medium-range forecast errors for an extratropical cyclone

    NASA Astrophysics Data System (ADS)

    Lamberson, William S.

    Medium-range forecasts of an extratropical cyclone set to impact Western Europe on 16 Dec 2011 consistently predicted that the cyclone would be intense, with a minimum pressure as low as 930-hPa and an expansive area of high winds. In actuality, the cyclone never deepened below 965-hPa and its impact on Western Europe was minimal. This study seeks to investigate and document sources of error in medium-range forecasts for this event. The dynamical processes responsible for an intense extratropical cyclone are diagnosed using five and six-day forecasts for the event from the 51-member European Center for Medium-Range Forecasts Ensemble Prediction System. To accomplish this, ensemble members with relatively accurate forecasts are compared against those that predicted an intense cyclone. The comparison indicates that accurate forecasts are characterized by a more intense ridge building event over Alaska on 11 Dec that occurred in associated with the Warm Conveyor Belt (WCB) of an intensifying extratropical cyclone over the Pacific Ocean. The ridge building event perturbed the waveguide and generated a Rossby wave packet, with a potent trough at the packet's leading edge. A more amplified ridge acted as a more intense perturbation on the waveguide and had the effect of increasing the eastward speed of the Rossby wave packet. The increased eastward speed of this wave packet did not allow the trough at its leading edge to phase with and enhance the development of the extratropical cyclone as it moved towards Western Europe. In contrast, the less amplified ridge building event seen in the members with a strong extratropical cyclone allowed the wave packet to move downstream at a slower speed and the trough at the packet's leading edge to phase with and enhance the development of the extratropical cyclone headed towards Western Europe.

  12. Technique-dependent Errors in the Realization of the ITRF Origin From Satellite Laser Ranging

    NASA Astrophysics Data System (ADS)

    Pavlis, E. C.; Kuzmicz-Cieslak, M.

    2012-12-01

    Over the past decade Satellite Laser Ranging (SLR) has focused on its unique strength of providing accurate observations of the origin and scale of the International Terrestrial Reference Frame (ITRF). The origin of the ITRF is defined to coincide with the center of mass of the Earth system (geocenter). SLR realizes this origin as the focal point of the tracked satellite orbits, and being the only (nominally) unbiased ranging technique, it provides the best realization for it. The goal of GGOS is to provide an ITRF with accuracy at epoch of 1 mm or better and a stability of 0.1 mm/y. In order to meet this stringent goal, Space Geodesy is taking a two-pronged approach: modernizing the engineering components (ground and space segments), and revising the modeling standards to take advantage of recent improvements in many areas of geophysical modeling for system Earth components. As we gain improved understanding of the Earth system components, space geodesy adjusts its underlying modeling of the system to better and more completely describe it. Similarly, from the engineering side we examine the observational process for improvement of the calibration and reduction procedures that will enhance the accuracy of the individual observations thence the final SLR products. Two areas that are currently under scrutiny are (a) the station-dependent and tracking-mode-dependent correction of the observations for the "center-of-mass-offset" of each satellite target, and (b) the station- and pass-dependent correction for the calibrated delay that refers each measurement to the nominal "zero" of the instrument. The former affects primarily the accuracy of the scale definition, while the latter affects both, the scale and the origin. However, because of the non-uniform data volume and non-symmetric geographic locations of the SLR stations, the major impact of the latter is on the definition of the origin. The ILRS is currently investigating the quality of models available for the

  13. Technique-Dependent Errors in the Satellite Laser Ranging Contributions to the ITRF

    NASA Astrophysics Data System (ADS)

    Pavlis, Erricos C.; Kuzmicz-Cieslak, Magdalena; König, Daniel

    2013-04-01

    Over the past decade Satellite Laser Ranging (SLR) has focused on its unique strength of providing accurate observations of the origin and scale of the International Terrestrial Reference Frame (ITRF). The origin of the ITRF is defined to coincide with the center of mass of the Earth system (geocenter). SLR realizes this origin as the focal point of the tracked satellite orbits, and being the only (nominally) unbiased ranging technique, it provides the best realization for it. The goal of GGOS is to provide an ITRF with accuracy at epoch of 1 mm or better and a stability of 0.1 mm/y. In order to meet this stringent goal, Space Geodesy is taking a two-pronged approach: modernizing the engineering components (ground and space segments), and revising the modeling standards to take advantage of recent improvements in many areas of geophysical modeling for system Earth components. As we gain improved understanding of the Earth system components, space geodesy adjusts its underlying modeling of the system to better and more completely describe it. Similarly, from the engineering side we examine the observational process for improvement of the calibration and reduction procedures that will enhance the accuracy of the individual observations thence the final SLR products. Two areas that are currently under scrutiny are (a) the station-dependent and tracking-mode-dependent correction of the observations for the "center-of-mass-offset" of each satellite target, and (b) the station- and pass-dependent correction for the calibrated delay that refers each measurement to the nominal "zero" of the instrument. The former affects primarily the accuracy of the scale definition, while the latter affects both, the scale and the origin. However, because of the non-uniform data volume and non-symmetric geographic locations of the SLR stations, the major impact of the latter is on the definition of the origin. The ILRS is currently investigating the quality of models available for the

  14. SU-E-T-550: Range Effects in Proton Therapy Caused by Systematic Errors in the Stoichiometric Calibration

    SciTech Connect

    Doolan, P; Dias, M; Collins Fekete, C; Seco, J

    2014-06-01

    Purpose: The procedure for proton treatment planning involves the conversion of the patient's X-ray CT from Hounsfield units into relative stopping powers (RSP), using a stoichiometric calibration curve (Schneider 1996). In clinical practice a 3.5% margin is added to account for the range uncertainty introduced by this process and other errors. RSPs for real tissues are calculated using composition data and the Bethe-Bloch formula (ICRU 1993). The purpose of this work is to investigate the impact that systematic errors in the stoichiometric calibration have on the proton range. Methods: Seven tissue inserts of the Gammex 467 phantom were imaged using our CT scanner. Their known chemical compositions (Watanabe 1999) were then used to calculate the theoretical RSPs, using the same formula as would be used for human tissues in the stoichiometric procedure. The actual RSPs of these inserts were measured using a Bragg peak shift measurement in the proton beam at our institution. Results: The theoretical calculation of the RSP was lower than the measured RSP values, by a mean/max error of - 1.5/-3.6%. For all seven inserts the theoretical approach underestimated the RSP, with errors variable across the range of Hounsfield units. Systematic errors for lung (average of two inserts), adipose and cortical bone were - 3.0/-2.1/-0.5%, respectively. Conclusion: There is a systematic underestimation caused by the theoretical calculation of RSP; a crucial step in the stoichiometric calibration procedure. As such, we propose that proton calibration curves should be based on measured RSPs. Investigations will be made to see if the same systematic errors exist for biological tissues. The impact of these differences on the range of proton beams, for phantoms and patient scenarios, will be investigated. This project was funded equally by the Engineering and Physical Sciences Research Council (UK) and Ion Beam Applications (Louvain-La-Neuve, Belgium)

  15. Bootstrap Standard Error and Confidence Intervals for the Correlation Corrected for Range Restriction: A Simulation Study

    ERIC Educational Resources Information Center

    Chan, Wai; Chan, Daniel W.-L.

    2004-01-01

    The standard Pearson correlation coefficient is a biased estimator of the true population correlation, ?, when the predictor and the criterion are range restricted. To correct the bias, the correlation corrected for range restriction, r-sub(c), has been recommended, and a standard formula based on asymptotic results for estimating its standard…

  16. Error Analysis of Combined Optical-Flow and Stereo Passive Ranging

    NASA Technical Reports Server (NTRS)

    Barniv, Yair

    1992-01-01

    The motion of an imaging sensor causes each imaged point of the seem to describe a time trajectory on the image plane. The trajectories of all imaged points are reminiscent of a flow (eg, of liquid) which is the source of the term "optical flow". Optical-flow ranging is a method by which the stream of two-dimensional images obtained from a forward-looking forward-moving passive sensor is used to compute range to points in the field of view. Another well-known ranging method consists of triangulation based on stereo images obtained from at least two stationary sensors. In this paper we analyze the potential accuracies of a combined optical flow and stereo passive-ranging system in the context of helicopter nap-of-the-earth obstacle avoidance. The Cramer-Rao lower bound (CRLB) is developed for the combined system under the assumption of a random angular misalignment common to both cameras of a stereo pair. It is shown that the range accuracy degradations caused by misalignment is negligible for a combined optical-flow and stereo system as compared with a monocular optical-flow system.

  17. Range camera calibration based on image sequences and dense comprehensive error statistics

    NASA Astrophysics Data System (ADS)

    Karel, Wilfried; Pfeifer, Norbert

    2009-01-01

    This article concentrates on the integrated self-calibration of both the interior orientation and the distance measurement system of a time-of-flght range camera (photonic mixer device). Unlike other approaches that investigate individual distortion factors separately, in the presented approach all calculations are based on the same data set that is captured without auxiliary devices serving as high-order reference, but with the camera being guided by hand. Flat, circular targets stuck on a planar whiteboard and with known positions are automatically tracked throughout the amplitude layer of long image sequences. These image observations are introduced into a bundle block adjustment, which on the one hand results in the determination of the interior orientation. Capitalizing the known planarity of the imaged board, the reconstructed exterior orientations furthermore allow for the derivation of reference values of the actual distance observations. Eased by the automatic reconstruction of the cameras trajectory and attitude, comprehensive statistics are generated, which are accumulated into a 5-dimensional matrix in order to be manageable. The marginal distributions of this matrix are inspected for the purpose of system identification, whereupon its elements are introduced into another least-squares adjustment, finally leading to clear range correction models and parameters.

  18. Coupled inverse and forward modelling to assess the range of acceptable thermal histories, a case study from SE Brazil

    NASA Astrophysics Data System (ADS)

    Cogné, N.; Gallagher, K.; Cobbold, P. R.

    2012-04-01

    We performed a new thermochronological study (fission track analysis and (U-Th)/He dating on apatite) in SE Brazil and integrate those data with inverse and forward modelling via QTQt software (Gallagher, 2012) to obtain thermal histories. The inversion results were used to characterize the general thermal histories and the associated uncertainties. For most of the samples we had a first phase of cooling during Late Cretaceous or Early Tertiary with subsequent reheating followed by Neogene cooling. The inverse modelling does not provide a unique solution and the associated uncertainties can be quite significant. Moreover the Tertiary parts of thermal histories were usually near the accepted resolution of the thermochronometric methods (~50-40°C). Therefore we performed deterministic forward modelling within the range of uncertainties to assess which solution is the most consistent with the data and independent geological information. These results are always conditional on the assumed kinetics for fission track annealing and diffusion of He, so we do not test the validity of that aspect. However, we can look at the range of predictions for the different forward models tested. This apporach implies that the reheating is required only for the samples around onshore Tertiary basins. For other samples we cannot conclude but geological information are against this hypothesis. However the Neogene cooling is required for all the samples.The combination of forward and inverse modelling allows us to better constrain the thermal histories for each sample in exploring the range of uncertainties and to reconcile a range of possible thermal histories with independent geological information. It also provides new information on the contrasting thermal evolution between different regions of the onshore SE Brazilian margin. Gallagher, K. 2012, Transdimensional Inverse thermal history modeling for quantitative thermochronology, Journal of Geophysical Research, in press.

  19. Building Energy Simulation Test for Existing Homes (BESTEST-EX): Instructions for Implementing the Test Procedure, Calibration Test Reference Results, and Example Acceptance-Range Criteria

    SciTech Connect

    Judkoff, R.; Polly, B.; Bianchi, M.; Neymark, J.; Kennedy, M.

    2011-08-01

    This publication summarizes building energy simulation test for existing homes (BESTEST-EX): instructions for implementing the test procedure, calibration tests reference results, and example acceptance-range criteria.

  20. Statistical model of the range-dependent error in radar-rainfall estimates due to the vertical profile of reflectivity

    NASA Astrophysics Data System (ADS)

    Krajewski, Witold F.; Vignal, Bertrand; Seo, Bong-Chul; Villarini, Gabriele

    2011-05-01

    SummaryThe authors developed an approach for deriving a statistical model of range-dependent error (RDE) in radar-rainfall estimates by parameterizing the structure of the non-uniform vertical profile of radar reflectivity (VPR). The proposed parameterization of the mean VPR and its expected variations are characterized by several climatological parameters that describe dominant atmospheric conditions related to vertical reflectivity variation. We have used four years of radar volume scan data from the Tulsa weather radar WSR-88D (Oklahoma) to illustrate this approach and have estimated the model parameters by minimizing the sum of the squared differences between the modeled and observed VPR influences that were computed using radar data. We evaluated the mean and standard deviation of the modeled RDE against rain gauge data from the Oklahoma Mesonet network. No rain gauge data were used in the model development. The authors used the three lowest antenna elevation angles to demonstrate the model performance for cold (November-April) and warm (May-October) seasons. The RDE derived from the parameterized models shows very good agreement with the observed differences between radar and rain gauge estimates of rainfall. For the third elevation angle and cold season, there are 82% and 42% improvements for the RDE and its standard deviation with respect to the no-VPR case. The results of this study indicate that VPR is a key factor in the characterization of the radar range-dependent bias, and the proposed models can be used to represent the radar RDE in the absence of rain gauge data.

  1. Restraint of range walk error in a Geiger-mode avalanche photodiode lidar to acquire high-precision depth and intensity information.

    PubMed

    Xu, Lu; Zhang, Yu; Zhang, Yong; Yang, Chenghua; Yang, Xu; Zhao, Yuan

    2016-03-01

    There exists a range walk error in a Geiger-mode avalanche photodiode (Gm-APD) lidar because of the fluctuation in the number of signal photoelectrons. To restrain this range walk error, we propose a new returning-wave signal processing technique based on the Poisson probability response model and the Gaussian functions fitting method. High-precision depth and intensity information of the target at the distance of 5 m is obtained by a Gm-APD lidar using a 6 ns wide pulsed laser. The experiment results show that the range and intensity precisions are 1.2 cm and 0.015 photoelectrons, respectively. PMID:26974630

  2. Post-processing of medium-range ensemble hydrological forecasting: impact of forcing, initial conditions and model errors

    NASA Astrophysics Data System (ADS)

    Roulin, Emmanuel; Vannitsem, Stéphane

    2015-04-01

    The impact of errors in the forcing, errors in the model structure and parameters, and errors in the initial conditions are investigated in a simple hydrological ensemble prediction system. The hydrological model is forced by precipitation forecasts from the ECMWF Ensemble Prediction System. The post-processing of the precipitation and/or the streamflow using information from the reforecasts performed by ECMWF are tested. For this purpose, hydrological reforecasts are obtained by forcing the hydrological model with the precipitation from the reforecast data. In the present case study, it is found that the post-processing of the hydrological ensembles with a statistical model fitted on the hydrological reforecasts improves the verification scores better than the use of post-processed precipitation ensembles. In the case of large biases in the precipitation, combining the post-processing of both precipitation and streamflow allows for further improvements. During winter, errors in the initial conditions have a larger impact on the scores than errors in the model structure as designed in the experiments. Errors in the parameter values are largely corrected with the post-processing.

  3. Assessment of the accuracy of global geodetic satellite laser ranging observations and estimated impact on ITRF scale: estimation of systematic errors in LAGEOS observations 1993-2014

    NASA Astrophysics Data System (ADS)

    Appleby, Graham; Rodríguez, José; Altamimi, Zuheir

    2016-06-01

    Satellite laser ranging (SLR) to the geodetic satellites LAGEOS and LAGEOS-2 uniquely determines the origin of the terrestrial reference frame and, jointly with very long baseline interferometry, its scale. Given such a fundamental role in satellite geodesy, it is crucial that any systematic errors in either technique are at an absolute minimum as efforts continue to realise the reference frame at millimetre levels of accuracy to meet the present and future science requirements. Here, we examine the intrinsic accuracy of SLR measurements made by tracking stations of the International Laser Ranging Service using normal point observations of the two LAGEOS satellites in the period 1993 to 2014. The approach we investigate in this paper is to compute weekly reference frame solutions solving for satellite initial state vectors, station coordinates and daily Earth orientation parameters, estimating along with these weekly average range errors for each and every one of the observing stations. Potential issues in any of the large number of SLR stations assumed to have been free of error in previous realisations of the ITRF may have been absorbed in the reference frame, primarily in station height. Likewise, systematic range errors estimated against a fixed frame that may itself suffer from accuracy issues will absorb network-wide problems into station-specific results. Our results suggest that in the past two decades, the scale of the ITRF derived from the SLR technique has been close to 0.7 ppb too small, due to systematic errors either or both in the range measurements and their treatment. We discuss these results in the context of preparations for ITRF2014 and additionally consider the impact of this work on the currently adopted value of the geocentric gravitational constant, GM.

  4. Study of and proposals for the correction of errors in a radar ranging device designed to facilitate docking of a teleoperator maneuvering system

    NASA Technical Reports Server (NTRS)

    Mcdonald, M. W.

    1982-01-01

    A frequency modulated continuous wave radar system was developed. The system operates in the 35 gigahertz frequency range and provides millimeter accuracy range and range rate measurements. This level of range resolution allows soft docking for the proposed teleoperator maneuvering system (TMS) or other autonomous or robotic space vehicles. Sources of error in the operation of the system which tend to limit its range resolution capabilities are identified. Alternative signal processing techniques are explored with emphasis on determination of the effects of inserting various signal filtering circuits in the system. The identification and elimination of an extraneous low frequency signal component created as a result of zero range immediate reflection of radar energy from the surface of the antenna dish back into the mixer of the system is described.

  5. Passive ranging errors due to multipath distortion of deterministic transient signals with application to the localization of small arms fire

    NASA Astrophysics Data System (ADS)

    Ferguson, Brian G.; Lo, Kam W.

    2002-01-01

    A passive ranging technique based on wavefront curvature is used to estimate the ranges and bearings of impulsive sound sources represented by small arms fire. The discharge of a firearm results in the generation of a transient acoustic signal whose energy propagates radially outwards from the omnidirectional source. The radius of curvature of the spherical wavefront at any instant is equal to the instantaneous range from the source. The curvature of the acoustic wavefront is sensed with a three-microphone linear array by first estimating the differential time of arrival (or time delay) of the acoustic wavefront at each of the two adjacent sensor pairs and then processing the time-delay information to extract the range and bearing of the source. However, modeling the passive ranging performance of the wavefront curvature method for a deterministic transient signal source in a multipath environment shows that when the multipath and direct path arrivals are unresolvable, the time-delay estimates are biased which, in turn, biases the range estimates. The model explains the observed under-ranging of small arms firing positions during a field experiment.

  6. Passive ranging errors due to multipath distortion of deterministic transient signals with application to the localization of small arms fire.

    PubMed

    Ferguson, Brian G; Lo, Kam W

    2002-01-01

    A passive ranging technique based on wavefront curvature is used to estimate the ranges and bearings of impulsive sound sources represented by small arms fire. The discharge of a firearm results in the generation of a transient acoustic signal whose energy propagates radially outwards from the omnidirectional source. The radius of curvature of the spherical wavefront at any instant is equal to the instantaneous range from the source. The curvature of the acoustic wavefront is sensed with a three-microphone linear array by first estimating the differential time of arrival (or time delay) of the acoustic wavefront at each of the two adjacent sensor pairs and then processing the time-delay information to extract the range and bearing of the source. However, modeling the passive ranging performance of the wavefront curvature method for a deterministic transient signal source in a multipath environment shows that when the multipath and direct path arrivals are unresolvable, the time-delay estimates are biased which, in turn, biases the range estimates. The model explains the observed under-ranging of small arms firing positions during a field experiment. PMID:11831787

  7. Influence of convective parameterization on the systematic errors of Climate Forecast System (CFS) model over the Indian monsoon region from an extended range forecast perspective

    NASA Astrophysics Data System (ADS)

    Pattnaik, S.; Abhilash, S.; De, S.; Sahai, A. K.; Phani, R.; Goswami, B. N.

    2013-07-01

    This study investigates the influence of Simplified Arakawa Schubert (SAS) and Relax Arakawa Schubert (RAS) cumulus parameterization schemes on coupled Climate Forecast System version.1 (CFS-1, T62L64) retrospective forecasts over Indian monsoon region from an extended range forecast perspective. The forecast data sets comprise 45 days of model integrations based on 31 different initial conditions at pentad intervals starting from 1 May to 28 September for the years 2001 to 2007. It is found that mean climatological features of Indian summer monsoon months (JJAS) are reasonably simulated by both the versions (i.e. SAS and RAS) of the model; however strong cross equatorial flow and excess stratiform rainfall are noted in RAS compared to SAS. Both the versions of the model overestimated apparent heat source and moisture sink compared to NCEP/NCAR reanalysis. The prognosis evaluation of daily forecast climatology reveals robust systematic warming (moistening) in RAS and cooling (drying) biases in SAS particularly at the middle and upper troposphere of the model respectively. Using error energy/variance and root mean square error methodology it is also established that major contribution to the model total error is coming from the systematic component of the model error. It is also found that the forecast error growth of temperature in RAS is less than that of SAS; however, the scenario is reversed for moisture errors, although the difference of moisture errors between these two forecasts is not very large compared to that of temperature errors. Broadly, it is found that both the versions of the model are underestimating (overestimating) the rainfall area and amount over the Indian land region (and neighborhood oceanic region). The rainfall forecast results at pentad interval exhibited that, SAS and RAS have good prediction skills over the Indian monsoon core zone and Arabian Sea. There is less excess rainfall particularly over oceanic region in RAS up to 30 days of

  8. Impact of a distance estimation error inducing a visualized zone gap on the target illuminance in range-gated active imaging.

    PubMed

    Matwyschuk, Alexis

    2014-01-01

    Some stand-alone airborne systems of target reconnaissance such as a missile seeker head use range-gated laser active imaging to visualize a target in the scene. To center the visualized zone on the target, it is important to know the distance between the active imaging system and the target. However, as this exact distance is not known before the detection of the target, it can be only estimated. This estimated distance can be erroneous (max≈500  m) with some technological drifts (gyrometric drift, accelerometric drift, missile position error, etc.). To be able to evaluate the impact of a distance estimation error on target illuminance in active imaging, the expressions of the illuminance attenuation ratio according to the decentered target position with regard to the visualized zone were determined. These different equations will be used to determine, in future stand-alone reconnaissance systems, the target signal-to-noise ratio as a function of the localization error. Generally speaking, two modes of visualization were used: first by using a fixed width of the visualized zone, and second by increasing the width of the visualized zone as a function of the distance. The defined different expressions allowed us to study the illuminance behavior of the target with regard to the value of the gap (difference between the estimated distance and the real distance) for each mode of visualization. The results showed that from a target distance of about 1 km, the visualization mode with variable zone width allowed us to decrease the target illuminance less during a gap caused by an estimation error of the target distance. PMID:24513988

  9. Stepwise error-prone PCR and DNA shuffling changed the pH activity range and product specificity of the cyclodextrin glucanotransferase from an alkaliphilic Bacillus sp.

    PubMed Central

    Melzer, Susanne; Sonnendecker, Christian; Föllner, Christina; Zimmermann, Wolfgang

    2015-01-01

    Cyclodextrin glucanotransferase (EC 2.4.1.19) from the alkaliphilic Bacillus sp. G-825-6 converts starch mainly to γ-cyclodextrin (CD8). A combination of error-prone PCR and DNA shuffling was used to obtain variants of this enzyme with higher product specificity for CD8 and a broad pH activity range. The variant S54 with seven amino acid substitutions showed a 1.2-fold increase in CD8-synthesizing activity and the product ratio of CD7:CD8 was shifted to 1:7 compared to 1:3 of the wild-type enzyme. Nine amino acid substitutions of the cyclodextrin glucanotransferase were performed to generate the variant S35 active in a pH range 4.0–10.0. Compared to the wild-type enzyme which is inactive below pH 6.0, S35 retained 70% of its CD8-synthesizing activity at pH 4.0. PMID:26155461

  10. Rain gauge - radar rainfall reanalysis of operational and research data in the Cévennes-Vivarais region, France, estimation error analysis over a wide range of scales.

    NASA Astrophysics Data System (ADS)

    Wijbrans, Annette; Delrieu, Guy; Nord, Guillaume; Boudevillain, Brice; Berne, Alexis; Grazioli, Jacopo; Confoland, Audrey

    2014-05-01

    In the Cévennes -Vivarais region in France, flash-flood events can occur due to high intensity precipitation events. These events are described in a detailed quantitative precipitation estimates, to be able to better characterize the hydrological response to these rain events in a number of small-scale nested watersheds (<100 km² typically), sampling various landscapes of the Mediterranean region. Radar - rain gauge merging methods described by Delrieu et al (2013) are applied to the 9 events of the autumn of 2012. Rainfall data is merged for both the operational networks in the Cévennes-Vivarais region in France on a 160 x 200 km window, as well as a research network, in the same region on a window of 15x30 km. The radar and rain gauge data of the operational network are collected from three organisms (Météo-France, Service de Prévision des Crues du Grand Delta and EdF/DTG). The research network contains high resolution data are from research rainfall observation systems deployed within the Enhanced Observation Period (autumn 2012-2015) of the HyMeX project (www.hymex.org). This project aims at studying the hydrological cycle in the Mediterranean with emphases on the hydro-meteorological extremes and their evolution in the coming decades. Rain gauge radar merging is performed using a kriging with external drift (KED) technique, and compared to the ordinary kriging (OK) of the rain gauges and the radar products on the same time scale using a cross-validation technique. Also a method is applied to quantify kriging estimation variances for both kriging techniques at the two spatial scales, in order to analyse the error characteristics of the interpolation methods at a scale range of 0.1 - 100 km² and 0.2 - 12 h. The combined information of the reanalysis of the data of the operational network and the research network gives a view on the error structure of rainfall estimations over several orders of magnitudes in spatial scale. This allows understanding of the

  11. Systematic errors in the simulation of the Asian summer monsoon: the role of rainfall variability on a range of time and space scales

    NASA Astrophysics Data System (ADS)

    Martin, Gill; Levine, Richard; Klingaman, Nicholas; Bush, Stephanie; Turner, Andrew; Woolnough, Steven

    2015-04-01

    -scale circulation than is indicated by the intensity and frequency of rainfall alone. Monsoon depressions and low pressure systems are important contributors to monsoon rainfall over central and northern India, areas where MetUM climate simulations typically show deficient monsoon rainfall. Analysis of MetUM climate simulations at resolutions ranging from N96 (~135km) to N512 (~25km) suggests that at lower resolution the numbers and intensities of monsoon depressions and low pressure systems and their associated rainfall are very low compared with re-analyses/observations. We show that there are substantial increases with horizontal resolution, but resolution is not the only factor. Idealised simulations, either using nudged atmospheric winds or initialised coupled hindcasts, which improve (strengthen) the mean state monsoon and cyclonic circulation over the Indian peninsula, also result in a substantial increase in monsoon depressions and associated rainfall. This suggests that a more realistic representation of monsoon depressions is possible even at lower resolution if the larger-scale systematic error pattern in the monsoon is improved.

  12. Comparing biomarker measurements to a normal range: when to use standard error of the mean (SEM) or standard deviation (SD) confidence intervals tests.

    PubMed

    Pleil, Joachim D

    2016-05-01

    This commentary is the second of a series outlining one specific concept in interpreting biomarkers data. In the first, an observational method was presented for assessing the distribution of measurements before making parametric calculations. Here, the discussion revolves around the next step, the choice of using standard error of the mean or the calculated standard deviation to compare or predict measurement results. PMID:27045663

  13. Acceptance threshold theory can explain occurrence of homosexual behaviour.

    PubMed

    Engel, Katharina C; Männer, Lisa; Ayasse, Manfred; Steiger, Sandra

    2015-01-01

    Same-sex sexual behaviour (SSB) has been documented in a wide range of animals, but its evolutionary causes are not well understood. Here, we investigated SSB in the light of Reeve's acceptance threshold theory. When recognition is not error-proof, the acceptance threshold used by males to recognize potential mating partners should be flexibly adjusted to maximize the fitness pay-off between the costs of erroneously accepting males and the benefits of accepting females. By manipulating male burying beetles' search time for females and their reproductive potential, we influenced their perceived costs of making an acceptance or rejection error. As predicted, when the costs of rejecting females increased, males exhibited more permissive discrimination decisions and showed high levels of SSB; when the costs of accepting males increased, males were more restrictive and showed low levels of SSB. Our results support the idea that in animal species, in which the recognition cues of females and males overlap to a certain degree, SSB is a consequence of an adaptive discrimination strategy to avoid the costs of making rejection errors. PMID:25631226

  14. Refractive Errors

    MedlinePlus

    ... and lens of your eye helps you focus. Refractive errors are vision problems that happen when the ... cornea, or aging of the lens. Four common refractive errors are Myopia, or nearsightedness - clear vision close ...

  15. Error-prone signalling.

    PubMed

    Johnstone, R A; Grafen, A

    1992-06-22

    The handicap principle of Zahavi is potentially of great importance to the study of biological communication. Existing models of the handicap principle, however, make the unrealistic assumption that communication is error free. It seems possible, therefore, that Zahavi's arguments do not apply to real signalling systems, in which some degree of error is inevitable. Here, we present a general evolutionarily stable strategy (ESS) model of the handicap principle which incorporates perceptual error. We show that, for a wide range of error functions, error-prone signalling systems must be honest at equilibrium. Perceptual error is thus unlikely to threaten the validity of the handicap principle. Our model represents a step towards greater realism, and also opens up new possibilities for biological signalling theory. Concurrent displays, direct perception of quality, and the evolution of 'amplifiers' and 'attenuators' are all probable features of real signalling systems, yet handicap models based on the assumption of error-free communication cannot accommodate these possibilities. PMID:1354361

  16. TU-C-BRE-08: IMRT QA: Selecting Meaningful Gamma Criteria Based On Error Detection Sensitivity

    SciTech Connect

    Steers, J; Fraass, B

    2014-06-15

    Purpose: To develop a strategy for defining meaningful tolerance limits and studying the sensitivity of IMRT QA gamma criteria by inducing known errors in QA plans. Methods: IMRT QA measurements (ArcCHECK, Sun Nuclear) were compared to QA plan calculations with induced errors. Many (>24) gamma comparisons between data and calculations were performed for each of several kinds of cases and classes of induced error types with varying magnitudes (e.g. MU errors ranging from -10% to +10%), resulting in over 3,000 comparisons. Gamma passing rates for each error class and case were graphed against error magnitude to create error curves in order to represent the range of missed errors in routine IMRT QA using various gamma criteria. Results: This study demonstrates that random, case-specific, and systematic errors can be detected by the error curve analysis. Depending on location of the peak of the error curve (e.g., not centered about zero), 3%/3mm threshold=10% criteria may miss MU errors of up to 10% and random MLC errors of up to 5 mm. Additionally, using larger dose thresholds for specific devices may increase error sensitivity (for the same X%/Ymm criteria) by up to a factor of two. This analysis will allow clinics to select more meaningful gamma criteria based on QA device, treatment techniques, and acceptable error tolerances. Conclusion: We propose a strategy for selecting gamma parameters based on the sensitivity of gamma criteria and individual QA devices to induced calculation errors in QA plans. Our data suggest large errors may be missed using conventional gamma criteria and that using stricter criteria with an increased dose threshold may reduce the range of missed errors. This approach allows quantification of gamma criteria sensitivity and is straightforward to apply to other combinations of devices and treatment techniques.

  17. Effect of roughness, deterministic and random errors in film thickness on the reflecting properties of aperiodic mirrors for the EUV range

    NASA Astrophysics Data System (ADS)

    Gaikovich, P. K.; Polkovnikov, V. N.; Salashchenko, N. N.; Chkhalo, N. I.; Schäfers, F.; Sokolov, A.

    2016-05-01

    By the example of three aperiodic multilayer Mo/Si mirrors (AMM) for the wavelength ranges 17–21 nm, 24–29 nm, and 28–33 nm we have studied numerically the effect of the linearly determinisctic and random fluctuations of the film thickness and the interlayer roughness on the spectral dependences of the reflection coefficient. The simulation results are used to solve the inverse problem of reconstructing the interlayer roughness and the thickness of individual films from the measured dependences of the extreme UV radiation reflection coefficients. It is shown that the 'asymmetry' of the boundaries affects the magnitude and slope of the reflection coefficient plateau. Random fluctuations of the film thickness with the variance of 1%–2% weakly influence the reflection characteristics of AMMs and allow reliable reconstruction of the thickness of individual films. The fluctuations with the variance 8%–10% allow the estimation of individual thicknesses, but the reflection curve in this case strongly differs from the desirable one. Larger fluctuations do not allow the reconstruction of the AMM structure. The basic criteria for high-quality AMM synthesis are formulated.

  18. UGV acceptance testing

    NASA Astrophysics Data System (ADS)

    Kramer, Jeffrey A.; Murphy, Robin R.

    2006-05-01

    With over 100 models of unmanned vehicles now available for military and civilian safety, security or rescue applications, it is important to for agencies to establish acceptance testing. However, there appears to be no general guidelines for what constitutes a reasonable acceptance test. This paper describes i) a preliminary method for acceptance testing by a customer of the mechanical and electrical components of an unmanned ground vehicle system, ii) how it has been applied to a man-packable micro-robot, and iii) discusses the value of testing both to ensure that the customer has a workable system and to improve design. The test method automated the operation of the robot to repeatedly exercise all aspects and combinations of components on the robot for 6 hours. The acceptance testing process uncovered many failures consistent with those shown to occur in the field, showing that testing by the user does predict failures. The process also demonstrated that the testing by the manufacturer can provide important design data that can be used to identify, diagnose, and prevent long-term problems. Also, the structured testing environment showed that sensor systems can be used to predict errors and changes in performance, as well as uncovering unmodeled behavior in subsystems.

  19. The Error in Total Error Reduction

    PubMed Central

    Witnauer, James E.; Urcelay, Gonzalo P.; Miller, Ralph R.

    2013-01-01

    Most models of human and animal learning assume that learning is proportional to the discrepancy between a delivered outcome and the outcome predicted by all cues present during that trial (i.e., total error across a stimulus compound). This total error reduction (TER) view has been implemented in connectionist and artificial neural network models to describe the conditions under which weights between units change. Electrophysiological work has revealed that the activity of dopamine neurons is correlated with the total error signal in models of reward learning. Similar neural mechanisms presumably support fear conditioning, human contingency learning, and other types of learning. Using a computational modelling approach, we compared several TER models of associative learning to an alternative model that rejects the TER assumption in favor of local error reduction (LER), which assumes that learning about each cue is proportional to the discrepancy between the delivered outcome and the outcome predicted by that specific cue on that trial. The LER model provided a better fit to the reviewed data than the TER models. Given the superiority of the LER model with the present data sets, acceptance of TER should be tempered. PMID:23891930

  20. Emperical Tests of Acceptance Sampling Plans

    NASA Technical Reports Server (NTRS)

    White, K. Preston, Jr.; Johnson, Kenneth L.

    2012-01-01

    Acceptance sampling is a quality control procedure applied as an alternative to 100% inspection. A random sample of items is drawn from a lot to determine the fraction of items which have a required quality characteristic. Both the number of items to be inspected and the criterion for determining conformance of the lot to the requirement are given by an appropriate sampling plan with specified risks of Type I and Type II sampling errors. In this paper, we present the results of empirical tests of the accuracy of selected sampling plans reported in the literature. These plans are for measureable quality characteristics which are known have either binomial, exponential, normal, gamma, Weibull, inverse Gaussian, or Poisson distributions. In the main, results support the accepted wisdom that variables acceptance plans are superior to attributes (binomial) acceptance plans, in the sense that these provide comparable protection against risks at reduced sampling cost. For the Gaussian and Weibull plans, however, there are ranges of the shape parameters for which the required sample sizes are in fact larger than the corresponding attributes plans, dramatically so for instances of large skew. Tests further confirm that the published inverse-Gaussian (IG) plan is flawed, as reported by White and Johnson (2011).

  1. Medication Errors

    MedlinePlus

    ... to reduce the risk of medication errors to industry and others at FDA. Additionally, DMEPA prospectively reviews ... List of Abbreviations Regulations and Guidances Guidance for Industry: Safety Considerations for Product Design to Minimize Medication ...

  2. Medication Errors

    MedlinePlus

    Medicines cure infectious diseases, prevent problems from chronic diseases, and ease pain. But medicines can also cause harmful reactions if not used ... You can help prevent errors by Knowing your medicines. Keep a list of the names of your ...

  3. Measurement error revisited

    NASA Astrophysics Data System (ADS)

    Henderson, Robert K.

    1999-12-01

    It is widely accepted in the electronics industry that measurement gauge error variation should be no larger than 10% of the related specification window. In a previous paper, 'What Amount of Measurement Error is Too Much?', the author used a framework from the process industries to evaluate the impact of measurement error variation in terms of both customer and supplier risk (i.e., Non-conformance and Yield Loss). Application of this framework in its simplest form suggested that in many circumstances the 10% criterion might be more stringent than is reasonably necessary. This paper reviews the framework and results of the earlier work, then examines some of the possible extensions to this framework suggested in that paper, including variance component models and sampling plans applicable in the photomask and semiconductor businesses. The potential impact of imperfect process control practices will be examined as well.

  4. Freeform solar concentrator with a highly asymmetric acceptance cone

    NASA Astrophysics Data System (ADS)

    Wheelwright, Brian; Angel, J. Roger P.; Coughenour, Blake; Hammer, Kimberly

    2014-10-01

    A solar concentrator with a highly asymmetric acceptance cone is investigated. Concentrating photovoltaic systems require dual-axis sun tracking to maintain nominal concentration throughout the day. In addition to collecting direct rays from the solar disk, which subtends ~0.53 degrees, concentrating optics must allow for in-field tracking errors due to mechanical misalignment of the module, wind loading, and control loop biases. The angular range over which the concentrator maintains <90% of on-axis throughput is defined as the optical acceptance angle. Concentrators with substantial rotational symmetry likewise exhibit rotationally symmetric acceptance angles. In the field, this is sometimes a poor match with azimuth-elevation trackers, which have inherently asymmetric tracking performance. Pedestal-mounted trackers with low torsional stiffness about the vertical axis have better elevation tracking than azimuthal tracking. Conversely, trackers which rotate on large-footprint circular tracks are often limited by elevation tracking performance. We show that a line-focus concentrator, composed of a parabolic trough primary reflector and freeform refractive secondary, can be tailored to have a highly asymmetric acceptance angle. The design is suitable for a tracker with excellent tracking accuracy in the elevation direction, and poor accuracy in the azimuthal direction. In the 1000X design given, when trough optical errors (2mrad rms slope deviation) are accounted for, the azimuthal acceptance angle is +/- 1.65°, while the elevation acceptance angle is only +/-0.29°. This acceptance angle does not include the angular width of the sun, which consumes nearly all of the elevation tolerance at this concentration level. By decreasing the average concentration, the elevation acceptance angle can be increased. This is well-suited for a pedestal alt-azimuth tracker with a low cost slew bearing (without anti-backlash features).

  5. Acceptance speech.

    PubMed

    Carpenter, M

    1994-01-01

    In Bangladesh, the assistant administrator of USAID gave an acceptance speech at an awards ceremony on the occasion of the 25th anniversary of oral rehydration solution (ORS). The ceremony celebrated the key role of the International Centre for Diarrhoeal Disease Research, Bangladesh (ICDDR,B) in the discovery of ORS. Its research activities over the last 25 years have brought ORS to every village in the world, preventing more than a million deaths each year. ORS is the most important medical advance of the 20th century. It is affordable and client-oriented, a true appropriate technology. USAID has provided more than US$ 40 million to ICDDR,B for diarrheal disease and measles research, urban and rural applied family planning and maternal and child health research, and vaccine development. ICDDR,B began as the relatively small Cholera Research Laboratory and has grown into an acclaimed international center for health, family planning, and population research. It leads the world in diarrheal disease research. ICDDR,B is the leading center for applied health research in South Asia. It trains public health specialists from around the world. The government of Bangladesh and the international donor community have actively joined in support of ICDDR,B. The government applies the results of ICDDR,B research to its programs to improve the health and well-being of Bangladeshis. ICDDR,B now also studies acute respiratory diseases and measles. Population and health comprise 1 of USAID's 4 strategic priorities, the others being economic growth, environment, and democracy, USAID promotes people's participation in these 4 areas and in the design and implementation of development projects. USAID is committed to the use and improvement of ORS and to complementary strategies that further reduce diarrhea-related deaths. Continued collaboration with a strong user perspective and integrated services will lead to sustainable development. PMID:12345470

  6. Acceptance speech.

    PubMed

    Yusuf, C K

    1994-01-01

    I am proud and honored to accept this award on behalf of the Government of Bangladesh, and the millions of Bangladeshi children saved by oral rehydration solution. The Government of Bangladesh is grateful for this recognition of its commitment to international health and population research and cost-effective health care for all. The Government of Bangladesh has already made remarkable strides forward in the health and population sector, and this was recognized in UNICEF's 1993 "State of the World's Children". The national contraceptive prevalence rate, at 40%, is higher than that of many developed countries. It is appropriate that Bangladesh, where ORS was discovered, has the largest ORS production capacity in the world. It was remarkable that after the devastating cyclone in 1991, the country was able to produce enough ORS to meet the needs and remain self-sufficient. Similarly, Bangladesh has one of the most effective, flexible and efficient control of diarrheal disease and epidemic response program in the world. Through the country, doctors have been trained in diarrheal disease management, and stores of ORS are maintained ready for any outbreak. Despite grim predictions after the 1991 cyclone and the 1993 floods, relatively few people died from diarrheal disease. This is indicative of the strength of the national program. I want to take this opportunity to acknowledge the contribution of ICDDR, B and the important role it plays in supporting the Government's efforts in the health and population sector. The partnership between the Government of Bangladesh and ICDDR, B has already borne great fruit, and I hope and believe that it will continue to do so for many years in the future. Thank you. PMID:12345479

  7. Error compensation for thermally induced errors on a machine tool

    SciTech Connect

    Krulewich, D.A.

    1996-11-08

    Heat flow from internal and external sources and the environment create machine deformations, resulting in positioning errors between the tool and workpiece. There is no industrially accepted method for thermal error compensation. A simple model has been selected that linearly relates discrete temperature measurements to the deflection. The biggest problem is how to locate the temperature sensors and to determine the number of required temperature sensors. This research develops a method to determine the number and location of temperature measurements.

  8. Estimating patient specific uncertainty parameters for adaptive treatment re-planning in proton therapy using in vivo range measurements and Bayesian inference: application to setup and stopping power errors.

    PubMed

    Labarbe, Rudi; Janssens, Guillaume; Sterpin, Edmond

    2016-09-01

    In proton therapy, quantification of the proton range uncertainty is important to achieve dose distribution compliance. The promising accuracy of prompt gamma imaging (PGI) suggests the development of a mathematical framework using the range measurements to convert population based estimates of uncertainties into patient specific estimates with the purpose of plan adaptation. We present here such framework using Bayesian inference. The sources of uncertainty were modeled by three parameters: setup bias m, random setup precision r and water equivalent path length bias u. The evolution of the expectation values E(m), E(r) and E(u) during the treatment was simulated. The expectation values converged towards the true simulation parameters after 5 and 10 fractions, for E(m) and E(u), respectively. E(r) settle on a constant value slightly lower than the true value after 10 fractions. In conclusion, the simulation showed that there is enough information in the frequency distribution of the range errors measured by PGI to estimate the expectation values and the confidence interval of the model parameters by Bayesian inference. The updated model parameters were used to compute patient specific lateral and local distal margins for adaptive re-planning. PMID:27494118

  9. Estimating patient specific uncertainty parameters for adaptive treatment re-planning in proton therapy using in vivo range measurements and Bayesian inference: application to setup and stopping power errors

    NASA Astrophysics Data System (ADS)

    Labarbe, Rudi; Janssens, Guillaume; Sterpin, Edmond

    2016-09-01

    In proton therapy, quantification of the proton range uncertainty is important to achieve dose distribution compliance. The promising accuracy of prompt gamma imaging (PGI) suggests the development of a mathematical framework using the range measurements to convert population based estimates of uncertainties into patient specific estimates with the purpose of plan adaptation. We present here such framework using Bayesian inference. The sources of uncertainty were modeled by three parameters: setup bias m, random setup precision r and water equivalent path length bias u. The evolution of the expectation values E(m), E(r) and E(u) during the treatment was simulated. The expectation values converged towards the true simulation parameters after 5 and 10 fractions, for E(m) and E(u), respectively. E(r) settle on a constant value slightly lower than the true value after 10 fractions. In conclusion, the simulation showed that there is enough information in the frequency distribution of the range errors measured by PGI to estimate the expectation values and the confidence interval of the model parameters by Bayesian inference. The updated model parameters were used to compute patient specific lateral and local distal margins for adaptive re-planning.

  10. High acceptance recoil polarimeter

    SciTech Connect

    The HARP Collaboration

    1992-12-05

    In order to detect neutrons and protons in the 50 to 600 MeV energy range and measure their polarization, an efficient, low-noise, self-calibrating device is being designed. This detector, known as the High Acceptance Recoil Polarimeter (HARP), is based on the recoil principle of proton detection from np[r arrow]n[prime]p[prime] or pp[r arrow]p[prime]p[prime] scattering (detected particles are underlined) which intrinsically yields polarization information on the incoming particle. HARP will be commissioned to carry out experiments in 1994.

  11. 27 CFR 46.120 - Errors discovered on inspection.

    Code of Federal Regulations, 2010 CFR

    2010-04-01

    ... Errors discovered on inspection. When a TTB officer discovers on a special tax stamp a material error in... amended return and an acceptable explanation for the error, the officer will make the proper correction on the stamp and return it to the taxpayer. However, if the error found by the TTB officer is on...

  12. 27 CFR 46.120 - Errors discovered on inspection.

    Code of Federal Regulations, 2011 CFR

    2011-04-01

    ... Errors discovered on inspection. When a TTB officer discovers on a special tax stamp a material error in... amended return and an acceptable explanation for the error, the officer will make the proper correction on the stamp and return it to the taxpayer. However, if the error found by the TTB officer is on...

  13. Operational Interventions to Maintenance Error

    NASA Technical Reports Server (NTRS)

    Kanki, Barbara G.; Walter, Diane; Dulchinos, VIcki

    1997-01-01

    A significant proportion of aviation accidents and incidents are known to be tied to human error. However, research of flight operational errors has shown that so-called pilot error often involves a variety of human factors issues and not a simple lack of individual technical skills. In aircraft maintenance operations, there is similar concern that maintenance errors which may lead to incidents and accidents are related to a large variety of human factors issues. Although maintenance error data and research are limited, industry initiatives involving human factors training in maintenance have become increasingly accepted as one type of maintenance error intervention. Conscientious efforts have been made in re-inventing the team7 concept for maintenance operations and in tailoring programs to fit the needs of technical opeRAtions. Nevertheless, there remains a dual challenge: 1) to develop human factors interventions which are directly supported by reliable human error data, and 2) to integrate human factors concepts into the procedures and practices of everyday technical tasks. In this paper, we describe several varieties of human factors interventions and focus on two specific alternatives which target problems related to procedures and practices; namely, 1) structured on-the-job training and 2) procedure re-design. We hope to demonstrate that the key to leveraging the impact of these solutions comes from focused interventions; that is, interventions which are derived from a clear understanding of specific maintenance errors, their operational context and human factors components.

  14. Rectifying calibration error of Goldmann applanation tonometer is easy!

    PubMed

    Choudhari, Nikhil S; Moorthy, Krishna P; Tungikar, Vinod B; Kumar, Mohan; George, Ronnie; Rao, Harsha L; Senthil, Sirisha; Vijaya, Lingam; Garudadri, Chandra Sekhar

    2014-11-01

    Purpose: Goldmann applanation tonometer (GAT) is the current Gold standard tonometer. However, its calibration error is common and can go unnoticed in clinics. Its company repair has limitations. The purpose of this report is to describe a self-taught technique of rectifying calibration error of GAT. Materials and Methods: Twenty-nine slit-lamp-mounted Haag-Streit Goldmann tonometers (Model AT 900 C/M; Haag-Streit, Switzerland) were included in this cross-sectional interventional pilot study. The technique of rectification of calibration error of the tonometer involved cleaning and lubrication of the instrument followed by alignment of weights when lubrication alone didn't suffice. We followed the South East Asia Glaucoma Interest Group's definition of calibration error tolerance (acceptable GAT calibration error within ±2, ±3 and ±4 mm Hg at the 0, 20 and 60-mm Hg testing levels, respectively). Results: Twelve out of 29 (41.3%) GATs were out of calibration. The range of positive and negative calibration error at the clinically most important 20-mm Hg testing level was 0.5 to 20 mm Hg and -0.5 to -18 mm Hg, respectively. Cleaning and lubrication alone sufficed to rectify calibration error of 11 (91.6%) faulty instruments. Only one (8.3%) faulty GAT required alignment of the counter-weight. Conclusions: Rectification of calibration error of GAT is possible in-house. Cleaning and lubrication of GAT can be carried out even by eye care professionals and may suffice to rectify calibration error in the majority of faulty instruments. Such an exercise may drastically reduce the downtime of the Gold standard tonometer. PMID:25494251

  15. Spatial frequency domain error budget

    SciTech Connect

    Hauschildt, H; Krulewich, D

    1998-08-27

    The aim of this paper is to describe a methodology for designing and characterizing machines used to manufacture or inspect parts with spatial-frequency-based specifications. At Lawrence Livermore National Laboratory, one of our responsibilities is to design or select the appropriate machine tools to produce advanced optical and weapons systems. Recently, many of the component tolerances for these systems have been specified in terms of the spatial frequency content of residual errors on the surface. We typically use an error budget as a sensitivity analysis tool to ensure that the parts manufactured by a machine will meet the specified component tolerances. Error budgets provide the formalism whereby we account for all sources of uncertainty in a process, and sum them to arrive at a net prediction of how "precisely" a manufactured component can meet a target specification. Using the error budget, we are able to minimize risk during initial stages by ensuring that the machine will produce components that meet specifications before the machine is actually built or purchased. However, the current error budgeting procedure provides no formal mechanism for designing machines that can produce parts with spatial-frequency-based specifications. The output from the current error budgeting procedure is a single number estimating the net worst case or RMS error on the work piece. This procedure has limited ability to differentiate between low spatial frequency form errors versus high frequency surface finish errors. Therefore the current error budgeting procedure can lead us to reject a machine that is adequate or accept a machine that is inadequate. This paper will describe a new error budgeting methodology to aid in the design and characterization of machines used to manufacture or inspect parts with spatial-frequency-based specifications. The output from this new procedure is the continuous spatial frequency content of errors that result on a machined part. If the machine

  16. [Diagnostic Errors in Medicine].

    PubMed

    Buser, Claudia; Bankova, Andriyana

    2015-12-01

    The recognition of diagnostic errors in everyday practice can help improve patient safety. The most common diagnostic errors are the cognitive errors, followed by system-related errors and no fault errors. The cognitive errors often result from mental shortcuts, known as heuristics. The rate of cognitive errors can be reduced by a better understanding of heuristics and the use of checklists. The autopsy as a retrospective quality assessment of clinical diagnosis has a crucial role in learning from diagnostic errors. Diagnostic errors occur more often in primary care in comparison to hospital settings. On the other hand, the inpatient errors are more severe than the outpatient errors. PMID:26649954

  17. TU-C-BRE-07: Quantifying the Clinical Impact of VMAT Delivery Errors Relative to Prior Patients’ Plans and Adjusted for Anatomical Differences

    SciTech Connect

    Stanhope, C; Wu, Q; Yuan, L; Liu, J; Hood, R; Yin, F; Adamson, J

    2014-06-15

    Purpose: There is increased interest in the Radiation Oncology Physics community regarding sensitivity of pre-treatment IMRT/VMAT QA to delivery errors. Consequently, tools mapping pre-treatment QA to the patient DVH have been developed. However, the quantity of plan degradation that is acceptable remains uncertain. Using DVHs adapted from prior patients’ plans, we developed a technique to determine the magnitude of various delivery errors required to degrade a treatment plan to outside the clinically accepted range. Methods: DVHs for relevant organs at risk were adapted from a population of prior patients’ plans using a machine learning algorithm to establish the clinically acceptable DVH range specific to the patient’s anatomy. We applied this technique to six low-risk prostate cancer patients treated with single-arc VMAT and compared error-induced DVH changes to the adapted DVHs to determine the magnitude of error required to push the plan outside of the acceptable range. The procedure follows: (1) Errors (systematic ' random shift of MLCs, gantry-MLC desynchronization, dose rate fluctuations, etc.) were simulated and degraded DVHs calculated using the Varian Eclipse TPS. (2) Adapted DVHs and acceptable ranges for DVHs were established. (3) Relevant dosimetric indices and corresponding acceptable ranges were calculated from the DVHs. Key indices included NTCP (Lyman-Kutcher-Burman Model) and QUANTEC’s dose-volume Objectives: s of V75Gy≤0.15 for the rectum and V75Gy≤0.25 for the bladder. Results: Degradations to the clinical plan became “unacceptable” for 19±29mm and 1.9±2.0mm systematic outward shifts of a single leaf and leaf bank, respectively. All other simulated errors fell within the acceptable range. Conclusion: Utilizing machine learning and prior patients’ plans one can predict a clinically acceptable range of DVH degradation for a specific patient. Comparing error-induced DVH degradations to this range, it is shown that single

  18. How does human error affect safety in anesthesia?

    PubMed

    Gravenstein, J S

    2000-01-01

    Anesthesia morbidity and mortality, while acceptable, are not zero. Most mishaps have a multifactorial cause in which human error plays a significant part. Good design of anesthesia machines, ventilators, and monitors can prevent some, but not all, human error. Attention to the system in which the errors occur is important. Modern training with simulators is designed to reduce the frequency of human errors and to teach anesthesiologists how to deal with the consequences of such errors. PMID:10601526

  19. The Study of Prescribing Errors Among General Dentists

    PubMed Central

    Araghi, Solmaz; Sharifi, Rohollah; Ahmadi, Goran; Esfehani, Mahsa; Rezaei, Fatemeh

    2016-01-01

    Introduction: In dentistry, medicine often prescribed to relieve pain and remove infections. Therefore, wrong prescription can lead to a range of problems including lack of pain, antimicrobial treatment failure and the development of resistance to antibiotics. Materials and Methods: In this cross-sectional study, the aim was to evaluate the common errors in written prescriptions by general dentists in Kermanshah in 2014. Dentists received a questionnaire describing five hypothetical patient and the appropriate prescription for the patient in question was asked. Information about age, gender, work experience and the admission in university was collected. The frequency of errors in prescriptions was determined. Data by SPSS 20 statistical software and using statistical t-test, chi-square and Pearson correlation were analyzed (0.05> P). Results: A total of 180 dentists (62.6% male and 37.4% female) with a mean age of 8.23 ± 39.199 participated in this study. Prescription errors include the wrong in pharmaceutical form (11%), not having to write therapeutic dose (13%), writing wrong dose (14%), typos (15%), error prescription (23%) and writing wrong number of drugs (24%). The most frequent errors in the administration of antiviral drugs (31%) and later stages of antifungal drugs (30%), analgesics (23%) and antibiotics (16%) was observed. Males dentists compared with females dentists showed more frequent errors (P=0.046). Error frequency among dentists with a long work history (P>0.001) and the acceptance in the university except for the entrance examination (P=0.041) had a statistically significant relationship. Conclusion: This study showed that the written prescription by general dentists examined contained significant errors and improve prescribing through continuing education of dentists is essential. PMID:26573049

  20. Sun compass error model

    NASA Technical Reports Server (NTRS)

    Blucker, T. J.; Ferry, W. W.

    1971-01-01

    An error model is described for the Apollo 15 sun compass, a contingency navigational device. Field test data are presented along with significant results of the test. The errors reported include a random error resulting from tilt in leveling the sun compass, a random error because of observer sighting inaccuracies, a bias error because of mean tilt in compass leveling, a bias error in the sun compass itself, and a bias error because the device is leveled to the local terrain slope.

  1. Computational Errors of Mentally Retarded Students.

    ERIC Educational Resources Information Center

    Janke, Robert W.

    1980-01-01

    Examined computational errors made by educable mentally retarded students on the arithmetic subtest of the Wide Range Achievement Test. Retarded students had a lower percent of grouping and inappropriate inversion errors and a higher percent of incorrect operation errors than regular students had in Engelhardt's study. (Author)

  2. Bayesian Error Estimation Functionals

    NASA Astrophysics Data System (ADS)

    Jacobsen, Karsten W.

    The challenge of approximating the exchange-correlation functional in Density Functional Theory (DFT) has led to the development of numerous different approximations of varying accuracy on different calculated properties. There is therefore a need for reliable estimation of prediction errors within the different approximation schemes to DFT. The Bayesian Error Estimation Functionals (BEEF) have been developed with this in mind. The functionals are constructed by fitting to experimental and high-quality computational databases for molecules and solids including chemisorption and van der Waals systems. This leads to reasonably accurate general-purpose functionals with particual focus on surface science. The fitting procedure involves considerations on how to combine different types of data, and applies Tikhonov regularization and bootstrap cross validation. The methodology has been applied to construct GGA and metaGGA functionals with and without inclusion of long-ranged van der Waals contributions. The error estimation is made possible by the generation of not only a single functional but through the construction of a probability distribution of functionals represented by a functional ensemble. The use of the functional ensemble is illustrated on compound heat of formation and by investigations of the reliability of calculated catalytic ammonia synthesis rates.

  3. Remediating Common Math Errors.

    ERIC Educational Resources Information Center

    Wagner, Rudolph F.

    1981-01-01

    Explanations and remediation suggestions for five types of mathematics errors due either to perceptual or cognitive difficulties are given. Error types include directionality problems, mirror writing, visually misperceived signs, diagnosed directionality problems, and mixed process errors. (CL)

  4. Errors, error detection, error correction and hippocampal-region damage: data and theories.

    PubMed

    MacKay, Donald G; Johnson, Laura W

    2013-11-01

    This review and perspective article outlines 15 observational constraints on theories of errors, error detection, and error correction, and their relation to hippocampal-region (HR) damage. The core observations come from 10 studies with H.M., an amnesic with cerebellar and HR damage but virtually no neocortical damage. Three studies examined the detection of errors planted in visual scenes (e.g., a bird flying in a fish bowl in a school classroom) and sentences (e.g., I helped themselves to the birthday cake). In all three experiments, H.M. detected reliably fewer errors than carefully matched memory-normal controls. Other studies examined the detection and correction of self-produced errors, with controls for comprehension of the instructions, impaired visual acuity, temporal factors, motoric slowing, forgetting, excessive memory load, lack of motivation, and deficits in visual scanning or attention. In these studies, H.M. corrected reliably fewer errors than memory-normal and cerebellar controls, and his uncorrected errors in speech, object naming, and reading aloud exhibited two consistent features: omission and anomaly. For example, in sentence production tasks, H.M. omitted one or more words in uncorrected encoding errors that rendered his sentences anomalous (incoherent, incomplete, or ungrammatical) reliably more often than controls. Besides explaining these core findings, the theoretical principles discussed here explain H.M.'s retrograde amnesia for once familiar episodic and semantic information; his anterograde amnesia for novel information; his deficits in visual cognition, sentence comprehension, sentence production, sentence reading, and object naming; and effects of aging on his ability to read isolated low frequency words aloud. These theoretical principles also explain a wide range of other data on error detection and correction and generate new predictions for future test. PMID:23999403

  5. ATLAS ACCEPTANCE TEST

    SciTech Connect

    J.C. COCHRANE; J.V. PARKER; ET AL

    2001-06-01

    The acceptance test program for Atlas, a 23 MJ pulsed power facility for use in the Los Alamos High Energy Density Hydrodynamics program, has been completed. Completion of this program officially releases Atlas from the construction phase and readies it for experiments. Details of the acceptance test program results and of machine capabilities for experiments will be presented.

  6. Surface measurement errors using commercial scanning white light interferometers

    NASA Astrophysics Data System (ADS)

    Gao, F.; Leach, R. K.; Petzing, J.; Coupland, J. M.

    2008-01-01

    This paper examines the performance of commercial scanning white light interferometers in a range of measurement tasks. A step height artefact is used to investigate the response of the instruments at a discontinuity, while gratings with sinusoidal and rectangular profiles are used to investigate the effects of surface gradient and spatial frequency. Results are compared with measurements made with tapping mode atomic force microscopy and discrepancies are discussed with reference to error mechanisms put forward in the published literature. As expected, it is found that most instruments report errors when used in regions close to a discontinuity or those with a surface gradient that is large compared to the acceptance angle of the objective lens. Amongst other findings, however, we report systematic errors that are observed when the surface gradient is considerably smaller. Although these errors are typically less than the mean wavelength, they are significant compared to the vertical resolution of the instrument and indicate that current scanning white light interferometers should be used with some caution if sub-wavelength accuracy is required.

  7. Error Field Correction in ITER

    SciTech Connect

    Park, Jong-kyu; Boozer, Allen H.; Menard, Jonathan E.; Schaffer, Michael J.

    2008-05-22

    A new method for correcting magnetic field errors in the ITER tokamak is developed using the Ideal Perturbed Equilibrium Code (IPEC). The dominant external magnetic field for driving islands is shown to be localized to the outboard midplane for three ITER equilibria that represent the projected range of operational scenarios. The coupling matrices between the poloidal harmonics of the external magnetic perturbations and the resonant fields on the rational surfaces that drive islands are combined for different equilibria and used to determine an ordered list of the dominant errors in the external magnetic field. It is found that efficient and robust error field correction is possible with a fixed setting of the correction currents relative to the currents in the main coils across the range of ITER operating scenarios that was considered.

  8. Structure and dating errors in the geologic time scale and periodicity in mass extinctions

    NASA Technical Reports Server (NTRS)

    Stothers, Richard B.

    1989-01-01

    Structure in the geologic time scale reflects a partly paleontological origin. As a result, ages of Cenozoic and Mesozoic stage boundaries exhibit a weak 28-Myr periodicity that is similar to the strong 26-Myr periodicity detected in mass extinctions of marine life by Raup and Sepkoski. Radiometric dating errors in the geologic time scale, to which the mass extinctions are stratigraphically tied, do not necessarily lessen the likelihood of a significant periodicity in mass extinctions, but do spread the acceptable values of the period over the range 25-27 Myr for the Harland et al. time scale or 25-30 Myr for the DNAG time scale. If the Odin time scale is adopted, acceptable periods fall between 24 and 33 Myr, but are not robust against dating errors. Some indirect evidence from independently-dated flood-basalt volcanic horizons tends to favor the Odin time scale.

  9. Learner Error, Affectual Stimulation, and Conceptual Change

    ERIC Educational Resources Information Center

    Allen, Michael

    2010-01-01

    Pupils' expectation-related errors oppose the development of an appropriate scientific attitude towards empirical evidence and the learning of accepted science content, representing a hitherto neglected area of research in science education. In spite of these apparent drawbacks, a pedagogy is described that "encourages" pupils to allow their…

  10. Advanced error-prediction LDPC with temperature compensation for highly reliable SSDs

    NASA Astrophysics Data System (ADS)

    Tokutomi, Tsukasa; Tanakamaru, Shuhei; Iwasaki, Tomoko Ogura; Takeuchi, Ken

    2015-09-01

    To improve the reliability of NAND Flash memory based solid-state drives (SSDs), error-prediction LDPC (EP-LDPC) has been proposed for multi-level-cell (MLC) NAND Flash memory (Tanakamaru et al., 2012, 2013), which is effective for long retention times. However, EP-LDPC is not as effective for triple-level cell (TLC) NAND Flash memory, because TLC NAND Flash has higher error rates and is more sensitive to program-disturb error. Therefore, advanced error-prediction LDPC (AEP-LDPC) has been proposed for TLC NAND Flash memory (Tokutomi et al., 2014). AEP-LDPC can correct errors more accurately by precisely describing the error phenomena. In this paper, the effects of AEP-LDPC are investigated in a 2×nm TLC NAND Flash memory with temperature characterization. Compared with LDPC-with-BER-only, the SSD's data-retention time is increased by 3.4× and 9.5× at room-temperature (RT) and 85 °C, respectively. Similarly, the acceptable BER is increased by 1.8× and 2.3×, respectively. Moreover, AEP-LDPC can correct errors with pre-determined tables made at higher temperatures to shorten the measurement time before shipping. Furthermore, it is found that one table can cover behavior over a range of temperatures in AEP-LDPC. As a result, the total table size can be reduced to 777 kBytes, which makes this approach more practical.

  11. Improved ranging systems

    NASA Technical Reports Server (NTRS)

    Young, Larry E.

    1989-01-01

    Spacecraft range measurements have provided the most accurate tests, to date, of some relativistic gravitational parameters, even though the measurements were made with ranging systems having error budgets of about 10 meters. Technology is now available to allow an improvement of two orders of magnitude in the accuracy of spacecraft ranging. The largest gains in accuracy result from the replacement of unstable analog components with high speed digital circuits having precisely known delays and phase shifts.

  12. Acceptance, values, and probability.

    PubMed

    Steel, Daniel

    2015-10-01

    This essay makes a case for regarding personal probabilities used in Bayesian analyses of confirmation as objects of acceptance and rejection. That in turn entails that personal probabilities are subject to the argument from inductive risk, which aims to show non-epistemic values can legitimately influence scientific decisions about which hypotheses to accept. In a Bayesian context, the argument from inductive risk suggests that value judgments can influence decisions about which probability models to accept for likelihoods and priors. As a consequence, if the argument from inductive risk is sound, then non-epistemic values can affect not only the level of evidence deemed necessary to accept a hypothesis but also degrees of confirmation themselves. PMID:26386533

  13. Newbery Medal Acceptance.

    ERIC Educational Resources Information Center

    Freedman, Russell

    1988-01-01

    Presents the Newbery Medal acceptance speech of Russell Freedman, writer of children's nonfiction. Discusses the place of nonfiction in the world of children's literature, the evolution of children's biographies, and the author's work on "Lincoln." (ARH)

  14. Newbery Medal Acceptance.

    ERIC Educational Resources Information Center

    Cleary, Beverly

    1984-01-01

    Reprints the text of Ms. Cleary's Newbery medal acceptance speech in which she gives personal history concerning her development as a writer and her response to the letters she receives from children. (CRH)

  15. Caldecott Medal Acceptance.

    ERIC Educational Resources Information Center

    Provensen, Alice; Provensen, Martin

    1984-01-01

    Reprints the text of the Provensens' Caldecott medal acceptance speech in which they describe their early interest in libraries and literature, the collaborative aspect of their work, and their current interest in aviation. (CRH)

  16. Field error lottery

    SciTech Connect

    Elliott, C.J.; McVey, B. ); Quimby, D.C. )

    1990-01-01

    The level of field errors in an FEL is an important determinant of its performance. We have computed 3D performance of a large laser subsystem subjected to field errors of various types. These calculations have been guided by simple models such as SWOOP. The technique of choice is utilization of the FELEX free electron laser code that now possesses extensive engineering capabilities. Modeling includes the ability to establish tolerances of various types: fast and slow scale field bowing, field error level, beam position monitor error level, gap errors, defocusing errors, energy slew, displacement and pointing errors. Many effects of these errors on relative gain and relative power extraction are displayed and are the essential elements of determining an error budget. The random errors also depend on the particular random number seed used in the calculation. The simultaneous display of the performance versus error level of cases with multiple seeds illustrates the variations attributable to stochasticity of this model. All these errors are evaluated numerically for comprehensive engineering of the system. In particular, gap errors are found to place requirements beyond mechanical tolerances of {plus minus}25{mu}m, and amelioration of these may occur by a procedure utilizing direct measurement of the magnetic fields at assembly time. 4 refs., 12 figs.

  17. Inborn errors of metabolism

    MedlinePlus

    Metabolism - inborn errors of ... Bodamer OA. Approach to inborn errors of metabolism. In: Goldman L, Schafer AI, eds. Goldman's Cecil Medicine . 25th ed. Philadelphia, PA: Elsevier Saunders; 2015:chap 205. Rezvani I, Rezvani G. An ...

  18. Detection and avoidance of errors in computer software

    NASA Technical Reports Server (NTRS)

    Kinsler, Les

    1989-01-01

    The acceptance test errors of a computer software project to determine if the errors could be detected or avoided in earlier phases of development. GROAGSS (Gamma Ray Observatory Attitude Ground Support System) was selected as the software project to be examined. The development of the software followed the standard Flight Dynamics Software Development methods. GROAGSS was developed between August 1985 and April 1989. The project is approximately 250,000 lines of code of which approximately 43,000 lines are reused from previous projects. GROAGSS had a total of 1715 Change Report Forms (CRFs) submitted during the entire development and testing. These changes contained 936 errors. Of these 936 errors, 374 were found during the acceptance testing. These acceptance test errors were first categorized into methods of avoidance including: more clearly written requirements; detail review; code reading; structural unit testing; and functional system integration testing. The errors were later broken down in terms of effort to detect and correct, class of error, and probability that the prescribed detection method would be successful. These determinations were based on Software Engineering Laboratory (SEL) documents and interviews with the project programmers. A summary of the results of the categorizations is presented. The number of programming errors at the beginning of acceptance testing can be significantly reduced. The results of the existing development methodology are examined for ways of improvements. A basis is provided for the definition is a new development/testing paradigm. Monitoring of the new scheme will objectively determine its effectiveness on avoiding and detecting errors.

  19. Immediate error correction process following sleep deprivation.

    PubMed

    Hsieh, Shulan; Cheng, I-Chen; Tsai, Ling-Ling

    2007-06-01

    Previous studies have suggested that one night of sleep deprivation decreases frontal lobe metabolic activity, particularly in the anterior cingulated cortex (ACC), resulting in decreased performance in various executive function tasks. This study thus attempted to address whether sleep deprivation impaired the executive function of error detection and error correction. Sixteen young healthy college students (seven women, nine men, with ages ranging from 18 to 23 years) participated in this study. Participants performed a modified letter flanker task and were instructed to make immediate error corrections on detecting performance errors. Event-related potentials (ERPs) during the flanker task were obtained using a within-subject, repeated-measure design. The error negativity or error-related negativity (Ne/ERN) and the error positivity (Pe) seen immediately after errors were analyzed. The results show that the amplitude of the Ne/ERN was reduced significantly following sleep deprivation. Reduction also occurred for error trials with subsequent correction, indicating that sleep deprivation influenced error correction ability. This study further demonstrated that the impairment in immediate error correction following sleep deprivation was confined to specific stimulus types, with both Ne/ERN and behavioral correction rates being reduced only for trials in which flanker stimuli were incongruent with the target stimulus, while the response to the target was compatible with that of the flanker stimuli following sleep deprivation. The results thus warrant future systematic investigation of the interaction between stimulus type and error correction following sleep deprivation. PMID:17542943

  20. Programming Errors in APL.

    ERIC Educational Resources Information Center

    Kearsley, Greg P.

    This paper discusses and provides some preliminary data on errors in APL programming. Data were obtained by analyzing listings of 148 complete and partial APL sessions collected from student terminal rooms at the University of Alberta. Frequencies of errors for the various error messages are tabulated. The data, however, are limited because they…

  1. Disruption of N terminus long range non covalent interactions shifted temp.opt 25°C to cold: Evolution of point mutant Bacillus lipase by error prone PCR.

    PubMed

    Goomber, Shelly; Kumar, Arbind; Kaur, Jagdeep

    2016-01-15

    Cold adapted enzymes have applications in detergent, textile, food, bioremediation and biotechnology processes. Bacillus lipases are 'generally recognized as safe' (GRAS) and hence are industrially attractive. Bacillus lipase of 1.4 subfamily are of lowest molecular weight and are reversibly unfolded due to absence of disulphide bonds. Therefore these are largely used to study energetic of protein stability that represents unfolding of native protein to fully unfolded state. In present study, metagenomically isolated Bacillus LipJ was laboratory evolved for cold adaptation by error Prone PCR. Library of variants were screened for high relative activity at low temperature of 10°C compared to native protein LipJ. Point mutant sequenced as Phe19→Leu was determined to be active at cold and was selected for extensive biochemical, biophysical characterization. Variant F19L showed its maximum activity at 10°C where parent protein LipJ had 20% relative activity. Psychrophilic nature of F19L was established with about 50% relative active at 5°C where native protein was frozen to act. Variant F19L showed no activity at temperature 40°C and above, establishing its thermolabile nature. Thermostability studies determined mutant to be unstable above 20°C and three fold decrease in its half life at 30°C compared to native protein. Far UV-CD and intrinsic fluorescence study demonstrated unstable tertiary structure of point variant F19L leading to its unfolding at low temperature of 20°C. Cold adaptation of mutant F19L is accompanied with increased specific activity. Mutant was catalytically more efficient with 1.3 fold increase in kcat. Homologue structure modelling predicted disruption of intersecondary hydrophobic core formed by aromatic ring of Phe19 with non polar residues placed at β3, β4, β5, β6, αF. Increased local flexibility of variant F19L explains molecular basis of its psychrophilic nature. PMID:26456196

  2. Reduction of Maintenance Error Through Focused Interventions

    NASA Technical Reports Server (NTRS)

    Kanki, Barbara G.; Walter, Diane; Rosekind, Mark R. (Technical Monitor)

    1997-01-01

    It is well known that a significant proportion of aviation accidents and incidents are tied to human error. In flight operations, research of operational errors has shown that so-called "pilot error" often involves a variety of human factors issues and not a simple lack of individual technical skills. In aircraft maintenance operations, there is similar concern that maintenance errors which may lead to incidents and accidents are related to a large variety of human factors issues. Although maintenance error data and research are limited, industry initiatives involving human factors training in maintenance have become increasingly accepted as one type of maintenance error intervention. Conscientious efforts have been made in re-inventing the "team" concept for maintenance operations and in tailoring programs to fit the needs of technical operations. Nevertheless, there remains a dual challenge: to develop human factors interventions which are directly supported by reliable human error data, and to integrate human factors concepts into the procedures and practices of everyday technical tasks. In this paper, we describe several varieties of human factors interventions and focus on two specific alternatives which target problems related to procedures and practices; namely, 1) structured on-the-job training and 2) procedure re-design. We hope to demonstrate that the key to leveraging the impact of these solutions comes from focused interventions; that is, interventions which are derived from a clear understanding of specific maintenance errors, their operational context and human factors components.

  3. Refractive errors in children.

    PubMed

    Tongue, A C

    1987-12-01

    Optical correction of refractive errors in infants and young children is indicated when the refractive errors are sufficiently large to cause unilateral or bilateral amblyopia, if they are impairing the child's ability to function normally, or if the child has accommodative strabismus. Screening for refractive errors is important and should be performed as part of the annual physical examination in all verbal children. Screening for significant refractive errors in preverbal children is more difficult; however, the red reflex test of Bruckner is useful for the detection of anisometropic refractive errors. The photorefraction test, which is an adaptation of Bruckner's red reflex test, may prove to be a useful screening device for detecting bilateral as well as unilateral refractive errors. Objective testing as well as subjective testing enables ophthalmologists to prescribe proper optical correction for refractive errors for infants and children of any age. PMID:3317238

  4. Aircraft system modeling error and control error

    NASA Technical Reports Server (NTRS)

    Kulkarni, Nilesh V. (Inventor); Kaneshige, John T. (Inventor); Krishnakumar, Kalmanje S. (Inventor); Burken, John J. (Inventor)

    2012-01-01

    A method for modeling error-driven adaptive control of an aircraft. Normal aircraft plant dynamics is modeled, using an original plant description in which a controller responds to a tracking error e(k) to drive the component to a normal reference value according to an asymptote curve. Where the system senses that (1) at least one aircraft plant component is experiencing an excursion and (2) the return of this component value toward its reference value is not proceeding according to the expected controller characteristics, neural network (NN) modeling of aircraft plant operation may be changed. However, if (1) is satisfied but the error component is returning toward its reference value according to expected controller characteristics, the NN will continue to model operation of the aircraft plant according to an original description.

  5. Systematic Error Modeling and Bias Estimation

    PubMed Central

    Zhang, Feihu; Knoll, Alois

    2016-01-01

    This paper analyzes the statistic properties of the systematic error in terms of range and bearing during the transformation process. Furthermore, we rely on a weighted nonlinear least square method to calculate the biases based on the proposed models. The results show the high performance of the proposed approach for error modeling and bias estimation. PMID:27213386

  6. Systematic Error Modeling and Bias Estimation.

    PubMed

    Zhang, Feihu; Knoll, Alois

    2016-01-01

    This paper analyzes the statistic properties of the systematic error in terms of range and bearing during the transformation process. Furthermore, we rely on a weighted nonlinear least square method to calculate the biases based on the proposed models. The results show the high performance of the proposed approach for error modeling and bias estimation. PMID:27213386

  7. Digital system detects binary code patterns containing errors

    NASA Technical Reports Server (NTRS)

    Muller, R. M.; Tharpe, H. M., Jr.

    1966-01-01

    System of square loop magnetic cores associated with code input registers to react to input code patterns by reference to a group of control cores in such a manner that errors are canceled and patterns containing errors are accepted for amplification and processing. This technique improves reception capabilities in PCM telemetry systems.

  8. Accept or divert?

    PubMed

    Angelucci, P A

    1999-09-01

    Stretching scarce resources is more than a managerial issue. Should you accept the patient to an understaffed ICU or divert him to another facility? The intense "medical utility" controversy focuses on a situation that critical care nurses now face every day. PMID:10614370

  9. Approaches to acceptable risk

    SciTech Connect

    Whipple, C.

    1997-04-30

    Several alternative approaches to address the question {open_quotes}How safe is safe enough?{close_quotes} are reviewed and an attempt is made to apply the reasoning behind these approaches to the issue of acceptability of radiation exposures received in space. The approaches to the issue of the acceptability of technological risk described here are primarily analytical, and are drawn from examples in the management of environmental health risks. These include risk-based approaches, in which specific quantitative risk targets determine the acceptability of an activity, and cost-benefit and decision analysis, which generally focus on the estimation and evaluation of risks, benefits and costs, in a framework that balances these factors against each other. These analytical methods tend by their quantitative nature to emphasize the magnitude of risks, costs and alternatives, and to downplay other factors, especially those that are not easily expressed in quantitative terms, that affect acceptance or rejection of risk. Such other factors include the issues of risk perceptions and how and by whom risk decisions are made.

  10. 1984 Newbery Acceptance Speech.

    ERIC Educational Resources Information Center

    Cleary, Beverly

    1984-01-01

    This acceptance speech for an award honoring "Dear Mr. Henshaw," a book about feelings of a lonely child of divorce intended for eight-, nine-, and ten-year-olds, highlights children's letters to author. Changes in society that affect children, the inception of "Dear Mr. Henshaw," and children's reactions to books are highlighted. (EJS)

  11. Why was Relativity Accepted?

    NASA Astrophysics Data System (ADS)

    Brush, S. G.

    Historians of science have published many studies of the reception of Einstein's special and general theories of relativity. Based on a review of these studies, and my own research on the role of the light-bending prediction in the reception of general relativity, I discuss the role of three kinds of reasons for accepting relativity (1) empirical predictions and explanations; (2) social-psychological factors; and (3) aesthetic-mathematical factors. According to the historical studies, acceptance was a three-stage process. First, a few leading scientists adopted the special theory for aesthetic-mathematical reasons. In the second stage, their enthusiastic advocacy persuaded other scientists to work on the theory and apply it to problems currently of interest in atomic physics. The special theory was accepted by many German physicists by 1910 and had begun to attract some interest in other countries. In the third stage, the confirmation of Einstein's light-bending prediction attracted much public attention and forced all physicists to take the general theory of relativity seriously. In addition to light-bending, the explanation of the advance of Mercury's perihelion was considered strong evidence by theoretical physicists. The American astronomers who conducted successful tests of general relativity became defenders of the theory. There is little evidence that relativity was `socially constructed' but its initial acceptance was facilitated by the prestige and resources of its advocates.

  12. Comparing Absolute Error with Squared Error for Evaluating Empirical Models of Continuous Variables: Compositions, Implications, and Consequences

    NASA Astrophysics Data System (ADS)

    Gao, J.

    2014-12-01

    Reducing modeling error is often a major concern of empirical geophysical models. However, modeling errors can be defined in different ways: When the response variable is continuous, the most commonly used metrics are squared (SQ) and absolute (ABS) errors. For most applications, ABS error is the more natural, but SQ error is mathematically more tractable, so is often used as a substitute with little scientific justification. Existing literature has not thoroughly investigated the implications of using SQ error in place of ABS error, especially not geospatially. This study compares the two metrics through the lens of bias-variance decomposition (BVD). BVD breaks down the expected modeling error of each model evaluation point into bias (systematic error), variance (model sensitivity), and noise (observation instability). It offers a way to probe the composition of various error metrics. I analytically derived the BVD of ABS error and compared it with the well-known SQ error BVD, and found that not only the two metrics measure the characteristics of the probability distributions of modeling errors differently, but also the effects of these characteristics on the overall expected error are different. Most notably, under SQ error all bias, variance, and noise increase expected error, while under ABS error certain parts of the error components reduce expected error. Since manipulating these subtractive terms is a legitimate way to reduce expected modeling error, SQ error can never capture the complete story embedded in ABS error. I then empirically compared the two metrics with a supervised remote sensing model for mapping surface imperviousness. Pair-wise spatially-explicit comparison for each error component showed that SQ error overstates all error components in comparison to ABS error, especially variance-related terms. Hence, substituting ABS error with SQ error makes model performance appear worse than it actually is, and the analyst would more likely accept a

  13. Burst error correction extensions for large Reed Solomon codes

    NASA Technical Reports Server (NTRS)

    Owsley, P.

    1990-01-01

    Reed Solomon codes are powerful error correcting codes that include some of the best random and burst correcting codes currently known. It is well known that an (n,k) Reed Solomon code can correct up to (n - k)/2 errors. Many applications utilizing Reed Solomon codes require corrections of errors consisting primarily of bursts. In this paper, it is shown that the burst correcting ability of Reed Solomon codes can be increased beyond (n - k)/2 with an acceptable probability of miscorrect.

  14. Error detection method

    DOEpatents

    Olson, Eric J.

    2013-06-11

    An apparatus, program product, and method that run an algorithm on a hardware based processor, generate a hardware error as a result of running the algorithm, generate an algorithm output for the algorithm, compare the algorithm output to another output for the algorithm, and detect the hardware error from the comparison. The algorithm is designed to cause the hardware based processor to heat to a degree that increases the likelihood of hardware errors to manifest, and the hardware error is observable in the algorithm output. As such, electronic components may be sufficiently heated and/or sufficiently stressed to create better conditions for generating hardware errors, and the output of the algorithm may be compared at the end of the run to detect a hardware error that occurred anywhere during the run that may otherwise not be detected by traditional methodologies (e.g., due to cooling, insufficient heat and/or stress, etc.).

  15. 14 CFR 415.35 - Acceptable flight risk.

    Code of Federal Regulations, 2014 CFR

    2014-01-01

    ... 14 Aeronautics and Space 4 2014-01-01 2014-01-01 false Acceptable flight risk. 415.35 Section 415... Launch Range § 415.35 Acceptable flight risk. (a) Flight risk through orbital insertion or impact. Acceptable flight risk through orbital insertion for an orbital launch vehicle, and through impact for...

  16. 14 CFR 415.35 - Acceptable flight risk.

    Code of Federal Regulations, 2013 CFR

    2013-01-01

    ... 14 Aeronautics and Space 4 2013-01-01 2013-01-01 false Acceptable flight risk. 415.35 Section 415... Launch Range § 415.35 Acceptable flight risk. (a) Flight risk through orbital insertion or impact. Acceptable flight risk through orbital insertion for an orbital launch vehicle, and through impact for...

  17. 14 CFR 415.35 - Acceptable flight risk.

    Code of Federal Regulations, 2012 CFR

    2012-01-01

    ... 14 Aeronautics and Space 4 2012-01-01 2012-01-01 false Acceptable flight risk. 415.35 Section 415... Launch Range § 415.35 Acceptable flight risk. (a) Flight risk through orbital insertion or impact. Acceptable flight risk through orbital insertion for an orbital launch vehicle, and through impact for...

  18. Error coding simulations

    NASA Technical Reports Server (NTRS)

    Noble, Viveca K.

    1993-01-01

    There are various elements such as radio frequency interference (RFI) which may induce errors in data being transmitted via a satellite communication link. When a transmission is affected by interference or other error-causing elements, the transmitted data becomes indecipherable. It becomes necessary to implement techniques to recover from these disturbances. The objective of this research is to develop software which simulates error control circuits and evaluate the performance of these modules in various bit error rate environments. The results of the evaluation provide the engineer with information which helps determine the optimal error control scheme. The Consultative Committee for Space Data Systems (CCSDS) recommends the use of Reed-Solomon (RS) and convolutional encoders and Viterbi and RS decoders for error correction. The use of forward error correction techniques greatly reduces the received signal to noise needed for a certain desired bit error rate. The use of concatenated coding, e.g. inner convolutional code and outer RS code, provides even greater coding gain. The 16-bit cyclic redundancy check (CRC) code is recommended by CCSDS for error detection.

  19. Model Error Budgets

    NASA Technical Reports Server (NTRS)

    Briggs, Hugh C.

    2008-01-01

    An error budget is a commonly used tool in design of complex aerospace systems. It represents system performance requirements in terms of allowable errors and flows these down through a hierarchical structure to lower assemblies and components. The requirements may simply be 'allocated' based upon heuristics or experience, or they may be designed through use of physics-based models. This paper presents a basis for developing an error budget for models of the system, as opposed to the system itself. The need for model error budgets arises when system models are a principle design agent as is increasingly more common for poorly testable high performance space systems.

  20. Error coding simulations

    NASA Astrophysics Data System (ADS)

    Noble, Viveca K.

    1993-11-01

    There are various elements such as radio frequency interference (RFI) which may induce errors in data being transmitted via a satellite communication link. When a transmission is affected by interference or other error-causing elements, the transmitted data becomes indecipherable. It becomes necessary to implement techniques to recover from these disturbances. The objective of this research is to develop software which simulates error control circuits and evaluate the performance of these modules in various bit error rate environments. The results of the evaluation provide the engineer with information which helps determine the optimal error control scheme. The Consultative Committee for Space Data Systems (CCSDS) recommends the use of Reed-Solomon (RS) and convolutional encoders and Viterbi and RS decoders for error correction. The use of forward error correction techniques greatly reduces the received signal to noise needed for a certain desired bit error rate. The use of concatenated coding, e.g. inner convolutional code and outer RS code, provides even greater coding gain. The 16-bit cyclic redundancy check (CRC) code is recommended by CCSDS for error detection.

  1. Drug errors: consequences, mechanisms, and avoidance.

    PubMed

    Glavin, R J

    2010-07-01

    Medication errors are common throughout healthcare and result in significant human and financial cost. Prospective studies suggest that the error rate in anaesthesia is around one error in every 133 anaesthetics. There are several categories of medication error ranging from slips and lapses to fixation errors and deliberate violations. Violations may be more likely in organizations with a tendency to blame front-line workers, a tendency to deny the existence of latent conditions, and a blinkered pursuit of productivity indicators. In these organizations, borderline-tolerated conditions of use may occur which blur the distinction between safe and unsafe practice. Latent conditions will also make the error at the 'sharp end' more likely to result in actual patient harm. Several complementary strategies are proposed which may result in fewer medication errors. At the organizational level, developing a safety culture and promoting robust error reporting systems is key. The individual anaesthetist can play a part in this, setting an example to other members of the team in vigilance for errors, creating a safety climate with psychological safety, and reporting and learning from errors. PMID:20507858

  2. Acceptability of human risk.

    PubMed Central

    Kasperson, R E

    1983-01-01

    This paper has three objectives: to explore the nature of the problem implicit in the term "risk acceptability," to examine the possible contributions of scientific information to risk standard-setting, and to argue that societal response is best guided by considerations of process rather than formal methods of analysis. Most technological risks are not accepted but are imposed. There is also little reason to expect consensus among individuals on their tolerance of risk. Moreover, debates about risk levels are often at base debates over the adequacy of the institutions which manage the risks. Scientific information can contribute three broad types of analyses to risk-setting deliberations: contextual analysis, equity assessment, and public preference analysis. More effective risk-setting decisions will involve attention to the process used, particularly in regard to the requirements of procedural justice and democratic responsibility. PMID:6418541

  3. Everyday Scale Errors

    ERIC Educational Resources Information Center

    Ware, Elizabeth A.; Uttal, David H.; DeLoache, Judy S.

    2010-01-01

    Young children occasionally make "scale errors"--they attempt to fit their bodies into extremely small objects or attempt to fit a larger object into another, tiny, object. For example, a child might try to sit in a dollhouse-sized chair or try to stuff a large doll into it. Scale error research was originally motivated by parents' and…

  4. Medical error and disclosure.

    PubMed

    White, Andrew A; Gallagher, Thomas H

    2013-01-01

    Errors occur commonly in healthcare and can cause significant harm to patients. Most errors arise from a combination of individual, system, and communication failures. Neurologists may be involved in harmful errors in any practice setting and should familiarize themselves with tools to prevent, report, and examine errors. Although physicians, patients, and ethicists endorse candid disclosure of harmful medical errors to patients, many physicians express uncertainty about how to approach these conversations. A growing body of research indicates physicians often fail to meet patient expectations for timely and open disclosure. Patients desire information about the error, an apology, and a plan for preventing recurrence of the error. To meet these expectations, physicians should participate in event investigations and plan thoroughly for each disclosure conversation, preferably with a disclosure coach. Physicians should also anticipate and attend to the ongoing medical and emotional needs of the patient. A cultural change towards greater transparency following medical errors is in motion. Substantial progress is still required, but neurologists can further this movement by promoting policies and environments conducive to open reporting, respectful disclosure to patients, and support for the healthcare workers involved. PMID:24182370

  5. Age and Acceptance of Euthanasia.

    ERIC Educational Resources Information Center

    Ward, Russell A.

    1980-01-01

    Study explores relationship between age (and sex and race) and acceptance of euthanasia. Women and non-Whites were less accepting because of religiosity. Among older people less acceptance was attributable to their lesser education and greater religiosity. Results suggest that quality of life in old age affects acceptability of euthanasia. (Author)

  6. Evaluation of Astrometry Errors due to the Optical Surface Distortions in Adaptive Optics Systems and Science Instruments

    NASA Astrophysics Data System (ADS)

    Ellerbroek, Brent; Herriot, Glen; Suzuki, Ryuji; Matthias, Schoeck

    2013-12-01

    The objectives for high precision astrometry on ELTs will be challenging, with requirements in the range from 10 to 50 micro-arc-seconds for some instruments and science cases. Reducing and correctly calibrating the systematic and quasi-static errors introduced by optical surface distortions will be an important part of meeting these goals. In a recently submitted paper, we described an analytical Fourier domain model for evaluating these effects as the sum of three terms: (i) under-sampling errors, due to measuring the effects of static surface distortions using a finite number of discrete reference sources; (ii) unknown beam wander across the static surface distortions due to line-of-sight jitter or boresighting errors, and (iii) quasi-static errors due to slowly varying surface distortions. In this paper, we apply these methods to evaluating this term in the astrometry error budgets for the TMT Infrared Imaging Spectrograph (IRIS) and the facility AO system, NFIRAOS. The inputs to this exercise include the original top-down allocations for this error term, the original optical surface specifications for IRIS and NFIRAOS as derived earlier on the basis of wavefront error requirements, our assessment of the feasible density and positioning accuracy for an array of calibration sources, and the expected beam wander due to tip/tilt jitter and bore-sighting errors between NFIRAOS and IRIS. The astrometry error computed for these initial parameters was considerably larger than the top-down allocation due to the contributions from the NFIRAOS double-pane entrance window, which is close to the system's input focal plane. The error can be reduced to fall within the allocation by defining tighter, but still feasible, specifications for these elements. We also evaluated the astrometry errors due to quasi-static drift of the figures of the NFIRAOS deformable mirrors, and determined that they are acceptable for RMS wavefront distortions of up to about 30 nm RMS.

  7. Reduced discretization error in HZETRN

    SciTech Connect

    Slaba, Tony C.; Blattnig, Steve R.; Tweed, John

    2013-02-01

    The deterministic particle transport code HZETRN is an efficient analysis tool for studying the effects of space radiation on humans, electronics, and shielding materials. In a previous work, numerical methods in the code were reviewed, and new methods were developed that further improved efficiency and reduced overall discretization error. It was also shown that the remaining discretization error could be attributed to low energy light ions (A < 4) with residual ranges smaller than the physical step-size taken by the code. Accurately resolving the spectrum of low energy light particles is important in assessing risk associated with astronaut radiation exposure. In this work, modifications to the light particle transport formalism are presented that accurately resolve the spectrum of low energy light ion target fragments. The modified formalism is shown to significantly reduce overall discretization error and allows a physical approximation to be removed. For typical step-sizes and energy grids used in HZETRN, discretization errors for the revised light particle transport algorithms are shown to be less than 4% for aluminum and water shielding thicknesses as large as 100 g/cm{sup 2} exposed to both solar particle event and galactic cosmic ray environments.

  8. Uncorrected refractive errors

    PubMed Central

    Naidoo, Kovin S; Jaggernath, Jyoti

    2012-01-01

    Global estimates indicate that more than 2.3 billion people in the world suffer from poor vision due to refractive error; of which 670 million people are considered visually impaired because they do not have access to corrective treatment. Refractive errors, if uncorrected, results in an impaired quality of life for millions of people worldwide, irrespective of their age, sex and ethnicity. Over the past decade, a series of studies using a survey methodology, referred to as Refractive Error Study in Children (RESC), were performed in populations with different ethnic origins and cultural settings. These studies confirmed that the prevalence of uncorrected refractive errors is considerably high for children in low-and-middle-income countries. Furthermore, uncorrected refractive error has been noted to have extensive social and economic impacts, such as limiting educational and employment opportunities of economically active persons, healthy individuals and communities. The key public health challenges presented by uncorrected refractive errors, the leading cause of vision impairment across the world, require urgent attention. To address these issues, it is critical to focus on the development of human resources and sustainable methods of service delivery. This paper discusses three core pillars to addressing the challenges posed by uncorrected refractive errors: Human Resource (HR) Development, Service Development and Social Entrepreneurship. PMID:22944755

  9. Uncorrected refractive errors.

    PubMed

    Naidoo, Kovin S; Jaggernath, Jyoti

    2012-01-01

    Global estimates indicate that more than 2.3 billion people in the world suffer from poor vision due to refractive error; of which 670 million people are considered visually impaired because they do not have access to corrective treatment. Refractive errors, if uncorrected, results in an impaired quality of life for millions of people worldwide, irrespective of their age, sex and ethnicity. Over the past decade, a series of studies using a survey methodology, referred to as Refractive Error Study in Children (RESC), were performed in populations with different ethnic origins and cultural settings. These studies confirmed that the prevalence of uncorrected refractive errors is considerably high for children in low-and-middle-income countries. Furthermore, uncorrected refractive error has been noted to have extensive social and economic impacts, such as limiting educational and employment opportunities of economically active persons, healthy individuals and communities. The key public health challenges presented by uncorrected refractive errors, the leading cause of vision impairment across the world, require urgent attention. To address these issues, it is critical to focus on the development of human resources and sustainable methods of service delivery. This paper discusses three core pillars to addressing the challenges posed by uncorrected refractive errors: Human Resource (HR) Development, Service Development and Social Entrepreneurship. PMID:22944755

  10. Baby-Crying Acceptance

    NASA Astrophysics Data System (ADS)

    Martins, Tiago; de Magalhães, Sérgio Tenreiro

    The baby's crying is his most important mean of communication. The crying monitoring performed by devices that have been developed doesn't ensure the complete safety of the child. It is necessary to join, to these technological resources, means of communicating the results to the responsible, which would involve the digital processing of information available from crying. The survey carried out, enabled to understand the level of adoption, in the continental territory of Portugal, of a technology that will be able to do such a digital processing. It was used the TAM as the theoretical referential. The statistical analysis showed that there is a good probability of acceptance of such a system.

  11. Insulin use: preventable errors.

    PubMed

    2014-01-01

    Insulin is vital for patients with type 1 diabetes and useful for certain patients with type 2 diabetes. The serious consequences of insulin-related medication errors are overdose, resulting in severe hypoglycaemia, causing seizures, coma and even death; or underdose, resulting in hyperglycaemia and sometimes ketoacidosis. Errors associated with the preparation and administration of insulin are often reported, both outside and inside the hospital setting. These errors are preventable. By analysing reports from organisations devoted to medication error prevention and from poison control centres, as well as a few studies and detailed case reports of medication errors, various types of error associated with insulin use have been identified, especially in the hospital setting. Generally, patients know more about the practicalities of their insulin treatment than healthcare professionals with intermittent involvement. Medication errors involving insulin can occur at each step of the medication-use process: prescribing, data entry, preparation, dispensing and administration. When prescribing insulin, wrong-dose errors have been caused by the use of abbreviations, especially "U" instead of the word "units" (often resulting in a 10-fold overdose because the "U" is read as a zero), or by failing to write the drug's name correctly or in full. In electronic prescribing, the sheer number of insulin products is a source of confusion and, ultimately, wrong-dose errors, and often overdose. Prescribing, dispensing or administration software is rarely compatible with insulin prescriptions in which the dose is adjusted on the basis of the patient's subsequent capillary blood glucose readings, and can therefore generate errors. When preparing and dispensing insulin, a tuberculin syringe is sometimes used instead of an insulin syringe, leading to overdose. Other errors arise from confusion created by similar packaging, between different insulin products or between insulin and other

  12. Error Prevention Aid

    NASA Technical Reports Server (NTRS)

    1987-01-01

    In a complex computer environment there is ample opportunity for error, a mistake by a programmer, or a software-induced undesirable side effect. In insurance, errors can cost a company heavily, so protection against inadvertent change is a must for the efficient firm. The data processing center at Transport Life Insurance Company has taken a step to guard against accidental changes by adopting a software package called EQNINT (Equations Interpreter Program). EQNINT cross checks the basic formulas in a program against the formulas that make up the major production system. EQNINT assures that formulas are coded correctly and helps catch errors before they affect the customer service or its profitability.

  13. The Rate of Physicochemical Incompatibilities, Administration Errors. Factors Correlating with Nurses' Errors.

    PubMed

    Fahimi, Fanak; Sefidani Forough, Aida; Taghikhani, Sepideh; Saliminejad, Leila

    2015-01-01

    Medication errors are commonly encountered in hospital setting. Intravenous medications pose particular risks because of their greater complexity and the multiple steps required in their preparation, administration and monitoring. We aimed to determine the rate of errors during the preparation and administration phase of intravenous medications and the correlation of these errors with the demographics of nurses involved in the process. One hundred patients who were receiving IV medications were monitored by a trained pharmacist. The researcher accompanied the nurses during the preparation and administration process of IV medications. Collected data were compared with the acceptable guidelines. A checklist was filled for each IV medication. Demographic data of the nurses were collected as well. A total of 454 IV medications were recorded. Inappropriate administration rate constituted a large proportion of errors in our study (35.3%). No significant or life threatening drug interaction was recorded during the study. Evaluating the impact of the nurses' demographic characteristics on the incidence of medication errors showed that there is a direct correlation between nurses' employment status and the rate of medication errors, while other characteristics did not show a significant impact on the rate of administration errors. Administration errors were significantly higher in temporary 1-year contract group than other groups (p-value < 0.0001). Study results show that there should be more vigilance on administration rate of IV medications to prevent negative consequences especially by pharmacists. Optimizing the working conditions of nurses may play a crucial role. PMID:26185509

  14. The Rate of Physicochemical Incompatibilities, Administration Errors. Factors Correlating with Nurses' Errors

    PubMed Central

    Fahimi, Fanak; Sefidani Forough, Aida; Taghikhani, Sepideh; Saliminejad, Leila

    2015-01-01

    Medication errors are commonly encountered in hospital setting. Intravenous medications pose particular risks because of their greater complexity and the multiple steps required in their preparation, administration and monitoring. We aimed to determine the rate of errors during the preparation and administration phase of intravenous medications and the correlation of these errors with the demographics of nurses involved in the process. One hundred patients who were receiving IV medications were monitored by a trained pharmacist. The researcher accompanied the nurses during the preparation and administration process of IV medications. Collected data were compared with the acceptable guidelines. A checklist was filled for each IV medication. Demographic data of the nurses were collected as well. A total of 454 IV medications were recorded. Inappropriate administration rate constituted a large proportion of errors in our study (35.3%). No significant or life threatening drug interaction was recorded during the study. Evaluating the impact of the nurses’ demographic characteristics on the incidence of medication errors showed that there is a direct correlation between nurses’ employment status and the rate of medication errors, while other characteristics did not show a significant impact on the rate of administration errors. Administration errors were significantly higher in temporary 1-year contract group than other groups (p-value < 0.0001). Study results show that there should be more vigilance on administration rate of IV medications to prevent negative consequences especially by pharmacists. Optimizing the working conditions of nurses may play a crucial role. PMID:26185509

  15. Facts about Refractive Errors

    MedlinePlus

    ... the lens can cause refractive errors. What is refraction? Refraction is the bending of light as it passes ... rays entering the eye, causing a more precise refraction or focus. In many cases, contact lenses provide ...

  16. Errors in prenatal diagnosis.

    PubMed

    Anumba, Dilly O C

    2013-08-01

    Prenatal screening and diagnosis are integral to antenatal care worldwide. Prospective parents are offered screening for common fetal chromosomal and structural congenital malformations. In most developed countries, prenatal screening is routinely offered in a package that includes ultrasound scan of the fetus and the assay in maternal blood of biochemical markers of aneuploidy. Mistakes can arise at any point of the care pathway for fetal screening and diagnosis, and may involve individual or corporate systemic or latent errors. Special clinical circumstances, such as maternal size, fetal position, and multiple pregnancy, contribute to the complexities of prenatal diagnosis and to the chance of error. Clinical interventions may lead to adverse outcomes not caused by operator error. In this review I discuss the scope of the errors in prenatal diagnosis, and highlight strategies for their prevention and diagnosis, as well as identify areas for further research and study to enhance patient safety. PMID:23725900

  17. Error mode prediction.

    PubMed

    Hollnagel, E; Kaarstad, M; Lee, H C

    1999-11-01

    The study of accidents ('human errors') has been dominated by efforts to develop 'error' taxonomies and 'error' models that enable the retrospective identification of likely causes. In the field of Human Reliability Analysis (HRA) there is, however, a significant practical need for methods that can predict the occurrence of erroneous actions--qualitatively and quantitatively. The present experiment tested an approach for qualitative performance prediction based on the Cognitive Reliability and Error Analysis Method (CREAM). Predictions of possible erroneous actions were made for operators using different types of alarm systems. The data were collected as part of a large-scale experiment using professional nuclear power plant operators in a full scope simulator. The analysis showed that the predictions were correct in more than 70% of the cases, and also that the coverage of the predictions depended critically on the comprehensiveness of the preceding task analysis. PMID:10582035

  18. Pronominal Case-Errors

    ERIC Educational Resources Information Center

    Kaper, Willem

    1976-01-01

    Contradicts a previous assertion by C. Tanz that children commit substitution errors usually using objective pronoun forms for nominative ones. Examples from Dutch and German provide evidence that substitutions are made in both directions. (CHK)

  19. Estimating Bias Error Distributions

    NASA Technical Reports Server (NTRS)

    Liu, Tian-Shu; Finley, Tom D.

    2001-01-01

    This paper formulates the general methodology for estimating the bias error distribution of a device in a measuring domain from less accurate measurements when a minimal number of standard values (typically two values) are available. A new perspective is that the bias error distribution can be found as a solution of an intrinsic functional equation in a domain. Based on this theory, the scaling- and translation-based methods for determining the bias error distribution arc developed. These methods are virtually applicable to any device as long as the bias error distribution of the device can be sufficiently described by a power series (a polynomial) or a Fourier series in a domain. These methods have been validated through computational simulations and laboratory calibration experiments for a number of different devices.

  20. Computer acceptance of older adults.

    PubMed

    Nägle, Sibylle; Schmidt, Ludger

    2012-01-01

    Even though computers play a massive role in everyday life of modern societies, older adults, and especially older women, are less likely to use a computer, and they perform fewer activities on it than younger adults. To get a better understanding of the factors affecting older adults' intention towards and usage of computers, the Unified Theory of Acceptance and Usage of Technology (UTAUT) was applied as part of a more extensive study with 52 users and non-users of computers, ranging in age from 50 to 90 years. The model covers various aspects of computer usage in old age via four key constructs, namely performance expectancy, effort expectancy, social influences, and facilitating conditions, as well as the variables gender, age, experience, and voluntariness it. Interestingly, next to performance expectancy, facilitating conditions showed the strongest correlation with use as well as with intention. Effort expectancy showed no significant correlation with the intention of older adults to use a computer. PMID:22317258

  1. Error-Compensated Telescope

    NASA Technical Reports Server (NTRS)

    Meinel, Aden B.; Meinel, Marjorie P.; Stacy, John E.

    1989-01-01

    Proposed reflecting telescope includes large, low-precision primary mirror stage and small, precise correcting mirror. Correcting mirror machined under computer control to compensate for error in primary mirror. Correcting mirror machined by diamond cutting tool. Computer analyzes interferometric measurements of primary mirror to determine shape of surface of correcting mirror needed to compensate for errors in wave front reflected from primary mirror and commands position and movement of cutting tool accordingly.

  2. Thermodynamics of Error Correction

    NASA Astrophysics Data System (ADS)

    Sartori, Pablo; Pigolotti, Simone

    2015-10-01

    Information processing at the molecular scale is limited by thermal fluctuations. This can cause undesired consequences in copying information since thermal noise can lead to errors that can compromise the functionality of the copy. For example, a high error rate during DNA duplication can lead to cell death. Given the importance of accurate copying at the molecular scale, it is fundamental to understand its thermodynamic features. In this paper, we derive a universal expression for the copy error as a function of entropy production and work dissipated by the system during wrong incorporations. Its derivation is based on the second law of thermodynamics; hence, its validity is independent of the details of the molecular machinery, be it any polymerase or artificial copying device. Using this expression, we find that information can be copied in three different regimes. In two of them, work is dissipated to either increase or decrease the error. In the third regime, the protocol extracts work while correcting errors, reminiscent of a Maxwell demon. As a case study, we apply our framework to study a copy protocol assisted by kinetic proofreading, and show that it can operate in any of these three regimes. We finally show that, for any effective proofreading scheme, error reduction is limited by the chemical driving of the proofreading reaction.

  3. Characterizing Radar Raingauge Errors for NWP Assimilation

    NASA Astrophysics Data System (ADS)

    Dance, S.; Seed, A.

    2012-04-01

    The statistical characterisation of errors in quantitative precipitation estimates (QPE) is needed when generating QPE ensembles, combining multiple radars into a single mosaic, and when assimilating QPE into numerical weather prediction (NWP) models. The first step in the analysis was to characterise the errors at pixel resolution (1 km) as a function of radar specification, geographical location under the radar, and meteorology using data from 18 radars and 1500 rain gauges over a two-year period. The probability distribution of the radar - rain gauge residuals was evaluated and, as expected, the log-Normal distribution was found to fit the data better than the Normal distribution. Therefore the subsequent analysis was performed on the residuals expressed as decibels. The impact of beam width on the estimation errors was evaluated by comparing the errors from a one-degree S band radar (S1) with a two-degree S band radar (S2) for the same location (Brisbane) and time period. The standard deviation of the errors was found to increase by 0.2 dB per km for the S2 radar while the standard deviation for the S1 radar was constant out to the maximum range of 150 km. When data from all the S1 radars over the two years were pooled and compared with the S2 radars the standard deviation of the errors for the S1 radars increased by 0.1 dB per km compared with 0.25 dB per km for the S2 radars. The mean of the errors was found to vary significantly with range for all radars with underestimation at close range (< 30 km) and at far range (> 100 km). We think that this points to artefacts in the data due to clutter suppression at close range and over shooting the echo tops at the far range. The spatial distribution of the errors as a function of the altitude and roughness of the topography was investigated using the data from the S1 and S2 radars in Brisbane, but no relationship was found although there is clearly structure in the field. We also attempted to quantify the

  4. Dose error analysis for a scanned proton beam delivery system

    NASA Astrophysics Data System (ADS)

    Coutrakon, G.; Wang, N.; Miller, D. W.; Yang, Y.

    2010-12-01

    All particle beam scanning systems are subject to dose delivery errors due to errors in position, energy and intensity of the delivered beam. In addition, finite scan speeds, beam spill non-uniformities, and delays in detector, detector electronics and magnet responses will all contribute errors in delivery. In this paper, we present dose errors for an 8 × 10 × 8 cm3 target of uniform water equivalent density with 8 cm spread out Bragg peak and a prescribed dose of 2 Gy. Lower doses are also analyzed and presented later in the paper. Beam energy errors and errors due to limitations of scanning system hardware have been included in the analysis. By using Gaussian shaped pencil beams derived from measurements in the research room of the James M Slater Proton Treatment and Research Center at Loma Linda, CA and executing treatment simulations multiple times, statistical dose errors have been calculated in each 2.5 mm cubic voxel in the target. These errors were calculated by delivering multiple treatments to the same volume and calculating the rms variation in delivered dose at each voxel in the target. The variations in dose were the result of random beam delivery errors such as proton energy, spot position and intensity fluctuations. The results show that with reasonable assumptions of random beam delivery errors, the spot scanning technique yielded an rms dose error in each voxel less than 2% or 3% of the 2 Gy prescribed dose. These calculated errors are within acceptable clinical limits for radiation therapy.

  5. Telemetry location error in a forested habitat

    USGS Publications Warehouse

    Chu, D.S.; Hoover, B.A.; Fuller, M.R.; Geissler, P.H.

    1989-01-01

    The error associated with locations estimated by radio-telemetry triangulation can be large and variable in a hardwood forest. We assessed the magnitude and cause of telemetry location errors in a mature hardwood forest by using a 4-element Yagi antenna and compass bearings toward four transmitters, from 21 receiving sites. The distance error from the azimuth intersection to known transmitter locations ranged from 0 to 9251 meters. Ninety-five percent of the estimated locations were within 16 to 1963 meters, and 50% were within 99 to 416 meters of actual locations. Angles with 20o of parallel had larger distance errors than other angles. While angle appeared most important, greater distances and the amount of vegetation between receivers and transmitters also contributed to distance error.

  6. Human error in aviation operations

    NASA Technical Reports Server (NTRS)

    Nagel, David C.

    1988-01-01

    The role of human error in commercial and general aviation accidents and the techniques used to evaluate it are reviewed from a human-factors perspective. Topics addressed include the general decline in accidents per million departures since the 1960s, the increase in the proportion of accidents due to human error, methods for studying error, theoretical error models, and the design of error-resistant systems. Consideration is given to information acquisition and processing errors, visually guided flight, disorientation, instrument-assisted guidance, communication errors, decision errors, debiasing, and action errors.

  7. Errata: Papers in Error Analysis.

    ERIC Educational Resources Information Center

    Svartvik, Jan, Ed.

    Papers presented at the symposium of error analysis in Lund, Sweden, in September 1972, approach error analysis specifically in its relation to foreign language teaching and second language learning. Error analysis is defined as having three major aspects: (1) the description of the errors, (2) the explanation of errors by means of contrastive…

  8. CO2 laser ranging systems study

    NASA Technical Reports Server (NTRS)

    Filippi, C. A.

    1975-01-01

    The conceptual design and error performance of a CO2 laser ranging system are analyzed. Ranging signal and subsystem processing alternatives are identified, and their comprehensive evaluation yields preferred candidate solutions which are analyzed to derive range and range rate error contributions. The performance results are presented in the form of extensive tables and figures which identify the ranging accuracy compromises as a function of the key system design parameters and subsystem performance indexes. The ranging errors obtained are noted to be within the high accuracy requirements of existing NASA/GSFC missions with a proper system design.

  9. Smoothing error pitfalls

    NASA Astrophysics Data System (ADS)

    von Clarmann, T.

    2014-09-01

    The difference due to the content of a priori information between a constrained retrieval and the true atmospheric state is usually represented by a diagnostic quantity called smoothing error. In this paper it is shown that, regardless of the usefulness of the smoothing error as a diagnostic tool in its own right, the concept of the smoothing error as a component of the retrieval error budget is questionable because it is not compliant with Gaussian error propagation. The reason for this is that the smoothing error does not represent the expected deviation of the retrieval from the true state but the expected deviation of the retrieval from the atmospheric state sampled on an arbitrary grid, which is itself a smoothed representation of the true state; in other words, to characterize the full loss of information with respect to the true atmosphere, the effect of the representation of the atmospheric state on a finite grid also needs to be considered. The idea of a sufficiently fine sampling of this reference atmospheric state is problematic because atmospheric variability occurs on all scales, implying that there is no limit beyond which the sampling is fine enough. Even the idealization of infinitesimally fine sampling of the reference state does not help, because the smoothing error is applied to quantities which are only defined in a statistical sense, which implies that a finite volume of sufficient spatial extent is needed to meaningfully discuss temperature or concentration. Smoothing differences, however, which play a role when measurements are compared, are still a useful quantity if the covariance matrix involved has been evaluated on the comparison grid rather than resulting from interpolation and if the averaging kernel matrices have been evaluated on a grid fine enough to capture all atmospheric variations that the instruments are sensitive to. This is, under the assumptions stated, because the undefined component of the smoothing error, which is the

  10. Representing model error in ensemble data assimilation

    NASA Astrophysics Data System (ADS)

    Cardinali, C.; Žagar, N.; Radnoti, G.; Buizza, R.

    2014-09-01

    The paper investigates a method to represent model error in the ensemble data assimilation (EDA) system. The ECMWF operational EDA simulates the effect of both observations and model uncertainties. Observation errors are represented by perturbations with statistics characterized by the observation error covariance matrix whilst the model uncertainties are represented by stochastic perturbations added to the physical tendencies to simulate the effect of random errors in the physical parameterizations (ST-method). In this work an alternative method (XB-method) is proposed to simulate model uncertainties by adding perturbations to the model background field. In this way the error represented is not just restricted to model error in the usual sense but potentially extends to any form of background error. The perturbations have the same correlation as the background error covariance matrix and their magnitude is computed from comparing the high-resolution operational innovation variances with the ensemble variances when the ensemble is obtained by perturbing only the observations (OBS-method). The XB-method has been designed to represent the short range model error relevant for the data assimilation window. Spread diagnostic shows that the XB-method generates a larger spread than the ST-method that is operationally used at ECMWF, in particular in the extratropics. Three-dimensional normal-mode diagnostics indicate that XB-EDA spread projects more than the spread from the other EDAs onto the easterly inertia-gravity modes associated with equatorial Kelvin waves, tropical dynamics and, in general, model error sources. The background error statistics from the above described EDAs have been employed in the assimilation system. The assimilation system performance showed that the XB-method background error statistics increase the observation influence in the analysis process. The other EDA background error statistics, when inflated by a global factor, generate analyses with

  11. Recognition errors by honey bee (Apis mellifera) guards demonstrate overlapping cues in conspecific recognition

    PubMed Central

    Couvillon, Margaret J; Roy, Gabrielle G F; Ratnieks, Francis L W

    2015-01-01

    Summary Honey bee (Apis mellifera) entrance guards discriminate nestmates from intruders. We tested the hypothesis that the recognition cues between nestmate bees and intruder bees overlap by comparing their acceptances with that of worker common wasps, Vespula vulgaris, by entrance guards. If recognition cues of nestmate and non-nestmate bees overlap, we would expect recognition errors. Conversely, we hypothesised that guards would not make errors in recognizing wasps because wasps and bees should have distinct, non-overlapping cues. We found both to be true. There was a negative correlation between errors in recognizing nestmates (error: reject nestmate) and nonnestmates (error: accept non-nestmate) bees such that when guards were likely to reject nestmates, they were less likely to accept a nonnestmate; conversely, when guards were likely to accept a non-nestmate, they were less likely to reject a nestmate. There was, however, no correlation between errors in the recognition of nestmate bees (error: reject nestmate) and wasps (error: accept wasp), demonstrating that guards were able to reject wasps categorically. Our results strongly support that overlapping cue distributions occur, resulting in errors and leading to adaptive shifts in guard acceptance thresholds PMID:26005220

  12. Compact disk error measurements

    NASA Technical Reports Server (NTRS)

    Howe, D.; Harriman, K.; Tehranchi, B.

    1993-01-01

    The objectives of this project are as follows: provide hardware and software that will perform simple, real-time, high resolution (single-byte) measurement of the error burst and good data gap statistics seen by a photoCD player read channel when recorded CD write-once discs of variable quality (i.e., condition) are being read; extend the above system to enable measurement of the hard decision (i.e., 1-bit error flags) and soft decision (i.e., 2-bit error flags) decoding information that is produced/used by the Cross Interleaved - Reed - Solomon - Code (CIRC) block decoder employed in the photoCD player read channel; construct a model that uses data obtained via the systems described above to produce meaningful estimates of output error rates (due to both uncorrected ECC words and misdecoded ECC words) when a CD disc having specific (measured) error statistics is read (completion date to be determined); and check the hypothesis that current adaptive CIRC block decoders are optimized for pressed (DAD/ROM) CD discs. If warranted, do a conceptual design of an adaptive CIRC decoder that is optimized for write-once CD discs.

  13. Radar range measurements in the atmosphere.

    SciTech Connect

    Doerry, Armin Walter

    2013-02-01

    The earth's atmosphere affects the velocity of propagation of microwave signals. This imparts a range error to radar range measurements that assume the typical simplistic model for propagation velocity. This range error is a function of atmospheric constituents, such as water vapor, as well as the geometry of the radar data collection, notably altitude and range. Models are presented for calculating atmospheric effects on radar range measurements, and compared against more elaborate atmospheric models.

  14. An error control system with multiple-stage forward error corrections

    NASA Technical Reports Server (NTRS)

    Takata, Toyoo; Fujiwara, Toru; Kasami, Tadao; Lin, Shu

    1990-01-01

    A robust error-control coding system is presented. This system is a cascaded FEC (forward error control) scheme supported by parity retransmissions for further error correction in the erroneous data words. The error performance and throughput efficiency of the system are analyzed. Two specific examples of the error-control system are studied. The first example does not use an inner code, and the outer code, which is not interleaved, is a shortened code of the NASA standard RS code over GF(28). The second example, as proposed for NASA, uses the same shortened RS code as the base outer code C2, except that it is interleaved to a depth of 2. It is shown that both examples provide high reliability and throughput efficiency even for high channel bit-error rates in the range of 0.01.

  15. Experimental Quantum Error Detection

    PubMed Central

    Jin, Xian-Min; Yi, Zhen-Huan; Yang, Bin; Zhou, Fei; Yang, Tao; Peng, Cheng-Zhi

    2012-01-01

    Faithful transmission of quantum information is a crucial ingredient in quantum communication networks. To overcome the unavoidable decoherence in a noisy channel, to date, many efforts have been made to transmit one state by consuming large numbers of time-synchronized ancilla states. However, such huge demands of quantum resources are hard to meet with current technology and this restricts practical applications. Here we experimentally demonstrate quantum error detection, an economical approach to reliably protecting a qubit against bit-flip errors. Arbitrary unknown polarization states of single photons and entangled photons are converted into time bins deterministically via a modified Franson interferometer. Noise arising in both 10 m and 0.8 km fiber, which induces associated errors on the reference frame of time bins, is filtered when photons are detected. The demonstrated resource efficiency and state independence make this protocol a promising candidate for implementing a real-world quantum communication network. PMID:22953047

  16. Error propagation in hydrodynamics of lowland rivers due to uncertainty in vegetation roughness parameterization

    NASA Astrophysics Data System (ADS)

    Straatsma, Menno

    2010-05-01

    density in the lookup table led to a difference in water levels of 0.20, 0.20, and 0.36 m for Upper Rhine- Waal, Pannerdensch Kanaal-Nederrijn-Lek and the IJssel river respectively. The discharge distribution at the Pannerdensche Kop bifurcation point is 165 m3/s for both error sources, classification and lookup table. The discharge distribution at the IJsselkop is more sensitive for classification error than for errors in the lookup table (160 vs. 70 m3/s for range in classification error and interquartile range in lookup table error). Priority should be given to increasing the classification accuracy as this generates the largest error for water levels as well as discharge distribution. The quantification of the uncertainty in water levels and discharge distribution will help to make decisions more realistically as the error bands are substantiated. It can also influence the assessment of the height of the embankments as insight is given in the variability of the outcome of the flow models at design discharge. Moreover, the error bands may serve as an incentive to quantify the desired accuracy in the vegetation structural characteristics. This means that an upper limit can be put on the variation in water levels that is accepted from errors in the roughness parameterization.

  17. Effect of Field Errors in Muon Collider IR Magnets on Beam Dynamics

    SciTech Connect

    Alexahin, Y.; Gianfelice-Wendt, E.; Kapin, V.V.; /Fermilab

    2012-05-01

    In order to achieve peak luminosity of a Muon Collider (MC) in the 10{sup 35} cm{sup -2}s{sup -1} range very small values of beta-function at the interaction point (IP) are necessary ({beta}* {le} 1 cm) while the distance from IP to the first quadrupole can not be made shorter than {approx}6 m as dictated by the necessity of detector protection from backgrounds. In the result the beta-function at the final focus quadrupoles can reach 100 km making beam dynamics very sensitive to all kind of errors. In the present report we consider the effects on momentum acceptance and dynamic aperture of multipole field errors in the body of IR dipoles as well as of fringe-fields in both dipoles and quadrupoles in the ase of 1.5 TeV (c.o.m.) MC. Analysis shows these effects to be strong but correctable with dedicated multipole correctors.

  18. Surprise beyond prediction error

    PubMed Central

    Chumbley, Justin R; Burke, Christopher J; Stephan, Klaas E; Friston, Karl J; Tobler, Philippe N; Fehr, Ernst

    2014-01-01

    Surprise drives learning. Various neural “prediction error” signals are believed to underpin surprise-based reinforcement learning. Here, we report a surprise signal that reflects reinforcement learning but is neither un/signed reward prediction error (RPE) nor un/signed state prediction error (SPE). To exclude these alternatives, we measured surprise responses in the absence of RPE and accounted for a host of potential SPE confounds. This new surprise signal was evident in ventral striatum, primary sensory cortex, frontal poles, and amygdala. We interpret these findings via a normative model of surprise. PMID:24700400

  19. Evolution of error diffusion

    NASA Astrophysics Data System (ADS)

    Knox, Keith T.

    1999-10-01

    As we approach the new millennium, error diffusion is approaching the 25th anniversary of its invention. Because of its exceptionally high image quality, it continues to be a popular choice among digital halftoning algorithms. Over the last 24 years, many attempts have been made to modify and improve the algorithm--to eliminate unwanted textures and to extend it to printing media and color. Some of these modifications have been very successful and are in use today. This paper will review the history of the algorithm and its modifications. Three watershed events in the development of error diffusion will be described, together with the lessons learned along the way.

  20. Evolution of error diffusion

    NASA Astrophysics Data System (ADS)

    Knox, Keith T.

    1998-12-01

    As we approach the new millennium, error diffusion is approaching the 25th anniversary of its invention. Because of its exceptionally high image quality, it continues to be a popular choice among digital halftoning algorithms. Over the last 24 years, many attempts have been made to modify and improve the algorithm - to eliminate unwanted textures and to extend it to printing media and color. Some of these modifications have been very successful and are in use today. This paper will review the history of the algorithm and its modifications. Three watershed events in the development of error diffusion will be described, together with the lesions learned along the way.

  1. Error Free Software

    NASA Technical Reports Server (NTRS)

    1985-01-01

    A mathematical theory for development of "higher order" software to catch computer mistakes resulted from a Johnson Space Center contract for Apollo spacecraft navigation. Two women who were involved in the project formed Higher Order Software, Inc. to develop and market the system of error analysis and correction. They designed software which is logically error-free, which, in one instance, was found to increase productivity by 600%. USE.IT defines its objectives using AXES -- a user can write in English and the system converts to computer languages. It is employed by several large corporations.

  2. Simulation of rare events in quantum error correction

    NASA Astrophysics Data System (ADS)

    Bravyi, Sergey; Vargo, Alexander

    2013-12-01

    We consider the problem of calculating the logical error probability for a stabilizer quantum code subject to random Pauli errors. To access the regime of large code distances where logical errors are extremely unlikely we adopt the splitting method widely used in Monte Carlo simulations of rare events and Bennett's acceptance ratio method for estimating the free energy difference between two canonical ensembles. To illustrate the power of these methods in the context of error correction, we calculate the logical error probability PL for the two-dimensional surface code on a square lattice with a pair of holes for all code distances d≤20 and all error rates p below the fault-tolerance threshold. Our numerical results confirm the expected exponential decay PL˜exp[-α(p)d] and provide a simple fitting formula for the decay rate α(p). Both noiseless and noisy syndrome readout circuits are considered.

  3. Sonic boom acceptability studies

    NASA Astrophysics Data System (ADS)

    Shepherd, Kevin P.; Sullivan, Brenda M.; Leatherwood, Jack D.; McCurdy, David A.

    1992-04-01

    The determination of the magnitude of sonic boom exposure which would be acceptable to the general population requires, as a starting point, a method to assess and compare individual sonic booms. There is no consensus within the scientific and regulatory communities regarding an appropriate sonic boom assessment metric. Loudness, being a fundamental and well-understood attribute of human hearing was chosen as a means of comparing sonic booms of differing shapes and amplitudes. The figure illustrates the basic steps which yield a calculated value of loudness. Based upon the aircraft configuration and its operating conditions, the sonic boom pressure signature which reaches the ground is calculated. This pressure-time history is transformed to the frequency domain and converted into a one-third octave band spectrum. The essence of the loudness method is to account for the frequency response and integration characteristics of the auditory system. The result of the calculation procedure is a numerical description (perceived level, dB) which represents the loudness of the sonic boom waveform.

  4. Sonic boom acceptability studies

    NASA Technical Reports Server (NTRS)

    Shepherd, Kevin P.; Sullivan, Brenda M.; Leatherwood, Jack D.; Mccurdy, David A.

    1992-01-01

    The determination of the magnitude of sonic boom exposure which would be acceptable to the general population requires, as a starting point, a method to assess and compare individual sonic booms. There is no consensus within the scientific and regulatory communities regarding an appropriate sonic boom assessment metric. Loudness, being a fundamental and well-understood attribute of human hearing was chosen as a means of comparing sonic booms of differing shapes and amplitudes. The figure illustrates the basic steps which yield a calculated value of loudness. Based upon the aircraft configuration and its operating conditions, the sonic boom pressure signature which reaches the ground is calculated. This pressure-time history is transformed to the frequency domain and converted into a one-third octave band spectrum. The essence of the loudness method is to account for the frequency response and integration characteristics of the auditory system. The result of the calculation procedure is a numerical description (perceived level, dB) which represents the loudness of the sonic boom waveform.

  5. Automatically generated acceptance test: A software reliability experiment

    NASA Technical Reports Server (NTRS)

    Protzel, Peter W.

    1988-01-01

    This study presents results of a software reliability experiment investigating the feasibility of a new error detection method. The method can be used as an acceptance test and is solely based on empirical data about the behavior of internal states of a program. The experimental design uses the existing environment of a multi-version experiment previously conducted at the NASA Langley Research Center, in which the launch interceptor problem is used as a model. This allows the controlled experimental investigation of versions with well-known single and multiple faults, and the availability of an oracle permits the determination of the error detection performance of the test. Fault interaction phenomena are observed that have an amplifying effect on the number of error occurrences. Preliminary results indicate that all faults examined so far are detected by the acceptance test. This shows promise for further investigations, and for the employment of this test method on other applications.

  6. Human Factors Process Task Analysis: Liquid Oxygen Pump Acceptance Test Procedure at the Advanced Technology Development Center

    NASA Technical Reports Server (NTRS)

    Diorio, Kimberly A.; Voska, Ned (Technical Monitor)

    2002-01-01

    This viewgraph presentation provides information on Human Factors Process Failure Modes and Effects Analysis (HF PFMEA). HF PFMEA includes the following 10 steps: Describe mission; Define System; Identify human-machine; List human actions; Identify potential errors; Identify factors that effect error; Determine likelihood of error; Determine potential effects of errors; Evaluate risk; Generate solutions (manage error). The presentation also describes how this analysis was applied to a liquid oxygen pump acceptance test.

  7. Help prevent hospital errors

    MedlinePlus

    ... A.D.A.M. Editorial team. Related MedlinePlus Health Topics Medication Errors Patient Safety Browse the Encyclopedia A.D.A.M., Inc. is accredited by URAC, also known as the American Accreditation HealthCare Commission ... for online health information and services. Learn more about A.D. ...

  8. Orwell's Instructive Errors

    ERIC Educational Resources Information Center

    Julian, Liam

    2009-01-01

    In this article, the author talks about George Orwell, his instructive errors, and the manner in which Orwell pierced worthless theory, faced facts and defended decency (with fluctuating success), and largely ignored the tradition of accumulated wisdom that has rendered him a timeless teacher--one whose inadvertent lessons, while infrequently…

  9. Analyzing temozolomide medication errors: potentially fatal.

    PubMed

    Letarte, Nathalie; Gabay, Michael P; Bressler, Linda R; Long, Katie E; Stachnik, Joan M; Villano, J Lee

    2014-10-01

    The EORTC-NCIC regimen for glioblastoma requires different dosing of temozolomide (TMZ) during radiation and maintenance therapy. This complexity is exacerbated by the availability of multiple TMZ capsule strengths. TMZ is an alkylating agent and the major toxicity of this class is dose-related myelosuppression. Inadvertent overdose can be fatal. The websites of the Institute for Safe Medication Practices (ISMP), and the Food and Drug Administration (FDA) MedWatch database were reviewed. We searched the MedWatch database for adverse events associated with TMZ and obtained all reports including hematologic toxicity submitted from 1st November 1997 to 30th May 2012. The ISMP describes errors with TMZ resulting from the positioning of information on the label of the commercial product. The strength and quantity of capsules on the label were in close proximity to each other, and this has been changed by the manufacturer. MedWatch identified 45 medication errors. Patient errors were the most common, accounting for 21 or 47% of errors, followed by dispensing errors, which accounted for 13 or 29%. Seven reports or 16% were errors in the prescribing of TMZ. Reported outcomes ranged from reversible hematological adverse events (13%), to hospitalization for other adverse events (13%) or death (18%). Four error reports lacked detail and could not be categorized. Although the FDA issued a warning in 2003 regarding fatal medication errors and the product label warns of overdosing, errors in TMZ dosing occur for various reasons and involve both healthcare professionals and patients. Overdosing errors can be fatal. PMID:25026995

  10. Challenge and Error: Critical Events and Attention-Related Errors

    ERIC Educational Resources Information Center

    Cheyne, James Allan; Carriere, Jonathan S. A.; Solman, Grayden J. F.; Smilek, Daniel

    2011-01-01

    Attention lapses resulting from reactivity to task challenges and their consequences constitute a pervasive factor affecting everyday performance errors and accidents. A bidirectional model of attention lapses (error [image omitted] attention-lapse: Cheyne, Solman, Carriere, & Smilek, 2009) argues that errors beget errors by generating attention…

  11. Range Ecosystems

    Technology Transfer Automated Retrieval System (TEKTRAN)

    After more than two hundred years, grazing remains California’s most extensive land use. The ‘Range Ecosystems’ chapter in the ‘Ecosystems of California’ sourcebook provides an integrated picture of the biophysical, social, and economic aspects of lands grazed by livestock in the state. Grazing mana...

  12. Functional Error Models to Accelerate Nested Sampling

    NASA Astrophysics Data System (ADS)

    Josset, L.; Elsheikh, A. H.; Demyanov, V.; Lunati, I.

    2014-12-01

    Sampling algorithm, the proposed geostatistical realization is first evaluated through the approximate model to decide whether it is useful or not to perform a full physics simulation. This improves the acceptance rate of full physics simulations and opens the door to iteratively test the performance and improve the quality of the error model.

  13. Correcting for Sequencing Error in Maximum Likelihood Phylogeny Inference

    PubMed Central

    Kuhner, Mary K.; McGill, James

    2014-01-01

    Accurate phylogenies are critical to taxonomy as well as studies of speciation processes and other evolutionary patterns. Accurate branch lengths in phylogenies are critical for dating and rate measurements. Such accuracy may be jeopardized by unacknowledged sequencing error. We use simulated data to test a correction for DNA sequencing error in maximum likelihood phylogeny inference. Over a wide range of data polymorphism and true error rate, we found that correcting for sequencing error improves recovery of the branch lengths, even if the assumed error rate is up to twice the true error rate. Low error rates have little effect on recovery of the topology. When error is high, correction improves topological inference; however, when error is extremely high, using an assumed error rate greater than the true error rate leads to poor recovery of both topology and branch lengths. The error correction approach tested here was proposed in 2004 but has not been widely used, perhaps because researchers do not want to commit to an estimate of the error rate. This study shows that correction with an approximate error rate is generally preferable to ignoring the issue. PMID:25378476

  14. [The error, source of learning].

    PubMed

    Joyeux, Stéphanie; Bohic, Valérie

    2016-05-01

    The error itself is not recognised as a fault. It is the intentionality which differentiates between an error and a fault. An error is unintentional while a fault is a failure to respect known rules. The risk of error is omnipresent in health institutions. Public authorities have therefore set out a series of measures to reduce this risk. PMID:27155272

  15. Imagery of Errors in Typing

    ERIC Educational Resources Information Center

    Rieger, Martina; Martinez, Fanny; Wenke, Dorit

    2011-01-01

    Using a typing task we investigated whether insufficient imagination of errors and error corrections is related to duration differences between execution and imagination. In Experiment 1 spontaneous error imagination was investigated, whereas in Experiment 2 participants were specifically instructed to imagine errors. Further, in Experiment 2 we…

  16. Neural Correlates of Reach Errors

    PubMed Central

    Hashambhoy, Yasmin; Rane, Tushar; Shadmehr, Reza

    2005-01-01

    Reach errors may be broadly classified into errors arising from unpredictable changes in target location, called target errors, and errors arising from miscalibration of internal models, called execution errors. Execution errors may be caused by miscalibration of dynamics (e.g.. when a force field alters limb dynamics) or by miscalibration of kinematics (e.g., when prisms alter visual feedback). While all types of errors lead to similar online corrections, we found that the motor system showed strong trial-by-trial adaptation in response to random execution errors but not in response to random target errors. We used fMRI and a compatible robot to study brain regions involved in processing each kind of error. Both kinematic and dynamic execution errors activated regions along the central and the post-central sulci and in lobules V, VI, and VIII of the cerebellum, making these areas possible sites of plastic changes in internal models for reaching. Only activity related to kinematic errors extended into parietal area 5. These results are inconsistent with the idea that kinematics and dynamics of reaching are computed in separate neural entities. In contrast, only target errors caused increased activity in the striatum and the posterior superior parietal lobule. The cerebellum and motor cortex were as strongly activated as with execution errors. These findings indicate a neural and behavioral dissociation between errors that lead to switching of behavioral goals, and errors that lead to adaptation of internal models of limb dynamics and kinematics. PMID:16251440

  17. Automatic Error Analysis Using Intervals

    ERIC Educational Resources Information Center

    Rothwell, E. J.; Cloud, M. J.

    2012-01-01

    A technique for automatic error analysis using interval mathematics is introduced. A comparison to standard error propagation methods shows that in cases involving complicated formulas, the interval approach gives comparable error estimates with much less effort. Several examples are considered, and numerical errors are computed using the INTLAB…

  18. The Insufficiency of Error Analysis

    ERIC Educational Resources Information Center

    Hammarberg, B.

    1974-01-01

    The position here is that error analysis is inadequate, particularly from the language-teaching point of view. Non-errors must be considered in specifying the learner's current command of the language, its limits, and his learning tasks. A cyclic procedure of elicitation and analysis, to secure evidence of errors and non-errors, is outlined.…

  19. Control by model error estimation

    NASA Technical Reports Server (NTRS)

    Likins, P. W.; Skelton, R. E.

    1976-01-01

    Modern control theory relies upon the fidelity of the mathematical model of the system. Truncated modes, external disturbances, and parameter errors in linear system models are corrected by augmenting to the original system of equations an 'error system' which is designed to approximate the effects of such model errors. A Chebyshev error system is developed for application to the Large Space Telescope (LST).

  20. Errors in thermochromic liquid crystal thermometry

    NASA Astrophysics Data System (ADS)

    Wiberg, Roland; Lior, Noam

    2004-09-01

    This article experimentally investigates and assesses the errors that may be incurred in the hue-based thermochromic liquid crystal thermochromic liquid crystal (TLC) method, and their causes. The errors include response time, hysteresis, aging, surrounding illumination disturbance, direct illumination and viewing angle, amount of light into the camera, TLC thickness, digital resolution of the image conversion system, and measurement noise. Some of the main conclusions are that: (1) The 3×8 bits digital representation of the red green and blue TLC color values produces a temperature measurement error of typically 1% of the TLC effective temperature range, (2) an eight-fold variation of the light intensity into the camera produced variations, which were not discernable from the digital resolution error, (3) this temperature depends on the TLC film thickness, and (4) thicker films are less susceptible to aging and thickness nonuniformities.

  1. Manson's triple error.

    PubMed

    F, Delaporte

    2008-09-01

    The author discusses the significance, implications and limitations of Manson's work. How did Patrick Manson resolve some of the major problems raised by the filarial worm life cycle? The Amoy physician showed that circulating embryos could only leave the blood via the percutaneous route, thereby requiring a bloodsucking insect. The discovery of a new autonomous, airborne, active host undoubtedly had a considerable impact on the history of parasitology, but the way in which Manson formulated and solved the problem of the transfer of filarial worms from the body of the mosquito to man resulted in failure. This article shows how the epistemological transformation operated by Manson was indissociably related to a series of errors and how a major breakthrough can be the result of a series of false proposals and, consequently, that the history of truth often involves a history of error. PMID:18814729

  2. Modular error embedding

    DOEpatents

    Sandford, II, Maxwell T.; Handel, Theodore G.; Ettinger, J. Mark

    1999-01-01

    A method of embedding auxiliary information into the digital representation of host data containing noise in the low-order bits. The method applies to digital data representing analog signals, for example digital images. The method reduces the error introduced by other methods that replace the low-order bits with auxiliary information. By a substantially reverse process, the embedded auxiliary data can be retrieved easily by an authorized user through use of a digital key. The modular error embedding method includes a process to permute the order in which the host data values are processed. The method doubles the amount of auxiliary information that can be added to host data values, in comparison with bit-replacement methods for high bit-rate coding. The invention preserves human perception of the meaning and content of the host data, permitting the addition of auxiliary data in the amount of 50% or greater of the original host data.

  3. Error-Free Software

    NASA Technical Reports Server (NTRS)

    1989-01-01

    001 is an integrated tool suited for automatically developing ultra reliable models, simulations and software systems. Developed and marketed by Hamilton Technologies, Inc. (HTI), it has been applied in engineering, manufacturing, banking and software tools development. The software provides the ability to simplify the complex. A system developed with 001 can be a prototype or fully developed with production quality code. It is free of interface errors, consistent, logically complete and has no data or control flow errors. Systems can be designed, developed and maintained with maximum productivity. Margaret Hamilton, President of Hamilton Technologies, also directed the research and development of USE.IT, an earlier product which was the first computer aided software engineering product in the industry to concentrate on automatically supporting the development of an ultrareliable system throughout its life cycle. Both products originated in NASA technology developed under a Johnson Space Center contract.

  4. 14 CFR 415.35 - Acceptable flight risk.

    Code of Federal Regulations, 2010 CFR

    2010-01-01

    ... explosive debris. (b) Hazard identification and risk assessment. To demonstrate compliance with paragraph (a... 14 Aeronautics and Space 4 2010-01-01 2010-01-01 false Acceptable flight risk. 415.35 Section 415... Launch Range § 415.35 Acceptable flight risk. (a) Flight risk through orbital insertion or...

  5. 14 CFR 415.35 - Acceptable flight risk.

    Code of Federal Regulations, 2011 CFR

    2011-01-01

    ... explosive debris. (b) Hazard identification and risk assessment. To demonstrate compliance with paragraph (a... 14 Aeronautics and Space 4 2011-01-01 2011-01-01 false Acceptable flight risk. 415.35 Section 415... Launch Range § 415.35 Acceptable flight risk. (a) Flight risk through orbital insertion or...

  6. Error-correction coding

    NASA Technical Reports Server (NTRS)

    Hinds, Erold W. (Principal Investigator)

    1996-01-01

    This report describes the progress made towards the completion of a specific task on error-correcting coding. The proposed research consisted of investigating the use of modulation block codes as the inner code of a concatenated coding system in order to improve the overall space link communications performance. The study proposed to identify and analyze candidate codes that will complement the performance of the overall coding system which uses the interleaved RS (255,223) code as the outer code.

  7. Human Error In Complex Systems

    NASA Technical Reports Server (NTRS)

    Morris, Nancy M.; Rouse, William B.

    1991-01-01

    Report presents results of research aimed at understanding causes of human error in such complex systems as aircraft, nuclear powerplants, and chemical processing plants. Research considered both slips (errors of action) and mistakes (errors of intention), and influence of workload on them. Results indicated that: humans respond to conditions in which errors expected by attempting to reduce incidence of errors; and adaptation to conditions potent influence on human behavior in discretionary situations.

  8. [The notion and classification of expert errors].

    PubMed

    Klevno, V A

    2012-01-01

    The author presents the analysis of the legal and forensic medical literature concerning currently accepted concepts and classification of expert malpractice. He proposes a new easy-to-remember definition of the expert error and considers the classification of such mistakes. The analysis of the cases of erroneous application of the medical criteria for estimation of the harm to health made it possible to reveal and systematize the causes accounting for the cases of expert malpractice committed by forensic medical experts and health providers when determining the degree of harm to human health. PMID:22686055

  9. Accepters and Rejecters of Counseling.

    ERIC Educational Resources Information Center

    Rose, Harriett A.; Elton, Charles F.

    Personality differences between students who accept or reject proffered counseling assistance were investigated by comparing personality traits of 116 male students at the University of Kentucky who accepted or rejected letters of invitation to group counseling. Factor analysis of Omnibus Personality Inventory (OPI) scores to two groups of 60 and…

  10. Cone penetrometer acceptance test report

    SciTech Connect

    Boechler, G.N.

    1996-09-19

    This Acceptance Test Report (ATR) documents the results of acceptance test procedure WHC-SD-WM-ATR-151. Included in this report is a summary of the tests, the results and issues, the signature and sign- off ATP pages, and a summarized table of the specification vs. ATP section that satisfied the specification.

  11. The characteristics of key analysis errors

    NASA Astrophysics Data System (ADS)

    Caron, Jean-Francois

    This thesis investigates the characteristics of the corrections to the initial state of the atmosphere. The technique employed is the key analysis error algorithm, recently developed to estimate the initial state errors responsible for poor short-range to medium-range numerical weather prediction (NWP) forecasts. The main goal of this work is to determine to which extent the initial corrections obtained with this method can be associated with analysis errors. A secondary goal is to understand their dynamics in improving the forecast. In the first part of the thesis, we examine the realism of the initial corrections obtained from the key analysis error algorithm in terms of dynamical balance and closeness to the observations. The result showed that the initial corrections are strongly out of balance and systematically increase the departure between the control analysis and the observations suggesting that the key analysis error algorithm produced initial corrections that represent more than analysis errors. Significant artificial correction to the initial state seems to be present. The second part of this work examines a few approaches to isolate the balanced component of the initial corrections from the key analysis error method. The best results were obtained with the nonlinear balance potential vorticity (PV) inversion technique. The removal of the imbalance part of the initial corrections makes the corrected analysis slightly closer to the observations, but remains systematically further away as compared to the control analysis. Thus the balanced part of the key analysis errors cannot justifiably be associated with analysis errors. In light of the results presented, some recommendations to improve the key analysis error algorithm were proposed. In the third and last part of the thesis, a diagnosis of the evolution of the initial corrections from the key analysis error method is presented using a PV approach. The initial corrections tend to grow rapidly in time

  12. Evaluating a medical error taxonomy.

    PubMed Central

    Brixey, Juliana; Johnson, Todd R.; Zhang, Jiajie

    2002-01-01

    Healthcare has been slow in using human factors principles to reduce medical errors. The Center for Devices and Radiological Health (CDRH) recognizes that a lack of attention to human factors during product development may lead to errors that have the potential for patient injury, or even death. In response to the need for reducing medication errors, the National Coordinating Council for Medication Errors Reporting and Prevention (NCC MERP) released the NCC MERP taxonomy that provides a standard language for reporting medication errors. This project maps the NCC MERP taxonomy of medication error to MedWatch medical errors involving infusion pumps. Of particular interest are human factors associated with medical device errors. The NCC MERP taxonomy of medication errors is limited in mapping information from MEDWATCH because of the focus on the medical device and the format of reporting. PMID:12463789

  13. GY SAMPLING THEORY IN ENVIRONMENTAL STUDIES 2: SUBSAMPLING ERROR MEASUREMENTS

    EPA Science Inventory

    Sampling can be a significant source of error in the measurement process. The characterization and cleanup of hazardous waste sites require data that meet site-specific levels of acceptable quality if scientifically supportable decisions are to be made. In support of this effort,...

  14. Systematic lossy forward error protection for error-resilient digital video broadcasting

    NASA Astrophysics Data System (ADS)

    Rane, Shantanu D.; Aaron, Anne; Girod, Bernd

    2004-01-01

    We present a novel scheme for error-resilient digital video broadcasting,using the Wyner-Ziv coding paradigm. We apply the general framework of systematic lossy source-channel coding to generate a supplementary bitstream that can correct transmission errors in the decoded video waveform up to a certain residual distortion. The systematic portion consists of a conventional MPEG-coded bitstream, which is transmitted over the error-prone channel without forward error correction.The supplementary bitstream is a low rate representation of the transmitted video sequence generated using Wyner-Ziv encoding. We use the conventionally decoded error-concealed MPEG video sequence as side information to decode the Wyner-Ziv bits. The decoder combines the error-prone side information and the Wyner-Ziv description to yield an improved decoded video signal. Our results indicate that, over a large range of channel error probabilities, this scheme yields superior video quality when compared with traditional forward error correction techniques employed in digital video broadcasting.

  15. Errors in the use of isokinetic dynamometers.

    PubMed

    Winter, D A; Wells, R P; Orr, G W

    1981-01-01

    The use of constant velocity dynamometers in functional testing and in exercise and therapy programs has been increasing in the past few years. For movements in the horizontal plane there are no gravitational errors. However, in vertical movements the limbs are not only working against the dynamometer but also are either aided or opposed by gravity. Far too often these gravitational forces have not been taken into account, and the error involved can be quite large. 1. A study on four subjects using knee extension (against gravity) and flexion (with gravity) showed the error in mechanical work to vary from 26-43% in extension and from 55-510% in flexion. 2. A relatively simple solution is offered to overcome the problem by compensating for the gravitational errors over the full range of movement. The time course of the error is such as to cause erroneous magnitudes of recorded moments which are a maximum at low level contractions and minimum at high level contractions. 3. The effect of gravitational and other acceleration errors are such as to cause false early peaks in the muscle moment curves resulting in erroneous conclusions about muscle function. PMID:7196329

  16. Design of Large Momentum Acceptance Transport Systems

    SciTech Connect

    D.R. Douglas

    2005-05-01

    The use of energy recovery to enable high power linac operation often gives rise to an attendant challenge--the transport of high power beams subtending large phase space volumes. In particular applications--such as FEL driver accelerators--this manifests itself as a requirement for beam transport systems with large momentum acceptance. We will discuss the design, implementation, and operation of such systems. Though at times counterintuitive in behavior (perturbative descriptions may, for example, be misleading), large acceptance systems have been successfully utilized for generations as spectrometers and accelerator recirculators [1]. Such systems are in fact often readily designed using appropriate geometric descriptions of beam behavior; insight provided using such a perspective may in addition reveal inherent symmetries that simplify construction and improve operability. Our discussion will focus on two examples: the Bates-clone recirculator used in the Jefferson Lab 10 kW IR U pgrade FEL (which has an observed acceptance of 10% or more) and a compaction-managed mirror-bend achromat concept with an acceptance ranging from 50 to 150 MeV.

  17. Range and range rate system

    NASA Technical Reports Server (NTRS)

    Graham, Olin L. (Inventor); Russell, Jim K. (Inventor); Epperly, Walter L. (Inventor)

    1988-01-01

    A video controlled solid state range finding system which requires no radar, high power laser, or sophisticated laser target is disclosed. The effective range of the system is from 1 to about 200 ft. The system includes an opto-electric camera such as a lens CCD array device. A helium neon laser produces a source beam of coherent light which is applied to a beam splitter. The beam splitter applies a reference beam to the camera and produces an outgoing beam applied to a first angularly variable reflector which directs the outgoing beam to the distant object. An incoming beam is reflected from the object to a second angularly variable reflector which reflects the incoming beam to the opto-electric camera via the beam splitter. The first reflector and the second reflector are configured so that the distance travelled by the outgoing beam from the beam splitter and the first reflector is the same as the distance travelled by the incoming beam from the second reflector to the beam splitter. The reference beam produces a reference signal in the geometric center of the camera. The incoming beam produces an object signal at the camera.

  18. Speech Errors, Error Correction, and the Construction of Discourse.

    ERIC Educational Resources Information Center

    Linde, Charlotte

    Speech errors have been used in the construction of production models of the phonological and semantic components of language, and for a model of interactional processes. Errors also provide insight into how speakers plan discourse and syntactic structure,. Different types of discourse exhibit different types of error. The present data are taken…

  19. Laser Ranging Simulation Program

    NASA Technical Reports Server (NTRS)

    Piazolla, Sabino; Hemmati, Hamid; Tratt, David

    2003-01-01

    Laser Ranging Simulation Program (LRSP) is a computer program that predicts selected aspects of the performances of a laser altimeter or other laser ranging or remote-sensing systems and is especially applicable to a laser-based system used to map terrain from a distance of several kilometers. Designed to run in a more recent version (5 or higher) of the MATLAB programming language, LRSP exploits the numerical and graphical capabilities of MATLAB. LRSP generates a graphical user interface that includes a pop-up menu that prompts the user for the input of data that determine the performance of a laser ranging system. Examples of input data include duration and energy of the laser pulse, the laser wavelength, the width of the laser beam, and several parameters that characterize the transmitting and receiving optics, the receiving electronic circuitry, and the optical properties of the atmosphere and the terrain. When the input data have been entered, LRSP computes the signal-to-noise ratio as a function of range, signal and noise currents, and ranging and pointing errors.

  20. Skylab water balance error analysis

    NASA Technical Reports Server (NTRS)

    Leonard, J. I.

    1977-01-01

    Estimates of the precision of the net water balance were obtained for the entire Skylab preflight and inflight phases as well as for the first two weeks of flight. Quantitative estimates of both total sampling errors and instrumentation errors were obtained. It was shown that measurement error is minimal in comparison to biological variability and little can be gained from improvement in analytical accuracy. In addition, a propagation of error analysis demonstrated that total water balance error could be accounted for almost entirely by the errors associated with body mass changes. Errors due to interaction between terms in the water balance equation (covariances) represented less than 10% of the total error. Overall, the analysis provides evidence that daily measurements of body water changes obtained from the indirect balance technique are reasonable, precise, and relaible. The method is not biased toward net retention or loss.

  1. Standard Errors for Matrix Correlations.

    ERIC Educational Resources Information Center

    Ogasawara, Haruhiko

    1999-01-01

    Derives the asymptotic standard errors and intercorrelations for several matrix correlations assuming multivariate normality for manifest variables and derives the asymptotic standard errors of the matrix correlations for two factor-loading matrices. (SLD)

  2. Preventing medication errors in cancer chemotherapy.

    PubMed

    Cohen, M R; Anderson, R W; Attilio, R M; Green, L; Muller, R J; Pruemer, J M

    1996-04-01

    Recommendations for preventing medication errors in cancer chemotherapy are made. Before a health care provider is granted privileges to prescribe, dispense, or administer antineoplastic agents, he or she should undergo a tailored educational program and possibly testing or certification. Appropriate reference materials should be developed. Each institution should develop a dose-verification process with as many independent checks as possible. A detailed checklist covering prescribing, transcribing, dispensing, and administration should be used. Oral orders are not acceptable. All doses should be calculated independently by the physician, the pharmacist, and the nurse. Dosage limits should be established and a review process set up for doses that exceed the limits. These limits should be entered into pharmacy computer systems, listed on preprinted order forms, stated on the product packaging, placed in strategic locations in the institution, and communicated to employees. The prescribing vocabulary must be standardized. Acronyms, abbreviations, and brand names must be avoided and steps taken to avoid other sources of confusion in the written orders, such as trailing zeros. Preprinted antineoplastic drug order forms containing checklists can help avoid errors. Manufacturers should be encouraged to avoid or eliminate ambiguities in drug names and dosing information. Patients must be educated about all aspects of their cancer chemotherapy, as patients represent a last line of defense against errors. An interdisciplinary team at each practice site should review every medication error reported. Pharmacists should be involved at all sites where antineoplastic agents are dispensed. Although it may not be possible to eliminate all medication errors in cancer chemotherapy, the risk can be minimized through specific steps. Because of their training and experience, pharmacists should take the lead in this effort. PMID:8697025

  3. Modeling Errors in Daily Precipitation Measurements: Additive or Multiplicative?

    NASA Technical Reports Server (NTRS)

    Tian, Yudong; Huffman, George J.; Adler, Robert F.; Tang, Ling; Sapiano, Matthew; Maggioni, Viviana; Wu, Huan

    2013-01-01

    The definition and quantification of uncertainty depend on the error model used. For uncertainties in precipitation measurements, two types of error models have been widely adopted: the additive error model and the multiplicative error model. This leads to incompatible specifications of uncertainties and impedes intercomparison and application.In this letter, we assess the suitability of both models for satellite-based daily precipitation measurements in an effort to clarify the uncertainty representation. Three criteria were employed to evaluate the applicability of either model: (1) better separation of the systematic and random errors; (2) applicability to the large range of variability in daily precipitation; and (3) better predictive skills. It is found that the multiplicative error model is a much better choice under all three criteria. It extracted the systematic errors more cleanly, was more consistent with the large variability of precipitation measurements, and produced superior predictions of the error characteristics. The additive error model had several weaknesses, such as non constant variance resulting from systematic errors leaking into random errors, and the lack of prediction capability. Therefore, the multiplicative error model is a better choice.

  4. Accepted scientific research works (abstracts).

    PubMed

    2014-01-01

    These are the 39 accepted abstracts for IAYT's Symposium on Yoga Research (SYR) September 24-24, 2014 at the Kripalu Center for Yoga & Health and published in the Final Program Guide and Abstracts. PMID:25645134

  5. L-286 Acceptance Test Record

    SciTech Connect

    HARMON, B.C.

    2000-01-14

    This document provides a detailed account of how the acceptance testing was conducted for Project L-286, ''200E Area Sanitary Water Plant Effluent Stream Reduction''. The testing of the L-286 instrumentation system was conducted under the direct supervision

  6. Dose error analysis for a scanned proton beam delivery system.

    PubMed

    Coutrakon, G; Wang, N; Miller, D W; Yang, Y

    2010-12-01

    All particle beam scanning systems are subject to dose delivery errors due to errors in position, energy and intensity of the delivered beam. In addition, finite scan speeds, beam spill non-uniformities, and delays in detector, detector electronics and magnet responses will all contribute errors in delivery. In this paper, we present dose errors for an 8 × 10 × 8 cm(3) target of uniform water equivalent density with 8 cm spread out Bragg peak and a prescribed dose of 2 Gy. Lower doses are also analyzed and presented later in the paper. Beam energy errors and errors due to limitations of scanning system hardware have been included in the analysis. By using Gaussian shaped pencil beams derived from measurements in the research room of the James M Slater Proton Treatment and Research Center at Loma Linda, CA and executing treatment simulations multiple times, statistical dose errors have been calculated in each 2.5 mm cubic voxel in the target. These errors were calculated by delivering multiple treatments to the same volume and calculating the rms variation in delivered dose at each voxel in the target. The variations in dose were the result of random beam delivery errors such as proton energy, spot position and intensity fluctuations. The results show that with reasonable assumptions of random beam delivery errors, the spot scanning technique yielded an rms dose error in each voxel less than 2% or 3% of the 2 Gy prescribed dose. These calculated errors are within acceptable clinical limits for radiation therapy. PMID:21076200

  7. Uncertainty quantification and error analysis

    SciTech Connect

    Higdon, Dave M; Anderson, Mark C; Habib, Salman; Klein, Richard; Berliner, Mark; Covey, Curt; Ghattas, Omar; Graziani, Carlo; Seager, Mark; Sefcik, Joseph; Stark, Philip

    2010-01-01

    UQ studies all sources of error and uncertainty, including: systematic and stochastic measurement error; ignorance; limitations of theoretical models; limitations of numerical representations of those models; limitations on the accuracy and reliability of computations, approximations, and algorithms; and human error. A more precise definition for UQ is suggested below.

  8. Grammatical Errors and Communication Breakdown.

    ERIC Educational Resources Information Center

    Tomiyama, Machiko

    This study investigated the relationship between grammatical errors and communication breakdown by examining native speakers' ability to correct grammatical errors. The assumption was that communication breakdown exists to a certain degree if a native speaker cannot correct the error or if the correction distorts the information intended to be…

  9. Design Optimization for the Measurement Accuracy Improvement of a Large Range Nanopositioning Stage.

    PubMed

    Torralba, Marta; Yagüe-Fabra, José Antonio; Albajez, José Antonio; Aguilar, Juan José

    2016-01-01

    Both an accurate machine design and an adequate metrology loop definition are critical factors when precision positioning represents a key issue for the final system performance. This article discusses the error budget methodology as an advantageous technique to improve the measurement accuracy of a 2D-long range stage during its design phase. The nanopositioning platform NanoPla is here presented. Its specifications, e.g., XY-travel range of 50 mm × 50 mm and sub-micrometric accuracy; and some novel designed solutions, e.g., a three-layer and two-stage architecture are described. Once defined the prototype, an error analysis is performed to propose improvement design features. Then, the metrology loop of the system is mathematically modelled to define the propagation of the different sources. Several simplifications and design hypothesis are justified and validated, including the assumption of rigid body behavior, which is demonstrated after a finite element analysis verification. The different error sources and their estimated contributions are enumerated in order to conclude with the final error values obtained from the error budget. The measurement deviations obtained demonstrate the important influence of the working environmental conditions, the flatness error of the plane mirror reflectors and the accurate manufacture and assembly of the components forming the metrological loop. Thus, a temperature control of ±0.1 °C results in an acceptable maximum positioning error for the developed NanoPla stage, i.e., 41 nm, 36 nm and 48 nm in X-, Y- and Z-axis, respectively. PMID:26761014

  10. Design Optimization for the Measurement Accuracy Improvement of a Large Range Nanopositioning Stage

    PubMed Central

    Torralba, Marta; Yagüe-Fabra, José Antonio; Albajez, José Antonio; Aguilar, Juan José

    2016-01-01

    Both an accurate machine design and an adequate metrology loop definition are critical factors when precision positioning represents a key issue for the final system performance. This article discusses the error budget methodology as an advantageous technique to improve the measurement accuracy of a 2D-long range stage during its design phase. The nanopositioning platform NanoPla is here presented. Its specifications, e.g., XY-travel range of 50 mm × 50 mm and sub-micrometric accuracy; and some novel designed solutions, e.g., a three-layer and two-stage architecture are described. Once defined the prototype, an error analysis is performed to propose improvement design features. Then, the metrology loop of the system is mathematically modelled to define the propagation of the different sources. Several simplifications and design hypothesis are justified and validated, including the assumption of rigid body behavior, which is demonstrated after a finite element analysis verification. The different error sources and their estimated contributions are enumerated in order to conclude with the final error values obtained from the error budget. The measurement deviations obtained demonstrate the important influence of the working environmental conditions, the flatness error of the plane mirror reflectors and the accurate manufacture and assembly of the components forming the metrological loop. Thus, a temperature control of ±0.1 °C results in an acceptable maximum positioning error for the developed NanoPla stage, i.e., 41 nm, 36 nm and 48 nm in X-, Y- and Z-axis, respectively. PMID:26761014

  11. Errors inducing radiation overdoses.

    PubMed

    Grammaticos, Philip C

    2013-01-01

    There is no doubt that equipments exposing radiation and used for therapeutic purposes should be often checked for possibly administering radiation overdoses to the patients. Technologists, radiation safety officers, radiologists, medical physicists, healthcare providers and administration should take proper care on this issue. "We must be beneficial and not harmful to the patients", according to the Hippocratic doctrine. Cases of radiation overdose are often reported. A series of cases of radiation overdoses have recently been reported. Doctors who were responsible, received heavy punishments. It is much better to prevent than to treat an error or a disease. A Personal Smart Card or Score Card has been suggested for every patient undergoing therapeutic and/or diagnostic procedures by the use of radiation. Taxonomy may also help. PMID:24251304

  12. Beta systems error analysis

    NASA Technical Reports Server (NTRS)

    1984-01-01

    The atmospheric backscatter coefficient, beta, measured with an airborne CO Laser Doppler Velocimeter (LDV) system operating in a continuous wave, focussed model is discussed. The Single Particle Mode (SPM) algorithm, was developed from concept through analysis of an extensive amount of data obtained with the system on board a NASA aircraft. The SPM algorithm is intended to be employed in situations where one particle at a time appears in the sensitive volume of the LDV. In addition to giving the backscatter coefficient, the SPM algorithm also produces as intermediate results the aerosol density and the aerosol backscatter cross section distribution. A second method, which measures only the atmospheric backscatter coefficient, is called the Volume Mode (VM) and was simultaneously employed. The results of these two methods differed by slightly less than an order of magnitude. The measurement uncertainties or other errors in the results of the two methods are examined.

  13. Medical device error.

    PubMed

    Goodman, Gerald R

    2002-12-01

    This article discusses principal concepts for the analysis, classification, and reporting of problems involving medical device technology. We define a medical device in regulatory terminology and define and discuss concepts and terminology used to distinguish the causes and sources of medical device problems. Database classification systems for medical device failure tracking are presented, as are sources of information on medical device failures. The importance of near-accident reporting is discussed to alert users that reported medical device errors are typically limited to those that have caused an injury or death. This can represent only a fraction of the true number of device problems. This article concludes with a summary of the most frequently reported medical device failures by technology type, clinical application, and clinical setting. PMID:12400632

  14. Medication error detection in two major teaching hospitals: What are the types of errors?

    PubMed Central

    Saghafi, Fatemeh; Zargarzadeh, Amir H

    2014-01-01

    Background: Increasing number of reports on medication errors and relevant subsequent damages, especially in medical centers has become a growing concern for patient safety in recent decades. Patient safety and in particular, medication safety is a major concern and challenge for health care professionals around the world. Our prospective study was designed to detect prescribing, transcribing, dispensing, and administering medication errors in two major university hospitals. Materials and Methods: After choosing 20 similar hospital wards in two large teaching hospitals in the city of Isfahan, Iran, the sequence was randomly selected. Diagrams for drug distribution were drawn by the help of pharmacy directors. Direct observation technique was chosen as the method for detecting the errors. A total of 50 doses were studied in each ward to detect prescribing, transcribing and administering errors in each ward. The dispensing error was studied on 1000 doses dispensed in each hospital pharmacy. Results: A total of 8162 number of doses of medications were studied during the four stages, of which 8000 were complete data to be analyzed. 73% of prescribing orders were incomplete and did not have all six parameters (name, dosage form, dose and measuring unit, administration route, and intervals of administration). We found 15% transcribing errors. One-third of administration of medications on average was erroneous in both hospitals. Dispensing errors ranged between 1.4% and 2.2%. Conclusion: Although prescribing and administrating compromise most of the medication errors, improvements are needed in all four stages with regard to medication errors. Clear guidelines must be written and executed in both hospitals to reduce the incidence of medication errors. PMID:25364360

  15. Validation and acceptance of synthetic infrared imagery

    NASA Astrophysics Data System (ADS)

    Smith, Moira I.; Bernhardt, Mark; Angell, Christopher R.; Hickman, Duncan; Whitehead, Philip; Patel, Dilip

    2004-08-01

    This paper describes the use of an image query database (IQ-DB) tool as a means of implementing a validation strategy for synthetic long-wave infrared images of sea clutter. Specifically it was required to determine the validity of the synthetic imagery for use in developing and testing automatic target detection algorithms. The strategy adopted for exploiting synthetic imagery is outlined and the key issues of validation and acceptance are discussed in detail. A wide range of image metrics has been developed to achieve pre-defined validation criteria. A number of these metrics, which include post processing algorithms, are presented. Furthermore, the IQ-DB provides a robust mechanism for configuration management and control of the large volume of data used. The implementation of the IQ-DB is reviewed in terms of its cardinal point specification and its central role in synthetic imagery validation and EOSS progressive acceptance.

  16. Agriculture, forestry, range resources

    NASA Technical Reports Server (NTRS)

    Crea, W. J.

    1974-01-01

    In the area of crop specie identification, it has been found that temporal data analysis, preliminary stratification, and unequal probability analysis were several of the factors that contributed to high identification accuracies. Single data set accuracies on fields of greater than 80,000 sq m (20 acres) are in the 70- to 90-percent range; however, with the use of temporal data, accuracies of 95 percent have been reported. Identification accuracy drops off significantly on areas of less than 80,000 sq m (20 acres) as does measurement accuracy. Forest stratification into coniferous and deciduous areas has been accomplished to a 90- to 95-percent accuracy level. Using multistage sampling techniques, the timber volume of a national forest district has been estimated to a confidence level and standard deviation acceptable to the Forest Service at a very favorable cost-benefit time ratio. Range specie/plant community vegetation mapping has been accomplished at various levels of success (69- to 90-percent accuracy). However, several investigators have obtained encouraging initial results in range biomass (forage production) estimation and range readiness predictions. Soil association map correction and soil association mapping in new area appear to have been proven feasible on large areas; however, testing in a complex soil area should be undertaken.

  17. Register file soft error recovery

    DOEpatents

    Fleischer, Bruce M.; Fox, Thomas W.; Wait, Charles D.; Muff, Adam J.; Watson, III, Alfred T.

    2013-10-15

    Register file soft error recovery including a system that includes a first register file and a second register file that mirrors the first register file. The system also includes an arithmetic pipeline for receiving data read from the first register file, and error detection circuitry to detect whether the data read from the first register file includes corrupted data. The system further includes error recovery circuitry to insert an error recovery instruction into the arithmetic pipeline in response to detecting the corrupted data. The inserted error recovery instruction replaces the corrupted data in the first register file with a copy of the data from the second register file.

  18. Rapid mapping of volumetric errors

    SciTech Connect

    Krulewich, D.; Hale, L.; Yordy, D.

    1995-09-13

    This paper describes a relatively inexpensive, fast, and easy to execute approach to mapping the volumetric errors of a machine tool, coordinate measuring machine, or robot. An error map is used to characterize a machine or to improve its accuracy by compensating for the systematic errors. The method consists of three steps: (1) modeling the relationship between the volumetric error and the current state of the machine; (2) acquiring error data based on length measurements throughout the work volume; and (3) optimizing the model to the particular machine.

  19. Preventing Communication Errors in Telephone Medicine

    PubMed Central

    Reisman, Anna B; Brown, Karen E

    2005-01-01

    Errors in telephone communication can result in outcomes ranging from inconvenience and anxiety to serious compromises in patient safety. Although 25% of interactions between physicians and patients take place on the telephone, little has been written about telephone communication and medical mishaps. Similarly, training in telephone medicine skills is limited; only 6% of residency programs teach any aspect of telephone medicine. Increasing familiarity with common telephone challenges with patients may help physicians decrease the likelihood of negative outcomes. We use case vignettes to highlight communication errors in common telephone scenarios. These scenarios include giving sensitive test results, requests for narcotics, managing ill patients who are not sick enough for the emergency room, dealing with late-night calls, communicating with unintelligible patients, and handling calls from family members. We provide management strategies to minimize the occurrence of these errors. PMID:16191150

  20. Social aspects of clinical errors.

    PubMed

    Richman, Joel; Mason, Tom; Mason-Whitehead, Elizabeth; McIntosh, Annette; Mercer, Dave

    2009-08-01

    Clinical errors, whether committed by doctors, nurses or other professions allied to healthcare, remain a sensitive issue requiring open debate and policy formulation in order to reduce them. The literature suggests that the issues underpinning errors made by healthcare professionals involve concerns about patient safety, professional disclosure, apology, litigation, compensation, processes of recording and policy development to enhance quality service. Anecdotally, we are aware of narratives of minor errors, which may well have been covered up and remain officially undisclosed whilst the major errors resulting in damage and death to patients alarm both professionals and public with resultant litigation and compensation. This paper attempts to unravel some of these issues by highlighting the historical nature of clinical errors and drawing parallels to contemporary times by outlining the 'compensation culture'. We then provide an overview of what constitutes a clinical error and review the healthcare professional strategies for managing such errors. PMID:19201405

  1. Human factors error and patient monitoring.

    PubMed

    Walsh, T; Beatty, P C W

    2002-08-01

    A wide range of studies have shown that human factors errors are the major cause of critical incidents that threaten patient safety in the medical environments where patient monitoring takes place, contributing to approximately 87% of all such incidents. Studies have also shown that good cognitively ergonomic design of monitoring equipment for use in these environments should reduce the human factors errors associated with the information they provide. The purpose of this review is to consider the current state of knowledge concerning human factors engineering in its application to patient monitoring. It considers the prevalence of human factors error, principles of good human factors design, the effect of specific design features and the problem of the measurement of the effectiveness of designs in reducing human factors error. The conclusion of the review is that whilst the focus of human factors studies has, in recent years, moved from instrument design to organizational issues, patient monitor designers still have an important contribution to make to improving the safety of the monitored patient. Further, whilst better psychological understanding of the causes of human factors errors will in future guide better human factors engineering, in this area there are still many practical avenues of research that need exploring from the current base of understanding. PMID:12214768

  2. Validation and Error Metrics for the Atmospheric Compensation for a Landsat Land Surface Temperature Product

    NASA Astrophysics Data System (ADS)

    Cook, M. J.; Schott, J. R.

    2013-12-01

    An automated process for the atmospheric compensation for a Landsat land surface temperature product has been developed. Landsat data are very attractive for a global land surface temperature product because the spatial and temporal resolution and range of the imagery make them well matched to applications for the study of agriculture, the environment, weather, and climate among others. However, Landsat's single thermal band requires per-pixel atmospheric compensation and emissivity; this work focuses on the atmospheric compensation aspect of the process and will be integrated with ASTER derived emissivity data to output a land surface temperature product. For the same reasons Landsat is attractive, an automated atmospheric compensation technique is challenging; it requires atmospheric characterization over a large area and long time scale at an acceptable resolution. Using North American Regional Reanalysis (NARR) data, MODTRAN radiative transfer code, and a number of interpolation techniques, a tool has been developed to generate the necessary radiative transfer parameters at each pixel for any radiometrically calibrated North American Landsat scene in the archive. Initial validation of predicted temperatures using ground truth water temperatures from platforms and buoys verifies the fidelity of the process with good performance when the atmosphere is accurately characterized. However, performance is poorer when the composition of the atmosphere is not as well understood. Because of the desired automation and extent of the tool, we are limited in the availability of acceptable atmospheric profile data. The goal is to understand sources of error in order to predict and characterize the uncertainty in the retrieved temperatures. While the performance has been extensively tested using a number of NOAA buoys with bulk temperature measurements corrected to skin temperature, traditional error analysis is complicated by the atmospheric reanalysis, radiative transfer

  3. The 13 errors.

    PubMed

    Flower, J

    1998-01-01

    The reality is that most change efforts fail. McKinsey & Company carried out a fascinating research project on change to "crack the code" on creating and managing change in large organizations. One of the questions they asked--and answered--is why most organizations fail in their efforts to manage change. They found that 80 percent of these failures could be traced to 13 common errors. They are: (1) No winning strategy; (2) failure to make a compelling and urgent case for change; (3) failure to distinguish between decision-driven and behavior-dependent change; (4) over-reliance on structure and systems to change behavior; (5) lack of skills and resources; (6) failure to experiment; (7) leaders' inability or unwillingness to confront how they and their roles must change; (8) failure to mobilize and engage pivotal groups; (9) failure to understand and shape the informal organization; (10) inability to integrate and align all the initiatives; (11) no performance focus; (12) excessively open-ended process; and (13) failure to make the whole process transparent and meaningful to individuals. PMID:10351717

  4. Contour Error Map Algorithm

    NASA Technical Reports Server (NTRS)

    Merceret, Francis; Lane, John; Immer, Christopher; Case, Jonathan; Manobianco, John

    2005-01-01

    The contour error map (CEM) algorithm and the software that implements the algorithm are means of quantifying correlations between sets of time-varying data that are binarized and registered on spatial grids. The present version of the software is intended for use in evaluating numerical weather forecasts against observational sea-breeze data. In cases in which observational data come from off-grid stations, it is necessary to preprocess the observational data to transform them into gridded data. First, the wind direction is gridded and binarized so that D(i,j;n) is the input to CEM based on forecast data and d(i,j;n) is the input to CEM based on gridded observational data. Here, i and j are spatial indices representing 1.25-km intervals along the west-to-east and south-to-north directions, respectively; and n is a time index representing 5-minute intervals. A binary value of D or d = 0 corresponds to an offshore wind, whereas a value of D or d = 1 corresponds to an onshore wind. CEM includes two notable subalgorithms: One identifies and verifies sea-breeze boundaries; the other, which can be invoked optionally, performs an image-erosion function for the purpose of attempting to eliminate river-breeze contributions in the wind fields.

  5. Parental Reports of Children's Scale Errors in Everyday Life

    ERIC Educational Resources Information Center

    Rosengren, Karl S.; Gutierrez, Isabel T.; Anderson, Kathy N.; Schein, Stevie S.

    2009-01-01

    Scale errors refer to behaviors where young children attempt to perform an action on an object that is too small to effectively accommodate the behavior. The goal of this study was to examine the frequency and characteristics of scale errors in everyday life. To do so, the researchers collected parental reports of children's (age range = 13-21…

  6. 5 CFR 846.724 - Belated elections and correction of administrative errors.

    Code of Federal Regulations, 2010 CFR

    2010-01-01

    ... administrative errors. 846.724 Section 846.724 Administrative Personnel OFFICE OF PERSONNEL MANAGEMENT (CONTINUED... Open Enrollment Elections Election Procedures § 846.724 Belated elections and correction of administrative errors. (a) Belated elections. The employing office may accept a belated election of FERS...

  7. 5 CFR 846.724 - Belated elections and correction of administrative errors.

    Code of Federal Regulations, 2011 CFR

    2011-01-01

    ... administrative errors. 846.724 Section 846.724 Administrative Personnel OFFICE OF PERSONNEL MANAGEMENT (CONTINUED... Open Enrollment Elections Election Procedures § 846.724 Belated elections and correction of administrative errors. (a) Belated elections. The employing office may accept a belated election of FERS...

  8. Error analysis in laparoscopic surgery

    NASA Astrophysics Data System (ADS)

    Gantert, Walter A.; Tendick, Frank; Bhoyrul, Sunil; Tyrrell, Dana; Fujino, Yukio; Rangel, Shawn; Patti, Marco G.; Way, Lawrence W.

    1998-06-01

    Iatrogenic complications in laparoscopic surgery, as in any field, stem from human error. In recent years, cognitive psychologists have developed theories for understanding and analyzing human error, and the application of these principles has decreased error rates in the aviation and nuclear power industries. The purpose of this study was to apply error analysis to laparoscopic surgery and evaluate its potential for preventing complications. Our approach is based on James Reason's framework using a classification of errors according to three performance levels: at the skill- based performance level, slips are caused by attention failures, and lapses result form memory failures. Rule-based mistakes constitute the second level. Knowledge-based mistakes occur at the highest performance level and are caused by shortcomings in conscious processing. These errors committed by the performer 'at the sharp end' occur in typical situations which often times are brought about by already built-in latent system failures. We present a series of case studies in laparoscopic surgery in which errors are classified and the influence of intrinsic failures and extrinsic system flaws are evaluated. Most serious technical errors in lap surgery stem from a rule-based or knowledge- based mistake triggered by cognitive underspecification due to incomplete or illusory visual input information. Error analysis in laparoscopic surgery should be able to improve human performance, and it should detect and help eliminate system flaws. Complication rates in laparoscopic surgery due to technical errors can thus be considerably reduced.

  9. Per-beam, planar IMRT QA passing rates do not predict clinically relevant patient dose errors

    SciTech Connect

    Nelms, Benjamin E.; Zhen Heming; Tome, Wolfgang A.

    2011-02-15

    Purpose: The purpose of this work is to determine the statistical correlation between per-beam, planar IMRT QA passing rates and several clinically relevant, anatomy-based dose errors for per-patient IMRT QA. The intent is to assess the predictive power of a common conventional IMRT QA performance metric, the Gamma passing rate per beam. Methods: Ninety-six unique data sets were created by inducing four types of dose errors in 24 clinical head and neck IMRT plans, each planned with 6 MV Varian 120-leaf MLC linear accelerators using a commercial treatment planning system and step-and-shoot delivery. The error-free beams/plans were used as ''simulated measurements'' (for generating the IMRT QA dose planes and the anatomy dose metrics) to compare to the corresponding data calculated by the error-induced plans. The degree of the induced errors was tuned to mimic IMRT QA passing rates that are commonly achieved using conventional methods. Results: Analysis of clinical metrics (parotid mean doses, spinal cord max and D1cc, CTV D95, and larynx mean) vs IMRT QA Gamma analysis (3%/3 mm, 2/2, 1/1) showed that in all cases, there were only weak to moderate correlations (range of Pearson's r-values: -0.295 to 0.653). Moreover, the moderate correlations actually had positive Pearson's r-values (i.e., clinically relevant metric differences increased with increasing IMRT QA passing rate), indicating that some of the largest anatomy-based dose differences occurred in the cases of high IMRT QA passing rates, which may be called ''false negatives.'' The results also show numerous instances of false positives or cases where low IMRT QA passing rates do not imply large errors in anatomy dose metrics. In none of the cases was there correlation consistent with high predictive power of planar IMRT passing rates, i.e., in none of the cases did high IMRT QA Gamma passing rates predict low errors in anatomy dose metrics or vice versa. Conclusions: There is a lack of correlation between

  10. Defining acceptable conditions in wilderness

    NASA Astrophysics Data System (ADS)

    Roggenbuck, J. W.; Williams, D. R.; Watson, A. E.

    1993-03-01

    The limits of acceptable change (LAC) planning framework recognizes that forest managers must decide what indicators of wilderness conditions best represent resource naturalness and high-quality visitor experiences and how much change from the pristine is acceptable for each indicator. Visitor opinions on the aspects of the wilderness that have great impact on their experience can provide valuable input to selection of indicators. Cohutta, Georgia; Caney Creek, Arkansas; Upland Island, Texas; and Rattlesnake, Montana, wilderness visitors have high shared agreement that littering and damage to trees in campsites, noise, and seeing wildlife are very important influences on wilderness experiences. Camping within sight or sound of other people influences experience quality more than do encounters on the trails. Visitors’ standards of acceptable conditions within wilderness vary considerably, suggesting a potential need to manage different zones within wilderness for different clientele groups and experiences. Standards across wildernesses, however, are remarkably similar.

  11. From requirements to acceptance tests

    NASA Technical Reports Server (NTRS)

    Baize, Lionel; Pasquier, Helene

    1993-01-01

    From user requirements definition to accepted software system, the software project management wants to be sure that the system will meet the requirements. For the development of a telecommunication satellites Control Centre, C.N.E.S. has used new rules to make the use of tracing matrix easier. From Requirements to Acceptance Tests, each item of a document must have an identifier. A unique matrix traces the system and allows the tracking of the consequences of a change in the requirements. A tool has been developed, to import documents into a relational data base. Each record of the data base corresponds to an item of a document, the access key is the item identifier. Tracing matrix is also processed, providing automatically links between the different documents. It enables the reading on the same screen of traced items. For example one can read simultaneously the User Requirements items, the corresponding Software Requirements items and the Acceptance Tests.

  12. Error correction for IFSAR

    DOEpatents

    Doerry, Armin W.; Bickel, Douglas L.

    2002-01-01

    IFSAR images of a target scene are generated by compensating for variations in vertical separation between collection surfaces defined for each IFSAR antenna by adjusting the baseline projection during image generation. In addition, height information from all antennas is processed before processing range and azimuth information in a normal fashion to create the IFSAR image.

  13. Computer-socket manufacturing error: How much before it is clinically apparent?

    PubMed Central

    Sanders, Joan E.; Severance, Michael R.; Allyn, Kathryn J.

    2015-01-01

    The purpose of this research was to pursue quality standards for computer-manufacturing of prosthetic sockets for people with transtibial limb loss. Thirty-three duplicates of study participants’ normally used sockets were fabricated using central fabrication facilities. Socket-manufacturing errors were compared with clinical assessments of socket fit. Of the 33 sockets tested, 23 were deemed clinically to need modification. All 13 sockets with mean radial error (MRE) greater than 0.25 mm were clinically unacceptable, and 11 of those were deemed in need of sizing reduction. Of the remaining 20 sockets, 5 sockets with interquartile range (IQR) greater than 0.40 mm were deemed globally or regionally oversized and in need of modification. Of the remaining 15 sockets, 5 sockets with closed contours of elevated surface normal angle error (SNAE) were deemed clinically to need shape modification at those closed contour locations. The remaining 10 sockets were deemed clinically acceptable and not in need modification. MRE, IQR, and SNAE may serve as effective metrics to characterize quality of computer-manufactured prosthetic sockets, helping facilitate the development of quality standards for the socket manufacturing industry. PMID:22773260

  14. The MAGNEX large acceptance spectrometer

    SciTech Connect

    Cavallaro, M.; Cappuzzello, F.; Cunsolo, A.; Carbone, D.; Foti, A.

    2010-03-01

    The main features of the MAGNEX large acceptance magnetic spectrometer are described. It has a quadrupole + dipole layout and a hybrid detector located at the focal plane. The aberrations due to the large angular (50 msr) and momentum (+- 13%) acceptance are reduced by an accurate hardware design and then compensated by an innovative software ray-reconstruction technique. The obtained resolution in energy, angle and mass are presented in the paper. MAGNEX has been used up to now for different experiments in nuclear physics and astrophysics confirming to be a multipurpose device.

  15. Development of quantitative risk acceptance criteria

    SciTech Connect

    Griesmeyer, J. M.; Okrent, D.

    1981-01-01

    Some of the major considerations for effective management of risk are discussed, with particular emphasis on risks due to nuclear power plant operations. Although there are impacts associated with the rest of the fuel cycle, they are not addressed here. Several previously published proposals for quantitative risk criteria are reviewed. They range from a simple acceptance criterion on individual risk of death to a quantitative risk management framework. The final section discussed some of the problems in the establishment of a framework for the quantitative management of risk.

  16. Acceptance Criteria Framework for Autonomous Biological Detectors

    SciTech Connect

    Dzenitis, J M

    2006-12-12

    The purpose of this study was to examine a set of user acceptance criteria for autonomous biological detection systems for application in high-traffic, public facilities. The test case for the acceptance criteria was the Autonomous Pathogen Detection System (APDS) operating in high-traffic facilities in New York City (NYC). However, the acceptance criteria were designed to be generally applicable to other biological detection systems in other locations. For such detection systems, ''users'' will include local authorities (e.g., facility operators, public health officials, and law enforcement personnel) and national authorities [including personnel from the Department of Homeland Security (DHS), the BioWatch Program, the Centers for Disease Control and Prevention (CDC), and the Federal Bureau of Investigation (FBI)]. The panel members brought expertise from a broad range of backgrounds to complete this picture. The goals of this document are: (1) To serve as informal guidance for users in considering the benefits and costs of these systems. (2) To serve as informal guidance for developers in understanding the needs of users. In follow-up work, this framework will be used to systematically document the APDS for appropriateness and readiness for use in NYC.

  17. A joint acceptance function for enclosed spaces

    NASA Astrophysics Data System (ADS)

    Koopmann, G. H.; Pollard, H. F.

    1980-12-01

    A method is proposed for quantifying the geometric coupling between the acoustic modes of an enclosure and the vibratory motion of the enclosing surfaces. A dimensionless quantity, called the joint acceptance function, is defined which gives the coupling efficiency of an enclosing surface within a range from zero to unity. The joint acceptance function is based on an integral solution to the wave equation which requires a knowledge of the Green function for an enclosed space. Section 2 of the paper is devoted to methods of generating Green functions for enclosed spaces as series expansions in terms of orthogonal eigenfunctions. Computer generated Green functions are shown to compare favorably with those obtained from experiments performed on a hard walled, rectangular box. Section 3 of the paper describes the method of calculating the joint acceptance function for arbitrarily shaped enclosures. Two applications of this function are presented: (a) a rectangular enclosure with surfaces vibrating as simply supported plates, and (b) a practical case involving the reduction of noise in a tractor cabin.

  18. Simulation of large acceptance LINAC for muons

    SciTech Connect

    Miyadera, H; Kurennoy, S; Jason, A J

    2010-01-01

    There has been a recent need for muon accelerators not only for future Neutrino Factories and Muon Colliders but also for other applications in industry and medical use. We carried out simulations on a large-acceptance muon linac with a new concept 'mixed buncher/acceleration'. The linac can accept pions/muons from a production target with large acceptance and accelerate muon without any beam cooling which makes the initial section of muon-linac system very compact. The linac has a high impact on Neutrino Factory and Muon Collider (NF/MC) scenario since the 300-m injector section can be replaced by the muon linac of only 10-m length. The current design of the linac consists of the following components: independent 805-MHz cavity structure with 6- or 8-cm-radius aperture window; injection of a broad range of pion/muon energies, 10-100 MeV, and acceleration to 150 - 200 MeV. Further acceleration of the muon beam are relatively easy since the beam is already bunched.

  19. Random errors in egocentric networks.

    PubMed

    Almquist, Zack W

    2012-10-01

    The systematic errors that are induced by a combination of human memory limitations and common survey design and implementation have long been studied in the context of egocentric networks. Despite this, little if any work exists in the area of random error analysis on these same networks; this paper offers a perspective on the effects of random errors on egonet analysis, as well as the effects of using egonet measures as independent predictors in linear models. We explore the effects of false-positive and false-negative error in egocentric networks on both standard network measures and on linear models through simulation analysis on a ground truth egocentric network sample based on facebook-friendships. Results show that 5-20% error rates, which are consistent with error rates known to occur in ego network data, can cause serious misestimation of network properties and regression parameters. PMID:23878412

  20. Random errors in egocentric networks

    PubMed Central

    Almquist, Zack W.

    2013-01-01

    The systematic errors that are induced by a combination of human memory limitations and common survey design and implementation have long been studied in the context of egocentric networks. Despite this, little if any work exists in the area of random error analysis on these same networks; this paper offers a perspective on the effects of random errors on egonet analysis, as well as the effects of using egonet measures as independent predictors in linear models. We explore the effects of false-positive and false-negative error in egocentric networks on both standard network measures and on linear models through simulation analysis on a ground truth egocentric network sample based on facebook-friendships. Results show that 5–20% error rates, which are consistent with error rates known to occur in ego network data, can cause serious misestimation of network properties and regression parameters. PMID:23878412

  1. Dopamine reward prediction error coding

    PubMed Central

    Schultz, Wolfram

    2016-01-01

    Reward prediction errors consist of the differences between received and predicted rewards. They are crucial for basic forms of learning about rewards and make us strive for more rewards—an evolutionary beneficial trait. Most dopamine neurons in the midbrain of humans, monkeys, and rodents signal a reward prediction error; they are activated by more reward than predicted (positive prediction error), remain at baseline activity for fully predicted rewards, and show depressed activity with less reward than predicted (negative prediction error). The dopamine signal increases nonlinearly with reward value and codes formal economic utility. Drugs of addiction generate, hijack, and amplify the dopamine reward signal and induce exaggerated, uncontrolled dopamine effects on neuronal plasticity. The striatum, amygdala, and frontal cortex also show reward prediction error coding, but only in subpopulations of neurons. Thus, the important concept of reward prediction errors is implemented in neuronal hardware. PMID:27069377

  2. [Error factors in spirometry].

    PubMed

    Quadrelli, S A; Montiel, G C; Roncoroni, A J

    1994-01-01

    Spirometry is the more frequently used method to estimate pulmonary function in the clinical laboratory. It is important to comply with technical requisites to approximate the real values sought as well as adequate interpretation of results. Recommendations are made to establish: 1--quality control 2--define abnormality 3--classify the change from normal and its degree 4--define reversibility. In relation to quality control several criteria are pointed out such as end of the test, back-extrapolation and extrapolated volume in order to delineate most common errors. Daily calibration is advised. Inspection of graphical records of the test is mandatory. The limitations to the common use of 80% of predicted values to establish abnormality is stressed. The reasons for employing 95% confidence limits are detailed. It is important to select the reference values equation (in view of the differences in predicted values). It is advisable to validate the selection with local population normal values. In relation to the definition of the defect as restrictive or obstructive, the limitations of vital capacity (VC) to establish restriction, when obstruction is also present, are defined. Also the limitations of maximal mid-expiratory flow 25-75 (FMF 25-75) as an isolated marker of obstruction. Finally the qualities of forced expiratory volume in 1 sec (VEF1) and the difficulties with other indicators (CVF, FMF 25-75, VEF1/CVF) to estimate reversibility after bronchodilators are evaluated. The value of different methods used to define reversibility (% of change in initial value, absolute change or % of predicted), is commented. Clinical spirometric studies in order to be valuable should be performed with the same technical rigour as any other more complex studies. PMID:7990690

  3. Statistical errors in Monte Carlo estimates of systematic errors

    NASA Astrophysics Data System (ADS)

    Roe, Byron P.

    2007-01-01

    For estimating the effects of a number of systematic errors on a data sample, one can generate Monte Carlo (MC) runs with systematic parameters varied and examine the change in the desired observed result. Two methods are often used. In the unisim method, the systematic parameters are varied one at a time by one standard deviation, each parameter corresponding to a MC run. In the multisim method (see ), each MC run has all of the parameters varied; the amount of variation is chosen from the expected distribution of each systematic parameter, usually assumed to be a normal distribution. The variance of the overall systematic error determination is derived for each of the two methods and comparisons are made between them. If one focuses not on the error in the prediction of an individual systematic error, but on the overall error due to all systematic errors in the error matrix element in data bin m, the number of events needed is strongly reduced because of the averaging effect over all of the errors. For simple models presented here the multisim model was far better if the statistical error in the MC samples was larger than an individual systematic error, while for the reverse case, the unisim model was better. Exact formulas and formulas for the simple toy models are presented so that realistic calculations can be made. The calculations in the present note are valid if the errors are in a linear region. If that region extends sufficiently far, one can have the unisims or multisims correspond to k standard deviations instead of one. This reduces the number of events required by a factor of k2. The specific terms unisim and multisim were coined by Peter Meyers and Steve Brice, respectively, for the MiniBooNE experiment. However, the concepts have been developed over time and have been in general use for some time.

  4. Acceptance Probability (P a) Analysis for Process Validation Lifecycle Stages.

    PubMed

    Alsmeyer, Daniel; Pazhayattil, Ajay; Chen, Shu; Munaretto, Francesco; Hye, Maksuda; Sanghvi, Pradeep

    2016-04-01

    This paper introduces an innovative statistical approach towards understanding how variation impacts the acceptance criteria of quality attributes. Because of more complex stage-wise acceptance criteria, traditional process capability measures are inadequate for general application in the pharmaceutical industry. The probability of acceptance concept provides a clear measure, derived from specific acceptance criteria for each quality attribute. In line with the 2011 FDA Guidance, this approach systematically evaluates data and scientifically establishes evidence that a process is capable of consistently delivering quality product. The probability of acceptance provides a direct and readily understandable indication of product risk. As with traditional capability indices, the acceptance probability approach assumes that underlying data distributions are normal. The computational solutions for dosage uniformity and dissolution acceptance criteria are readily applicable. For dosage uniformity, the expected AV range may be determined using the s lo and s hi values along with the worst case estimates of the mean. This approach permits a risk-based assessment of future batch performance of the critical quality attributes. The concept is also readily applicable to sterile/non sterile liquid dose products. Quality attributes such as deliverable volume and assay per spray have stage-wise acceptance that can be converted into an acceptance probability. Accepted statistical guidelines indicate processes with C pk > 1.33 as performing well within statistical control and those with C pk < 1.0 as "incapable" (1). A C pk > 1.33 is associated with a centered process that will statistically produce less than 63 defective units per million. This is equivalent to an acceptance probability of >99.99%. PMID:26024723

  5. Teratogenic inborn errors of metabolism.

    PubMed Central

    Leonard, J. V.

    1986-01-01

    Most children with inborn errors of metabolism are born healthy without malformations as the fetus is protected by the metabolic activity of the placenta. However, certain inborn errors of the fetus have teratogenic effects although the mechanisms responsible for the malformations are not generally understood. Inborn errors in the mother may also be teratogenic. The adverse effects of these may be reduced by improved metabolic control of the biochemical disorder. PMID:3540927

  6. Confidence limits and their errors

    SciTech Connect

    Rajendran Raja

    2002-03-22

    Confidence limits are common place in physics analysis. Great care must be taken in their calculation and use especially in cases of limited statistics. We introduce the concept of statistical errors of confidence limits and argue that not only should limits be calculated but also their errors in order to represent the results of the analysis to the fullest. We show that comparison of two different limits from two different experiments becomes easier when their errors are also quoted. Use of errors of confidence limits will lead to abatement of the debate on which method is best suited to calculate confidence limits.

  7. Compensating For GPS Ephemeris Error

    NASA Technical Reports Server (NTRS)

    Wu, Jiun-Tsong

    1992-01-01

    Method of computing position of user station receiving signals from Global Positioning System (GPS) of navigational satellites compensates for most of GPS ephemeris error. Present method enables user station to reduce error in its computed position substantially. User station must have access to two or more reference stations at precisely known positions several hundred kilometers apart and must be in neighborhood of reference stations. Based on fact that when GPS data used to compute baseline between reference station and user station, vector error in computed baseline is proportional ephemeris error and length of baseline.

  8. Retransmission error control with memory

    NASA Technical Reports Server (NTRS)

    Sindhu, P. S.

    1977-01-01

    In this paper, an error control technique that is a basic improvement over automatic-repeat-request ARQ is presented. Erroneously received blocks in an ARQ system are used for error control. The technique is termed ARQ-with-memory (MRQ). The general MRQ system is described, and simple upper and lower bounds are derived on the throughput achievable by MRQ. The performance of MRQ with respect to throughput, message delay and probability of error is compared to that of ARQ by simulating both systems using error data from a VHF satellite channel being operated in the ALOHA packet broadcasting mode.

  9. Medication Errors in Outpatient Pediatrics.

    PubMed

    Berrier, Kyla

    2016-01-01

    Medication errors may occur during parental administration of prescription and over-the-counter medications in the outpatient pediatric setting. Misinterpretation of medication labels and dosing errors are two types of errors in medication administration. Health literacy may play an important role in parents' ability to safely manage their child's medication regimen. There are several proposed strategies for decreasing these medication administration errors, including using standardized dosing instruments, using strictly metric units for medication dosing, and providing parents and caregivers with picture-based dosing instructions. Pediatric healthcare providers should be aware of these strategies and seek to implement many of them into their practices. PMID:27537086

  10. Physical examination. Frequently observed errors.

    PubMed

    Wiener, S; Nathanson, M

    1976-08-16

    A method allowing for direct observation of intern and resident physicians while interviewing and examining patients has been in use on our medical wards for the last five years. A large number of errors in the performance of the medical examination by young physicians were noted and a classification of these errors into those of technique, omission, detection, interpretation, and recording was made. An approach to detection and correction of each of these kinds of errors is presented, as well as a discussion of possible reasons for the occurrence of these errors in physician performance. PMID:947266

  11. A theory of human error

    NASA Technical Reports Server (NTRS)

    Mcruer, D. T.; Clement, W. F.; Allen, R. W.

    1980-01-01

    Human error, a significant contributing factor in a very high proportion of civil transport, general aviation, and rotorcraft accidents is investigated. Correction of the sources of human error requires that one attempt to reconstruct underlying and contributing causes of error from the circumstantial causes cited in official investigative reports. A validated analytical theory of the input-output behavior of human operators involving manual control, communication, supervisory, and monitoring tasks which are relevant to aviation operations is presented. This theory of behavior, both appropriate and inappropriate, provides an insightful basis for investigating, classifying, and quantifying the needed cause-effect relationships governing propagation of human error.

  12. Nitrogen trailer acceptance test report

    SciTech Connect

    Kostelnik, A.J.

    1996-02-12

    This Acceptance Test Report documents compliance with the requirements of specification WHC-S-0249. The equipment was tested according to WHC-SD-WM-ATP-108 Rev.0. The equipment being tested is a portable contained nitrogen supply. The test was conducted at Norco`s facility.

  13. Helping Our Children Accept Themselves.

    ERIC Educational Resources Information Center

    Gamble, Mae

    1984-01-01

    Parents of a child with muscular dystrophy recount their reactions to learning of the diagnosis, their gradual acceptance, and their son's resistance, which was gradually lessened when he was provided with more information and treated more normally as a member of the family. (CL)

  14. Acceptability of Treatments for Plagiarism

    ERIC Educational Resources Information Center

    Carter, Stacy L.; Punyanunt-Carter, Narissra Maria

    2007-01-01

    This study focused on various treatments for addressing incidents of plagiarism by college students. College students rated the acceptability of different responses by college faculty to a case description of a college student who engaged in plagiarism. The findings revealed that students found some methods of addressing this problem behavior by…

  15. Euthanasia Acceptance: An Attitudinal Inquiry.

    ERIC Educational Resources Information Center

    Klopfer, Fredrick J.; Price, William F.

    The study presented was conducted to examine potential relationships between attitudes regarding the dying process, including acceptance of euthanasia, and other attitudinal or demographic attributes. The data of the survey was comprised of responses given by 331 respondents to a door-to-door interview. Results are discussed in terms of preferred…

  16. Critical evidence for the prediction error theory in associative learning

    PubMed Central

    Terao, Kanta; Matsumoto, Yukihisa; Mizunami, Makoto

    2015-01-01

    In associative learning in mammals, it is widely accepted that the discrepancy, or error, between actual and predicted reward determines whether learning occurs. Complete evidence for the prediction error theory, however, has not been obtained in any learning systems: Prediction error theory stems from the finding of a blocking phenomenon, but blocking can also be accounted for by other theories, such as the attentional theory. We demonstrated blocking in classical conditioning in crickets and obtained evidence to reject the attentional theory. To obtain further evidence supporting the prediction error theory and rejecting alternative theories, we constructed a neural model to match the prediction error theory, by modifying our previous model of learning in crickets, and we tested a prediction from the model: the model predicts that pharmacological intervention of octopaminergic transmission during appetitive conditioning impairs learning but not formation of reward prediction itself, and it thus predicts no learning in subsequent training. We observed such an “auto-blocking”, which could be accounted for by the prediction error theory but not by other competitive theories to account for blocking. This study unambiguously demonstrates validity of the prediction error theory in associative learning. PMID:25754125

  17. Critical evidence for the prediction error theory in associative learning.

    PubMed

    Terao, Kanta; Matsumoto, Yukihisa; Mizunami, Makoto

    2015-01-01

    In associative learning in mammals, it is widely accepted that the discrepancy, or error, between actual and predicted reward determines whether learning occurs. Complete evidence for the prediction error theory, however, has not been obtained in any learning systems: Prediction error theory stems from the finding of a blocking phenomenon, but blocking can also be accounted for by other theories, such as the attentional theory. We demonstrated blocking in classical conditioning in crickets and obtained evidence to reject the attentional theory. To obtain further evidence supporting the prediction error theory and rejecting alternative theories, we constructed a neural model to match the prediction error theory, by modifying our previous model of learning in crickets, and we tested a prediction from the model: the model predicts that pharmacological intervention of octopaminergic transmission during appetitive conditioning impairs learning but not formation of reward prediction itself, and it thus predicts no learning in subsequent training. We observed such an "auto-blocking", which could be accounted for by the prediction error theory but not by other competitive theories to account for blocking. This study unambiguously demonstrates validity of the prediction error theory in associative learning. PMID:25754125

  18. A posteriori error estimator and error control for contact problems

    NASA Astrophysics Data System (ADS)

    Weiss, Alexander; Wohlmuth, Barbara I.

    2009-09-01

    In this paper, we consider two error estimators for one-body contact problems. The first error estimator is defined in terms of H( div ) -conforming stress approximations and equilibrated fluxes while the second is a standard edge-based residual error estimator without any modification with respect to the contact. We show reliability and efficiency for both estimators. Moreover, the error is bounded by the first estimator with a constant one plus a higher order data oscillation term plus a term arising from the contact that is shown numerically to be of higher order. The second estimator is used in a control-based AFEM refinement strategy, and the decay of the error in the energy is shown. Several numerical tests demonstrate the performance of both estimators.

  19. Precise Orbit Determination for GEOSAT Follow-On Using Satellite Laser Ranging Data and Intermission Altimeter Crossovers

    NASA Technical Reports Server (NTRS)

    Lemoine, Frank G.; Rowlands, David D.; Luthcke, Scott B.; Zelensky, Nikita P.; Chinn, Douglas S.; Pavlis, Despina E.; Marr, Gregory

    2001-01-01

    The US Navy's GEOSAT Follow-On Spacecraft was launched on February 10, 1998 with the primary objective of the mission to map the oceans using a radar altimeter. Following an extensive set of calibration campaigns in 1999 and 2000, the US Navy formally accepted delivery of the satellite on November 29, 2000. Satellite laser ranging (SLR) and Doppler (Tranet-style) beacons track the spacecraft. Although limited amounts of GPS data were obtained, the primary mode of tracking remains satellite laser ranging. The GFO altimeter measurements are highly precise, with orbit error the largest component in the error budget. We have tuned the non-conservative force model for GFO and the gravity model using SLR, Doppler and altimeter crossover data sampled over one year. Gravity covariance projections to 70x70 show the radial orbit error on GEOSAT was reduced from 2.6 cm in EGM96 to 1.3 cm with the addition of SLR, GFO/GFO and TOPEX/GFO crossover data. Evaluation of the gravity fields using SLR and crossover data support the covariance projections and also show a dramatic reduction in geographically-correlated error for the tuned fields. In this paper, we report on progress in orbit determination for GFO using GFO/GFO and TOPEX/GFO altimeter crossovers. We will discuss improvements in satellite force modeling and orbit determination strategy, which allows reduction in GFO radial orbit error from 10-15 cm to better than 5 cm.

  20. Error coding simulations in C

    NASA Technical Reports Server (NTRS)

    Noble, Viveca K.

    1994-01-01

    When data is transmitted through a noisy channel, errors are produced within the data rendering it indecipherable. Through the use of error control coding techniques, the bit error rate can be reduced to any desired level without sacrificing the transmission data rate. The Astrionics Laboratory at Marshall Space Flight Center has decided to use a modular, end-to-end telemetry data simulator to simulate the transmission of data from flight to ground and various methods of error control. The simulator includes modules for random data generation, data compression, Consultative Committee for Space Data Systems (CCSDS) transfer frame formation, error correction/detection, error generation and error statistics. The simulator utilizes a concatenated coding scheme which includes CCSDS standard (255,223) Reed-Solomon (RS) code over GF(2(exp 8)) with interleave depth of 5 as the outermost code, (7, 1/2) convolutional code as an inner code and CCSDS recommended (n, n-16) cyclic redundancy check (CRC) code as the innermost code, where n is the number of information bits plus 16 parity bits. The received signal-to-noise for a desired bit error rate is greatly reduced through the use of forward error correction techniques. Even greater coding gain is provided through the use of a concatenated coding scheme. Interleaving/deinterleaving is necessary to randomize burst errors which may appear at the input of the RS decoder. The burst correction capability length is increased in proportion to the interleave depth. The modular nature of the simulator allows for inclusion or exclusion of modules as needed. This paper describes the development and operation of the simulator, the verification of a C-language Reed-Solomon code, and the possibility of using Comdisco SPW(tm) as a tool for determining optimal error control schemes.

  1. Error coding simulations in C

    NASA Astrophysics Data System (ADS)

    Noble, Viveca K.

    1994-10-01

    When data is transmitted through a noisy channel, errors are produced within the data rendering it indecipherable. Through the use of error control coding techniques, the bit error rate can be reduced to any desired level without sacrificing the transmission data rate. The Astrionics Laboratory at Marshall Space Flight Center has decided to use a modular, end-to-end telemetry data simulator to simulate the transmission of data from flight to ground and various methods of error control. The simulator includes modules for random data generation, data compression, Consultative Committee for Space Data Systems (CCSDS) transfer frame formation, error correction/detection, error generation and error statistics. The simulator utilizes a concatenated coding scheme which includes CCSDS standard (255,223) Reed-Solomon (RS) code over GF(2(exp 8)) with interleave depth of 5 as the outermost code, (7, 1/2) convolutional code as an inner code and CCSDS recommended (n, n-16) cyclic redundancy check (CRC) code as the innermost code, where n is the number of information bits plus 16 parity bits. The received signal-to-noise for a desired bit error rate is greatly reduced through the use of forward error correction techniques. Even greater coding gain is provided through the use of a concatenated coding scheme. Interleaving/deinterleaving is necessary to randomize burst errors which may appear at the input of the RS decoder. The burst correction capability length is increased in proportion to the interleave depth. The modular nature of the simulator allows for inclusion or exclusion of modules as needed. This paper describes the development and operation of the simulator, the verification of a C-language Reed-Solomon code, and the possibility of using Comdisco SPW(tm) as a tool for determining optimal error control schemes.

  2. Error Models of the Analog to Digital Converters

    NASA Astrophysics Data System (ADS)

    Michaeli, Linus; Šaliga, Ján

    2014-04-01

    Error models of the Analog to Digital Converters describe metrological properties of the signal conversion from analog to digital domain in a concise form using few dominant error parameters. Knowledge of the error models allows the end user to provide fast testing in the crucial points of the full input signal range and to use identified error models for post correction in the digital domain. The imperfections of the internal ADC structure determine the error characteristics represented by the nonlinearities as a function of the output code. Progress in the microelectronics and missing information about circuital details together with the lack of knowledge about interfering effects caused by ADC installation prefers another modeling approach based on the input-output behavioral characterization by the input-output error box. Internal links in the ADC structure cause that the input-output error function could be described in a concise form by suitable function. Modeled functional parameters allow determining the integral error parameters of ADC. Paper is a survey of error models starting from the structural models for the most common architectures and their linkage with the behavioral models represented by the simple look up table or the functional description of nonlinear errors for the output codes.

  3. Inductively Coupled Plasma Mass Spectrometry Uranium Error Propagation

    SciTech Connect

    Hickman, D P; Maclean, S; Shepley, D; Shaw, R K

    2001-07-01

    The Hazards Control Department at Lawrence Livermore National Laboratory (LLNL) uses Inductively Coupled Plasma Mass Spectrometer (ICP/MS) technology to analyze uranium in urine. The ICP/MS used by the Hazards Control Department is a Perkin-Elmer Elan 6000 ICP/MS. The Department of Energy Laboratory Accreditation Program requires that the total error be assessed for bioassay measurements. A previous evaluation of the errors associated with the ICP/MS measurement of uranium demonstrated a {+-} 9.6% error in the range of 0.01 to 0.02 {micro}g/l. However, the propagation of total error for concentrations above and below this level have heretofore been undetermined. This document is an evaluation of the errors associated with the current LLNL ICP/MS method for a more expanded range of uranium concentrations.

  4. Error Estimation for Reduced Order Models of Dynamical Systems

    SciTech Connect

    Homescu, C; Petzold, L; Serban, R

    2004-01-22

    The use of reduced order models to describe a dynamical system is pervasive in science and engineering. Often these models are used without an estimate of their error or range of validity. In this paper we consider dynamical systems and reduced models built using proper orthogonal decomposition. We show how to compute estimates and bounds for these errors, by a combination of small sample statistical condition estimation and error estimation using the adjoint method. Most importantly, the proposed approach allows the assessment of regions of validity for reduced models, i.e., ranges of perturbations in the original system over which the reduced model is still appropriate. Numerical examples validate our approach: the error norm estimates approximate well the forward error while the derived bounds are within an order of magnitude.

  5. Explaining Errors in Children's Questions

    ERIC Educational Resources Information Center

    Rowland, Caroline F.

    2007-01-01

    The ability to explain the occurrence of errors in children's speech is an essential component of successful theories of language acquisition. The present study tested some generativist and constructivist predictions about error on the questions produced by ten English-learning children between 2 and 5 years of age. The analyses demonstrated that,…

  6. Dyslexia and Oral Reading Errors

    ERIC Educational Resources Information Center

    Singleton, Chris

    2005-01-01

    Thomson was the first of very few researchers to have studied oral reading errors as a means of addressing the question: Are dyslexic readers different to other readers? Using the Neale Analysis of Reading Ability and Goodman's taxonomy of oral reading errors, Thomson concluded that dyslexic readers are different, but he found that they do not…

  7. Children's Scale Errors with Tools

    ERIC Educational Resources Information Center

    Casler, Krista; Eshleman, Angelica; Greene, Kimberly; Terziyan, Treysi

    2011-01-01

    Children sometimes make "scale errors," attempting to interact with tiny object replicas as though they were full size. Here, we demonstrate that instrumental tools provide special insight into the origins of scale errors and, moreover, into the broader nature of children's purpose-guided reasoning and behavior with objects. In Study 1, 1.5- to…

  8. Robustness and modeling error characterization

    NASA Technical Reports Server (NTRS)

    Lehtomaki, N. A.; Castanon, D. A.; Sandell, N. R., Jr.; Levy, B. C.; Athans, M.; Stein, G.

    1984-01-01

    The results on robustness theory presented here are extensions of those given in Lehtomaki et al., (1981). The basic innovation in these new results is that they utilize minimal additional information about the structure of the modeling error, as well as its magnitude, to assess the robustness of feedback systems for which robustness tests based on the magnitude of modeling error alone are inconclusive.

  9. Human Error: A Concept Analysis

    NASA Technical Reports Server (NTRS)

    Hansen, Frederick D.

    2007-01-01

    Human error is the subject of research in almost every industry and profession of our times. This term is part of our daily language and intuitively understood by most people however, it would be premature to assume that everyone's understanding of human error s the same. For example, human error is used to describe the outcome or consequence of human action, the causal factor of an accident, deliberate violations,a nd the actual action taken by a human being. As a result, researchers rarely agree on the either a specific definition or how to prevent human error. The purpose of this article is to explore the specific concept of human error using Concept Analysis as described by Walker and Avant (1995). The concept of human error is examined as currently used in the literature of a variety of industries and professions. Defining attributes and examples of model, borderline, and contrary cases are described. The antecedents and consequences of human error are also discussed and a definition of human error is offered.

  10. Dual Processing and Diagnostic Errors

    ERIC Educational Resources Information Center

    Norman, Geoff

    2009-01-01

    In this paper, I review evidence from two theories in psychology relevant to diagnosis and diagnostic errors. "Dual Process" theories of thinking, frequently mentioned with respect to diagnostic error, propose that categorization decisions can be made with either a fast, unconscious, contextual process called System 1 or a slow, analytical,…

  11. Measurement Errors in Organizational Surveys.

    ERIC Educational Resources Information Center

    Dutka, Solomon; Frankel, Lester R.

    1993-01-01

    Describes three classes of measurement techniques: (1) interviewing methods; (2) record retrieval procedures; and (3) observation methods. Discusses primary reasons for measurement error. Concludes that, although measurement error can be defined and controlled for, there are other design factors that also must be considered. (CFR)

  12. Barriers to Medical Error Reporting

    PubMed Central

    Poorolajal, Jalal; Rezaie, Shirin; Aghighi, Negar

    2015-01-01

    Background: This study was conducted to explore the prevalence of medical error underreporting and associated barriers. Methods: This cross-sectional study was performed from September to December 2012. Five hospitals, affiliated with Hamadan University of Medical Sciences, in Hamedan, Iran were investigated. A self-administered questionnaire was used for data collection. Participants consisted of physicians, nurses, midwives, residents, interns, and staffs of radiology and laboratory departments. Results: Overall, 50.26% of subjects had committed but not reported medical errors. The main reasons mentioned for underreporting were lack of effective medical error reporting system (60.0%), lack of proper reporting form (51.8%), lack of peer supporting a person who has committed an error (56.0%), and lack of personal attention to the importance of medical errors (62.9%). The rate of committing medical errors was higher in men (71.4%), age of 50–40 years (67.6%), less-experienced personnel (58.7%), educational level of MSc (87.5%), and staff of radiology department (88.9%). Conclusions: This study outlined the main barriers to reporting medical errors and associated factors that may be helpful for healthcare organizations in improving medical error reporting as an essential component for patient safety enhancement. PMID:26605018

  13. Reducing latent errors, drift errors, and stakeholder dissonance.

    PubMed

    Samaras, George M

    2012-01-01

    Healthcare information technology (HIT) is being offered as a transformer of modern healthcare delivery systems. Some believe that it has the potential to improve patient safety, increase the effectiveness of healthcare delivery, and generate significant cost savings. In other industrial sectors, information technology has dramatically influenced quality and profitability - sometimes for the better and sometimes not. Quality improvement efforts in healthcare delivery have not yet produced the dramatic results obtained in other industrial sectors. This may be that previously successful quality improvement experts do not possess the requisite domain knowledge (clinical experience and expertise). It also appears related to a continuing misconception regarding the origins and meaning of work errors in healthcare delivery. The focus here is on system use errors rather than individual user errors. System use errors originate in both the development and the deployment of technology. Not recognizing stakeholders and their conflicting needs, wants, and desires (NWDs) may lead to stakeholder dissonance. Mistakes translating stakeholder NWDs into development or deployment requirements may lead to latent errors. Mistakes translating requirements into specifications may lead to drift errors. At the sharp end, workers encounter system use errors or, recognizing the risk, expend extensive and unanticipated resources to avoid them. PMID:22317001

  14. A theory of human error

    NASA Technical Reports Server (NTRS)

    Mcruer, D. T.; Clement, W. F.; Allen, R. W.

    1981-01-01

    Human errors tend to be treated in terms of clinical and anecdotal descriptions, from which remedial measures are difficult to derive. Correction of the sources of human error requires an attempt to reconstruct underlying and contributing causes of error from the circumstantial causes cited in official investigative reports. A comprehensive analytical theory of the cause-effect relationships governing propagation of human error is indispensable to a reconstruction of the underlying and contributing causes. A validated analytical theory of the input-output behavior of human operators involving manual control, communication, supervisory, and monitoring tasks which are relevant to aviation, maritime, automotive, and process control operations is highlighted. This theory of behavior, both appropriate and inappropriate, provides an insightful basis for investigating, classifying, and quantifying the needed cause-effect relationships governing propagation of human error.

  15. Onorbit IMU alignment error budget

    NASA Technical Reports Server (NTRS)

    Corson, R. W.

    1980-01-01

    The Star Tracker, Crew Optical Alignment Sight (COAS), and Inertial Measurement Unit (IMU) from a complex navigation system with a multitude of error sources were combined. A complete list of the system errors is presented. The errors were combined in a rational way to yield an estimate of the IMU alignment accuracy for STS-1. The expected standard deviation in the IMU alignment error for STS-1 type alignments was determined to be 72 arc seconds per axis for star tracker alignments and 188 arc seconds per axis for COAS alignments. These estimates are based on current knowledge of the star tracker, COAS, IMU, and navigation base error specifications, and were partially verified by preliminary Monte Carlo analysis.

  16. Error decomposition and estimation of inherent optical properties.

    PubMed

    Salama, Mhd Suhyb; Stein, Alfred

    2009-09-10

    We describe a methodology to quantify and separate the errors of inherent optical properties (IOPs) derived from ocean-color model inversion. Their total error is decomposed into three different sources, namely, model approximations and inversion, sensor noise, and atmospheric correction. Prior information on plausible ranges of observation, sensor noise, and inversion goodness-of-fit are employed to derive the posterior probability distribution of the IOPs. The relative contribution of each error component to the total error budget of the IOPs, all being of stochastic nature, is then quantified. The method is validated with the International Ocean Colour Coordinating Group (IOCCG) data set and the NASA bio-Optical Marine Algorithm Data set (NOMAD). The derived errors are close to the known values with correlation coefficients of 60-90% and 67-90% for IOCCG and NOMAD data sets, respectively. Model-induced errors inherent to the derived IOPs are between 10% and 57% of the total error, whereas atmospheric-induced errors are in general above 43% and up to 90% for both data sets. The proposed method is applied to synthesized and in situ measured populations of IOPs. The mean relative errors of the derived values are between 2% and 20%. A specific error table to the Medium Resolution Imaging Spectrometer (MERIS) sensor is constructed. It serves as a benchmark to evaluate the performance of the atmospheric correction method and to compute atmospheric-induced errors. Our method has a better performance and is more appropriate to estimate actual errors of ocean-color derived products than the previously suggested methods. Moreover, it is generic and can be applied to quantify the error of any derived biogeophysical parameter regardless of the used derivation. PMID:19745859

  17. Estimating errors in least-squares fitting

    NASA Technical Reports Server (NTRS)

    Richter, P. H.

    1995-01-01

    While least-squares fitting procedures are commonly used in data analysis and are extensively discussed in the literature devoted to this subject, the proper assessment of errors resulting from such fits has received relatively little attention. The present work considers statistical errors in the fitted parameters, as well as in the values of the fitted function itself, resulting from random errors in the data. Expressions are derived for the standard error of the fit, as a function of the independent variable, for the general nonlinear and linear fitting problems. Additionally, closed-form expressions are derived for some examples commonly encountered in the scientific and engineering fields, namely ordinary polynomial and Gaussian fitting functions. These results have direct application to the assessment of the antenna gain and system temperature characteristics, in addition to a broad range of problems in data analysis. The effects of the nature of the data and the choice of fitting function on the ability to accurately model the system under study are discussed, and some general rules are deduced to assist workers intent on maximizing the amount of information obtained form a given set of measurements.

  18. Accuracy analysis of optical ranging in atmosphere

    NASA Astrophysics Data System (ADS)

    Yuan, Hong-wu; Huang, Yin-bo; Mei, Hai-ping; Rao, Rui-zhong

    2009-07-01

    Optical ranging is one of the most precise techniques for distance measurement. The effects of the density variation of atmosphere, aerosols and clouds on optical ranging precision are generally considered, a new method is proposed for calculating the ranging precision in the presence of aerosol particles and clouds. The size distribution spectrum models for aerosols and clouds in the Optical Properties of Aerosols and Clouds Package (OPAC) are adopted. Results show that aerosols and clouds could introduce errors of several centimeters to several ten meters to the ranging. The relationship between the ranging precision and the relative humidity, the zenith angle of ranging direction and the optical wavelength is also analyzed. The ranging error doesn't have an obvious relationship with the wavelength, but depends on the zenith angle, especially for the angle larger than 70 degree. The ranging error depends on the relative humidity as well. The ranging error induced by aerosols increases gradually with the increase of the relative humidity when the relative humidity is less than 80%, but it increases rapidly when the relative humidity is larger than 80%. Our results could provide a theoretical basis and reference for the application of optical ranging.

  19. Error diffusion with a more symmetric error distribution

    NASA Astrophysics Data System (ADS)

    Fan, Zhigang

    1994-05-01

    In this paper a new error diffusion algorithm is presented that effectively eliminates the `worm' artifacts appearing in the standard methods. The new algorithm processes each scanline of the image in two passes, a forward pass followed by a backward one. This enables the error made at one pixel to be propagated to all the `future' pixels. A much more symmetric error distribution is achieved than that of the standard methods. The frequency response of the noise shaping filter associated with the new algorithm is mirror-symmetric in magnitude.

  20. Acceptability of reactors in space

    SciTech Connect

    Buden, D.

    1981-04-01

    Reactors are the key to our future expansion into space. However, there has been some confusion in the public as to whether they are a safe and acceptable technology for use in space. The answer to these questions is explored. The US position is that when reactors are the preferred technical choice, that they can be used safely. In fact, it dies not appear that reactors add measurably to the risk associated with the Space Transportation System.

  1. Acceptability of reactors in space

    SciTech Connect

    Buden, D.

    1981-01-01

    Reactors are the key to our future expansion into space. However, there has been some confusion in the public as to whether they are a safe and acceptable technology for use in space. The answer to these questions is explored. The US position is that when reactors are the preferred technical choice, that they can be used safely. In fact, it does not appear that reactors add measurably to the risk associated with the Space Transportation System.

  2. Reactor tank UT acceptance criteria

    SciTech Connect

    Daugherty, W.L.

    1990-01-30

    The SRS reactor tanks are constructed of type 304 stainless steel, with 0.5 inch thick walls. An ultrasonic (UT) in-service inspection program has been developed for examination of these tanks, in accordance with the ISI Plan for the Savannah River Production Reactors Process Water System (DPSTM-88-100-1). Prior to initiation of these inspections, criteria for the disposition of any indications that might be found are required. A working group has been formed to review available information on the SRS reactor tanks and develop acceptance criteria. This working group includes nationally recognized experts in the nuclear industry. The working group has met three times and produced three documents describing the proposed acceptance criteria, the technical basis for the criteria and a proposed initial sampling plan. This report transmits these three documents, which were prepared in accordance with the technical task plan and quality assurance plan for this task, task 88-001-A- 1. In addition, this report summarizes the acceptance criteria and proposed sampling plan, and provides further interpretation of the intent of these three documents where necessary.

  3. 12 CFR 7.1007 - Acceptances.

    Code of Federal Regulations, 2010 CFR

    2010-01-01

    ... 12 Banks and Banking 1 2010-01-01 2010-01-01 false Acceptances. 7.1007 Section 7.1007 Banks and... § 7.1007 Acceptances. A national bank is not limited in the character of acceptances it may make in financing credit transactions. Bankers' acceptances may be used for such purpose, since the making...

  4. 12 CFR 7.1007 - Acceptances.

    Code of Federal Regulations, 2011 CFR

    2011-01-01

    ... 12 Banks and Banking 1 2011-01-01 2011-01-01 false Acceptances. 7.1007 Section 7.1007 Banks and... § 7.1007 Acceptances. A national bank is not limited in the character of acceptances it may make in financing credit transactions. Bankers' acceptances may be used for such purpose, since the making...

  5. 21 CFR 820.86 - Acceptance status.

    Code of Federal Regulations, 2010 CFR

    2010-04-01

    ... 21 Food and Drugs 8 2010-04-01 2010-04-01 false Acceptance status. 820.86 Section 820.86 Food and... QUALITY SYSTEM REGULATION Acceptance Activities § 820.86 Acceptance status. Each manufacturer shall identify by suitable means the acceptance status of product, to indicate the conformance or...

  6. 12 CFR 615.5550 - Bankers' acceptances.

    Code of Federal Regulations, 2011 CFR

    2011-01-01

    ... 12 Banks and Banking 6 2011-01-01 2011-01-01 false Bankers' acceptances. 615.5550 Section 615.5550... POLICIES AND OPERATIONS, AND FUNDING OPERATIONS Bankers' Acceptances § 615.5550 Bankers' acceptances. Banks for cooperatives may rediscount with other purchasers the acceptances they have created. The bank...

  7. 48 CFR 245.606-3 - Acceptance.

    Code of Federal Regulations, 2010 CFR

    2010-10-01

    ... 48 Federal Acquisition Regulations System 3 2010-10-01 2010-10-01 false Acceptance. 245.606-3... Contractor Inventory 245.606-3 Acceptance. (a) If the schedules are acceptable, the plant clearance officer shall, within 15 days, complete and send the contractor a DD Form 1637, Notice of Acceptance...

  8. 48 CFR 12.402 - Acceptance.

    Code of Federal Regulations, 2011 CFR

    2011-10-01

    ... 48 Federal Acquisition Regulations System 1 2011-10-01 2011-10-01 false Acceptance. 12.402 Section... Acceptance. (a) The acceptance paragraph in 52.212-4 is based upon the assumption that the Government will rely on the contractor's assurances that the commercial item tendered for acceptance conforms to...

  9. 21 CFR 820.86 - Acceptance status.

    Code of Federal Regulations, 2011 CFR

    2011-04-01

    ... 21 Food and Drugs 8 2011-04-01 2011-04-01 false Acceptance status. 820.86 Section 820.86 Food and... QUALITY SYSTEM REGULATION Acceptance Activities § 820.86 Acceptance status. Each manufacturer shall identify by suitable means the acceptance status of product, to indicate the conformance or...

  10. 12 CFR 615.5550 - Bankers' acceptances.

    Code of Federal Regulations, 2010 CFR

    2010-01-01

    ... 12 Banks and Banking 6 2010-01-01 2010-01-01 false Bankers' acceptances. 615.5550 Section 615.5550... POLICIES AND OPERATIONS, AND FUNDING OPERATIONS Bankers' Acceptances § 615.5550 Bankers' acceptances. Banks for cooperatives may rediscount with other purchasers the acceptances they have created. The bank...

  11. 48 CFR 12.402 - Acceptance.

    Code of Federal Regulations, 2010 CFR

    2010-10-01

    ... 48 Federal Acquisition Regulations System 1 2010-10-01 2010-10-01 false Acceptance. 12.402 Section... Acceptance. (a) The acceptance paragraph in 52.212-4 is based upon the assumption that the Government will rely on the contractor's assurances that the commercial item tendered for acceptance conforms to...

  12. Hubble space telescope: Pointing error effects on nonlinear ball joints

    NASA Technical Reports Server (NTRS)

    Farmer, J. E.; Grissett, F. R.

    1985-01-01

    The Hubble Space Telescope pointing error produced by optical benches mounted on free ball joints is examined. Spacecraft cable connections are assumed to produce translational and rotational damping and restoring forces which act through the optical bench center of mass. The nonlinear dynamics are modeled and then implemented using an existing computer program for simulating the vehicle dynamics and pointing control system algorithm. Results are presented for the test case which indicate acceptable performance.

  13. Designing to Control Flight Crew Errors

    NASA Technical Reports Server (NTRS)

    Schutte, Paul C.; Willshire, Kelli F.

    1997-01-01

    It is widely accepted that human error is a major contributing factor in aircraft accidents. There has been a significant amount of research in why these errors occurred, and many reports state that the design of flight deck can actually dispose humans to err. This research has led to the call for changes in design according to human factors and human-centered principles. The National Aeronautics and Space Administration's (NASA) Langley Research Center has initiated an effort to design a human-centered flight deck from a clean slate (i.e., without constraints of existing designs.) The effort will be based on recent research in human-centered design philosophy and mission management categories. This design will match the human's model of the mission and function of the aircraft to reduce unnatural or non-intuitive interfaces. The product of this effort will be a flight deck design description, including training and procedures, and a cross reference or paper trail back to design hypotheses, and an evaluation of the design. The present paper will discuss the philosophy, process, and status of this design effort.

  14. The Second Law of Thermodynamics: Mathematical Error

    NASA Astrophysics Data System (ADS)

    Kalanov, Temur Z.

    2006-04-01

    The critical analysis of the generally accepted foundations of thermodynamics is proposed. Within the framework of the work [1], the following statement is proved: Gibbs's quantum canonical distribution fn =f0 exp (-{En } / {{En } {T)}}} {T)} (whereEn , n=0,;1,;... , fn , T are the energy of the subsystem, probability, and temperature, respectively) defines the correct relation of the thermal energy Q of the subsystem to the entropy S of the subsystem and the temperature T. This relation has the form: S=Q / {Q T}} } T and {lim }limits {T-> 0 } S=0 (where Q≡ ∑ limits n=0∞ {En } fn , S≡ ∑ limits n=0∞ {Sn fn } , Sn ≡ {En } / {{En } {T=-ln ({fn } / {{fn } {f0 )}}} {f0 )}}}} {T=-ln ({fn } {{fn } {f0 )}}} {f0 )}}). Consequence: the second law (i.e. dS={dQ} / {{dQ} T}}. (T) of thermodynamics represents mathematical error. Ref.: [1] T.Z. Kalanov, ``On the main errors underlying statistical physics.'' Bulletin of the APS, Vol. 47, No. 2 (2005), p. 164.

  15. Quantum rms error and Heisenberg's error-disturbance relation

    NASA Astrophysics Data System (ADS)

    Busch, Paul

    2014-09-01

    Reports on experiments recently performed in Vienna [Erhard et al, Nature Phys. 8, 185 (2012)] and Toronto [Rozema et al, Phys. Rev. Lett. 109, 100404 (2012)] include claims of a violation of Heisenberg's error-disturbance relation. In contrast, a Heisenberg-type tradeoff relation for joint measurements of position and momentum has been formulated and proven in [Phys. Rev. Lett. 111, 160405 (2013)]. Here I show how the apparent conflict is resolved by a careful consideration of the quantum generalization of the notion of root-mean-square error. The claim of a violation of Heisenberg's principle is untenable as it is based on a historically wrong attribution of an incorrect relation to Heisenberg, which is in fact trivially violated. We review a new general trade-off relation for the necessary errors in approximate joint measurements of incompatible qubit observables that is in the spirit of Heisenberg's intuitions. The experiments mentioned may directly be used to test this new error inequality.

  16. Stochastic Models of Human Errors

    NASA Technical Reports Server (NTRS)

    Elshamy, Maged; Elliott, Dawn M. (Technical Monitor)

    2002-01-01

    Humans play an important role in the overall reliability of engineering systems. More often accidents and systems failure are traced to human errors. Therefore, in order to have meaningful system risk analysis, the reliability of the human element must be taken into consideration. Describing the human error process by mathematical models is a key to analyzing contributing factors. Therefore, the objective of this research effort is to establish stochastic models substantiated by sound theoretic foundation to address the occurrence of human errors in the processing of the space shuttle.

  17. Noncompliance pattern due to medication errors at a Teaching Hospital in Srikot, India

    PubMed Central

    Thakur, Heenopama; Thawani, Vijay; Raina, Rangeel Singh; Kothiyal, Gitanjali; Chakarabarty, Mrinmoy

    2013-01-01

    Objective: To study the medication errors leading to noncompliance in a tertiary care teaching hospital. Materials and Methods: This study was conducted in a tertiary care hospital of a teaching institution from Srikot, Garhwal, Uttarakhand to analyze the medication errors in 500 indoor prescriptions from medicine, surgery, obstetrics and gynecology, pediatrics and ENT departments over five months and 100 outdoor patients of medicine department. Results: Medication error rate for indoor patients was found to be 22.4 % and 11.4% for outdoor patients as against the standard acceptable error rate 3%. Maximum errors were observed in the indoor prescriptions of the surgery department accounting for 44 errors followed by medicine 32 and gynecology 25 in the 500 cases studied leading to faulty administration of medicines. Conclusion: Many medication errors were noted which go against the practice of rational therapeutics. Such studies can be directed to usher in the rational use of medicines for increasing compliance and therapeutic benefits. PMID:23833376

  18. Error analysis of large aperture static interference imaging spectrometer

    NASA Astrophysics Data System (ADS)

    Li, Fan; Zhang, Guo

    2015-12-01

    Large Aperture Static Interference Imaging Spectrometer is a new type of spectrometer with light structure, high spectral linearity, high luminous flux and wide spectral range, etc ,which overcomes the contradiction between high flux and high stability so that enables important values in science studies and applications. However, there're different error laws in imaging process of LASIS due to its different imaging style from traditional imaging spectrometers, correspondingly, its data processing is complicated. In order to improve accuracy of spectrum detection and serve for quantitative analysis and monitoring of topographical surface feature, the error law of LASIS imaging is supposed to be learned. In this paper, the LASIS errors are classified as interferogram error, radiometric correction error and spectral inversion error, and each type of error is analyzed and studied. Finally, a case study of Yaogan-14 is proposed, in which the interferogram error of LASIS by time and space combined modulation is mainly experimented and analyzed, as well as the errors from process of radiometric correction and spectral inversion.

  19. Sensitivity of planetary cruise navigation to earth orientation calibration errors

    NASA Technical Reports Server (NTRS)

    Estefan, J. A.; Folkner, W. M.

    1995-01-01

    A detailed analysis was conducted to determine the sensitivity of spacecraft navigation errors to the accuracy and timeliness of Earth orientation calibrations. Analyses based on simulated X-band (8.4-GHz) Doppler and ranging measurements acquired during the interplanetary cruise segment of the Mars Pathfinder heliocentric trajectory were completed for the nominal trajectory design and for an alternative trajectory with a longer transit time. Several error models were developed to characterize the effect of Earth orientation on navigational accuracy based on current and anticipated Deep Space Network calibration strategies. The navigational sensitivity of Mars Pathfinder to calibration errors in Earth orientation was computed for each candidate calibration strategy with the Earth orientation parameters included as estimated parameters in the navigation solution. In these cases, the calibration errors contributed 23 to 58% of the total navigation error budget, depending on the calibration strategy being assessed. Navigation sensitivity calculations were also performed for cases in which Earth orientation calibration errors were not adjusted in the navigation solution. In these cases, Earth orientation calibration errors contributed from 26 to as much as 227% of the total navigation error budget. The final analysis suggests that, not only is the method used to calibrate Earth orientation vitally important for precision navigation of Mars Pathfinder, but perhaps equally important is the method for inclusion of the calibration errors in the navigation solutions.

  20. Neural markers of errors as endophenotypes in neuropsychiatric disorders

    PubMed Central

    Manoach, Dara S.; Agam, Yigal

    2013-01-01

    Learning from errors is fundamental to adaptive human behavior. It requires detecting errors, evaluating what went wrong, and adjusting behavior accordingly. These dynamic adjustments are at the heart of behavioral flexibility and accumulating evidence suggests that deficient error processing contributes to maladaptively rigid and repetitive behavior in a range of neuropsychiatric disorders. Neuroimaging and electrophysiological studies reveal highly reliable neural markers of error processing. In this review, we evaluate the evidence that abnormalities in these neural markers can serve as sensitive endophenotypes of neuropsychiatric disorders. We describe the behavioral and neural hallmarks of error processing, their mediation by common genetic polymorphisms, and impairments in schizophrenia, obsessive-compulsive disorder, and autism spectrum disorders. We conclude that neural markers of errors meet several important criteria as endophenotypes including heritability, established neuroanatomical and neurochemical substrates, association with neuropsychiatric disorders, presence in syndromally-unaffected family members, and evidence of genetic mediation. Understanding the mechanisms of error processing deficits in neuropsychiatric disorders may provide novel neural and behavioral targets for treatment and sensitive surrogate markers of treatment response. Treating error processing deficits may improve functional outcome since error signals provide crucial information for flexible adaptation to changing environments. Given the dearth of effective interventions for cognitive deficits in neuropsychiatric disorders, this represents a potentially promising approach. PMID:23882201

  1. Altimeter error sources at the 10-cm performance level

    NASA Technical Reports Server (NTRS)

    Martin, C. F.

    1977-01-01

    Error sources affecting the calibration and operational use of a 10 cm altimeter are examined to determine the magnitudes of current errors and the investigations necessary to reduce them to acceptable bounds. Errors considered include those affecting operational data pre-processing, and those affecting altitude bias determination, with error budgets developed for both. The most significant error sources affecting pre-processing are bias calibration, propagation corrections for the ionosphere, and measurement noise. No ionospheric models are currently validated at the required 10-25% accuracy level. The optimum smoothing to reduce the effects of measurement noise is investigated and found to be on the order of one second, based on the TASC model of geoid undulations. The 10 cm calibrations are found to be feasible only through the use of altimeter passes that are very high elevation for a tracking station which tracks very close to the time of altimeter track, such as a high elevation pass across the island of Bermuda. By far the largest error source, based on the current state-of-the-art, is the location of the island tracking station relative to mean sea level in the surrounding ocean areas.

  2. Algorithmic Error Correction of Impedance Measuring Sensors

    PubMed Central

    Starostenko, Oleg; Alarcon-Aquino, Vicente; Hernandez, Wilmar; Sergiyenko, Oleg; Tyrsa, Vira

    2009-01-01

    This paper describes novel design concepts and some advanced techniques proposed for increasing the accuracy of low cost impedance measuring devices without reduction of operational speed. The proposed structural method for algorithmic error correction and iterating correction method provide linearization of transfer functions of the measuring sensor and signal conditioning converter, which contribute the principal additive and relative measurement errors. Some measuring systems have been implemented in order to estimate in practice the performance of the proposed methods. Particularly, a measuring system for analysis of C-V, G-V characteristics has been designed and constructed. It has been tested during technological process control of charge-coupled device CCD manufacturing. The obtained results are discussed in order to define a reasonable range of applied methods, their utility, and performance. PMID:22303177

  3. Nevada Livestock Grazing and Range Management

    Technology Transfer Automated Retrieval System (TEKTRAN)

    The range livestock industry had a very slow start in Nevada because it was commonly accepted that the environment would not support livestock production. Freighters discovered that oxen could winter on the dry herbage of desert bunchgrasses and come off the range in excellent condition in the spri...

  4. Total error vs. measurement uncertainty: revolution or evolution?

    PubMed

    Oosterhuis, Wytze P; Theodorsson, Elvar

    2016-02-01

    The first strategic EFLM conference "Defining analytical performance goals, 15 years after the Stockholm Conference" was held in the autumn of 2014 in Milan. It maintained the Stockholm 1999 hierarchy of performance goals but rearranged them and established five task and finish groups to work on topics related to analytical performance goals including one on the "total error" theory. Jim Westgard recently wrote a comprehensive overview of performance goals and of the total error theory critical of the results and intentions of the Milan 2014 conference. The "total error" theory originated by Jim Westgard and co-workers has a dominating influence on the theory and practice of clinical chemistry but is not accepted in other fields of metrology. The generally accepted uncertainty theory, however, suffers from complex mathematics and conceived impracticability in clinical chemistry. The pros and cons of the total error theory need to be debated, making way for methods that can incorporate all relevant causes of uncertainty when making medical diagnoses and monitoring treatment effects. This development should preferably proceed not as a revolution but as an evolution. PMID:26540227

  5. Robust characterization of leakage errors

    NASA Astrophysics Data System (ADS)

    Wallman, Joel J.; Barnhill, Marie; Emerson, Joseph

    2016-04-01

    Leakage errors arise when the quantum state leaks out of some subspace of interest, for example, the two-level subspace of a multi-level system defining a computational ‘qubit’, the logical code space of a quantum error-correcting code, or a decoherence-free subspace. Leakage errors pose a distinct challenge to quantum control relative to the more well-studied decoherence errors and can be a limiting factor to achieving fault-tolerant quantum computation. Here we present a scalable and robust randomized benchmarking protocol for quickly estimating the leakage rate due to an arbitrary Markovian noise process on a larger system. We illustrate the reliability of the protocol through numerical simulations.

  6. Static Detection of Disassembly Errors

    SciTech Connect

    Krishnamoorthy, Nithya; Debray, Saumya; Fligg, Alan K

    2009-10-13

    Static disassembly is a crucial first step in reverse engineering executable files, and there is a consider- able body of work in reverse-engineering of binaries, as well as areas such as semantics-based security anal- ysis, that assumes that the input executable has been correctly disassembled. However, disassembly errors, e.g., arising from binary obfuscations, can render this assumption invalid. This work describes a machine- learning-based approach, using decision trees, for stat- ically identifying possible errors in a static disassem- bly; such potential errors may then be examined more closely, e.g., using dynamic analyses. Experimental re- sults using a variety of input executables indicate that our approach performs well, correctly identifying most disassembly errors with relatively few false positives.

  7. Orbital and Geodetic Error Analysis

    NASA Technical Reports Server (NTRS)

    Felsentreger, T.; Maresca, P.; Estes, R.

    1985-01-01

    Results that previously required several runs determined in more computer-efficient manner. Multiple runs performed only once with GEODYN and stored on tape. ERODYN then performs matrix partitioning and linear algebra required for each individual error-analysis run.

  8. Prospective errors determine motor learning.

    PubMed

    Takiyama, Ken; Hirashima, Masaya; Nozaki, Daichi

    2015-01-01

    Diverse features of motor learning have been reported by numerous studies, but no single theoretical framework concurrently accounts for these features. Here, we propose a model for motor learning to explain these features in a unified way by extending a motor primitive framework. The model assumes that the recruitment pattern of motor primitives is determined by the predicted movement error of an upcoming movement (prospective error). To validate this idea, we perform a behavioural experiment to examine the model's novel prediction: after experiencing an environment in which the movement error is more easily predictable, subsequent motor learning should become faster. The experimental results support our prediction, suggesting that the prospective error might be encoded in the motor primitives. Furthermore, we demonstrate that this model has a strong explanatory power to reproduce a wide variety of motor-learning-related phenomena that have been separately explained by different computational models. PMID:25635628

  9. Human errors and measurement uncertainty

    NASA Astrophysics Data System (ADS)

    Kuselman, Ilya; Pennecchi, Francesca

    2015-04-01

    Evaluating the residual risk of human errors in a measurement and testing laboratory, remaining after the error reduction by the laboratory quality system, and quantifying the consequences of this risk for the quality of the measurement/test results are discussed based on expert judgments and Monte Carlo simulations. A procedure for evaluation of the contribution of the residual risk to the measurement uncertainty budget is proposed. Examples are provided using earlier published sets of expert judgments on human errors in pH measurement of groundwater, elemental analysis of geological samples by inductively coupled plasma mass spectrometry, and multi-residue analysis of pesticides in fruits and vegetables. The human error contribution to the measurement uncertainty budget in the examples was not negligible, yet also not dominant. This was assessed as a good risk management result.

  10. Quantum error correction beyond qubits

    NASA Astrophysics Data System (ADS)

    Aoki, Takao; Takahashi, Go; Kajiya, Tadashi; Yoshikawa, Jun-Ichi; Braunstein, Samuel L.; van Loock, Peter; Furusawa, Akira

    2009-08-01

    Quantum computation and communication rely on the ability to manipulate quantum states robustly and with high fidelity. To protect fragile quantum-superposition states from corruption through so-called decoherence noise, some form of error correction is needed. Therefore, the discovery of quantum error correction (QEC) was a key step to turn the field of quantum information from an academic curiosity into a developing technology. Here, we present an experimental implementation of a QEC code for quantum information encoded in continuous variables, based on entanglement among nine optical beams. This nine-wave-packet adaptation of Shor's original nine-qubit scheme enables, at least in principle, full quantum error correction against an arbitrary single-beam error.

  11. Controlling type-1 error rates in whole effluent toxicity testing

    SciTech Connect

    Smith, R.; Johnson, S.C.

    1995-12-31

    A form of variability, called the dose x test interaction, has been found to affect the variability of the mean differences from control in the statistical tests used to evaluate Whole Effluent Toxicity Tests for compliance purposes. Since the dose x test interaction is not included in these statistical tests, the assumed type-1 and type-2 error rates can be incorrect. The accepted type-1 error rate for these tests is 5%. Analysis of over 100 Ceriodaphnia, fathead minnow and sea urchin fertilization tests showed that when the test x dose interaction term was not included in the calculations the type-1 error rate was inflated to as high as 20%. In a compliance setting, this problem may lead to incorrect regulatory decisions. Statistical tests are proposed that properly incorporate the dose x test interaction variance.

  12. Multicenter Assessment of Gram Stain Error Rates.

    PubMed

    Samuel, Linoj P; Balada-Llasat, Joan-Miquel; Harrington, Amanda; Cavagnolo, Robert

    2016-06-01

    Gram stains remain the cornerstone of diagnostic testing in the microbiology laboratory for the guidance of empirical treatment prior to availability of culture results. Incorrectly interpreted Gram stains may adversely impact patient care, and yet there are no comprehensive studies that have evaluated the reliability of the technique and there are no established standards for performance. In this study, clinical microbiology laboratories at four major tertiary medical care centers evaluated Gram stain error rates across all nonblood specimen types by using standardized criteria. The study focused on several factors that primarily contribute to errors in the process, including poor specimen quality, smear preparation, and interpretation of the smears. The number of specimens during the evaluation period ranged from 976 to 1,864 specimens per site, and there were a total of 6,115 specimens. Gram stain results were discrepant from culture for 5% of all specimens. Fifty-eight percent of discrepant results were specimens with no organisms reported on Gram stain but significant growth on culture, while 42% of discrepant results had reported organisms on Gram stain that were not recovered in culture. Upon review of available slides, 24% (63/263) of discrepant results were due to reader error, which varied significantly based on site (9% to 45%). The Gram stain error rate also varied between sites, ranging from 0.4% to 2.7%. The data demonstrate a significant variability between laboratories in Gram stain performance and affirm the need for ongoing quality assessment by laboratories. Standardized monitoring of Gram stains is an essential quality control tool for laboratories and is necessary for the establishment of a quality benchmark across laboratories. PMID:26888900

  13. Interpolation Errors in Spectrum Analyzers

    NASA Technical Reports Server (NTRS)

    Martin, J. L.

    1996-01-01

    To obtain the proper measurement amplitude with a spectrum analyzer, the correct frequency-dependent transducer factor must be added to the voltage measured by the transducer. This report examines how entering transducer factors into a spectrum analyzer can cause significant errors in field amplitude due to the misunderstanding of the analyzer's interpolation methods. It also discusses how to reduce these errors to obtain a more accurate field amplitude reading.

  14. An Empirical State Error Covariance Matrix Orbit Determination Example

    NASA Technical Reports Server (NTRS)

    Frisbee, Joseph H., Jr.

    2015-01-01

    is suspect. In its most straight forward form, the technique only requires supplemental calculations to be added to existing batch estimation algorithms. In the current problem being studied a truth model making use of gravity with spherical, J2 and J4 terms plus a standard exponential type atmosphere with simple diurnal and random walk components is used. The ability of the empirical state error covariance matrix to account for errors is investigated under four scenarios during orbit estimation. These scenarios are: exact modeling under known measurement errors, exact modeling under corrupted measurement errors, inexact modeling under known measurement errors, and inexact modeling under corrupted measurement errors. For this problem a simple analog of a distributed space surveillance network is used. The sensors in this network make only range measurements and with simple normally distributed measurement errors. The sensors are assumed to have full horizon to horizon viewing at any azimuth. For definiteness, an orbit at the approximate altitude and inclination of the International Space Station is used for the study. The comparison analyses of the data involve only total vectors. No investigation of specific orbital elements is undertaken. The total vector analyses will look at the chisquare values of the error in the difference between the estimated state and the true modeled state using both the empirical and theoretical error covariance matrices for each of scenario.

  15. Medical Error and Moral Luck.

    PubMed

    Hubbeling, Dieneke

    2016-09-01

    This paper addresses the concept of moral luck. Moral luck is discussed in the context of medical error, especially an error of omission that occurs frequently, but only rarely has adverse consequences. As an example, a failure to compare the label on a syringe with the drug chart results in the wrong medication being administered and the patient dies. However, this error may have previously occurred many times with no tragic consequences. Discussions on moral luck can highlight conflicting intuitions. Should perpetrators receive a harsher punishment because of an adverse outcome, or should they be dealt with in the same way as colleagues who have acted similarly, but with no adverse effects? An additional element to the discussion, specifically with medical errors, is that according to the evidence currently available, punishing individual practitioners does not seem to be effective in preventing future errors. The following discussion, using relevant philosophical and empirical evidence, posits a possible solution for the moral luck conundrum in the context of medical error: namely, making a distinction between the duty to make amends and assigning blame. Blame should be assigned on the basis of actual behavior, while the duty to make amends is dependent on the outcome. PMID:26662613

  16. Error image aware content restoration

    NASA Astrophysics Data System (ADS)

    Choi, Sungwoo; Lee, Moonsik; Jung, Byunghee

    2015-12-01

    As the resolution of TV significantly increased, content consumers have become increasingly sensitive to the subtlest defect in TV contents. This rising standard in quality demanded by consumers has posed a new challenge in today's context where the tape-based process has transitioned to the file-based process: the transition necessitated digitalizing old archives, a process which inevitably produces errors such as disordered pixel blocks, scattered white noise, or totally missing pixels. Unsurprisingly, detecting and fixing such errors require a substantial amount of time and human labor to meet the standard demanded by today's consumers. In this paper, we introduce a novel, automated error restoration algorithm which can be applied to different types of classic errors by utilizing adjacent images while preserving the undamaged parts of an error image as much as possible. We tested our method to error images detected from our quality check system in KBS(Korean Broadcasting System) video archive. We are also implementing the algorithm as a plugin of well-known NLE(Non-linear editing system), which is a familiar tool for quality control agent.

  17. Verification and acceptance tests of the PRIMA DDL optics

    NASA Astrophysics Data System (ADS)

    Bizenberger, Peter; Baumeister, Harald; Graser, Uwe; Henning, Thomas; Krause, Nathalie; Launhardt, Ralf; Naranjo, Vianak; Queloz, Didier; Quirrenbach, Andreas

    2008-07-01

    The last step in designing and building instruments are the verification and acceptance tests of the assembled units and of the final instrument. For instruments, which are engineered to work at the limit of feasibility, these tests must be accurate and stable at a level much better than the expected performance of the instrument. Particularly for interferometric instruments, this requires special care for the test planning and implementation in order to achieve the necessary performance. This paper describes the verification and acceptance tests of the PRIMA DDL optics in terms of wavefront error and tilt requirements as well as the assembling and aligning accuracy. We demonstrate the conformity of the optics and point out the limitations of the test methods.

  18. Lunar orbiter ranging data: initial results.

    PubMed

    Mulholland, J D; Sjogren, W L

    1967-01-01

    Data from two Lunar Orbiter spacecraft have been used to test the significance of corrections to the lunar ephemeris. Range residuals of up to 1700 meters were reduced by an order of magnitude by application of the corrections, with most of the residuals reduced to less than 100 meters. Removal of gross errors in the ephemeris reveals residual patterns that may indicate errors in location of observing stations, as well as the expected effects of Lunar nonsphericity. PMID:17799149

  19. Estimation of rod scale errors in geodetic leveling

    USGS Publications Warehouse

    Craymer, Michael R.; Vaníček, Petr; Castle, Robert O.

    1995-01-01

    Comparisons among repeated geodetic levelings have often been used for detecting and estimating residual rod scale errors in leveled heights. Individual rod-pair scale errors are estimated by a two-step procedure using a model based on either differences in heights, differences in section height differences, or differences in section tilts. It is shown that the estimated rod-pair scale errors derived from each model are identical only when the data are correctly weighted, and the mathematical correlations are accounted for in the model based on heights. Analyses based on simple regressions of changes in height versus height can easily lead to incorrect conclusions. We also show that the statistically estimated scale errors are not a simple function of height, height difference, or tilt. The models are valid only when terrain slope is constant over adjacent pairs of setups (i.e., smoothly varying terrain). In order to discriminate between rod scale errors and vertical displacements due to crustal motion, the individual rod-pairs should be used in more than one leveling, preferably in areas of contrasting tectonic activity. From an analysis of 37 separately calibrated rod-pairs used in 55 levelings in southern California, we found eight statistically significant coefficients that could be reasonably attributed to rod scale errors, only one of which was larger than the expected random error in the applied calibration-based scale correction. However, significant differences with other independent checks indicate that caution should be exercised before accepting these results as evidence of scale error. Further refinements of the technique are clearly needed if the results are to be routinely applied in practice.

  20. Error Growth Rate in the MM5 Model

    NASA Astrophysics Data System (ADS)

    Ivanov, S.; Palamarchuk, J.

    2006-12-01

    The goal of this work is to estimate model error growth rates in simulations of the atmospheric circulation by the MM5 model all the way from the short range to the medium range and beyond. The major topics are addressed to: (i) search the optimal set of parameterization schemes; (ii) evaluate the spatial structure and scales of the model error for various atmospheric fields; (iii) determine geographical regions where model errors are largest; (iv) define particular atmospheric patterns contributing to the fast and significant model error growth. Results are presented for geopotential, temperature, relative humidity and horizontal wind components fields on standard surfaces over the Atlantic-European region during winter 2002. Various combinations of parameterization schemes for cumulus, PBL, moisture and radiation are used to identify which one provides a lesser difference between the model state and analysis. The comparison of the model fields is carried out versus ERA-40 reanalysis of the ECMWF. Results show that the rate, at which the model error grows as well as its magnitude, varies depending on the forecast range, atmospheric variable and level. The typical spatial scale and structure of the model error also depends on the particular atmospheric variable. The distribution of the model error over the domain can be separated in two parts: the steady and transient. The first part is associated with a few high mountain regions including Greenland, where model error is larger. The transient model error mainly moves along with areas of high gradients in the atmospheric flow. Acknowledgement: This study has been supported by NATO Science for Peace grant #981044. The MM5 modelling system used in this study has been provided by UCAR. ERA-40 re-analysis data have been obtained from the ECMWF data server.

  1. Error Detection Processes in Problem Solving.

    ERIC Educational Resources Information Center

    Allwood, Carl Martin

    1984-01-01

    Describes a study which analyzed problem solvers' error detection processes by instructing subjects to think aloud when solving statistical problems. Effects of evaluative episodes on error detection, detection of different error types, error detection processes per se, and relationship of error detection behavior to problem-solving proficiency…

  2. Error-associated behaviors and error rates for robotic geology

    NASA Technical Reports Server (NTRS)

    Anderson, Robert C.; Thomas, Geb; Wagner, Jacob; Glasgow, Justin

    2004-01-01

    This study explores human error as a function of the decision-making process. One of many models for human decision-making is Rasmussen's decision ladder [9]. The decision ladder identifies the multiple tasks and states of knowledge involved in decision-making. The tasks and states of knowledge can be classified by the level of cognitive effort required to make the decision, leading to the skill, rule, and knowledge taxonomy (Rasmussen, 1987). Skill based decisions require the least cognitive effort and knowledge based decisions require the greatest cognitive effort. Errors can occur at any of the cognitive levels.

  3. Acceptance and Commitment Therapy (ACT): An Overview for Practitioners

    ERIC Educational Resources Information Center

    Bowden, Tim; Bowden, Sandra

    2012-01-01

    Acceptance and Commitment Therapy (ACT) offers school counsellors a practical and meaningful approach to helping students deal with a range of issues. This is achieved through encouraging psychological flexibility through the application of six key principles. This article describes our introduction to ACT, ACT's application to children and…

  4. Higher-order ionospheric error at Arecibo, Millstone, and Jicamarca

    NASA Astrophysics Data System (ADS)

    Matteo, N. A.; Morton, Y. T.

    2010-12-01

    The ionosphere is a dominant source of Global Positioning System receiver range measurement error. Although dual-frequency receivers can eliminate the first-order ionospheric error, most second- and third-order errors remain in the range measurements. Higher-order ionospheric error is a function of both electron density distribution and the magnetic field vector along the GPS signal propagation path. This paper expands previous efforts by combining incoherent scatter radar (ISR) electron density measurements, the International Reference Ionosphere model, exponential decay extensions of electron densities, the International Geomagnetic Reference Field, and total electron content maps to compute higher-order error at ISRs in Arecibo, Puerto Rico; Jicamarca, Peru; and Millstone Hill, Massachusetts. Diurnal patterns, dependency on signal direction, seasonal variation, and geomagnetic activity dependency are analyzed. Higher-order error is largest at Arecibo with code phase maxima circa 7 cm for low-elevation southern signals. The maximum variation of the error over all angles of arrival is circa 8 cm.

  5. Verification of the Forecast Errors Based on Ensemble Spread

    NASA Astrophysics Data System (ADS)

    Vannitsem, S.; Van Schaeybroeck, B.

    2014-12-01

    The use of ensemble prediction systems allows for an uncertainty estimation of the forecast. Most end users do not require all the information contained in an ensemble and prefer the use of a single uncertainty measure. This measure is the ensemble spread which serves to forecast the forecast error. It is however unclear how best the quality of these forecasts can be performed, based on spread and forecast error only. The spread-error verification is intricate for two reasons: First for each probabilistic forecast only one observation is substantiated and second, the spread is not meant to provide an exact prediction for the error. Despite these facts several advances were recently made, all based on traditional deterministic verification of the error forecast. In particular, Grimit and Mass (2007) and Hopson (2014) considered in detail the strengths and weaknesses of the spread-error correlation, while Christensen et al (2014) developed a proper-score extension of the mean squared error. However, due to the strong variance of the error given a certain spread, the error forecast should be preferably considered as probabilistic in nature. In the present work, different probabilistic error models are proposed depending on the spread-error metrics used. Most of these models allow for the discrimination of a perfect forecast from an imperfect one, independent of the underlying ensemble distribution. The new spread-error scores are tested on the ensemble prediction system of the European Centre of Medium-range forecasts (ECMWF) over Europe and Africa. ReferencesChristensen, H. M., Moroz, I. M. and Palmer, T. N., 2014, Evaluation of ensemble forecast uncertainty using a new proper score: application to medium-range and seasonal forecasts. In press, Quarterly Journal of the Royal Meteorological Society. Grimit, E. P., and C. F. Mass, 2007: Measuring the ensemble spread-error relationship with a probabilistic approach: Stochastic ensemble results. Mon. Wea. Rev., 135, 203

  6. Masking of errors in transmission of VAPC-coded speech

    NASA Technical Reports Server (NTRS)

    Cox, Neil B.; Froese, Edwin L.

    1990-01-01

    A subjective evaluation is provided of the bit error sensitivity of the message elements of a Vector Adaptive Predictive (VAPC) speech coder, along with an indication of the amenability of these elements to a popular error masking strategy (cross frame hold over). As expected, a wide range of bit error sensitivity was observed. The most sensitive message components were the short term spectral information and the most significant bits of the pitch and gain indices. The cross frame hold over strategy was found to be useful for pitch and gain information, but it was not beneficial for the spectral information unless severe corruption had occurred.

  7. An Empirical State Error Covariance Matrix for Batch State Estimation

    NASA Technical Reports Server (NTRS)

    Frisbee, Joseph H., Jr.

    2011-01-01

    state estimate, regardless as to the source of the uncertainty. Also, in its most straight forward form, the technique only requires supplemental calculations to be added to existing batch algorithms. The generation of this direct, empirical form of the state error covariance matrix is independent of the dimensionality of the observations. Mixed degrees of freedom for an observation set are allowed. As is the case with any simple, empirical sample variance problems, the presented approach offers an opportunity (at least in the case of weighted least squares) to investigate confidence interval estimates for the error covariance matrix elements. The diagonal or variance terms of the error covariance matrix have a particularly simple form to associate with either a multiple degree of freedom chi-square distribution (more approximate) or with a gamma distribution (less approximate). The off diagonal or covariance terms of the matrix are less clear in their statistical behavior. However, the off diagonal covariance matrix elements still lend themselves to standard confidence interval error analysis. The distributional forms associated with the off diagonal terms are more varied and, perhaps, more approximate than those associated with the diagonal terms. Using a simple weighted least squares sample problem, results obtained through use of the proposed technique are presented. The example consists of a simple, two observer, triangulation problem with range only measurements. Variations of this problem reflect an ideal case (perfect knowledge of the range errors) and a mismodeled case (incorrect knowledge of the range errors).

  8. Microdensitometer errors: Their effect on photometric data reduction

    NASA Technical Reports Server (NTRS)

    Bozyan, E. P.; Opal, C. B.

    1984-01-01

    The performance of densitometers used for photometric data reduction of high dynamic range electrographic plate material is analyzed. Densitometer repeatability is tested by comparing two scans of one plate. Internal densitometer errors are examined by constructing histograms of digitized densities and finding inoperative bits and differential nonlinearity in the analog to digital converter. Such problems appear common to the four densitometers used in this investigation and introduce systematic algorithm dependent errors in the results. Strategies to improve densitometer performance are suggested.

  9. Aerial measurement error with a dot planimeter: Some experimental estimates

    NASA Technical Reports Server (NTRS)

    Yuill, R. S.

    1971-01-01

    A shape analysis is presented which utilizes a computer to simulate a multiplicity of dot grids mathematically. Results indicate that the number of dots placed over an area to be measured provides the entire correlation with accuracy of measurement, the indices of shape being of little significance. Equations and graphs are provided from which the average expected error, and the maximum range of error, for various numbers of dot points can be read.

  10. Error Cost Escalation Through the Project Life Cycle

    NASA Technical Reports Server (NTRS)

    Stecklein, Jonette M.; Dabney, Jim; Dick, Brandon; Haskins, Bill; Lovell, Randy; Moroney, Gregory

    2004-01-01

    It is well known that the costs to fix errors increase as the project matures, but how fast do those costs build? A study was performed to determine the relative cost of fixing errors discovered during various phases of a project life cycle. This study used three approaches to determine the relative costs: the bottom-up cost method, the total cost breakdown method, and the top-down hypothetical project method. The approaches and results described in this paper presume development of a hardware/software system having project characteristics similar to those used in the development of a large, complex spacecraft, a military aircraft, or a small communications satellite. The results show the degree to which costs escalate, as errors are discovered and fixed at later and later phases in the project life cycle. If the cost of fixing a requirements error discovered during the requirements phase is defined to be 1 unit, the cost to fix that error if found during the design phase increases to 3 - 8 units; at the manufacturing/build phase, the cost to fix the error is 7 - 16 units; at the integration and test phase, the cost to fix the error becomes 21 - 78 units; and at the operations phase, the cost to fix the requirements error ranged from 29 units to more than 1500 units

  11. Photocephalometry: errors of projection and landmark location.

    PubMed

    Phillips, C; Greer, J; Vig, P; Matteson, S

    1984-09-01

    A method called photocephalometry was recently described for the possible soft-tissue evaluation of orthognathic surgery patients by the superimposition of coordinated cephalographs and photographs. A grid analysis was performed to determine the accuracy of the superimposition method. In addition, the reliability of landmark identification was analyzed by the method error of Baumrind and Frantz, using three replicates of twelve patients' photographs. Comparison of twenty-one grid intervals showed that the magnification of the photographic image for any given grid plane is not correlated to that of the radiographic image. Accurate comparisons between soft- and hard-tissue anatomy by simply superimposing the images are not feasible because of the difference in the enlargement factors between the photographs and x-ray films. As was noted by Baumrind and Frantz, a wide range exists in the variability of estimating the location of landmarks. Sixty-six percent of the lateral photographic landmarks and 57% of the frontal landmarks had absolute mean errors for all twelve patients that were less than or equal to 2.0 mm. In general, the envelope of error for most landmarks was not circular. Although the photocephalometric apparatus as described by Hohl and colleagues does not yield the desired quantitative correlation between hard and soft tissues, valuable quantitative information on soft tissue can be easily obtained with the standardization and replication possible with the camera setup and enlarged photographs. PMID:6591803

  12. 48 CFR 811.103 - Market acceptance.

    Code of Federal Regulations, 2011 CFR

    2011-10-01

    ... 48 Federal Acquisition Regulations System 5 2011-10-01 2011-10-01 false Market acceptance. 811.103 Section 811.103 Federal Acquisition Regulations System DEPARTMENT OF VETERANS AFFAIRS COMPETITION AND... acceptance....

  13. 48 CFR 811.103 - Market acceptance.

    Code of Federal Regulations, 2010 CFR

    2010-10-01

    ... 48 Federal Acquisition Regulations System 5 2010-10-01 2010-10-01 false Market acceptance. 811.103 Section 811.103 Federal Acquisition Regulations System DEPARTMENT OF VETERANS AFFAIRS COMPETITION AND... acceptance....

  14. Spacecraft and propulsion technician error

    NASA Astrophysics Data System (ADS)

    Schultz, Daniel Clyde

    Commercial aviation and commercial space similarly launch, fly, and land passenger vehicles. Unlike aviation, the U.S. government has not established maintenance policies for commercial space. This study conducted a mixed methods review of 610 U.S. space launches from 1984 through 2011, which included 31 failures. An analysis of the failure causal factors showed that human error accounted for 76% of those failures, which included workmanship error accounting for 29% of the failures. With the imminent future of commercial space travel, the increased potential for the loss of human life demands that changes be made to the standardized procedures, training, and certification to reduce human error and failure rates. Several recommendations were made by this study to the FAA's Office of Commercial Space Transportation, space launch vehicle operators, and maintenance technician schools in an effort to increase the safety of the space transportation passengers.

  15. Error analysis using organizational simulation.

    PubMed Central

    Fridsma, D. B.

    2000-01-01

    Organizational simulations have been used by project organizations in civil and aerospace industries to identify work processes and organizational structures that are likely to fail under certain conditions. Using a simulation system based on Galbraith's information-processing theory and Simon's notion of bounded-rationality, we retrospectively modeled a chemotherapy administration error that occurred in a hospital setting. Our simulation suggested that when there is a high rate of unexpected events, the oncology fellow was differentially backlogged with work when compared with other organizational members. Alternative scenarios suggested that providing more knowledge resources to the oncology fellow improved her performance more effectively than adding additional staff to the organization. Although it is not possible to know whether this might have prevented the error, organizational simulation may be an effective tool to prospectively evaluate organizational "weak links", and explore alternative scenarios to correct potential organizational problems before they generate errors. PMID:11079885

  16. Synthetic aperture interferometry: error analysis

    SciTech Connect

    Biswas, Amiya; Coupland, Jeremy

    2010-07-10

    Synthetic aperture interferometry (SAI) is a novel way of testing aspherics and has a potential for in-process measurement of aspherics [Appl. Opt.42, 701 (2003)].APOPAI0003-693510.1364/AO.42.000701 A method to measure steep aspherics using the SAI technique has been previously reported [Appl. Opt.47, 1705 (2008)].APOPAI0003-693510.1364/AO.47.001705 Here we investigate the computation of surface form using the SAI technique in different configurations and discuss the computational errors. A two-pass measurement strategy is proposed to reduce the computational errors, and a detailed investigation is carried out to determine the effect of alignment errors on the measurement process.

  17. Reward positivity: Reward prediction error or salience prediction error?

    PubMed

    Heydari, Sepideh; Holroyd, Clay B

    2016-08-01

    The reward positivity is a component of the human ERP elicited by feedback stimuli in trial-and-error learning and guessing tasks. A prominent theory holds that the reward positivity reflects a reward prediction error signal that is sensitive to outcome valence, being larger for unexpected positive events relative to unexpected negative events (Holroyd & Coles, 2002). Although the theory has found substantial empirical support, most of these studies have utilized either monetary or performance feedback to test the hypothesis. However, in apparent contradiction to the theory, a recent study found that unexpected physical punishments also elicit the reward positivity (Talmi, Atkinson, & El-Deredy, 2013). The authors of this report argued that the reward positivity reflects a salience prediction error rather than a reward prediction error. To investigate this finding further, in the present study participants navigated a virtual T maze and received feedback on each trial under two conditions. In a reward condition, the feedback indicated that they would either receive a monetary reward or not and in a punishment condition the feedback indicated that they would receive a small shock or not. We found that the feedback stimuli elicited a typical reward positivity in the reward condition and an apparently delayed reward positivity in the punishment condition. Importantly, this signal was more positive to the stimuli that predicted the omission of a possible punishment relative to stimuli that predicted a forthcoming punishment, which is inconsistent with the salience hypothesis. PMID:27184070

  18. 7 CFR 966.29 - Acceptance.

    Code of Federal Regulations, 2011 CFR

    2011-01-01

    ... 7 Agriculture 8 2011-01-01 2011-01-01 false Acceptance. 966.29 Section 966.29 Agriculture... Regulating Handling Committee § 966.29 Acceptance. Any person selected as a committee member or alternate shall qualify by filing a written acceptance with the Secretary within ten days after being notified...

  19. 7 CFR 924.25 - Acceptance.

    Code of Federal Regulations, 2011 CFR

    2011-01-01

    ... 7 Agriculture 8 2011-01-01 2011-01-01 false Acceptance. 924.25 Section 924.25 Agriculture....25 Acceptance. Any person selected by the Secretary as a member or as an alternate member of the committee shall qualify by filing a written acceptance with the Secretary promptly after being notified...

  20. 7 CFR 924.25 - Acceptance.

    Code of Federal Regulations, 2010 CFR

    2010-01-01

    ... 7 Agriculture 8 2010-01-01 2010-01-01 false Acceptance. 924.25 Section 924.25 Agriculture....25 Acceptance. Any person selected by the Secretary as a member or as an alternate member of the committee shall qualify by filing a written acceptance with the Secretary promptly after being notified...

  1. 7 CFR 923.25 - Acceptance.

    Code of Federal Regulations, 2010 CFR

    2010-01-01

    ... 7 Agriculture 8 2010-01-01 2010-01-01 false Acceptance. 923.25 Section 923.25 Agriculture... COUNTIES IN WASHINGTON Order Regulating Handling Administrative Body § 923.25 Acceptance. Any person prior... written acceptance of willingness to serve on the committee....

  2. 7 CFR 993.31 - Acceptance.

    Code of Federal Regulations, 2011 CFR

    2011-01-01

    ... 7 Agriculture 8 2011-01-01 2011-01-01 false Acceptance. 993.31 Section 993.31 Agriculture... Order Regulating Handling Prune Marketing Committee § 993.31 Acceptance. Each person selected as a... with the Secretary a written acceptance within 15 days after receiving notice of his selection....

  3. 7 CFR 1215.23 - Acceptance.

    Code of Federal Regulations, 2010 CFR

    2010-01-01

    ... 7 Agriculture 10 2010-01-01 2010-01-01 false Acceptance. 1215.23 Section 1215.23 Agriculture... Acceptance. Each individual nominated for membership of the Board shall qualify by filing a written acceptance with the Secretary at the time of nomination....

  4. 7 CFR 906.25 - Acceptance.

    Code of Federal Regulations, 2011 CFR

    2011-01-01

    ... 7 Agriculture 8 2011-01-01 2011-01-01 false Acceptance. 906.25 Section 906.25 Agriculture... RIO GRANDE VALLEY IN TEXAS Order Regulating Handling Committee § 906.25 Acceptance. Any person selected as a committee member or alternate shall qualify by filing a written acceptance with the...

  5. 7 CFR 946.26 - Acceptance.

    Code of Federal Regulations, 2011 CFR

    2011-01-01

    ... 7 Agriculture 8 2011-01-01 2011-01-01 false Acceptance. 946.26 Section 946.26 Agriculture... Order Regulating Handling Administrative Committee § 946.26 Acceptance. Any person selected by the Secretary as a committee member or as an alternate shall qualify by filing a written acceptance with...

  6. 7 CFR 959.29 - Acceptance.

    Code of Federal Regulations, 2010 CFR

    2010-01-01

    ... 7 Agriculture 8 2010-01-01 2010-01-01 false Acceptance. 959.29 Section 959.29 Agriculture... Regulating Handling Committee § 959.29 Acceptance. Any person selected as a committee member or alternate shall qualify by filing a written acceptance within ten days after being notified of such selection....

  7. 12 CFR 250.164 - Bankers' acceptances.

    Code of Federal Regulations, 2010 CFR

    2010-01-01

    ... 12 Banks and Banking 3 2010-01-01 2010-01-01 false Bankers' acceptances. 250.164 Section 250.164... MISCELLANEOUS INTERPRETATIONS Interpretations § 250.164 Bankers' acceptances. (a) Section 207 of the Bank Export... bankers' acceptances (“BAs”) that may be created by an individual member bank from 50 per cent (or 100...

  8. 24 CFR 3282.355 - Submission acceptance.

    Code of Federal Regulations, 2011 CFR

    2011-04-01

    ... 24 Housing and Urban Development 5 2011-04-01 2011-04-01 false Submission acceptance. 3282.355... § 3282.355 Submission acceptance. (a) A party whose submission is determined by the Department to be adequate shall be granted provisional acceptance until December 15, 1976, or for a six month period...

  9. 7 CFR 906.25 - Acceptance.

    Code of Federal Regulations, 2010 CFR

    2010-01-01

    ... 7 Agriculture 8 2010-01-01 2010-01-01 false Acceptance. 906.25 Section 906.25 Agriculture... RIO GRANDE VALLEY IN TEXAS Order Regulating Handling Committee § 906.25 Acceptance. Any person selected as a committee member or alternate shall qualify by filing a written acceptance with the...

  10. 24 CFR 3282.355 - Submission acceptance.

    Code of Federal Regulations, 2010 CFR

    2010-04-01

    ... 24 Housing and Urban Development 5 2010-04-01 2010-04-01 false Submission acceptance. 3282.355... § 3282.355 Submission acceptance. (a) A party whose submission is determined by the Department to be adequate shall be granted provisional acceptance until December 15, 1976, or for a six month period...

  11. 48 CFR 11.103 - Market acceptance.

    Code of Federal Regulations, 2011 CFR

    2011-10-01

    ... 48 Federal Acquisition Regulations System 1 2011-10-01 2011-10-01 false Market acceptance. 11.103... DESCRIBING AGENCY NEEDS Selecting and Developing Requirements Documents 11.103 Market acceptance. (a) Section... either— (i) Achieved commercial market acceptance; or (ii) Been satisfactorily supplied to an...

  12. 7 CFR 1250.330 - Acceptance.

    Code of Federal Regulations, 2011 CFR

    2011-01-01

    ... 7 Agriculture 10 2011-01-01 2011-01-01 false Acceptance. 1250.330 Section 1250.330 Agriculture... Research and Promotion Order Egg Board § 1250.330 Acceptance. Any person appointed by the Secretary as a member, or as an alternate member, of the Board shall qualify by filing a written acceptance with...

  13. 7 CFR 923.25 - Acceptance.

    Code of Federal Regulations, 2011 CFR

    2011-01-01

    ... 7 Agriculture 8 2011-01-01 2011-01-01 false Acceptance. 923.25 Section 923.25 Agriculture... COUNTIES IN WASHINGTON Order Regulating Handling Administrative Body § 923.25 Acceptance. Any person prior... written acceptance of willingness to serve on the committee....

  14. 7 CFR 1205.326 - Acceptance.

    Code of Federal Regulations, 2011 CFR

    2011-01-01

    ... 7 Agriculture 10 2011-01-01 2011-01-01 false Acceptance. 1205.326 Section 1205.326 Agriculture... Research and Promotion Order Cotton Board § 1205.326 Acceptance. Any person selected by the Secretary as a member or as an alternate member of the Board shall qualify by filing a written acceptance with...

  15. 7 CFR 953.21 - Acceptance.

    Code of Federal Regulations, 2010 CFR

    2010-01-01

    ... 7 Agriculture 8 2010-01-01 2010-01-01 false Acceptance. 953.21 Section 953.21 Agriculture... STATES Order Regulating Handling Administrative Committee § 953.21 Acceptance. Any person selected by the... acceptance with the Secretary within the time specified by the Secretary....

  16. 12 CFR 250.164 - Bankers' acceptances.

    Code of Federal Regulations, 2011 CFR

    2011-01-01

    ... 12 Banks and Banking 3 2011-01-01 2011-01-01 false Bankers' acceptances. 250.164 Section 250.164... MISCELLANEOUS INTERPRETATIONS Interpretations § 250.164 Bankers' acceptances. (a) Section 207 of the Bank Export... bankers' acceptances (“BAs”) that may be created by an individual member bank from 50 per cent (or 100...

  17. 7 CFR 966.29 - Acceptance.

    Code of Federal Regulations, 2010 CFR

    2010-01-01

    ... 7 Agriculture 8 2010-01-01 2010-01-01 false Acceptance. 966.29 Section 966.29 Agriculture... Regulating Handling Committee § 966.29 Acceptance. Any person selected as a committee member or alternate shall qualify by filing a written acceptance with the Secretary within ten days after being notified...

  18. 48 CFR 2911.103 - Market acceptance.

    Code of Federal Regulations, 2010 CFR

    2010-10-01

    ... 48 Federal Acquisition Regulations System 7 2010-10-01 2010-10-01 false Market acceptance. 2911... DESCRIBING AGENCY NEEDS Selecting And Developing Requirements Documents 2911.103 Market acceptance. The... offered have either achieved commercial market acceptance or been satisfactorily supplied to an...

  19. 7 CFR 1215.23 - Acceptance.

    Code of Federal Regulations, 2011 CFR

    2011-01-01

    ... 7 Agriculture 10 2011-01-01 2011-01-01 false Acceptance. 1215.23 Section 1215.23 Agriculture... Acceptance. Each individual nominated for membership of the Board shall qualify by filing a written acceptance with the Secretary at the time of nomination....

  20. 7 CFR 946.26 - Acceptance.

    Code of Federal Regulations, 2010 CFR

    2010-01-01

    ... 7 Agriculture 8 2010-01-01 2010-01-01 false Acceptance. 946.26 Section 946.26 Agriculture... Order Regulating Handling Administrative Committee § 946.26 Acceptance. Any person selected by the Secretary as a committee member or as an alternate shall qualify by filing a written acceptance with...

  1. 7 CFR 1205.326 - Acceptance.

    Code of Federal Regulations, 2010 CFR

    2010-01-01

    ... 7 Agriculture 10 2010-01-01 2010-01-01 false Acceptance. 1205.326 Section 1205.326 Agriculture... Research and Promotion Order Cotton Board § 1205.326 Acceptance. Any person selected by the Secretary as a member or as an alternate member of the Board shall qualify by filing a written acceptance with...

  2. 7 CFR 953.21 - Acceptance.

    Code of Federal Regulations, 2011 CFR

    2011-01-01

    ... 7 Agriculture 8 2011-01-01 2011-01-01 false Acceptance. 953.21 Section 953.21 Agriculture... STATES Order Regulating Handling Administrative Committee § 953.21 Acceptance. Any person selected by the... acceptance with the Secretary within the time specified by the Secretary....

  3. 7 CFR 1250.330 - Acceptance.

    Code of Federal Regulations, 2010 CFR

    2010-01-01

    ... 7 Agriculture 10 2010-01-01 2010-01-01 false Acceptance. 1250.330 Section 1250.330 Agriculture... Research and Promotion Order Egg Board § 1250.330 Acceptance. Any person appointed by the Secretary as a member, or as an alternate member, of the Board shall qualify by filing a written acceptance with...

  4. 48 CFR 2911.103 - Market acceptance.

    Code of Federal Regulations, 2011 CFR

    2011-10-01

    ... 48 Federal Acquisition Regulations System 7 2011-10-01 2011-10-01 false Market acceptance. 2911... DESCRIBING AGENCY NEEDS Selecting And Developing Requirements Documents 2911.103 Market acceptance. The... offered have either achieved commercial market acceptance or been satisfactorily supplied to an...

  5. 48 CFR 11.103 - Market acceptance.

    Code of Federal Regulations, 2010 CFR

    2010-10-01

    ... 48 Federal Acquisition Regulations System 1 2010-10-01 2010-10-01 false Market acceptance. 11.103... DESCRIBING AGENCY NEEDS Selecting and Developing Requirements Documents 11.103 Market acceptance. (a) Section... either— (i) Achieved commercial market acceptance; or (ii) Been satisfactorily supplied to an...

  6. 7 CFR 1210.323 - Acceptance.

    Code of Federal Regulations, 2011 CFR

    2011-01-01

    ... 7 Agriculture 10 2011-01-01 2011-01-01 false Acceptance. 1210.323 Section 1210.323 Agriculture... PLAN Watermelon Research and Promotion Plan National Watermelon Promotion Board § 1210.323 Acceptance. Each person nominated for membership on the Board shall qualify by filing a written acceptance with...

  7. 7 CFR 959.29 - Acceptance.

    Code of Federal Regulations, 2011 CFR

    2011-01-01

    ... 7 Agriculture 8 2011-01-01 2011-01-01 false Acceptance. 959.29 Section 959.29 Agriculture... Regulating Handling Committee § 959.29 Acceptance. Any person selected as a committee member or alternate shall qualify by filing a written acceptance within ten days after being notified of such selection....

  8. 7 CFR 993.31 - Acceptance.

    Code of Federal Regulations, 2010 CFR

    2010-01-01

    ... 7 Agriculture 8 2010-01-01 2010-01-01 false Acceptance. 993.31 Section 993.31 Agriculture... Order Regulating Handling Prune Marketing Committee § 993.31 Acceptance. Each person selected as a... with the Secretary a written acceptance within 15 days after receiving notice of his selection....

  9. 7 CFR 915.25 - Acceptance.

    Code of Federal Regulations, 2011 CFR

    2011-01-01

    ... 7 Agriculture 8 2011-01-01 2011-01-01 false Acceptance. 915.25 Section 915.25 Agriculture... Regulating Handling Administrative Body § 915.25 Acceptance. Any person selected by the Secretary as a member or as an alternate member of the committee shall qualify by filing a written acceptance with...

  10. 7 CFR 1210.323 - Acceptance.

    Code of Federal Regulations, 2010 CFR

    2010-01-01

    ... 7 Agriculture 10 2010-01-01 2010-01-01 false Acceptance. 1210.323 Section 1210.323 Agriculture... PLAN Watermelon Research and Promotion Plan National Watermelon Promotion Board § 1210.323 Acceptance. Each person nominated for membership on the Board shall qualify by filing a written acceptance with...

  11. 46 CFR 28.73 - Accepted organizations.

    Code of Federal Regulations, 2011 CFR

    2011-10-01

    ... 46 Shipping 1 2011-10-01 2011-10-01 false Accepted organizations. 28.73 Section 28.73 Shipping... INDUSTRY VESSELS General Provisions § 28.73 Accepted organizations. An organization desiring to be designated by the Commandant as an accepted organization must request such designation in writing. As...

  12. Older Adults' Acceptance of Information Technology

    ERIC Educational Resources Information Center

    Wang, Lin; Rau, Pei-Luen Patrick; Salvendy, Gavriel

    2011-01-01

    This study investigated variables contributing to older adults' information technology acceptance through a survey, which was used to find factors explaining and predicting older adults' information technology acceptance behaviors. Four factors, including needs satisfaction, perceived usability, support availability, and public acceptance, were…

  13. 48 CFR 11.103 - Market acceptance.

    Code of Federal Regulations, 2012 CFR

    2012-10-01

    ... 48 Federal Acquisition Regulations System 1 2012-10-01 2012-10-01 false Market acceptance. 11.103... DESCRIBING AGENCY NEEDS Selecting and Developing Requirements Documents 11.103 Market acceptance. (a) Section... either— (i) Achieved commercial market acceptance; or (ii) Been satisfactorily supplied to an...

  14. 48 CFR 11.103 - Market acceptance.

    Code of Federal Regulations, 2014 CFR

    2014-10-01

    ... 48 Federal Acquisition Regulations System 1 2014-10-01 2014-10-01 false Market acceptance. 11.103... DESCRIBING AGENCY NEEDS Selecting and Developing Requirements Documents 11.103 Market acceptance. (a) 41 U.S...) Achieved commercial market acceptance; or (ii) Been satisfactorily supplied to an agency under current...

  15. 48 CFR 2911.103 - Market acceptance.

    Code of Federal Regulations, 2012 CFR

    2012-10-01

    ... 48 Federal Acquisition Regulations System 7 2012-10-01 2012-10-01 false Market acceptance. 2911... DESCRIBING AGENCY NEEDS Selecting And Developing Requirements Documents 2911.103 Market acceptance. The... offered have either achieved commercial market acceptance or been satisfactorily supplied to an...

  16. 48 CFR 11.103 - Market acceptance.

    Code of Federal Regulations, 2013 CFR

    2013-10-01

    ... 48 Federal Acquisition Regulations System 1 2013-10-01 2013-10-01 false Market acceptance. 11.103... DESCRIBING AGENCY NEEDS Selecting and Developing Requirements Documents 11.103 Market acceptance. (a) Section... either— (i) Achieved commercial market acceptance; or (ii) Been satisfactorily supplied to an...

  17. 48 CFR 2911.103 - Market acceptance.

    Code of Federal Regulations, 2013 CFR

    2013-10-01

    ... 48 Federal Acquisition Regulations System 7 2013-10-01 2012-10-01 true Market acceptance. 2911.103... DESCRIBING AGENCY NEEDS Selecting And Developing Requirements Documents 2911.103 Market acceptance. The... offered have either achieved commercial market acceptance or been satisfactorily supplied to an...

  18. 48 CFR 2911.103 - Market acceptance.

    Code of Federal Regulations, 2014 CFR

    2014-10-01

    ... 48 Federal Acquisition Regulations System 7 2014-10-01 2014-10-01 false Market acceptance. 2911... DESCRIBING AGENCY NEEDS Selecting And Developing Requirements Documents 2911.103 Market acceptance. The... offered have either achieved commercial market acceptance or been satisfactorily supplied to an...

  19. The Relationship between Treatment Acceptability and Familism

    ERIC Educational Resources Information Center

    Pemberton, Joy R.; Borrego, Joaquin, Jr.

    2005-01-01

    Many studies have examined the acceptability of treatments for children with disruptive behaviors. However, few studies to date have tested the effects of home environment variables such as family support on treatment acceptability. In the current study, parents' level of familism was used to predict their willingness to accept several behavioral…

  20. 46 CFR 28.73 - Accepted organizations.

    Code of Federal Regulations, 2010 CFR

    2010-10-01

    ... 46 Shipping 1 2010-10-01 2010-10-01 false Accepted organizations. 28.73 Section 28.73 Shipping... INDUSTRY VESSELS General Provisions § 28.73 Accepted organizations. An organization desiring to be designated by the Commandant as an accepted organization must request such designation in writing. As...

  1. 20 Tips to Help Prevent Medical Errors

    MedlinePlus

    ... Prevent Medical Errors 20 Tips to Help Prevent Medical Errors: Patient Fact Sheet This information is for ... current information. Select to Download PDF (295 KB). Medical errors can occur anywhere in the health care ...

  2. Analysis of Medication Error Reports

    SciTech Connect

    Whitney, Paul D.; Young, Jonathan; Santell, John; Hicks, Rodney; Posse, Christian; Fecht, Barbara A.

    2004-11-15

    In medicine, as in many areas of research, technological innovation and the shift from paper based information to electronic records has created a climate of ever increasing availability of raw data. There has been, however, a corresponding lag in our abilities to analyze this overwhelming mass of data, and classic forms of statistical analysis may not allow researchers to interact with data in the most productive way. This is true in the emerging area of patient safety improvement. Traditionally, a majority of the analysis of error and incident reports has been carried out based on an approach of data comparison, and starts with a specific question which needs to be answered. Newer data analysis tools have been developed which allow the researcher to not only ask specific questions but also to “mine” data: approach an area of interest without preconceived questions, and explore the information dynamically, allowing questions to be formulated based on patterns brought up by the data itself. Since 1991, United States Pharmacopeia (USP) has been collecting data on medication errors through voluntary reporting programs. USP’s MEDMARXsm reporting program is the largest national medication error database and currently contains well over 600,000 records. Traditionally, USP has conducted an annual quantitative analysis of data derived from “pick-lists” (i.e., items selected from a list of items) without an in-depth analysis of free-text fields. In this paper, the application of text analysis and data analysis tools used by Battelle to analyze the medication error reports already analyzed in the traditional way by USP is described. New insights and findings were revealed including the value of language normalization and the distribution of error incidents by day of the week. The motivation for this effort is to gain additional insight into the nature of medication errors to support improvements in medication safety.

  3. How psychotherapists handle treatment errors – an ethical analysis

    PubMed Central

    2013-01-01

    Background Dealing with errors in psychotherapy is challenging, both ethically and practically. There is almost no empirical research on this topic. We aimed (1) to explore psychotherapists’ self-reported ways of dealing with an error made by themselves or by colleagues, and (2) to reconstruct their reasoning according to the two principle-based ethical approaches that are dominant in the ethics discourse of psychotherapy, Beauchamp & Childress (B&C) and Lindsay et al. (L). Methods We conducted 30 semi-structured interviews with 30 psychotherapists (physicians and non-physicians) and analysed the transcripts using qualitative content analysis. Answers were deductively categorized according to the two principle-based ethical approaches. Results Most psychotherapists reported that they preferred to an disclose error to the patient. They justified this by spontaneous intuitions and common values in psychotherapy, rarely using explicit ethical reasoning. The answers were attributed to the following categories with descending frequency: 1. Respect for patient autonomy (B&C; L), 2. Non-maleficence (B&C) and Responsibility (L), 3. Integrity (L), 4. Competence (L) and Beneficence (B&C). Conclusions Psychotherapists need specific ethical and communication training to complement and articulate their moral intuitions as a support when disclosing their errors to the patients. Principle-based ethical approaches seem to be useful for clarifying the reasons for disclosure. Further research should help to identify the most effective and acceptable ways of error disclosure in psychotherapy. PMID:24321503

  4. High Precision Ranging and Range-Rate Measurements over Free-Space-Laser Communication Link

    NASA Technical Reports Server (NTRS)

    Yang, Guangning; Lu, Wei; Krainak, Michael; Sun, Xiaoli

    2016-01-01

    We present a high-precision ranging and range-rate measurement system via an optical-ranging or combined ranging-communication link. A complete bench-top optical communication system was built. It included a ground terminal and a space terminal. Ranging and range rate tests were conducted in two configurations. In the communication configuration with 622 data rate, we achieved a two-way range-rate error of 2 microns/s, or a modified Allan deviation of 9 x 10 (exp -15) with 10 second averaging time. Ranging and range-rate as a function of Bit Error Rate of the communication link is reported. They are not sensitive to the link error rate. In the single-frequency amplitude modulation mode, we report a two-way range rate error of 0.8 microns/s, or a modified Allan deviation of 2.6 x 10 (exp -15) with 10 second averaging time. We identified the major noise sources in the current system as the transmitter modulation injected noise and receiver electronics generated noise. A new improved system will be constructed to further improve the system performance for both operating modes.

  5. Ligation errors in DNA computing.

    PubMed

    Aoi, Y; Yoshinobu, T; Tanizawa, K; Kinoshita, K; Iwasaki, H

    1999-10-01

    DNA computing is a novel method of computing proposed by Adleman (1994), in which the data is encoded in the sequences of oligonucleotides. Massively parallel reactions between oligonucleotides are expected to make it possible to solve huge problems. In this study, reliability of the ligation process employed in the DNA computing is tested by estimating the error rate at which wrong oligonucleotides are ligated. Ligation of wrong oligonucleotides would result in a wrong answer in the DNA computing. The dependence of the error rate on the number of mismatches between oligonucleotides and on the combination of bases is investigated. PMID:10636043

  6. Toward a theoretical approach to medical error reporting system research and design.

    PubMed

    Karsh, Ben-Tzion; Escoto, Kamisha Hamilton; Beasley, John W; Holden, Richard J

    2006-05-01

    The release of the Institute of Medicine (Kohn et al., 2000) report "To Err is Human", brought attention to the problem of medical errors, which led to a concerted effort to study and design medical error reporting systems for the purpose of capturing and analyzing error data so that safety interventions could be designed. However, to make real gains in the efficacy of medical error or event reporting systems, it is necessary to begin developing a theory of reporting systems adoption and use and to understand how existing theories may play a role in explaining adoption and use. This paper presents the results of a 9-month study exploring the barriers and facilitators for the design of a statewide medical error reporting system and discusses how several existing theories of technology acceptance, adoption and implementation fit with many of the results. In addition we present an integrated theoretical model of medical error reporting system design and implementation. PMID:16182233

  7. Toward a theoretical approach to medical error reporting system research and design

    PubMed Central

    Karsh, Ben-Tzion; Escoto, Kamisha Hamilton; Beasley, John W.; Holden, Richard J.

    2014-01-01

    The release of the Institute of Medicine (Kohn et al., 2000) report “To Err is Human”, brought attention to the problem of medical errors, which led to a concerted effort to study and design medical error reporting systems for the purpose of capturing and analyzing error data so that safety interventions could be designed. However, to make real gains in the efficacy of medical error or event reporting systems, it is necessary to begin developing a theory of reporting systems adoption and use and to understand how existing theories may play a role in explaining adoption and use. This paper presents the results of a 9-month study exploring the barriers and facilitators for the design of a statewide medical error reporting system and discusses how several existing theories of technology acceptance, adoption and implementation fit with many of the results. In addition we present an integrated theoretical model of medical error reporting system design and implementation. PMID:16182233

  8. Automatic-repeat-request error control schemes

    NASA Technical Reports Server (NTRS)

    Lin, S.; Costello, D. J., Jr.; Miller, M. J.

    1983-01-01

    Error detection incorporated with automatic-repeat-request (ARQ) is widely used for error control in data communication systems. This method of error control is simple and provides high system reliability. If a properly chosen code is used for error detection, virtually error-free data transmission can be attained. Various types of ARQ and hybrid ARQ schemes, and error detection using linear block codes are surveyed.

  9. 20 CFR 30.515 - Is a recipient responsible for an overpayment that resulted from an error made by OWCP?

    Code of Federal Regulations, 2011 CFR

    2011-04-01

    ... that resulted from an error made by OWCP? 30.515 Section 30.515 Employees' Benefits OFFICE OF WORKERS... recipient responsible for an overpayment that resulted from an error made by OWCP? (a) The fact that OWCP... from liability for repayment if the recipient also was at fault in accepting the overpayment....

  10. Flight Test Results of an Angle of Attack and Angle of Sideslip Calibration Method Using Output-Error Optimization

    NASA Technical Reports Server (NTRS)

    Siu, Marie-Michele; Martos, Borja; Foster, John V.

    2013-01-01

    As part of a joint partnership between the NASA Aviation Safety Program (AvSP) and the University of Tennessee Space Institute (UTSI), research on advanced air data calibration methods has been in progress. This research was initiated to expand a novel pitot-static calibration method that was developed to allow rapid in-flight calibration for the NASA Airborne Subscale Transport Aircraft Research (AirSTAR) facility. This approach uses Global Positioning System (GPS) technology coupled with modern system identification methods that rapidly computes optimal pressure error models over a range of airspeed with defined confidence bounds. Subscale flight tests demonstrated small 2-s error bounds with significant reduction in test time compared to other methods. Recent UTSI full scale flight tests have shown airspeed calibrations with the same accuracy or better as the Federal Aviation Administration (FAA) accepted GPS 'four-leg' method in a smaller test area and in less time. The current research was motivated by the desire to extend this method for inflight calibration of angle of attack (AOA) and angle of sideslip (AOS) flow vanes. An instrumented Piper Saratoga research aircraft from the UTSI was used to collect the flight test data and evaluate flight test maneuvers. Results showed that the output-error approach produces good results for flow vane calibration. In addition, maneuvers for pitot-static and flow vane calibration can be integrated to enable simultaneous and efficient testing of each system.

  11. Having Fun with Error Analysis

    ERIC Educational Resources Information Center

    Siegel, Peter

    2007-01-01

    We present a fun activity that can be used to introduce students to error analysis: the M&M game. Students are told to estimate the number of individual candies plus uncertainty in a bag of M&M's. The winner is the group whose estimate brackets the actual number with the smallest uncertainty. The exercise produces enthusiastic discussions and…

  12. Multichannel error correction code decoder

    NASA Technical Reports Server (NTRS)

    Wagner, Paul K.; Ivancic, William D.

    1993-01-01

    A brief overview of a processing satellite for a mesh very-small-aperture (VSAT) communications network is provided. The multichannel error correction code (ECC) decoder system, the uplink signal generation and link simulation equipment, and the time-shared decoder are described. The testing is discussed. Applications of the time-shared decoder are recommended.

  13. Theory of Test Translation Error

    ERIC Educational Resources Information Center

    Solano-Flores, Guillermo; Backhoff, Eduardo; Contreras-Nino, Luis Angel

    2009-01-01

    In this article, we present a theory of test translation whose intent is to provide the conceptual foundation for effective, systematic work in the process of test translation and test translation review. According to the theory, translation error is multidimensional; it is not simply the consequence of defective translation but an inevitable fact…

  14. RM2: rms error comparisons

    NASA Technical Reports Server (NTRS)

    Rice, R. F.

    1976-01-01

    The root-mean-square error performance measure is used to compare the relative performance of several widely known source coding algorithms with the RM2 image data compression system. The results demonstrate that RM2 has a uniformly significant performance advantage.

  15. What Is a Reading Error?

    ERIC Educational Resources Information Center

    Labov, William; Baker, Bettina

    2010-01-01

    Early efforts to apply knowledge of dialect differences to reading stressed the importance of the distinction between differences in pronunciation and mistakes in reading. This study develops a method of estimating the probability that a given oral reading that deviates from the text is a true reading error by observing the semantic impact of the…

  16. Amplify Errors to Minimize Them

    ERIC Educational Resources Information Center

    Stewart, Maria Shine

    2009-01-01

    In this article, the author offers her experience of modeling mistakes and writing spontaneously in the computer classroom to get students' attention and elicit their editorial response. She describes how she taught her class about major sentence errors--comma splices, run-ons, and fragments--through her Sentence Meditation exercise, a rendition…

  17. Typical errors of ESP users

    NASA Astrophysics Data System (ADS)

    Eremina, Svetlana V.; Korneva, Anna A.

    2004-07-01

    The paper presents analysis of the errors made by ESP (English for specific purposes) users which have been considered as typical. They occur as a result of misuse of resources of English grammar and tend to resist. Their origin and places of occurrence have also been discussed.

  18. Cascade Error Projection Learning Algorithm

    NASA Technical Reports Server (NTRS)

    Duong, T. A.; Stubberud, A. R.; Daud, T.

    1995-01-01

    A detailed mathematical analysis is presented for a new learning algorithm termed cascade error projection (CEP) and a general learning frame work. This frame work can be used to obtain the cascade correlation learning algorithm by choosing a particular set of parameters.

  19. Input/output error analyzer

    NASA Technical Reports Server (NTRS)

    Vaughan, E. T.

    1977-01-01

    Program aids in equipment assessment. Independent assembly-language utility program is designed to operate under level 27 or 31 of EXEC 8 Operating System. It scans user-selected portions of system log file, whether located on tape or mass storage, and searches for and processes 1/0 error (type 6) entries.

  20. A brief history of error.

    PubMed

    Murray, Andrew W

    2011-10-01

    The spindle checkpoint monitors chromosome alignment on the mitotic and meiotic spindle. When the checkpoint detects errors, it arrests progress of the cell cycle while it attempts to correct the mistakes. This perspective will present a brief history summarizing what we know about the checkpoint, and a list of questions we must answer before we understand it. PMID:21968991

  1. Measurement error in geometric morphometrics.

    PubMed

    Fruciano, Carmelo

    2016-06-01

    Geometric morphometrics-a set of methods for the statistical analysis of shape once saluted as a revolutionary advancement in the analysis of morphology -is now mature and routinely used in ecology and evolution. However, a factor often disregarded in empirical studies is the presence and the extent of measurement error. This is potentially a very serious issue because random measurement error can inflate the amount of variance and, since many statistical analyses are based on the amount of "explained" relative to "residual" variance, can result in loss of statistical power. On the other hand, systematic bias can affect statistical analyses by biasing the results (i.e. variation due to bias is incorporated in the analysis and treated as biologically-meaningful variation). Here, I briefly review common sources of error in geometric morphometrics. I then review the most commonly used methods to measure and account for both random and non-random measurement error, providing a worked example using a real dataset. PMID:27038025

  2. Evaluating the Effect of Global Positioning System (GPS) Satellite Clock Error via GPS Simulation

    NASA Astrophysics Data System (ADS)

    Sathyamoorthy, Dinesh; Shafii, Shalini; Amin, Zainal Fitry M.; Jusoh, Asmariah; Zainun Ali, Siti

    2016-06-01

    This study is aimed at evaluating the effect of Global Positioning System (GPS) satellite clock error using GPS simulation. Two conditions of tests are used; Case 1: All the GPS satellites have clock errors within the normal range of 0 to 7 ns, corresponding to pseudorange error range of 0 to 2.1 m; Case 2: One GPS satellite suffers from critical failure, resulting in clock error in the pseudorange of up to 1 km. It is found that increase of GPS satellite clock error causes increase of average positional error due to increase of pseudorange error in the GPS satellite signals, which results in increasing error in the coordinates computed by the GPS receiver. Varying average positional error patterns are observed for the each of the readings. This is due to the GPS satellite constellation being dynamic, causing varying GPS satellite geometry over location and time, resulting in GPS accuracy being location / time dependent. For Case 1, in general, the highest average positional error values are observed for readings with the highest PDOP values, while the lowest average positional error values are observed for readings with the lowest PDOP values. For Case 2, no correlation is observed between the average positional error values and PDOP, indicating that the error generated is random.

  3. Sampling Errors in Satellite-derived Infrared Sea Surface Temperatures

    NASA Astrophysics Data System (ADS)

    Liu, Y.; Minnett, P. J.

    2014-12-01

    Sea Surface Temperature (SST) measured from satellites has been playing a crucial role in understanding geophysical phenomena. Generating SST Climate Data Records (CDRs) is considered to be the one that imposes the most stringent requirements on data accuracy. For infrared SSTs, sampling uncertainties caused by cloud presence and persistence generate errors. In addition, for sensors with narrow swaths, the swath gap will act as another sampling error source. This study is concerned with quantifying and understanding such sampling errors, which are important for SST CDR generation and for a wide range of satellite SST users. In order to quantify these errors, a reference Level 4 SST field (Multi-scale Ultra-high Resolution SST) is sampled by using realistic swath and cloud masks of Moderate Resolution Imaging Spectroradiometer (MODIS) and Advanced Along Track Scanning Radiometer (AATSR). Global and regional SST uncertainties are studied by assessing the sampling error at different temporal and spatial resolutions (7 spatial resolutions from 4 kilometers to 5.0° at the equator and 5 temporal resolutions from daily to monthly). Global annual and seasonal mean sampling errors are large in the high latitude regions, especially the Arctic, and have geographical distributions that are most likely related to stratus clouds occurrence and persistence. The region between 30°N and 30°S has smaller errors compared to higher latitudes, except for the Tropical Instability Wave area, where persistent negative errors are found. Important differences in sampling errors are also found between the broad and narrow swath scan patterns and between day and night fields. This is the first time that realistic magnitudes of the sampling errors are quantified. Future improvement in the accuracy of SST products will benefit from this quantification.

  4. Hybrid Models for Trajectory Error Modelling in Urban Environments

    NASA Astrophysics Data System (ADS)

    Angelatsa, E.; Parés, M. E.; Colomina, I.

    2016-06-01

    This paper tackles the first step of any strategy aiming to improve the trajectory of terrestrial mobile mapping systems in urban environments. We present an approach to model the error of terrestrial mobile mapping trajectories, combining deterministic and stochastic models. Due to urban specific environment, the deterministic component will be modelled with non-continuous functions composed by linear shifts, drifts or polynomial functions. In addition, we will introduce a stochastic error component for modelling residual noise of the trajectory error function. First step for error modelling requires to know the actual trajectory error values for several representative environments. In order to determine as accurately as possible the trajectories error, (almost) error less trajectories should be estimated using extracted nonsemantic features from a sequence of images collected with the terrestrial mobile mapping system and from a full set of ground control points. Once the references are estimated, they will be used to determine the actual errors in terrestrial mobile mapping trajectory. The rigorous analysis of these data sets will allow us to characterize the errors of a terrestrial mobile mapping system for a wide range of environments. This information will be of great use in future campaigns to improve the results of the 3D points cloud generation. The proposed approach has been evaluated using real data. The data originate from a mobile mapping campaign over an urban and controlled area of Dortmund (Germany), with harmful GNSS conditions. The mobile mapping system, that includes two laser scanner and two cameras, was mounted on a van and it was driven over a controlled area around three hours. The results show the suitability to decompose trajectory error with non-continuous deterministic and stochastic components.

  5. Acceptability of GM foods among Pakistani consumers.

    PubMed

    Ali, Akhter; Rahut, Dil Bahadur; Imtiaz, Muhammad

    2016-04-01

    In Pakistan majority of the consumers do not have information about genetically modified (GM) foods. In developing countries particularly in Pakistan few studies have focused on consumers' acceptability about GM foods. Using comprehensive primary dataset collected from 320 consumers in 2013 from Pakistan, this study analyzes the determinants of consumers' acceptability of GM foods. The data was analyzed by employing the bivariate probit model and censored least absolute deviation (CLAD) models. The empirical results indicated that urban consumers are more aware of GM foods compared to rural consumers. The acceptance of GM foods was more among females' consumers as compared to male consumers. In addition, the older consumers were more willing to accept GM food compared to young consumers. The acceptability of GM foods was also higher among wealthier households. Low price is the key factor leading to the acceptability of GM foods. The acceptability of the GM foods also reduces the risks among Pakistani consumers. PMID:27494790

  6. Prospective, multidisciplinary recording of perioperative errors in cerebrovascular surgery: is error in the eye of the beholder?

    PubMed

    Michalak, Suzanne M; Rolston, John D; Lawton, Michael T

    2016-06-01

    OBJECT Surgery requires careful coordination of multiple team members, each playing a vital role in mitigating errors. Previous studies have focused on eliciting errors from only the attending surgeon, likely missing events observed by other team members. METHODS Surveys were administered to the attending surgeon, resident surgeon, anesthesiologist, and nursing staff immediately following each of 31 cerebrovascular surgeries; participants were instructed to record any deviation from optimal course (DOC). DOCs were categorized and sorted by reporter and perioperative timing, then correlated with delays and outcome measures. RESULTS Errors were recorded in 93.5% of the 31 cases surveyed. The number of errors recorded per case ranged from 0 to 8, with an average of 3.1 ± 2.1 errors (± SD). Overall, technical errors were most common (24.5%), followed by communication (22.4%), management/judgment (16.0%), and equipment (11.7%). The resident surgeon reported the most errors (52.1%), followed by the circulating nurse (31.9%), the attending surgeon (26.6%), and the anesthesiologist (14.9%). The attending and resident surgeons were most likely to report technical errors (52% and 30.6%, respectively), while anesthesiologists and circulating nurses mostly reported anesthesia errors (36%) and communication errors (50%), respectively. The overlap in reported errors was 20.3%. If this study had used only the surveys completed by the attending surgeon, as in prior studies, 72% of equipment errors, 90% of anesthesia and communication errors, and 100% of nursing errors would have been missed. In addition, it would have been concluded that errors occurred in only 45.2% of cases (rather than 93.5%) and that errors resulting in a delay occurred in 3.2% of cases instead of the 74.2% calculated using data from 4 team members. Compiled results from all team members yielded significant correlations between technical DOCs and prolonged hospital stays and reported and actual delays (p = 0

  7. Toward a cognitive taxonomy of medical errors.

    PubMed Central

    Zhang, Jiajie; Patel, Vimla L.; Johnson, Todd R.; Shortliffe, Edward H.

    2002-01-01

    One critical step in addressing and resolving the problems associated with human errors is the development of a cognitive taxonomy of such errors. In the case of errors, such a taxonomy may be developed (1) to categorize all types of errors along cognitive dimensions, (2) to associate each type of error with a specific underlying cognitive mechanism, (3) to explain why, and even predict when and where, a specific error will occur, and (4) to generate intervention strategies for each type of error. Based on Reason's (1992) definition of human errors and Norman's (1986) cognitive theory of human action, we have developed a preliminary action-based cognitive taxonomy of errors that largely satisfies these four criteria in the domain of medicine. We discuss initial steps for applying this taxonomy to develop an online medical error reporting system that not only categorizes errors but also identifies problems and generates solutions. PMID:12463962

  8. Acceptance in Romantic Relationships: The Frequency and Acceptability of Partner Behavior Inventory

    ERIC Educational Resources Information Center

    Doss, Brian D.; Christensen, Andrew

    2006-01-01

    Despite the recent emphasis on acceptance in romantic relationships, no validated measure of relationship acceptance presently exists. To fill this gap, the 20-item Frequency and Acceptability of Partner Behavior Inventory (FAPBI; A. Christensen & N. S. Jacobson, 1997) was created to assess separately the acceptability and frequency of both…

  9. Least Squares Evaluations for Form and Profile Errors of Ellipse Using Coordinate Data

    NASA Astrophysics Data System (ADS)

    Liu, Fei; Xu, Guanghua; Liang, Lin; Zhang, Qing; Liu, Dan

    2016-04-01

    To improve the measurement and evaluation of form error of an elliptic section, an evaluation method based on least squares fitting is investigated to analyze the form and profile errors of an ellipse using coordinate data. Two error indicators for defining ellipticity are discussed, namely the form error and the profile error, and the difference between both is considered as the main parameter for evaluating machining quality of surface and profile. Because the form error and the profile error rely on different evaluation benchmarks, the major axis and the foci rather than the centre of an ellipse are used as the evaluation benchmarks and can accurately evaluate a tolerance range with the separated form error and profile error of workpiece. Additionally, an evaluation program based on the LS model is developed to extract the form error and the profile error of the elliptic section, which is well suited for separating the two errors by a standard program. Finally, the evaluation method about the form and profile errors of the ellipse is applied to the measurement of skirt line of the piston, and results indicate the effectiveness of the evaluation. This approach provides the new evaluation indicators for the measurement of form and profile errors of ellipse, which is found to have better accuracy and can thus be used to solve the difficult of the measurement and evaluation of the piston in industrial production.

  10. Aliasing errors in measurements of beam position and ellipticity

    NASA Astrophysics Data System (ADS)

    Ekdahl, Carl

    2005-09-01

    Beam position monitors (BPMs) are used in accelerators and ion experiments to measure currents, position, and azimuthal asymmetry. These usually consist of discrete arrays of electromagnetic field detectors, with detectors located at several equally spaced azimuthal positions at the beam tube wall. The discrete nature of these arrays introduces systematic errors into the data, independent of uncertainties resulting from signal noise, lack of recording dynamic range, etc. Computer simulations were used to understand and quantify these aliasing errors. If required, aliasing errors can be significantly reduced by employing more than the usual four detectors in the BPMs. These simulations show that the error in measurements of the centroid position of a large beam is indistinguishable from the error in the position of a filament. The simulations also show that aliasing errors in the measurement of beam ellipticity are very large unless the beam is accurately centered. The simulations were used to quantify the aliasing errors in beam parameter measurements during early experiments on the DARHT-II accelerator, demonstrating that they affected the measurements only slightly, if at all.

  11. Cloud retrieval using infrared sounder data - Error analysis

    NASA Technical Reports Server (NTRS)

    Wielicki, B. A.; Coakley, J. A., Jr.

    1981-01-01

    An error analysis is presented for cloud-top pressure and cloud-amount retrieval using infrared sounder data. Rms and bias errors are determined for instrument noise (typical of the HIRS-2 instrument on Tiros-N) and for uncertainties in the temperature profiles and water vapor profiles used to estimate clear-sky radiances. Errors are determined for a range of test cloud amounts (0.1-1.0) and cloud-top pressures (920-100 mb). Rms errors vary by an order of magnitude depending on the cloud height and cloud amount within the satellite's field of view. Large bias errors are found for low-altitude clouds. These bias errors are shown to result from physical constraints placed on retrieved cloud properties, i.e., cloud amounts between 0.0 and 1.0 and cloud-top pressures between the ground and tropopause levels. Middle-level and high-level clouds (above 3-4 km) are retrieved with low bias and rms errors.

  12. Errors and correction of precipitation measurements in China

    NASA Astrophysics Data System (ADS)

    Ren, Zhihua; Li, Mingqin

    2007-05-01

    In order to discover the range of various errors in Chinese precipitation measurements and seek a correction method, 30 precipitation evaluation stations were set up countrywide before 1993. All the stations are reference stations in China. To seek a correction method for wind-induced error, a precipitation correction instrument called the “horizontal precipitation gauge” was devised beforehand. Field intercomparison observations regarding 29,000 precipitation events have been conducted using one pit gauge, two elevated operational gauges and one horizontal gauge at the above 30 stations. The range of precipitation measurement errors in China is obtained by analysis of intercomparison measurement results. The distribution of random errors and systematic errors in precipitation measurements are studied in this paper. A correction method, especially for wind-induced errors, is developed. The results prove that a correlation of power function exists between the precipitation amount caught by the horizontal gauge and the absolute difference of observations implemented by the operational gauge and pit gauge. The correlation coefficient is 0.99. For operational observations, precipitation correction can be carried out only by parallel observation with a horizontal precipitation gauge. The precipitation accuracy after correction approaches that of the pit gauge. The correction method developed is simple and feasible.

  13. 2013 SYR Accepted Poster Abstracts.

    PubMed

    2013-01-01

    SYR 2013 Accepted Poster abstracts: 1. Benefits of Yoga as a Wellness Practice in a Veterans Affairs (VA) Health Care Setting: If You Build It, Will They Come? 2. Yoga-based Psychotherapy Group With Urban Youth Exposed to Trauma. 3. Embodied Health: The Effects of a Mind�Body Course for Medical Students. 4. Interoceptive Awareness and Vegetable Intake After a Yoga and Stress Management Intervention. 5. Yoga Reduces Performance Anxiety in Adolescent Musicians. 6. Designing and Implementing a Therapeutic Yoga Program for Older Women With Knee Osteoarthritis. 7. Yoga and Life Skills Eating Disorder Prevention Among 5th Grade Females: A Controlled Trial. 8. A Randomized, Controlled Trial Comparing the Impact of Yoga and Physical Education on the Emotional and Behavioral Functioning of Middle School Children. 9. Feasibility of a Multisite, Community based Randomized Study of Yoga and Wellness Education for Women With Breast Cancer Undergoing Chemotherapy. 10. A Delphi Study for the Development of Protocol Guidelines for Yoga Interventions in Mental Health. 11. Impact Investigation of Breathwalk Daily Practice: Canada�India Collaborative Study. 12. Yoga Improves Distress, Fatigue, and Insomnia in Older Veteran Cancer Survivors: Results of a Pilot Study. 13. Assessment of Kundalini Mantra and Meditation as an Adjunctive Treatment With Mental Health Consumers. 14. Kundalini Yoga Therapy Versus Cognitive Behavior Therapy for Generalized Anxiety Disorder and Co-Occurring Mood Disorder. 15. Baseline Differences in Women Versus Men Initiating Yoga Programs to Aid Smoking Cessation: Quitting in Balance Versus QuitStrong. 16. Pranayam Practice: Impact on Focus and Everyday Life of Work and Relationships. 17. Participation in a Tailored Yoga Program is Associated With Improved Physical Health in Persons With Arthritis. 18. Effects of Yoga on Blood Pressure: Systematic Review and Meta-analysis. 19. A Quasi-experimental Trial of a Yoga based Intervention to Reduce Stress and

  14. Space Saving Statistics: An Introduction to Constant Error, Variable Error, and Absolute Error.

    ERIC Educational Resources Information Center

    Guth, David

    1990-01-01

    Article discusses research on orientation and mobility (O&M) for individuals with visual impairments, examining constant, variable, and absolute error (descriptive statistics that quantify fundamentally different characteristics of distributions of spatially directed behavior). It illustrates the statistics with examples, noting their application…

  15. Discretization vs. Rounding Error in Euler's Method

    ERIC Educational Resources Information Center

    Borges, Carlos F.

    2011-01-01

    Euler's method for solving initial value problems is an excellent vehicle for observing the relationship between discretization error and rounding error in numerical computation. Reductions in stepsize, in order to decrease discretization error, necessarily increase the number of steps and so introduce additional rounding error. The problem is…

  16. Medical Errors: Tips to Help Prevent Them

    MedlinePlus

    ... to Web version Medical Errors: Tips to Help Prevent Them Medical Errors: Tips to Help Prevent Them Medical errors are one of the nation's ... single most important way you can help to prevent errors is to be an active member of ...

  17. Error-Related Psychophysiology and Negative Affect

    ERIC Educational Resources Information Center

    Hajcak, G.; McDonald, N.; Simons, R.F.

    2004-01-01

    The error-related negativity (ERN/Ne) and error positivity (Pe) have been associated with error detection and response monitoring. More recently, heart rate (HR) and skin conductance (SC) have also been shown to be sensitive to the internal detection of errors. An enhanced ERN has consistently been observed in anxious subjects and there is some…

  18. Bidirectional synchronization and hierarchical error correction for robust image transmission

    NASA Astrophysics Data System (ADS)

    Li, HongZhi; Chen, Chang W.

    1998-12-01

    In this paper, we present a novel joint source and channel image coding scheme for noisy channel transmission. The proposed scheme consists of two innovative components: (1) Intelligent bi-directional synchronization, and (2) Layered bit-plane error protection. The bi-directional synchronization is able to recover the coding synchronization when any single or even when two consecutive synchronization codes are corrupted by the channel noise. With synchronized partition, unequal error protection for each bit-plane can be designed to suit for a wide range of channel environments. The hierarchical error protection strategy is based on the analysis of bit-plane error sensitivity, aiming at achieving an optimal joint source and channel coding when the compressed image data are transmitted over noisy channels. Experimental results over extensive channel simulations show that the proposed scheme outperforms the approach proposed by Sherwood and Zeger who have reported the best numerical results in the literature.

  19. Sensitivity of SLR baselines to errors in Earth orientation

    NASA Technical Reports Server (NTRS)

    Smith, D. E.; Christodoulidis, D. C.

    1984-01-01

    The sensitivity of inter station distances derived from Satellite Laser Ranging (SLR) to errors in Earth orientation is discussed. An analysis experiment is performed which imposes a known polar motion error on all of the arcs used over this interval. The effect of the averaging of the errors over the tracking periods of individual sites is assessed. Baselines between stations that are supported by a global network of tracking stations are only marginally affected by errors in Earth orientation. The global network of stations retains its integrity even in the presence of systematic changes in the coordinate frame. The effect of these coordinate frame changes on the relative locations of the stations is minimal.

  20. Acceptance dependence of fluctuation measures near the QCD critical point

    NASA Astrophysics Data System (ADS)

    Ling, Bo; Stephanov, Mikhail A.

    2016-03-01

    We argue that a crucial determinant of the acceptance dependence of fluctuation measures in heavy-ion collisions is the range of correlations in the momentum space, e.g., in rapidity, Δ ycorr . The value of Δ ycorr˜1 for critical thermal fluctuations is determined by the thermal rapidity spread of the particles at freeze-out, and has little to do with position space correlations, even near the critical point where the spatial correlation length ξ becomes as large as 2-3 fm (this is in contrast to the magnitudes of the cumulants, which are sensitive to ξ ). When the acceptance window is large, Δ y ≫Δ ycorr , the cumulants of a given particle multiplicity, κk, scale linearly with Δ y , or mean multiplicity in acceptance, , and cumulant ratios are acceptance independent. In the opposite regime, Δ y ≪Δ ycorr , the factorial cumulants, κ̂k, scale as (Δy ) k, or k. We demonstrate this general behavior quantitatively in a model for critical point fluctuations, which also shows that the dependence on transverse momentum acceptance is very significant. We conclude that the extension of rapidity coverage as proposed by the STAR Collaboration should significantly increase the magnitude of the critical point fluctuation signatures.

  1. Comparative analysis of planetary laser ranging concepts

    NASA Astrophysics Data System (ADS)

    Dirkx, D.; Bauer, S.; Noomen, R.; Vermeersen, B. L. A.; Visser, P. N.

    2014-12-01

    Laser ranging is an emerging technology for tracking interplanetary missions, offering improved range accuracy and precision (mm-cm), compared to existing DSN tracking. The ground segment uses existing Satellite Laser Ranging (SLR) technology, whereas the space segment is modified with an active system. In a one-way system, such as that currently being used on the LRO spacecraft (Zuber et al., 2010), only an active detector is required on the spacecraft. For a two-way system, such as that tested by using the laser altimeter system on the MESSENGER spacecraft en route to Mercury (Smith et al., 2006), a laser transmitter system is additionally placed on the space segment, which will asynchronously fire laser pulses towards the ground stations. Although the one-way system requires less hardware, clock errors on both the space and ground segments will accumulate over time, polluting the range measurements. For a two-way system, the range measurements are only sensitive to clock errors integrated over the the two-way light time.We investigate the performance of both one- and two-way laser range systems by simulating their operation. We generate realizations of clock error time histories from Allan variance profiles, and use them to create range measurement error profiles. We subsequently perform the orbit determination process from this data to quanitfy the system's performance. For our simulations, we use two test cases: a lunar orbiter similar to LRO and a Phobos lander similar to the Phobos Laser Ranging concept (Turyshev et al., 2010). For the lunar orbiter, we include an empirical model for unmodelled non-gravitational accelerations in our truth model to include errors ihe dynamics. We include the estimation of clock parameters over a number of arc lengths for our simulations of the one-way range system and use a variety of state arc durations for the lunar orbiter simulations.We perform Monte Carlo simulations and generate true error distributions for both

  2. ERROR ANALYSIS OF COMPOSITE SHOCK INTERACTION PROBLEMS.

    SciTech Connect

    LEE,T.MU,Y.ZHAO,M.GLIMM,J.LI,X.YE,K.

    2004-07-26

    We propose statistical models of uncertainty and error in numerical solutions. To represent errors efficiently in shock physics simulations we propose a composition law. The law allows us to estimate errors in the solutions of composite problems in terms of the errors from simpler ones as discussed in a previous paper. In this paper, we conduct a detailed analysis of the errors. One of our goals is to understand the relative magnitude of the input uncertainty vs. the errors created within the numerical solution. In more detail, we wish to understand the contribution of each wave interaction to the errors observed at the end of the simulation.

  3. Study of an instrument for sensing errors in a telescope wavefront

    NASA Technical Reports Server (NTRS)

    Golden, L. J.; Shack, R. V.; Slater, D. N.

    1973-01-01

    Partial results are presented of theoretical and experimental investigations of different focal plane sensor configurations for determining the error in a telescope wavefront. The coarse range sensor and fine range sensors are used in the experimentation. The design of a wavefront error simulator is presented along with the Hartmann test, the shearing polarization interferometer, the Zernike test, and the Zernike polarization test.

  4. Radar ranging to Ganymede and Callisto

    NASA Technical Reports Server (NTRS)

    Harmon, J. K.; Ostro, S. J.; Chandler, J. F.; Hudson, R. S.

    1994-01-01

    Arecibo observations from 1992 February to March have yielded the first successful radar range measurements to the Galilean satellites. Round-up time delays were measured for Ganymede and Callisto with accuracies of 20 to 50 micrometer (3 to 7 km) and 90 micrometer (14 km), respectively. Both satellites showed round-trip delay residuals (relative to the E-3 ephemeris) of about a millisecond, most of which can be attributed to errors in the predicted along-track positions (orbital phases). Using a simple model that assumed that all of the ephemeris error was due to constant orbital phase and Jupiter range errors we estimate that Ganymede was leading its ephemeris by 122 +/- 4 km, Callisto was lagging its ephemeris by 307 +/- 14 km, and Jupiter was 11 +/- 4 km more distant than predicted by the PEP740 planetary ephemeris.

  5. Radar ranging to Ganymede and Callisto

    NASA Astrophysics Data System (ADS)

    Harmon, J. K.; Ostro, S. J.; Chandler, J. F.; Hudson, R. S.

    1994-03-01

    Arecibo observations from 1992 February to March have yielded the first successful radar range measurements to the Galilean satellites. Round-up time delays were measured for Ganymede and Callisto with accuracies of 20 to 50 micrometer (3 to 7 km) and 90 micrometer (14 km), respectively. Both satellites showed round-trip delay residuals (relative to the E-3 ephemeris) of about a millisecond, most of which can be attributed to errors in the predicted along-track positions (orbital phases). Using a simple model that assumed that all of the ephemeris error was due to constant orbital phase and Jupiter range errors we estimate that Ganymede was leading its ephemeris by 122 +/- 4 km, Callisto was lagging its ephemeris by 307 +/- 14 km, and Jupiter was 11 +/- 4 km more distant than predicted by the PEP740 planetary ephemeris.

  6. Report on errors in pretransfusion testing from a tertiary care center: A step toward transfusion safety

    PubMed Central

    Sidhu, Meena; Meenia, Renu; Akhter, Naveen; Sawhney, Vijay; Irm, Yasmeen

    2016-01-01

    Introduction: Errors in the process of pretransfusion testing for blood transfusion can occur at any stage from collection of the sample to administration of the blood component. The present study was conducted to analyze the errors that threaten patients’ transfusion safety and actual harm/serious adverse events that occurred to the patients due to these errors. Materials and Methods: The prospective study was conducted in the Department Of Transfusion Medicine, Shri Maharaja Gulab Singh Hospital, Government Medical College, Jammu, India from January 2014 to December 2014 for a period of 1 year. Errors were defined as any deviation from established policies and standard operating procedures. A near-miss event was defined as those errors, which did not reach the patient. Location and time of occurrence of the events/errors were also noted. Results: A total of 32,672 requisitions for the transfusion of blood and blood components were received for typing and cross-matching. Out of these, 26,683 products were issued to the various clinical departments. A total of 2,229 errors were detected over a period of 1 year. Near-miss events constituted 53% of the errors and actual harmful events due to errors occurred in 0.26% of the patients. Sample labeling errors were 2.4%, inappropriate request for blood components 2%, and information on requisition forms not matching with that on the sample 1.5% of all the requisitions received were the most frequent errors in clinical services. In transfusion services, the most common event was accepting sample in error with the frequency of 0.5% of all requisitions. ABO incompatible hemolytic reactions were the most frequent harmful event with the frequency of 2.2/10,000 transfusions. Conclusion: Sample labeling, inappropriate request, and sample received in error were the most frequent high-risk errors. PMID:27011670

  7. Entropic error-disturbance relations

    NASA Astrophysics Data System (ADS)

    Coles, Patrick; Furrer, Fabian

    2014-03-01

    We derive an entropic error-disturbance relation for a sequential measurement scenario as originally considered by Heisenberg, and we discuss how our relation could be tested using existing experimental setups. Our relation is valid for discrete observables, such as spin, as well as continuous observables, such as position and momentum. The novel aspect of our relation compared to earlier versions is its clear operational interpretation and the quantification of error and disturbance using entropic quantities. This directly relates the measurement uncertainty, a fundamental property of quantum mechanics, to information theoretical limitations and offers potential applications in for instance quantum cryptography. PC is funded by National Research Foundation Singapore and Ministry of Education Tier 3 Grant ``Random numbers from quantum processes'' (MOE2012-T3-1-009). FF is funded by Japan Society for the Promotion of Science, KAKENHI grant No. 24-02793.

  8. Robot learning and error correction

    NASA Technical Reports Server (NTRS)

    Friedman, L.

    1977-01-01

    A model of robot learning is described that associates previously unknown perceptions with the sensed known consequences of robot actions. For these actions, both the categories of outcomes and the corresponding sensory patterns are incorporated in a knowledge base by the system designer. Thus the robot is able to predict the outcome of an action and compare the expectation with the experience. New knowledge about what to expect in the world may then be incorporated by the robot in a pre-existing structure whether it detects accordance or discrepancy between a predicted consequence and experience. Errors committed during plan execution are detected by the same type of comparison process and learning may be applied to avoiding the errors.

  9. Negligence, genuine error, and litigation.

    PubMed

    Sohn, David H

    2013-01-01

    Not all medical injuries are the result of negligence. In fact, most medical injuries are the result either of the inherent risk in the practice of medicine, or due to system errors, which cannot be prevented simply through fear of disciplinary action. This paper will discuss the differences between adverse events, negligence, and system errors; the current medical malpractice tort system in the United States; and review current and future solutions, including medical malpractice reform, alternative dispute resolution, health courts, and no-fault compensation systems. The current political environment favors investigation of non-cap tort reform remedies; investment into more rational oversight systems, such as health courts or no-fault systems may reap both quantitative and qualitative benefits for a less costly and safer health system. PMID:23426783

  10. Human error in aviation operations

    NASA Technical Reports Server (NTRS)

    Billings, C. E.; Lanber, J. K.; Cooper, G. E.

    1974-01-01

    This report is a brief description of research being undertaken by the National Aeronautics and Space Administration. The project is designed to seek out factors in the aviation system which contribute to human error, and to search for ways of minimizing the potential threat posed by these factors. The philosophy and assumptions underlying the study are discussed, together with an outline of the research plan.

  11. Clinical review: Medication errors in critical care

    PubMed Central

    Moyen, Eric; Camiré, Eric; Stelfox, Henry Thomas

    2008-01-01

    Medication errors in critical care are frequent, serious, and predictable. Critically ill patients are prescribed twice as many medications as patients outside of the intensive care unit (ICU) and nearly all will suffer a potentially life-threatening error at some point during their stay. The aim of this article is to provide a basic review of medication errors in the ICU, identify risk factors for medication errors, and suggest strategies to prevent errors and manage their consequences. PMID:18373883

  12. Tilt error in cryospheric surface radiation measurements at high latitudes: a model study

    NASA Astrophysics Data System (ADS)

    Bogren, W. S.; Burkhart, J. F.; Kylling, A.

    2015-08-01

    We have evaluated the magnitude and makeup of error in cryospheric radiation observations due to small sensor misalignment in in-situ measurements of solar irradiance. This error is examined through simulation of diffuse and direct irradiance arriving at a detector with a cosine-response foreoptic. Emphasis is placed on assessing total error over the solar shortwave spectrum from 250 to 4500 nm, as well as supporting investigation over other relevant shortwave spectral ranges. The total measurement error introduced by sensor tilt is dominated by the direct component. For a typical high latitude albedo measurement with a solar zenith angle of 60°, a sensor tilted by 1, 3, and 5° can respectively introduce up to 2.6, 7.7, and 12.8 % error into the measured irradiance and similar errors in the derived albedo. Depending on the daily range of solar azimuth and zenith angles, significant measurement error can persist also in integrated daily irradiance and albedo.

  13. A Posteriori Correction of Forecast and Observation Error Variances

    NASA Technical Reports Server (NTRS)

    Rukhovets, Leonid

    2005-01-01

    Proposed method of total observation and forecast error variance correction is based on the assumption about normal distribution of "observed-minus-forecast" residuals (O-F), where O is an observed value and F is usually a short-term model forecast. This assumption can be accepted for several types of observations (except humidity) which are not grossly in error. Degree of nearness to normal distribution can be estimated by the symmetry or skewness (luck of symmetry) a(sub 3) = mu(sub 3)/sigma(sup 3) and kurtosis a(sub 4) = mu(sub 4)/sigma(sup 4) - 3 Here mu(sub i) = i-order moment, sigma is a standard deviation. It is well known that for normal distribution a(sub 3) = a(sub 4) = 0.

  14. Error detection and correction in an optoelectronic memory system

    NASA Astrophysics Data System (ADS)

    Hofmann, Robert; Pandey, Madhulima; Levitan, Steven P.; Chiarulli, Donald M.

    1998-11-01

    This paper describes the implementation of error detection and correction logic in the optoelectronic cache memory prototype at the University of Pittsburgh. In this project, our goal is to integrate a 3-D optical memory directly into the memory hierarchy of a personal computer. As with any optical storage system, error correction is essential to maintaining acceptable system performance. We have implemented a fully pipelined, real time decoder for 60-bit Spectral Reed-Solomon code words. The decoder is implemented in reconfigurable logic, using a single Xilinx 4000-series FPGA per code word and is fully scalable using multiple FPGA's. The current implementation operates at 33 Mhz, and processes two code words in parallel per clock cycle for an aggregate data rate of 4 Gb/s. We present a brief overview of the project and of Spectral Reed-Solomon codes followed by a description of our implementation and performance data.

  15. Calculation of magnetic error fields in hybrid insertion devices

    SciTech Connect

    Savoy, R.; Halbach, K.; Hassenzahl, W.; Hoyer, E.; Humphries, D.; Kincaid, B.

    1989-08-01

    The Advanced Light Source (ALS) at the Lawrence Berkeley Laboratory requires insertion devices with fields sufficiently accurate to take advantage of the small emittance of the ALS electron beam. To maintain the spectral performance of the synchrotron radiation and to limit steering effects on the electron beam these errors must be smaller than 0.25%. This paper develops a procedure for calculating the steering error due to misalignment of the easy axis of the permanent magnet material. The procedure is based on a three dimensional theory of the design of hybrid insertion devices developed by one of us. The acceptable tolerance for easy axis misalignment is found for a 5 cm period undulator proposed for the ALS. 11 refs., 5 figs.

  16. Systematic Errors of the Fsu Global Spectral Model

    NASA Astrophysics Data System (ADS)

    Surgi, Naomi

    Three 20 day winter forecasts have been carried out using the Florida State University Global Spectral Model to examine the systematic errors of the model. Most GCM's and global forecast models exhibit the same kind of error patterns even though the model formulations vary somewhat between them. Some of the dominant errors are a breakdown of the trade winds in the low latitudes, an over-prediction of the subtropical jets accompanied by an upward and poleward shift of the jets, an error in the mean sea-level pressure with over-intensification of the quasi-stationary oceanic lows and continental highs and a warming of the tropical mid and upper troposphere. In this study, a number of sensitivity experiments have been performed for which orography, model physics and initialization are considered as possible causes of these errors. A parameterization of the vertical distribution of momentum due to the sub-grid scale orography has been implemented in the model to address the model deficiencies associated with orographic forcing. This scheme incorporates the effects of moisture on the wave induced stress. The parameterization of gravity wave drag is shown to substantially reduce the large-scale wind and height errors in regions of direct forcing and well downstream of the mountainous regions. Also, a parameterization of the heat and moisture transport associated with shallow convection is found to have a positive impact on the errors particularly in the tropics. This is accomplished by the increase of moisture supply from the subtropics into the deep tropics and a subsequent enhancement of the secondary circulations. A dynamic relaxation was carried out to examine the impact of the long wave errors on the shorter wave. By constraining the long wave error, improvement is shown for wavenumbers 5-7 on medium to extended range time intervals. Thus, improved predictability of the transient flow is expected by applying this initialization procedure.

  17. Hybrid Projectile Body Angle Estimation for Selectable Range Increase

    NASA Astrophysics Data System (ADS)

    Gioia, Christopher J.

    A Hybrid Projectile (HP) is a tube launched munition that transforms into a gliding UAV, and is currently being researched at West Virginia University. A simple launch timer was first envisioned to control the transformation point in order to achieve maximum distance. However, this timer would need to be reprogrammed for any distance less than maximum range due to the nominal time to deployment varying with launch angle. A method was sought for automatic wing deployment that would not require reprogramming the round. A body angle estimation system was used to estimate the pitch of the HP relative to the Earth to determine when the HP is properly oriented for the designed glide slope angle. It was also necessary to filter out noise from a simulated inertial measurement unit (IMU), GPS receiver, and magnetometer. An Extended Kalman Filter (EKF) was chosen to estimate the Euler angles, position and velocity of the HP while an algorithm determined when to deploy the wings. A parametric study was done to verify the optimum deployment condition using a Simulink aerodynamic model. Because range is directly related to launch angle, various launch angles were simulated in the model. By fixing the glide slope angle to -10° as a deployment condition for all launch angles, the range differed only by a maximum of 6.1% from the maximum possible range. Based on these findings, the body angle deployment condition provides the most flexible option to maintain maximum distance without the need of reprogramming. Position and velocity estimates were also determined from the EKF using the GPS measurements. Simulations showed that the EKF estimates exhibited low root mean squared error values, corresponding to less than 3% of the total position values. Because the HP was in flight for less than a minute in this experiment, the drift encountered was acceptable.

  18. Error control in the GCF: An information-theoretic model for error analysis and coding

    NASA Technical Reports Server (NTRS)

    Adeyemi, O.

    1974-01-01

    The structure of data-transmission errors within the Ground Communications Facility is analyzed in order to provide error control (both forward error correction and feedback retransmission) for improved communication. Emphasis is placed on constructing a theoretical model of errors and obtaining from it all the relevant statistics for error control. No specific coding strategy is analyzed, but references to the significance of certain error pattern distributions, as predicted by the model, to error correction are made.

  19. Modelling and characterization of thin film planar capacitors: inherent errors and limits of applicability of partial capacitance methods

    NASA Astrophysics Data System (ADS)

    Vukadinovic, M.; Malic, B.; Kosec, M.; Krizaj, D.

    2009-11-01

    The analytical partial capacitance methods (PCM) widely accepted for calculation of properties of capacitors with planar electrodes, coplanar strip waveguides (CPS) and coplanar waveguides (CPW) are reviewed based on the challenges met during the development and tailoring of (Ba, Sr)TiO3 thin films fabricated on different types of substrates. An alternative view on the conformal mapping method is given, a correction for electrodes of finite thickness is introduced and the applicability of easy-to-use simplified analytical equations is reviewed and extended. Calculation results obtained with different models are compared against the results of FEM numerical simulation in the parameter range relevant for development of tuneable ferroelectric films. The models were found to be less accurate than previously reported. Accordingly, guidelines for extending the applicability domain and for minimizing the inherent errors of the analytical models are presented. The discussion is focused on the possible improvements of the measurement techniques.

  20. Uncertainty in measurement and total error - are they so incompatible?

    PubMed

    Farrance, Ian; Badrick, Tony; Sikaris, Kenneth A

    2016-08-01

    There appears to be a growing debate with regard to the use of "Westgard style" total error and "GUM style" uncertainty in measurement. Some may argue that the two approaches are irreconcilable. The recent appearance of an article "Quality goals at the crossroads: growing, going, or gone" on the well-regarded Westgard Internet site requires some comment. In particular, a number of assertions which relate to ISO 15189 and uncertainty in measurement appear misleading. An alternate view of the key issues raised by Westergard may serve to guide and enlighten others who may accept such statements at face value. PMID:27227711

  1. An investigation of error correcting techniques for OMV data

    NASA Technical Reports Server (NTRS)

    Ingels, Frank; Fryer, John

    1992-01-01

    Papers on the following topics are presented: considerations of testing the Orbital Maneuvering Vehicle (OMV) system with CLASS; OMV CLASS test results (first go around); equivalent system gain available from R-S encoding versus a desire to lower the power amplifier from 25 watts to 20 watts for OMV; command word acceptance/rejection rates for OMV; a memo concerning energy-to-noise ratio for the Viterbi-BSC Channel and the impact of Manchester coding loss; and an investigation of error correcting techniques for OMV and Advanced X-ray Astrophysics Facility (AXAF).

  2. 7 CFR 932.32 - Acceptance.

    Code of Federal Regulations, 2013 CFR

    2013-01-01

    ... AND ORDERS; FRUITS, VEGETABLES, NUTS), DEPARTMENT OF AGRICULTURE OLIVES GROWN IN CALIFORNIA Order Regulating Handling Olive Administrative Committee § 932.32 Acceptance. Any person selected by the...

  3. 7 CFR 932.32 - Acceptance.

    Code of Federal Regulations, 2014 CFR

    2014-01-01

    ... AND ORDERS; FRUITS, VEGETABLES, NUTS), DEPARTMENT OF AGRICULTURE OLIVES GROWN IN CALIFORNIA Order Regulating Handling Olive Administrative Committee § 932.32 Acceptance. Any person selected by the...

  4. 7 CFR 932.32 - Acceptance.

    Code of Federal Regulations, 2012 CFR

    2012-01-01

    ... and Orders; Fruits, Vegetables, Nuts), DEPARTMENT OF AGRICULTURE OLIVES GROWN IN CALIFORNIA Order Regulating Handling Olive Administrative Committee § 932.32 Acceptance. Any person selected by the...

  5. 7 CFR 1207.323 - Acceptance.

    Code of Federal Regulations, 2011 CFR

    2011-01-01

    ... AND ORDERS; MISCELLANEOUS COMMODITIES), DEPARTMENT OF AGRICULTURE POTATO RESEARCH AND PROMOTION PLAN Potato Research and Promotion Plan National Potato Promotion Board § 1207.323 Acceptance. Each...

  6. 7 CFR 1207.323 - Acceptance.

    Code of Federal Regulations, 2012 CFR

    2012-01-01

    ... AND ORDERS; MISCELLANEOUS COMMODITIES), DEPARTMENT OF AGRICULTURE POTATO RESEARCH AND PROMOTION PLAN Potato Research and Promotion Plan National Potato Promotion Board § 1207.323 Acceptance. Each...

  7. 7 CFR 1207.323 - Acceptance.

    Code of Federal Regulations, 2014 CFR

    2014-01-01

    ... AND ORDERS; MISCELLANEOUS COMMODITIES), DEPARTMENT OF AGRICULTURE POTATO RESEARCH AND PROMOTION PLAN Potato Research and Promotion Plan National Potato Promotion Board § 1207.323 Acceptance. Each...

  8. 7 CFR 1207.323 - Acceptance.

    Code of Federal Regulations, 2010 CFR

    2010-01-01

    ... AND ORDERS; MISCELLANEOUS COMMODITIES), DEPARTMENT OF AGRICULTURE POTATO RESEARCH AND PROMOTION PLAN Potato Research and Promotion Plan National Potato Promotion Board § 1207.323 Acceptance. Each...

  9. 7 CFR 1207.323 - Acceptance.

    Code of Federal Regulations, 2013 CFR

    2013-01-01

    ... AND ORDERS; MISCELLANEOUS COMMODITIES), DEPARTMENT OF AGRICULTURE POTATO RESEARCH AND PROMOTION PLAN Potato Research and Promotion Plan National Potato Promotion Board § 1207.323 Acceptance. Each...

  10. Writing biomedical manuscripts part II: standard elements and common errors.

    PubMed

    Ohwovoriole, A E

    2011-01-01

    It is incumbent on, satisfying to, and rewarding for, researchers to have their work published. Many workers are denied this satisfaction because of their inability to secure acceptance after what they consider a good research. Several reasons account for rejection or delay of manuscripts submitted to biomedical journals. A research poorly conceptualised and/or conducted will fail to fly but poor writing up of the completed work accounts for a greater majority of manuscripts that get rejected. The chances of manuscript acceptance can be increased by paying attention to the standard elements and avoiding correcting the common errors that make for the rejection of manuscripts. Cultivating the habit of structuring every department of the manuscript greatly improves chances of acceptance. The final paper should follow the universally accepted pattern of aim , introduction , methods, results, and discussion. The sequence of putting the paper together is different from the order in the final form. Follow a pattern that starts with the Tables and figures for the results section , followed by final version of the methods section. The title and abstract should be about the last to be written in the final version of the manuscript. You need to have results sorted out early as the rest of what you will write is largely dictated by your results. Revise the work several times and get co - authors and third parties to help read it over. To succeed follow the universal rules of writing and those of the target journal rules while avoiding those errors that are easily amenable to correction before you submit your manuscript. PMID:22786852

  11. Resources and Long-Range Forecasts

    ERIC Educational Resources Information Center

    Smith, Waldo E.

    1973-01-01

    The author argues that forecasts of quick depletion of resources in the environment as a result of overpopulation and increased usage may not be free from error. Ignorance still exists in understanding the recovery mechanisms of nature. Long-range forecasts are likely to be wrong in such situations. (PS)

  12. In acceptance we trust? Conceptualising acceptance as a viable approach to NGO security management.

    PubMed

    Fast, Larissa A; Freeman, C Faith; O'Neill, Michael; Rowley, Elizabeth

    2013-04-01

    This paper documents current understanding of acceptance as a security management approach and explores issues and challenges non-governmental organisations (NGOs) confront when implementing an acceptance approach to security management. It argues that the failure of organisations to systematise and clearly articulate acceptance as a distinct security management approach and a lack of organisational policies and procedures concerning acceptance hinder its efficacy as a security management approach. The paper identifies key and cross-cutting components of acceptance that are critical to its effective implementation in order to advance a comprehensive and systematic concept of acceptance. The key components of acceptance illustrate how organisational and staff functions affect positively or negatively an organisation's acceptance, and include: an organisation's principles and mission, communications, negotiation, programming, relationships and networks, stakeholder and context analysis, staffing, and image. The paper contends that acceptance is linked not only to good programming, but also to overall organisational management and structures. PMID:23278470

  13. Orbit determination by range-only data.

    NASA Technical Reports Server (NTRS)

    Duong, N.; Winn, C. B.

    1973-01-01

    The determination of satellite orbits for use in geodesy using range-only data has been examined. A recently developed recursive algorithm for rectification of the nominal orbit after processing each observation has been tested. It is shown that when a synchronous satellite is tracked simultaneously with a subsynchronous geodetic target satellite, the orbits of each may be readily determined by processing the range information. Random data errors and satellite perturbations are included in the examples presented.

  14. Public acceptance of urban rotorcraft operations

    NASA Astrophysics Data System (ADS)

    Marcolini, Michael A.; Powell, Clemans A.; Posey, Joe W.

    2001-05-01

    Even though tiltrotor operations from city center to city center could greatly shorten travel times over moderate ranges, public opposition to intense urban rotorcraft activity has kept this possibility from being realized. One significant factor in this opposition is rotorcraft noise. Over the last 25 years, NASA has explored the subjective response to rotorcraft noise and developed low noise design concepts and noise abatement flight procedures. While low noise designs can be applied for future rotorcraft, this is not an effective near-term means of reducing rotorcraft noise, because of the costs associated with replacement of helicopter rotor blades. Recent noise abatement research, which has been focusing on the development of tools and techniques to facilitate the design of quieter flight procedures for existing vehicles, has much more immediate application. While very little subjective response work has occurred recently, prior work at NASA in this area from the 1970s and 1980s is discussed. Lastly, thoughts on future research areas that might help improve the public acceptance of rotorcraft will be described.

  15. Quantum error correction of photon-scattering errors

    NASA Astrophysics Data System (ADS)

    Akerman, Nitzan; Glickman, Yinnon; Kotler, Shlomi; Ozeri, Roee

    2011-05-01

    Photon scattering by an atomic ground-state superposition is often considered as a source of decoherence. The same process also results in atom-photon entanglement which had been directly observed in various experiments using single atom, ion or a diamond nitrogen-vacancy center. Here we combine these two aspects to implement a quantum error correction protocol. We encode a qubit in the two Zeeman-splitted ground states of a single trapped 88 Sr+ ion. Photons are resonantly scattered on the S1 / 2 -->P1 / 2 transition. We study the process of single photon scattering i.e. the excitation of the ion to the excited manifold followed by a spontaneous emission and decay. In the absence of any knowledge on the emitted photon, the ion-qubit coherence is lost. However the joined ion-photon system still maintains coherence. We show that while scattering events where spin population is preserved (Rayleigh scattering) do not affect coherence, spin-changing (Raman) scattering events result in coherent amplitude exchange between the two qubit states. By applying a unitary spin rotation that is dependent on the detected photon polarization we retrieve the ion-qubit initial state. We characterize this quantum error correction protocol by process tomography and demonstrate an ability to preserve ion-qubit coherence with high fidelity.

  16. Performance analysis of a hybrid ARQ error control scheme for near earth satellite communications

    NASA Technical Reports Server (NTRS)

    Lin, Shu

    1987-01-01

    A robust error control coding scheme is presented. The scheme is a cascaded forward error correction (FEC) scheme supported by parity retransmissions for further error correction in the erroneous data words. The error performance and throughput efficiency of the scheme are analyzed. Two specific schemes are proposed for NASA near earth satellite communications. It is shown that both schemes provide high reliability and throughput efficiency even for high channel bit error rates in the range of .002. The schemes are suitable for high data rate file transfer.

  17. Error analysis of friction drive elements

    NASA Astrophysics Data System (ADS)

    Wang, Guomin; Yang, Shihai; Wang, Daxing

    2008-07-01

    Friction drive is used in some large astronomical telescopes in recent years. Comparing to the direct drive, friction drive train consists of more buildup parts. Usually, the friction drive train consists of motor-tachometer unit, coupling, reducer, driving roller, big wheel, encoder and encoder coupling. Normally, these buildup parts will introduce somewhat errors to the drive system. Some of them are random error and some of them are systematic error. For the random error, the effective way is to estimate their contributions and try to find proper way to decrease its influence. For the systematic error, the useful way is to analyse and test them quantitively, and then feedback the error to the control system to correct them. The main task of this paper is to analyse these error sources and find out their characteristics, such as random error, systematic error and contributions. The methods or equations used in the analysis will be also presented detail in this paper.

  18. Ranging/tracking system for proximity operations

    NASA Technical Reports Server (NTRS)

    Nilsen, P.; Udalov, S.

    1982-01-01

    The hardware development and testing phase of a hand held radar for the ranging and tracking for Shuttle proximity operations are considered. The radar is to measure range to a 3 sigma accuracy of 1 m (3.28 ft) to a maximum range of 1850 m (6000 ft) and velocity to a 3 sigma accuracy of 0.03 m/s (0.1 ft/s). Size and weight are similar to hand held radars, frequently seen in use by motorcycle police officers. Meeting these goals for a target in free space was very difficult to obtain in the testing program; however, at a range of approximately 700 m, the 3 sigma range error was found to be 0.96 m. It is felt that much of this error is due to clutter in the test environment. As an example of the velocity accuracy, at a range of 450 m, a 3 sigma velocity error of 0.02 m/s was measured. The principles of the radar and recommended changes to its design are given. Analyses performed in support of the design process, the actual circuit diagrams, and the software listing are included.

  19. Is preputioplasty effective and acceptable?

    PubMed

    Barber, N J; Chappell, B; Carter, P G; Britton, J P

    2003-09-01

    Foreskin complaints in childhood, if not manageable conservatively, are usually treated by circumcision. A less radical surgical option, when balanitis xerotica obliterans is absent, is preputioplasty. We sent questionnaires to the parents of 23 boys who had had this procedure and 22 replied. Mean interval since operation was 20 months (range 3-36). The main indications for surgery had been irretractable foreskin in 9, recurrent balanoposthitis in 10 and ballooning on voiding in 3 and the operation had dealt successfully with these in 7, 7, and 3, respectively. In all but one case the parents were satisfied with the cosmetic result. However, in 8 cases (36%) the parents said they would have preferred circumcision and 3 of the boys had been listed for further surgery. Preputioplasty is a satisfactory alternative to circumcision in selected cases. PMID:12949202

  20. Effects of Listening Conditions, Error Types, and Ensemble Textures on Error Detection Skills

    ERIC Educational Resources Information Center

    Waggoner, Dori T.

    2011-01-01

    This study was designed with three main purposes: (a) to investigate the effects of two listening conditions on error detection accuracy, (b) to compare error detection responses for rhythm errors and pitch errors, and (c) to examine the influences of texture on error detection accuracy. Undergraduate music education students (N = 18) listened to…

  1. Processing In A GPS Receiver To Reduce Multipath Errors

    NASA Technical Reports Server (NTRS)

    Meehan, Thomas K.

    1994-01-01

    Four techniques of ancillary real-time digital processing of signals in Global Positioning System, GPS, receiver introduced reducing effects of multipath propagation of signals on position estimates produced by receiver. Multipath range errors halved. Applied in addition to other signal-processing techniques and to other techniques designing as receiving antenna to make it insensitive to reflections of GPS signals from nearby objects.

  2. Error Gravity: Perceptions of Native-Speaking and Non-Native Speaking Faculty in EFL.

    ERIC Educational Resources Information Center

    Kresovich, Brant M.

    1988-01-01

    A survey of teachers of composition in English as a Second Language in Japan addressed the perceptions of native-English-speaking and non-native-English-speaking teachers of the acceptability of specific error types within sentences. The native speakers of English were one British and 16 Americans. The non-native group was comprised of 26 Japanese…

  3. 48 CFR 3011.103 - Market acceptance.

    Code of Federal Regulations, 2011 CFR

    2011-10-01

    ... 48 Federal Acquisition Regulations System 7 2011-10-01 2011-10-01 false Market acceptance. 3011.103 Section 3011.103 Federal Acquisition Regulations System DEPARTMENT OF HOMELAND SECURITY, HOMELAND... Developing Requirements Documents 3011.103 Market acceptance. (a) Contracting officers may act on behalf...

  4. 7 CFR 930.26 - Acceptance.

    Code of Federal Regulations, 2010 CFR

    2010-01-01

    ... 7 Agriculture 8 2010-01-01 2010-01-01 false Acceptance. 930.26 Section 930.26 Agriculture Regulations of the Department of Agriculture (Continued) AGRICULTURAL MARKETING SERVICE (Marketing Agreements... Administrative Body § 930.26 Acceptance. Each person to be appointed by the Secretary as a member or as...

  5. 48 CFR 2811.103 - Market acceptance.

    Code of Federal Regulations, 2011 CFR

    2011-10-01

    ... 48 Federal Acquisition Regulations System 6 2011-10-01 2011-10-01 false Market acceptance. 2811.103 Section 2811.103 Federal Acquisition Regulations System DEPARTMENT OF JUSTICE COMPETITION AND... acceptance. Pursuant to FAR 11.103, the HCA or designee at a level not lower than the BPC has the...

  6. 48 CFR 1311.103 - Market acceptance.

    Code of Federal Regulations, 2011 CFR

    2011-10-01

    ... 48 Federal Acquisition Regulations System 5 2011-10-01 2011-10-01 false Market acceptance. 1311.103 Section 1311.103 Federal Acquisition Regulations System DEPARTMENT OF COMMERCE COMPETITION AND... acceptance. The designee authorized as the head of the agency is set forth in CAM 1301.70....

  7. 36 CFR 251.62 - Acceptance.

    Code of Federal Regulations, 2011 CFR

    2011-07-01

    ... 36 Parks, Forests, and Public Property 2 2011-07-01 2011-07-01 false Acceptance. 251.62 Section 251.62 Parks, Forests, and Public Property FOREST SERVICE, DEPARTMENT OF AGRICULTURE LAND USES Special Uses § 251.62 Acceptance. Except for an easement, a special use authorization shall become...

  8. 48 CFR 3011.103 - Market acceptance.

    Code of Federal Regulations, 2010 CFR

    2010-10-01

    ... 48 Federal Acquisition Regulations System 7 2010-10-01 2010-10-01 false Market acceptance. 3011.103 Section 3011.103 Federal Acquisition Regulations System DEPARTMENT OF HOMELAND SECURITY, HOMELAND... Developing Requirements Documents 3011.103 Market acceptance. (a) Contracting officers may act on behalf...

  9. 48 CFR 1011.103 - Market Acceptance.

    Code of Federal Regulations, 2010 CFR

    2010-10-01

    ... 48 Federal Acquisition Regulations System 5 2010-10-01 2010-10-01 false Market Acceptance. 1011.103 Section 1011.103 Federal Acquisition Regulations System DEPARTMENT OF THE TREASURY COMPETITION AND... Acceptance. (a) BCPOs can act on behalf of the head of the agency in this subpart only. BCPOs,...

  10. 7 CFR 916.25 - Acceptance.

    Code of Federal Regulations, 2010 CFR

    2010-01-01

    ... 7 Agriculture 8 2010-01-01 2010-01-01 false Acceptance. 916.25 Section 916.25 Agriculture Regulations of the Department of Agriculture (Continued) AGRICULTURAL MARKETING SERVICE (Marketing Agreements... Regulating Handling Administrative Body § 916.25 Acceptance. Each person to be selected by the Secretary as...

  11. 43 CFR 3452.1-3 - Acceptance.

    Code of Federal Regulations, 2011 CFR

    2011-10-01

    ... 43 Public Lands: Interior 2 2011-10-01 2011-10-01 false Acceptance. 3452.1-3 Section 3452.1-3 Public Lands: Interior Regulations Relating to Public Lands (Continued) BUREAU OF LAND MANAGEMENT..., Cancellation, and Termination § 3452.1-3 Acceptance. The effective date of the lease relinquishment shall,...

  12. 48 CFR 1311.103 - Market acceptance.

    Code of Federal Regulations, 2010 CFR

    2010-10-01

    ... 48 Federal Acquisition Regulations System 5 2010-10-01 2010-10-01 false Market acceptance. 1311.103 Section 1311.103 Federal Acquisition Regulations System DEPARTMENT OF COMMERCE COMPETITION AND... acceptance. The designee authorized as the head of the agency is set forth in CAM 1301.70....

  13. 7 CFR 916.25 - Acceptance.

    Code of Federal Regulations, 2011 CFR

    2011-01-01

    ... 7 Agriculture 8 2011-01-01 2011-01-01 false Acceptance. 916.25 Section 916.25 Agriculture Regulations of the Department of Agriculture (Continued) AGRICULTURAL MARKETING SERVICE (Marketing Agreements... Regulating Handling Administrative Body § 916.25 Acceptance. Each person to be selected by the Secretary as...

  14. 7 CFR 920.25 - Acceptance.

    Code of Federal Regulations, 2010 CFR

    2010-01-01

    ... 7 Agriculture 8 2010-01-01 2010-01-01 false Acceptance. 920.25 Section 920.25 Agriculture Regulations of the Department of Agriculture (Continued) AGRICULTURAL MARKETING SERVICE (Marketing Agreements... Administrative Body § 920.25 Acceptance. Each person to be selected by the Secretary as a member or as...

  15. 5 CFR 1655.11 - Loan acceptance.

    Code of Federal Regulations, 2010 CFR

    2010-01-01

    ... CFR 1653.3(c); or (f) The participant has received a taxable loan distribution from the TSP within the... 5 Administrative Personnel 3 2010-01-01 2010-01-01 false Loan acceptance. 1655.11 Section 1655.11 Administrative Personnel FEDERAL RETIREMENT THRIFT INVESTMENT BOARD LOAN PROGRAM § 1655.11 Loan acceptance....

  16. 7 CFR 930.26 - Acceptance.

    Code of Federal Regulations, 2011 CFR

    2011-01-01

    ... 7 Agriculture 8 2011-01-01 2011-01-01 false Acceptance. 930.26 Section 930.26 Agriculture Regulations of the Department of Agriculture (Continued) AGRICULTURAL MARKETING SERVICE (Marketing Agreements... Administrative Body § 930.26 Acceptance. Each person to be appointed by the Secretary as a member or as...

  17. 7 CFR 945.27 - Acceptance.

    Code of Federal Regulations, 2010 CFR

    2010-01-01

    ... 7 Agriculture 8 2010-01-01 2010-01-01 false Acceptance. 945.27 Section 945.27 Agriculture Regulations of the Department of Agriculture (Continued) AGRICULTURAL MARKETING SERVICE (Marketing Agreements... § 945.27 Acceptance. Any person nominated to serve on the committee as a member or as an alternate...

  18. 19 CFR 114.21 - Acceptance.

    Code of Federal Regulations, 2010 CFR

    2010-04-01

    ... 19 Customs Duties 1 2010-04-01 2010-04-01 false Acceptance. 114.21 Section 114.21 Customs Duties U.S. CUSTOMS AND BORDER PROTECTION, DEPARTMENT OF HOMELAND SECURITY; DEPARTMENT OF THE TREASURY CARNETS Processing of Carnets § 114.21 Acceptance. A carnet executed in accordance with § 114.3 shall...

  19. 19 CFR 114.21 - Acceptance.

    Code of Federal Regulations, 2011 CFR

    2011-04-01

    ... 19 Customs Duties 1 2011-04-01 2011-04-01 false Acceptance. 114.21 Section 114.21 Customs Duties U.S. CUSTOMS AND BORDER PROTECTION, DEPARTMENT OF HOMELAND SECURITY; DEPARTMENT OF THE TREASURY CARNETS Processing of Carnets § 114.21 Acceptance. A carnet executed in accordance with § 114.3 shall...

  20. 7 CFR 920.25 - Acceptance.

    Code of Federal Regulations, 2011 CFR

    2011-01-01

    ... 7 Agriculture 8 2011-01-01 2011-01-01 false Acceptance. 920.25 Section 920.25 Agriculture Regulations of the Department of Agriculture (Continued) AGRICULTURAL MARKETING SERVICE (Marketing Agreements... Administrative Body § 920.25 Acceptance. Each person to be selected by the Secretary as a member or as...

  1. 7 CFR 932.32 - Acceptance.

    Code of Federal Regulations, 2011 CFR

    2011-01-01

    ... 7 Agriculture 8 2011-01-01 2011-01-01 false Acceptance. 932.32 Section 932.32 Agriculture... Regulating Handling Olive Administrative Committee § 932.32 Acceptance. Any person selected by the Secretary as a member or as an alternate member of the committee shall qualify by filing a written...

  2. 5 CFR 1655.11 - Loan acceptance.

    Code of Federal Regulations, 2011 CFR

    2011-01-01

    ... CFR 1653.3(c); or (f) The participant has received a taxable loan distribution from the TSP within the... 5 Administrative Personnel 3 2011-01-01 2011-01-01 false Loan acceptance. 1655.11 Section 1655.11 Administrative Personnel FEDERAL RETIREMENT THRIFT INVESTMENT BOARD LOAN PROGRAM § 1655.11 Loan acceptance....

  3. 48 CFR 411.103 - Market acceptance.

    Code of Federal Regulations, 2011 CFR

    2011-10-01

    ... 48 Federal Acquisition Regulations System 4 2011-10-01 2011-10-01 false Market acceptance. 411.103 Section 411.103 Federal Acquisition Regulations System DEPARTMENT OF AGRICULTURE COMPETITION AND... acceptance. (a) The head of the contracting activity (HCA) may determine that offerors must demonstrate,...

  4. 7 CFR 989.32 - Acceptance.

    Code of Federal Regulations, 2011 CFR

    2011-01-01

    ... 7 Agriculture 8 2011-01-01 2011-01-01 false Acceptance. 989.32 Section 989.32 Agriculture Regulations of the Department of Agriculture (Continued) AGRICULTURAL MARKETING SERVICE (Marketing Agreements... CALIFORNIA Order Regulating Handling Raisin Administrative Committee § 989.32 Acceptance. Each person to...

  5. 7 CFR 932.32 - Acceptance.

    Code of Federal Regulations, 2010 CFR

    2010-01-01

    ... 7 Agriculture 8 2010-01-01 2010-01-01 false Acceptance. 932.32 Section 932.32 Agriculture... Regulating Handling Olive Administrative Committee § 932.32 Acceptance. Any person selected by the Secretary as a member or as an alternate member of the committee shall qualify by filing a written...

  6. 7 CFR 945.27 - Acceptance.

    Code of Federal Regulations, 2011 CFR

    2011-01-01

    ... 7 Agriculture 8 2011-01-01 2011-01-01 false Acceptance. 945.27 Section 945.27 Agriculture Regulations of the Department of Agriculture (Continued) AGRICULTURAL MARKETING SERVICE (Marketing Agreements... § 945.27 Acceptance. Any person nominated to serve on the committee as a member or as an alternate...

  7. 48 CFR 411.103 - Market acceptance.

    Code of Federal Regulations, 2010 CFR

    2010-10-01

    ... 48 Federal Acquisition Regulations System 4 2010-10-01 2010-10-01 false Market acceptance. 411.103 Section 411.103 Federal Acquisition Regulations System DEPARTMENT OF AGRICULTURE COMPETITION AND... acceptance. (a) The head of the contracting activity (HCA) may determine that offerors must demonstrate,...

  8. 36 CFR 251.62 - Acceptance.

    Code of Federal Regulations, 2010 CFR

    2010-07-01

    ... 36 Parks, Forests, and Public Property 2 2010-07-01 2010-07-01 false Acceptance. 251.62 Section 251.62 Parks, Forests, and Public Property FOREST SERVICE, DEPARTMENT OF AGRICULTURE LAND USES Special Uses § 251.62 Acceptance. Except for an easement, a special use authorization shall become...

  9. 49 CFR 193.2303 - Construction acceptance.

    Code of Federal Regulations, 2013 CFR

    2013-10-01

    ... 49 Transportation 3 2013-10-01 2013-10-01 false Construction acceptance. 193.2303 Section 193.2303 Transportation Other Regulations Relating to Transportation (Continued) PIPELINE AND HAZARDOUS MATERIALS SAFETY...: FEDERAL SAFETY STANDARDS Construction § 193.2303 Construction acceptance. No person may place in...

  10. 49 CFR 193.2303 - Construction acceptance.

    Code of Federal Regulations, 2011 CFR

    2011-10-01

    ... 49 Transportation 3 2011-10-01 2011-10-01 false Construction acceptance. 193.2303 Section 193.2303 Transportation Other Regulations Relating to Transportation (Continued) PIPELINE AND HAZARDOUS MATERIALS SAFETY...: FEDERAL SAFETY STANDARDS Construction § 193.2303 Construction acceptance. No person may place in...

  11. Why Do Women Accept the Rape Myth?

    ERIC Educational Resources Information Center

    Tabone, Christopher; And Others

    The rape myth, defined as prejudicial, stereotyped, or false beliefs about rape, rape victims, and rapists, is accepted by individuals from varied walks of life, including women. It has been suggested that rape myth acceptance (RMA) among women serves a protective function by enabling women to dissociate themselves from a rape victim's experience.…

  12. Heavy Metal, Religiosity, and Suicide Acceptability.

    ERIC Educational Resources Information Center

    Stack, Steven

    1998-01-01

    Reports on data taken from the General Social Survey that found a link between "heavy metal" rock fanship and suicide acceptability. Finds that relationship becomes nonsignificant once level of religiosity is controlled. Heavy metal fans are low in religiosity, which contributes to greater suicide acceptability. (Author/JDM)

  13. Nevada Test Site Waste Acceptance Criteria (NTSWAC)

    SciTech Connect

    NNSA /NSO Waste Management Project

    2008-06-01

    This document establishes the U.S. Department of Energy, National Nuclear Security Administration Nevada Site Office, Nevada Test Site Waste Acceptance Criteria (NTSWAC). The NTSWAC provides the requirements, terms, and conditions under which the Nevada Test Site will accept low-level radioactive (LLW) and LLW Mixed Waste (MW) for disposal.

  14. 48 CFR 411.103 - Market acceptance.

    Code of Federal Regulations, 2013 CFR

    2013-10-01

    ... 48 Federal Acquisition Regulations System 4 2013-10-01 2013-10-01 false Market acceptance. 411.103... ACQUISITION PLANNING DESCRIBING AGENCY NEEDS Selecting and Developing Requirements Documents 411.103 Market... accordance with FAR 11.103(a), the market acceptability of their items to be offered. (b) The...

  15. 48 CFR 411.103 - Market acceptance.

    Code of Federal Regulations, 2014 CFR

    2014-10-01

    ... 48 Federal Acquisition Regulations System 4 2014-10-01 2014-10-01 false Market acceptance. 411.103... ACQUISITION PLANNING DESCRIBING AGENCY NEEDS Selecting and Developing Requirements Documents 411.103 Market... accordance with FAR 11.103(a), the market acceptability of their items to be offered. (b) The...

  16. 48 CFR 3011.103 - Market acceptance.

    Code of Federal Regulations, 2013 CFR

    2013-10-01

    ... 48 Federal Acquisition Regulations System 7 2013-10-01 2012-10-01 true Market acceptance. 3011.103 Section 3011.103 Federal Acquisition Regulations System DEPARTMENT OF HOMELAND SECURITY, HOMELAND SECURITY... Requirements Documents 3011.103 Market acceptance. (a) Contracting officers may act on behalf of the head...

  17. 48 CFR 411.103 - Market acceptance.

    Code of Federal Regulations, 2012 CFR

    2012-10-01

    ... 48 Federal Acquisition Regulations System 4 2012-10-01 2012-10-01 false Market acceptance. 411.103... ACQUISITION PLANNING DESCRIBING AGENCY NEEDS Selecting and Developing Requirements Documents 411.103 Market... accordance with FAR 11.103(a), the market acceptability of their items to be offered. (b) The...

  18. 48 CFR 3011.103 - Market acceptance.

    Code of Federal Regulations, 2012 CFR

    2012-10-01

    ... 48 Federal Acquisition Regulations System 7 2012-10-01 2012-10-01 false Market acceptance. 3011.103 Section 3011.103 Federal Acquisition Regulations System DEPARTMENT OF HOMELAND SECURITY, HOMELAND... Developing Requirements Documents 3011.103 Market acceptance. (a) Contracting officers may act on behalf...

  19. 48 CFR 3011.103 - Market acceptance.

    Code of Federal Regulations, 2014 CFR

    2014-10-01

    ... 48 Federal Acquisition Regulations System 7 2014-10-01 2014-10-01 false Market acceptance. 3011.103 Section 3011.103 Federal Acquisition Regulations System DEPARTMENT OF HOMELAND SECURITY, HOMELAND... Developing Requirements Documents 3011.103 Market acceptance. (a) Contracting officers may act on behalf...

  20. Four therapeutic diets: adherence and acceptability.

    PubMed

    Berkow, Susan E; Barnard, Neal; Eckart, Jill; Katcher, Heather

    2010-01-01

    Many health conditions are treated, at least in part, by therapeutic diets. Although the success of any intervention depends on its acceptability to the patient, the acceptability of therapeutic diets and factors that influence it have been largely neglected in nutrition research. A working definition of acceptability is proposed and an examination and summary are provided of available data on the acceptability of common diet regimens used for medical conditions. The goal is to suggest ways to improve the success of therapeutic diets. The proposed working definition of "acceptability" refers to the user's judgment of the advantages and disadvantages of a therapeutic diet-in relation to palatability, costs, and effects on eating behaviour and health-that influence the likelihood of adherence. Very low-calorie, reduced-fat omnivorous, vegetarian and vegan, and low-carbohydrate diets all achieve acceptability among the majority of users in studies of up to one year, in terms of attrition and adherence rates and results of questionnaires assessing eating behaviours. Longer studies are fewer, but they suggest that vegetarian, vegan, and reduced-fat diets are acceptable, as indicated by sustained changes in nutrient intake. Few studies of this length have been published for very low-calorie or low-carbohydrate diets. Long-term studies of adherence and acceptability of these and other therapeutic diets are warranted. PMID:21144137