Science.gov

Sample records for based empirical survey

  1. An Empirical Pixel-Based Correction for Imperfect CTE. I. HST's Advanced Camera for Surveys

    NASA Astrophysics Data System (ADS)

    Anderson, Jay; Bedin, Luigi

    2010-09-01

    We use an empirical approach to characterize the effect of charge-transfer efficiency (CTE) losses in images taken with the Wide-Field Channel of the Advanced Camera for Surveys (ACS). The study is based on profiles of warm pixels in 168 dark exposures taken between 2009 September and October. The dark exposures allow us to explore charge traps that affect electrons when the background is extremely low. We develop a model for the readout process that reproduces the observed trails out to 70 pixels. We then invert the model to convert the observed pixel values in an image into an estimate of the original pixel values. We find that when we apply this image-restoration process to science images with a variety of stars on a variety of background levels, it restores flux, position, and shape. This means that the observed trails contain essentially all of the flux lost to inefficient CTE. The Space Telescope Science Institute is currently evaluating this algorithm with the aim of optimizing it and eventually providing enhanced data products. The empirical procedure presented here should also work for other epochs (e.g., pre-SM4), though the parameters may have to be recomputed for the time when ACS was operated at a higher temperature than the current -81°C. Finally, this empirical approach may also hold promise for other instruments, such as WFPC2, STIS, the ASC's HRC, and even WFC3/UVIS.

  2. An Empirical Pixel-Based Correction for Imperfect CTE. I. HST's Advanced Camera for Surveys

    NASA Astrophysics Data System (ADS)

    Anderson, Jay; Bedin, Luigi R.

    2010-09-01

    We use an empirical approach to characterize the effect of charge-transfer efficiency (CTE) losses in images taken with the Wide-Field Channel of the Advanced Camera for Surveys (ACS). The study is based on profiles of warm pixels in 168 dark exposures taken between 2009 September and October. The dark exposures allow us to explore charge traps that affect electrons when the background is extremely low. We develop a model for the readout process that reproduces the observed trails out to 70 pixels. We then invert the model to convert the observed pixel values in an image into an estimate of the original pixel values. We find that when we apply this image-restoration process to science images with a variety of stars on a variety of background levels, it restores flux, position, and shape. This means that the observed trails contain essentially all of the flux lost to inefficient CTE. The Space Telescope Science Institute is currently evaluating this algorithm with the aim of optimizing it and eventually providing enhanced data products. The empirical procedure presented here should also work for other epochs (e.g., pre-SM4), though the parameters may have to be recomputed for the time when ACS was operated at a higher temperature than the current -81°C. Finally, this empirical approach may also hold promise for other instruments, such as WFPC2, STIS, the ACS's HRC, and even WFC3/UVIS. Based on observations with the NASA/ESA Hubble Space Telescope, obtained at the Space Telescope Science Institute, which is operated by AURA, Inc., under NASA contract NAS 5-26555.

  3. Monitoring of Qualifications and Employment in Austria: An Empirical Approach Based on the Labour Force Survey

    ERIC Educational Resources Information Center

    Lassnigg, Lorenz; Vogtenhuber, Stefan

    2011-01-01

    The empirical approach referred to in this article describes the relationship between education and training (ET) supply and employment in Austria; the use of the new ISCED (International Standard Classification of Education) fields of study variable makes this approach applicable abroad. The purpose is to explore a system that produces timely…

  4. The ALHAMBRA survey: An empirical estimation of the cosmic variance for merger fraction studies based on close pairs

    NASA Astrophysics Data System (ADS)

    López-Sanjuan, C.; Cenarro, A. J.; Hernández-Monteagudo, C.; Varela, J.; Molino, A.; Arnalte-Mur, P.; Ascaso, B.; Castander, F. J.; Fernández-Soto, A.; Huertas-Company, M.; Márquez, I.; Martínez, V. J.; Masegosa, J.; Moles, M.; Pović, M.; Aguerri, J. A. L.; Alfaro, E.; Aparicio-Villegas, T.; Benítez, N.; Broadhurst, T.; Cabrera-Caño, J.; Cepa, J.; Cerviño, M.; Cristóbal-Hornillos, D.; Del Olmo, A.; González Delgado, R. M.; Husillos, C.; Infante, L.; Perea, J.; Prada, F.; Quintana, J. M.

    2014-04-01

    Aims: Our goal is to estimate empirically the cosmic variance that affects merger fraction studies based on close pairs for the first time. Methods: We compute the merger fraction from photometric redshift close pairs with 10 h-1 kpc ≤ rp ≤ 50 h-1 kpc and Δv ≤ 500 km s-1 and measure it in the 48 sub-fields of the ALHAMBRA survey. We study the distribution of the measured merger fractions that follow a log-normal function and estimate the cosmic variance σv as the intrinsic dispersion of the observed distribution. We develop a maximum likelihood estimator to measure a reliable σv and avoid the dispersion due to the observational errors (including the Poisson shot noise term). Results: The cosmic variance σv of the merger fraction depends mainly on (i) the number density of the populations under study for both the principal (n1) and the companion (n2) galaxy in the close pair and (ii) the probed cosmic volume Vc. We do not find a significant dependence on either the search radius used to define close companions, the redshift, or the physical selection (luminosity or stellar mass) of the samples. Conclusions: We have estimated the cosmic variance that affects the measurement of the merger fraction by close pairs from observations. We provide a parametrisation of the cosmic variance with n1, n2, and Vc, σv ∝ n1-0.54Vc-0.48 (n_2/n_1)-0.37 . Thanks to this prescription, future merger fraction studies based on close pairs could properly account for the cosmic variance on their results. Based on observations collected at the German-Spanish Astronomical Center, Calar Alto, jointly operated by the Max-Planck-Institut für Astronomie (MPIA) at Heidelberg and the Instituto de Astrofísica de Andalucía (IAA-CSIC).Appendix is available in electronic form at http://www.aanda.org

  5. Social stratification, development and health in Pakistan: an empirical exploration of relationships in population-based national health examination survey data.

    PubMed

    Hadden, Wilbur C; Pappas, Gregory; Khan, Abdul Qayyum

    2003-11-01

    There is a growing literature on an overall direct relationship between health and social position in developed countries. The relationship, however, is less well documented in developing countries where social hierarchies are changing more rapidly, demographic and health transitions are less advanced, and this topic has received less attention from researchers than in some developed countries. This paper presents an empirical investigation of the relationship between social stratification and social development and population health using data on over 6000 adults from the National Health Survey of Pakistan, a nationally representative health examination survey of people in that country. We analyze four indicators of poor nutrition in adults from this data set. The findings reveal complex relationships among social development, social stratification and the consequences for the health of the people of Pakistan. Underweight is related to economic status, anemia to education and social development, and both severe dental caries and a monotonous diet are related to both development and economic status which interact with each other. These results suggest that continued conceptual refinement and the development of standardized measures of stratification and development would contribute to building cross-nationally comparable data sets addressing issues of the relationship among health and economic development and health transitions.

  6. GIS Teacher Training: Empirically-Based Indicators of Effectiveness

    ERIC Educational Resources Information Center

    Höhnle, Steffen; Fögele, Janis; Mehren, Rainer; Schubert, Jan Christoph

    2016-01-01

    In spite of various actions, the implementation of GIS (geographic information systems) in German schools is still very low. In the presented research, teaching experts as well as teaching novices were presented with empirically based constraints for implementation stemming from an earlier survey. In the process of various group discussions, the…

  7. An empirical Bayes approach to analyzing recurring animal surveys

    USGS Publications Warehouse

    Johnson, D.H.

    1989-01-01

    Recurring estimates of the size of animal populations are often required by biologists or wildlife managers. Because of cost or other constraints, estimates frequently lack the accuracy desired but cannot readily be improved by additional sampling. This report proposes a statistical method employing empirical Bayes (EB) estimators as alternatives to those customarily used to estimate population size, and evaluates them by a subsampling experiment on waterfowl surveys. EB estimates, especially a simple limited-translation version, were more accurate and provided shorter confidence intervals with greater coverage probabilities than customary estimates.

  8. Empirical estimates of cumulative refraction errors associated with procedurally constrained levelings based on the Gaithersburg- Tucson refraction tests of the National Geodetic Survey.

    USGS Publications Warehouse

    Castle, R.O.; Gilmore, T.D.; Mark, R.K.; Shaw, R.H.

    1985-01-01

    Analyses of results of the National Geodetic Survey's leveling refraction tests indicate that the standard deviation about the mean (sigma) for high-scale minus low-scale rod readings closely correlates with measured refraction error. Use of this relation in conjunction with values for sigma obtained from routinely constrainted surveys provides a basis for estimating the refraction error associated with levelings of stipulated order and class. -Authors

  9. Empirically Based Play Interventions for Children

    ERIC Educational Resources Information Center

    Reddy, Linda A., Ed.; Files-Hall, Tara M., Ed.; Schaefer, Charles E., Ed.

    2005-01-01

    "Empirically Based Play Interventions for Children" is a compilation of innovative, well-designed play interventions, presented for the first time in one text. Play therapy is the oldest and most popular form of child therapy in clinical practice and is widely considered by practitioners to be uniquely responsive to children's developmental needs.…

  10. Searching for Evidence-Based Practice: A Survey of Empirical Studies on Curricular Interventions Measuring and Reporting Fidelity of Implementation Published during 2004-2013

    ERIC Educational Resources Information Center

    Missett, Tracy C.; Foster, Lisa H.

    2015-01-01

    In an environment of accountability, the development of evidence-based practices is expected. To demonstrate that a practice is evidence based, quality indicators of rigorous methodology should be present including indications that teachers implementing an intervention have done so with fidelity to its design. Because evidence-based practices…

  11. Watershed-based survey designs.

    PubMed

    Detenbeck, Naomi E; Cincotta, Dan; Denver, Judith M; Greenlee, Susan K; Olsen, Anthony R; Pitchford, Ann M

    2005-04-01

    Watershed-based sampling design and assessment tools help serve the multiple goals for water quality monitoring required under the Clean Water Act, including assessment of regional conditions to meet Section 305(b), identification of impaired water bodies or watersheds to meet Section 303(d), and development of empirical relationships between causes or sources of impairment and biological responses. Creation of GIS databases for hydrography, hydrologically corrected digital elevation models, and hydrologic derivatives such as watershed boundaries and upstream-downstream topology of subcatchments would provide a consistent seamless nationwide framework for these designs. The elements of a watershed-based sample framework can be represented either as a continuous infinite set defined by points along a linear stream network, or as a discrete set of watershed polygons. Watershed-based designs can be developed with existing probabilistic survey methods, including the use of unequal probability weighting, stratification, and two-stage frames for sampling. Case studies for monitoring of Atlantic Coastal Plain streams, West Virginia wadeable streams, and coastal Oregon streams illustrate three different approaches for selecting sites for watershed-based survey designs. PMID:15861987

  12. Watershed-based survey designs

    USGS Publications Warehouse

    Detenbeck, N.E.; Cincotta, D.; Denver, J.M.; Greenlee, S.K.; Olsen, A.R.; Pitchford, A.M.

    2005-01-01

    Watershed-based sampling design and assessment tools help serve the multiple goals for water quality monitoring required under the Clean Water Act, including assessment of regional conditions to meet Section 305(b), identification of impaired water bodies or watersheds to meet Section 303(d), and development of empirical relationships between causes or sources of impairment and biological responses. Creation of GIS databases for hydrography, hydrologically corrected digital elevation models, and hydrologic derivatives such as watershed boundaries and upstream-downstream topology of subcatchments would provide a consistent seamless nationwide framework for these designs. The elements of a watershed-based sample framework can be represented either as a continuous infinite set defined by points along a linear stream network, or as a discrete set of watershed polygons. Watershed-based designs can be developed with existing probabilistic survey methods, including the use of unequal probability weighting, stratification, and two-stage frames for sampling. Case studies for monitoring of Atlantic Coastal Plain streams, West Virginia wadeable streams, and coastal Oregon streams illustrate three different approaches for selecting sites for watershed-based survey designs. ?? Springer Science + Business Media, Inc. 2005.

  13. Issues and Controversies that Surround Recent Texts on Empirically Supported and Empirically Based Treatments

    ERIC Educational Resources Information Center

    Paul, Howard A.

    2004-01-01

    Since the 1993 APA task force of the Society of Clinical Psychology developed guidelines to apply data-based psychology to the identification of effective psychotherapy, there has been an increasing number of texts focussing on Empirically based Psychotherapy and Empirically Supported Treatments. This manuscript examines recent key texts and…

  14. Optimized curvelet-based empirical mode decomposition

    NASA Astrophysics Data System (ADS)

    Wu, Renjie; Zhang, Qieshi; Kamata, Sei-ichiro

    2015-02-01

    The recent years has seen immense improvement in the development of signal processing based on Curvelet transform. The Curvelet transform provide a new multi-resolution representation. The frame elements of Curvelets exhibit higher direction sensitivity and anisotropic than the Wavelets, multi-Wavelets, steerable pyramids, and so on. These features are based on the anisotropic notion of scaling. In practical instances, time series signals processing problem is often encountered. To solve this problem, the time-frequency analysis based methods are studied. However, the time-frequency analysis cannot always be trusted. Many of the new methods were proposed. The Empirical Mode Decomposition (EMD) is one of them, and widely used. The EMD aims to decompose into their building blocks functions that are the superposition of a reasonably small number of components, well separated in the time-frequency plane. And each component can be viewed as locally approximately harmonic. However, it cannot solve the problem of directionality of high-dimensional. A reallocated method of Curvelet transform (optimized Curvelet-based EMD) is proposed in this paper. We introduce a definition for a class of functions that can be viewed as a superposition of a reasonably small number of approximately harmonic components by optimized Curvelet family. We analyze this algorithm and demonstrate its results on data. The experimental results prove the effectiveness of our method.

  15. Empirical Validation and Application of the Computing Attitudes Survey

    ERIC Educational Resources Information Center

    Dorn, Brian; Elliott Tew, Allison

    2015-01-01

    Student attitudes play an important role in shaping learning experiences. However, few validated instruments exist for measuring student attitude development in a discipline-specific way. In this paper, we present the design, development, and validation of the computing attitudes survey (CAS). The CAS is an extension of the Colorado Learning…

  16. Empirically Based Myths: Astrology, Biorhythms, and ATIs.

    ERIC Educational Resources Information Center

    Ragsdale, Ronald G.

    1980-01-01

    A myth may have an empirical basis through chance occurrence; perhaps Aptitude Treatment Interactions (ATIs) are in this category. While ATIs have great utility in describing, planning, and implementing instruction, few disordinal interactions have been found. Article suggests narrowing of ATI research with replications and estimates of effect…

  17. A survey on hematology-oncology pediatric AIEOP centers: prophylaxis, empirical therapy and nursing prevention procedures of infectious complications.

    PubMed

    Livadiotti, Susanna; Milano, Giuseppe Maria; Serra, Annalisa; Folgori, Laura; Jenkner, Alessandro; Castagnola, Elio; Cesaro, Simone; Rossi, Mario R; Barone, Angelica; Zanazzo, Giulio; Nesi, Francesca; Licciardello, Maria; De Santis, Raffaella; Ziino, Ottavio; Cellini, Monica; Porta, Fulvio; Caselli, Desiree; Pontrelli, Giuseppe

    2012-01-01

    A nationwide questionnaire-based survey was designed to evaluate the management and prophylaxis of febrile neutropenia in pediatric patients admitted to hematology-oncology and hematopoietic stem cell transplant units. Of the 34 participating centers, 40 and 63%, respectively, continue to prescribe antibacterial and antimycotic prophylaxis in low-risk subjects and 78 and 94% in transplant patients. Approximately half of the centers prescribe a combination antibiotic regimen as first-line therapy in low-risk patients and up to 81% in high-risk patients. When initial empirical therapy fails after seven days, 63% of the centers add empirical antimycotic therapy in low-and 81% in high-risk patients. Overall management varies significantly across centers. Preventive nursing procedures are in accordance with international guidelines. This survey is the first to focus on prescribing practices in children with cancer and could help to implement practice guidelines.

  18. School-Based Management and Paradigm Shift in Education an Empirical Study

    ERIC Educational Resources Information Center

    Cheng, Yin Cheong; Mok, Magdalena Mo Ching

    2007-01-01

    Purpose: This paper aims to report empirical research investigating how school-based management (SBM) and paradigm shift (PS) in education are closely related to teachers' student-centered teaching and students' active learning in a sample of Hong Kong secondary schools. Design/methodology/approach: It is a cross-sectional survey research…

  19. Web-Based Surveys: Not Your Basic Survey Anymore

    ERIC Educational Resources Information Center

    Bertot, John Carlo

    2009-01-01

    Web-based surveys are not new to the library environment. Although such surveys began as extensions of print surveys, the Web-based environment offers a number of approaches to conducting a survey that the print environment cannot duplicate easily. Since 1994, the author and others have conducted national surveys of public library Internet…

  20. WATERSHED BASED SURVEY DESIGNS

    EPA Science Inventory

    The development of watershed-based design and assessment tools will help to serve the multiple goals for water quality monitoring required under the Clean Water Act, including assessment of regional condition to meet Section 305(b), identification of impaired water bodies or wate...

  1. Survey-Based Measurement of Public Management and Policy Networks

    ERIC Educational Resources Information Center

    Henry, Adam Douglas; Lubell, Mark; McCoy, Michael

    2012-01-01

    Networks have become a central concept in the policy and public management literature; however, theoretical development is hindered by a lack of attention to the empirical properties of network measurement methods. This paper compares three survey-based methods for measuring organizational networks: the roster, the free-recall name generator, and…

  2. An Empirical Evaluation of Puzzle-Based Learning as an Interest Approach for Teaching Introductory Computer Science

    ERIC Educational Resources Information Center

    Merrick, K. E.

    2010-01-01

    This correspondence describes an adaptation of puzzle-based learning to teaching an introductory computer programming course. Students from two offerings of the course--with and without the puzzle-based learning--were surveyed over a two-year period. Empirical results show that the synthesis of puzzle-based learning concepts with existing course…

  3. Empirically Based Strategies for Preventing Juvenile Delinquency.

    PubMed

    Pardini, Dustin

    2016-04-01

    Juvenile crime is a serious public health problem that results in significant emotional and financial costs for victims and society. Using etiologic models as a guide, multiple interventions have been developed to target risk factors thought to perpetuate the emergence and persistence of delinquent behavior. Evidence suggests that the most effective interventions tend to have well-defined treatment protocols, focus on therapeutic approaches as opposed to external control techniques, and use multimodal cognitive-behavioral treatment strategies. Moving forward, there is a need to develop effective policies and procedures that promote the widespread adoption of evidence-based delinquency prevention practices across multiple settings. PMID:26980128

  4. Image-Based Empirical Modeling of the Plasmasphere

    NASA Technical Reports Server (NTRS)

    Adrian, Mark L.; Gallagher, D. L.

    2008-01-01

    A new suite of empirical models of plasmaspheric plasma based on remote, global images from the IMAGE EUV instrument is proposed for development. The purpose of these empirical models is to establish the statistical properties of the plasmasphere as a function of conditions. This suite of models will mark the first time the plasmaspheric plume is included in an empirical model. Development of these empirical plasmaspheric models will support synoptic studies (such as for wave propagation and growth, energetic particle loss through collisions and dust transport as influenced by charging) and serves as a benchmark against which physical models can be tested. The ability to know that a specific global density distribution occurs in response to specific magnetospheric and solar wind factors is a huge advantage over all previous in-situ based empirical models. The consequence of creating these new plasmaspheric models will be to provide much higher fidelity and much richer quantitative descriptions of the statistical properties of plasmaspheric plasma in the inner magnetosphere, whether that plasma is in the main body of the plasmasphere, nearby during recovery or in the plasmaspheric plume. Model products to be presented include statistical probabilities for being in the plasmasphere, near thermal He+ density boundaries and the complexity of its spatial structure.

  5. Space Based Dark Energy Surveys

    NASA Astrophysics Data System (ADS)

    Dore, Olivier

    2016-03-01

    Dark energy, the name given to the cause of the accelerating expansion of the Universe, is one of the most tantalizing mystery in modern physics. Current cosmological models hold that dark energy is currently the dominant component of the Universe, but the exact nature of DE remains poorly understood. There are ambitious ground-based surveys underway that seek to understand DE and NASA is participating in the development of significantly more ambitious space-based surveys planned for the next decade. NASA has provided mission enabling technology to the European Space Agency's (ESA) Euclid mission in exchange for US scientists to participate in the Euclid mission. NASA is also developing the Wide Field Infrared Survey Telescope-Astrophysics Focused Telescope Asset (WFIRST) mission for possible launch in 2024. WFIRST was the highest ranked space mission in the Astro2010 Decadal Survey and the current design uses a 2.4m space telescope to go beyond what was then envisioned. Understanding DE is one of the primary science goals of WFIRST-AFTA. This talk will review the state of DE, the relevant activities of the Cosmic Structure Interest Group (CoSSIG) of the PhyPAG, and detail the status and complementarity between Euclid, WFIRST and ot ambitious ground-based efforts.

  6. Responses to Commentaries on Advances in Empirically Based Assessment.

    ERIC Educational Resources Information Center

    McConaughy, Stephanie H.

    1993-01-01

    Author of article (this issue) describing research program to advance assessment of children's behavioral and emotional problems; presenting conceptual framework for multiaxial empirically based assessment; and summarizing research efforts to develop cross-informant scales for scoring parent, teacher, and self-reports responds to commentaries on…

  7. Determining VA physician requirements through empirically based models.

    PubMed Central

    Lipscomb, J; Kilpatrick, K E; Lee, K L; Pieper, K S

    1995-01-01

    OBJECTIVE: As part of a project to estimate physician requirements for the Department of Veterans Affairs, the Institute of Medicine (IOM) developed and tested empirically based models of physician staffing, by specialty, that could be applied to each VA facility. DATA SOURCE/STUDY SETTING. These analyses used selected data on all patient encounters and all facilities in VA's management information systems for FY 1989. STUDY DESIGN. Production functions (PFs), with patient workload dependent on physicians, other providers, and nonpersonnel factors, were estimated for each of 14 patient care areas in a VA medical center. Inverse production functions (IPFs), with physician staffing levels dependent on workload and other factors, were estimated for each of 11 specialty groupings. These models provide complementary approaches to deriving VA physician requirements for patient care and medical education. DATA COLLECTION/EXTRACTION METHODS. All data were assembled by VA and put in analyzable SAS data sets containing FY 1989 workload and staffing variables used in the PFs and IPFs. All statistical analyses reported here were conducted by the IOM. PRINCIPAL FINDINGS. Existing VA data can be used to develop statistically strong, clinically plausible, empirically based models for calculating physician requirements, by specialty. These models can (1) compare current physician staffing in a given setting with systemwide norms and (2) yield estimates of future staffing requirements conditional on future workload. CONCLUSIONS. Empirically based models can play an important role in determining VA physician staffing requirements. VA should test, evaluate, and revise these models on an ongoing basis. PMID:7860320

  8. Empirically Based Psychosocial Therapies for Schizophrenia: The Disconnection between Science and Practice.

    PubMed

    Shean, Glenn D

    2013-01-01

    Empirically validated psychosocial therapies for individuals diagnosed with schizophrenia were described in the report of the Schizophrenia Patient Outcomes Research Team (PORT, 2009). The PORT team identified eight psychosocial treatments: assertive community treatment, supported employment, cognitive behavioral therapy, family-based services, token economy, skills training, psychosocial interventions for alcohol and substance use disorders, and psychosocial interventions for weight management. PORT listings of empirically validated psychosocial therapies provide a useful template for the design of effective recovery-oriented mental health care systems. Unfortunately, surveys indicate that PORT listings have not been implemented in clinical settings. Obstacles to the implementation of PORT psychosocial therapy listings and suggestions for changes needed to foster implementation are discussed. Limitations of PORT therapy listings that are based on therapy outcome efficacy studies are discussed, and cross-cultural and course and outcome studies of correlates of recovery are summarized. PMID:23738068

  9. An empirical analysis of the demand for sleep: Evidence from the American Time Use Survey.

    PubMed

    Ásgeirsdóttir, Tinna Laufey; Ólafsson, Sigurður Páll

    2015-12-01

    Using data from the American Time Use Survey, this paper empirically examined the demand for sleep, with special attention to its opportunity cost represented by wages. Variation in the unemployment rate by state was also used to investigate the cyclical nature of sleep duration. We conducted separate estimations for males and females, as well as for those who received a fixed salary and hourly wages. The findings predominantly revealed no relationship between sleep duration and the business cycle. However, an inverse relationship between sleep duration and wages was detected. This is in accordance with sleep duration being an economic choice variable, rather than a predetermined subtraction of the 24-h day. Although the inverse relationship was not significant in all the estimations for salaried subjects, it was consistent and strong for subjects who received hourly wages. For instance, elasticity measures were −.03 for those who received hourly wages and −.003 for those who received a fixed salary. PMID:26603429

  10. An Empirically Based Shaped Charge Jet Break-Up Model

    NASA Astrophysics Data System (ADS)

    Baker, Ernest; Pham, James; Vuong, Tan

    2013-06-01

    This paper discusses an empirically based shaped charge jet break-up model based around Walsh's breakup theory and provides significant experimental confirmation over a broad range of velocity gradients. The parameters which affect jet length and breakup times are fairly well known, but there is some controversy over the exact nature of the dependencies. Walsh theorized that the dependence of jet length would take a particular form, based on his determination of a dimensionless parameter for the problem and numerical experiments in which initial perturbation strengths were varied. Walsh did not present comparisons with experimental results. Chou has presented a variety of different jet break-up models with some data comparisons. Mostert [3] has suggested that breakup time is proportional to (Δm/Δv) 1/3. It is shown here that the parameter (Δm/Δv)1/2 or (dm/dv)1/3, closely related to Walsh's dimensionless parameter, whose values were obtained from either experiments or simulations correlates quite well with jet breakup times for a very wide variety of shaped charge devices. The values of Δm and Δv are respectively the jet mass and the velocity difference of the portion of jet in question. For a typical shaped charge Δm/Δv is essentially invariant with respect to time. In this paper, we present the mathematical basis for an empirically based break-up model with a similar basis to Walsh and Mostert, as well as supporting empirical data for a broad range of shaped charge geometries.

  11. Recommended survey designs for occupancy modelling using motion-activated cameras: insights from empirical wildlife data

    PubMed Central

    Lewis, Jesse S.; Gerber, Brian D.

    2014-01-01

    Motion-activated cameras are a versatile tool that wildlife biologists can use for sampling wild animal populations to estimate species occurrence. Occupancy modelling provides a flexible framework for the analysis of these data; explicitly recognizing that given a species occupies an area the probability of detecting it is often less than one. Despite the number of studies using camera data in an occupancy framework, there is only limited guidance from the scientific literature about survey design trade-offs when using motion-activated cameras. A fuller understanding of these trade-offs will allow researchers to maximise available resources and determine whether the objectives of a monitoring program or research study are achievable. We use an empirical dataset collected from 40 cameras deployed across 160 km2 of the Western Slope of Colorado, USA to explore how survey effort (number of cameras deployed and the length of sampling period) affects the accuracy and precision (i.e., error) of the occupancy estimate for ten mammal and three virtual species. We do this using a simulation approach where species occupancy and detection parameters were informed by empirical data from motion-activated cameras. A total of 54 survey designs were considered by varying combinations of sites (10–120 cameras) and occasions (20–120 survey days). Our findings demonstrate that increasing total sampling effort generally decreases error associated with the occupancy estimate, but changing the number of sites or sampling duration can have very different results, depending on whether a species is spatially common or rare (occupancy = ψ) and easy or hard to detect when available (detection probability = p). For rare species with a low probability of detection (i.e., raccoon and spotted skunk) the required survey effort includes maximizing the number of sites and the number of survey days, often to a level that may be logistically unrealistic for many studies. For common species with

  12. Recommended survey designs for occupancy modelling using motion-activated cameras: insights from empirical wildlife data.

    PubMed

    Shannon, Graeme; Lewis, Jesse S; Gerber, Brian D

    2014-01-01

    Motion-activated cameras are a versatile tool that wildlife biologists can use for sampling wild animal populations to estimate species occurrence. Occupancy modelling provides a flexible framework for the analysis of these data; explicitly recognizing that given a species occupies an area the probability of detecting it is often less than one. Despite the number of studies using camera data in an occupancy framework, there is only limited guidance from the scientific literature about survey design trade-offs when using motion-activated cameras. A fuller understanding of these trade-offs will allow researchers to maximise available resources and determine whether the objectives of a monitoring program or research study are achievable. We use an empirical dataset collected from 40 cameras deployed across 160 km(2) of the Western Slope of Colorado, USA to explore how survey effort (number of cameras deployed and the length of sampling period) affects the accuracy and precision (i.e., error) of the occupancy estimate for ten mammal and three virtual species. We do this using a simulation approach where species occupancy and detection parameters were informed by empirical data from motion-activated cameras. A total of 54 survey designs were considered by varying combinations of sites (10-120 cameras) and occasions (20-120 survey days). Our findings demonstrate that increasing total sampling effort generally decreases error associated with the occupancy estimate, but changing the number of sites or sampling duration can have very different results, depending on whether a species is spatially common or rare (occupancy = ψ) and easy or hard to detect when available (detection probability = p). For rare species with a low probability of detection (i.e., raccoon and spotted skunk) the required survey effort includes maximizing the number of sites and the number of survey days, often to a level that may be logistically unrealistic for many studies. For common species with

  13. AGENT-BASED MODELS IN EMPIRICAL SOCIAL RESEARCH*

    PubMed Central

    Bruch, Elizabeth; Atwell, Jon

    2014-01-01

    Agent-based modeling has become increasingly popular in recent years, but there is still no codified set of recommendations or practices for how to use these models within a program of empirical research. This article provides ideas and practical guidelines drawn from sociology, biology, computer science, epidemiology, and statistics. We first discuss the motivations for using agent-based models in both basic science and policy-oriented social research. Next, we provide an overview of methods and strategies for incorporating data on behavior and populations into agent-based models, and review techniques for validating and testing the sensitivity of agent-based models. We close with suggested directions for future research. PMID:25983351

  14. Video watermarking with empirical PCA-based decoding.

    PubMed

    Khalilian, Hanieh; Bajic, Ivan V

    2013-12-01

    A new method for video watermarking is presented in this paper. In the proposed method, data are embedded in the LL subband of wavelet coefficients, and decoding is performed based on the comparison among the elements of the first principal component resulting from empirical principal component analysis (PCA). The locations for data embedding are selected such that they offer the most robust PCA-based decoding. Data are inserted in the LL subband in an adaptive manner based on the energy of high frequency subbands and visual saliency. Extensive testing was performed under various types of attacks, such as spatial attacks (uniform and Gaussian noise and median filtering), compression attacks (MPEG-2, H. 263, and H. 264), and temporal attacks (frame repetition, frame averaging, frame swapping, and frame rate conversion). The results show that the proposed method offers improved performance compared with several methods from the literature, especially under additive noise and compression attacks.

  15. State of Research on Giftedness and Gifted Education: A Survey of Empirical Studies Published during 1998-2010 (April)

    ERIC Educational Resources Information Center

    Dai, David Yun; Swanson, Joan Ann; Cheng, Hongyu

    2011-01-01

    This study surveyed 1,234 empirical studies on giftedness, gifted education, and creativity during 1998-2010 (April), using PsycINFO database and targeted journals as main sources, with respect to main topics these studies focused on, methods they used for investigation, and the conceptual spaces they traversed. Four main research topics emerged…

  16. Comparison between empirical and physically based models of atmospheric correction

    NASA Astrophysics Data System (ADS)

    Mandanici, E.; Franci, F.; Bitelli, G.; Agapiou, A.; Alexakis, D.; Hadjimitsis, D. G.

    2015-06-01

    A number of methods have been proposed for the atmospheric correction of the multispectral satellite images, based on either atmosphere modelling or images themselves. Full radiative transfer models require a lot of ancillary information about the atmospheric conditions at the acquisition time. Whereas, image based methods cannot account for all the involved phenomena. Therefore, the aim of this paper is the comparison of different atmospheric correction methods for multispectral satellite images. The experimentation was carried out on a study area located in the catchment area of Yialias river, 20 km South of Nicosia, the Cyprus capital. The following models were tested, both empirical and physically based: Dark object subtraction, QUAC, Empirical line, 6SV, and FLAASH. They were applied on a Landsat 8 multispectral image. The spectral signatures of ten different land cover types were measured during a field campaign in 2013 and 15 samples were collected for laboratory measurements in a second campaign in 2014. GER 1500 spectroradiometer was used; this instrument can record electromagnetic radiation from 350 up to 1050 nm, includes 512 different channels and each channel covers about 1.5 nm. The spectral signatures measured were used to simulate the reflectance values for the multispectral sensor bands by applying relative spectral response filters. These data were considered as ground truth to assess the accuracy of the different image correction models. Results do not allow to establish which method is the most accurate. The physics-based methods describe better the shape of the signatures, whereas the image-based models perform better regarding the overall albedo.

  17. The Gaia-ESO Survey: Empirical determination of the precision of stellar radial velocities and projected rotation velocities

    NASA Astrophysics Data System (ADS)

    Jackson, R. J.; Jeffries, R. D.; Lewis, J.; Koposov, S. E.; Sacco, G. G.; Randich, S.; Gilmore, G.; Asplund, M.; Binney, J.; Bonifacio, P.; Drew, J. E.; Feltzing, S.; Ferguson, A. M. N.; Micela, G.; Neguerela, I.; Prusti, T.; Rix, H.-W.; Vallenari, A.; Alfaro, E. J.; Allende Prieto, C.; Babusiaux, C.; Bensby, T.; Blomme, R.; Bragaglia, A.; Flaccomio, E.; Francois, P.; Hambly, N.; Irwin, M.; Korn, A. J.; Lanzafame, A. C.; Pancino, E.; Recio-Blanco, A.; Smiljanic, R.; Van Eck, S.; Walton, N.; Bayo, A.; Bergemann, M.; Carraro, G.; Costado, M. T.; Damiani, F.; Edvardsson, B.; Franciosini, E.; Frasca, A.; Heiter, U.; Hill, V.; Hourihane, A.; Jofré, P.; Lardo, C.; de Laverny, P.; Lind, K.; Magrini, L.; Marconi, G.; Martayan, C.; Masseron, T.; Monaco, L.; Morbidelli, L.; Prisinzano, L.; Sbordone, L.; Sousa, S. G.; Worley, C. C.; Zaggia, S.

    2015-08-01

    Context. The Gaia-ESO Survey (GES) is a large public spectroscopic survey at the European Southern Observatory Very Large Telescope. Aims: A key aim is to provide precise radial velocities (RVs) and projected equatorial velocities (vsini) for representative samples of Galactic stars, which will complement information obtained by the Gaia astrometry satellite. Methods: We present an analysis to empirically quantify the size and distribution of uncertainties in RV and vsini using spectra from repeated exposures of the same stars. Results: We show that the uncertainties vary as simple scaling functions of signal-to-noise ratio (S/N) and vsini, that the uncertainties become larger with increasing photospheric temperature, but that the dependence on stellar gravity, metallicity and age is weak. The underlying uncertainty distributions have extended tails that are better represented by Student's t-distributions than by normal distributions. Conclusions: Parametrised results are provided, which enable estimates of the RV precision for almost all GES measurements, and estimates of the vsini precision for stars in young clusters, as a function of S/N, vsini and stellar temperature. The precision of individual high S/N GES RV measurements is 0.22-0.26 km s-1, dependent on instrumental configuration. Based on observations collected with the FLAMES spectrograph at VLT/UT2 telescope (Paranal Observatory, ESO, Chile), for the Gaia- ESO Large Public Survey (188.B-3002).Full Table 2 is only available at the CDS via anonymous ftp to http://cdsarc.u-strasbg.fr (ftp://130.79.128.5) or via http://cdsarc.u-strasbg.fr/viz-bin/qcat?J/A+A/580/A75

  18. Development of an empirically based dynamic biomechanical strength model

    NASA Technical Reports Server (NTRS)

    Pandya, A.; Maida, J.; Aldridge, A.; Hasson, S.; Woolford, B.

    1992-01-01

    The focus here is on the development of a dynamic strength model for humans. Our model is based on empirical data. The shoulder, elbow, and wrist joints are characterized in terms of maximum isolated torque, position, and velocity in all rotational planes. This information is reduced by a least squares regression technique into a table of single variable second degree polynomial equations determining the torque as a function of position and velocity. The isolated joint torque equations are then used to compute forces resulting from a composite motion, which in this case is a ratchet wrench push and pull operation. What is presented here is a comparison of the computed or predicted results of the model with the actual measured values for the composite motion.

  19. Developing an empirical base for clinical nurse specialist education.

    PubMed

    Stahl, Arleen M; Nardi, Deena; Lewandowski, Margaret A

    2008-01-01

    This article reports on the design of a clinical nurse specialist (CNS) education program using National Association of Clinical Nurse Specialists (NACNS) CNS competencies to guide CNS program clinical competency expectations and curriculum outcomes. The purpose is to contribute to the development of an empirical base for education and credentialing of CNSs. The NACNS CNS core competencies and practice competencies in all 3 spheres of influence guided the creation of clinical competency grids for this university's practicum courses. This project describes the development, testing, and application of these clinical competency grids that link the program's CNS clinical courses with the NACNS CNS competencies. These documents guide identification, tracking, measurement, and evaluation of the competencies throughout the clinical practice portion of the CNS program. This ongoing project will continue to provide data necessary to the benchmarking of CNS practice competencies, which is needed to evaluate the effectiveness of direct practice performance and the currency of graduate nursing education. PMID:18438164

  20. Arduino based radiation survey meter

    NASA Astrophysics Data System (ADS)

    Rahman, Nur Aira Abd; Lombigit, Lojius; Abdullah, Nor Arymaswati; Azman, Azraf; Dolah, Taufik; Muzakkir, Amir; Jaafar, Zainudin; Mohamad, Glam Hadzir Patai; Ramli, Abd Aziz Mhd; Zain, Rasif Mohd; Said, Fazila; Khalid, Mohd Ashhar; Taat, Muhamad Zahidee

    2016-01-01

    This paper presents the design of new digital radiation survey meter with LND7121 Geiger Muller tube detector and Atmega328P microcontroller. Development of the survey meter prototype is carried out on Arduino Uno platform. 16-bit Timer1 on the microcontroller is utilized as external pulse counter to produce count per second or CPS measurement. Conversion from CPS to dose rate technique is also performed by Arduino to display results in micro Sievert per hour (μSvhr-1). Conversion factor (CF) value for conversion of CPM to μSvhr-1 determined from manufacturer data sheet is compared with CF obtained from calibration procedure. The survey meter measurement results are found to be linear for dose rates below 3500 µSv/hr.

  1. Are smokers rational addicts? Empirical evidence from the Indonesian Family Life Survey

    PubMed Central

    2011-01-01

    Background Indonesia is one of the largest consumers of tobacco in the world, however there has been little work done on the economics addiction of tobacco. This study provides an empirical test of a rational addiction (henceforth RA) hypothesis of cigarette demand in Indonesia. Methods Four estimators (OLS, 2SLS, GMM, and System-GMM) were explored to test the RA hypothesis. The author adopted several diagnostics tests to select the best estimator to overcome econometric problems faced in presence of the past and future cigarette consumption (suspected endogenous variables). A short-run and long-run price elasticities of cigarettes demand was then calculated. The model was applied to individuals pooled data derived from three-waves a panel of the Indonesian Family Life Survey spanning the period 1993-2000. Results The past cigarette consumption coefficients turned out to be a positive with a p-value < 1%, implying that cigarettes indeed an addictive goods. The rational addiction hypothesis was rejected in favour of myopic ones. The short-run cigarette price elasticity for male and female was estimated to be-0.38 and -0.57, respectively, and the long-run one was -0.4 and -3.85, respectively. Conclusions Health policymakers should redesign current public health campaign against cigarette smoking in the country. Given the demand for cigarettes to be more prices sensitive for the long run (and female) than the short run (and male), an increase in the price of cigarettes could lead to a significant fall in cigarette consumption in the long run rather than as a constant source of government revenue. PMID:21345229

  2. Ontology-Based Empirical Knowledge Verification for Professional Virtual Community

    ERIC Educational Resources Information Center

    Chen, Yuh-Jen

    2011-01-01

    A professional virtual community provides an interactive platform for enterprise experts to create and share their empirical knowledge cooperatively, and the platform contains a tremendous amount of hidden empirical knowledge that knowledge experts have preserved in the discussion process. Therefore, enterprise knowledge management highly…

  3. Empirical Likelihood-Based ANOVA for Trimmed Means

    PubMed Central

    Velina, Mara; Valeinis, Janis; Greco, Luca; Luta, George

    2016-01-01

    In this paper, we introduce an alternative to Yuen’s test for the comparison of several population trimmed means. This nonparametric ANOVA type test is based on the empirical likelihood (EL) approach and extends the results for one population trimmed mean from Qin and Tsao (2002). The results of our simulation study indicate that for skewed distributions, with and without variance heterogeneity, Yuen’s test performs better than the new EL ANOVA test for trimmed means with respect to control over the probability of a type I error. This finding is in contrast with our simulation results for the comparison of means, where the EL ANOVA test for means performs better than Welch’s heteroscedastic F test. The analysis of a real data example illustrates the use of Yuen’s test and the new EL ANOVA test for trimmed means for different trimming levels. Based on the results of our study, we recommend the use of Yuen’s test for situations involving the comparison of population trimmed means between groups of interest. PMID:27690063

  4. Developing a Comprehensive, Empirically Based Research Framework for Classroom-Based Assessment

    ERIC Educational Resources Information Center

    Hill, Kathryn; McNamara, Tim

    2012-01-01

    This paper presents a comprehensive framework for researching classroom-based assessment (CBA) processes, and is based on a detailed empirical study of two Australian school classrooms where students aged 11 to 13 were studying Indonesian as a foreign language. The framework can be considered innovative in several respects. It goes beyond the…

  5. Children's Experiences of Completing a Computer-Based Violence Survey: Ethical Implications

    ERIC Educational Resources Information Center

    Ellonen, Noora; Poso, Tarja

    2011-01-01

    This article aims to contribute to the discussion about the ethics of research on children when studying sensitive issues such as violence. The empirical analysis is based on the accounts given by children (11 377) who completed a computer-based questionnaire about their experiences of violence ("The Finnish Child Victim Survey 2008") and their…

  6. Evidence-based ethics? On evidence-based practice and the "empirical turn" from normative bioethics

    PubMed Central

    Goldenberg, Maya J

    2005-01-01

    Background The increase in empirical methods of research in bioethics over the last two decades is typically perceived as a welcomed broadening of the discipline, with increased integration of social and life scientists into the field and ethics consultants into the clinical setting, however it also represents a loss of confidence in the typical normative and analytic methods of bioethics. Discussion The recent incipiency of "Evidence-Based Ethics" attests to this phenomenon and should be rejected as a solution to the current ambivalence toward the normative resolution of moral problems in a pluralistic society. While "evidence-based" is typically read in medicine and other life and social sciences as the empirically-adequate standard of reasonable practice and a means for increasing certainty, I propose that the evidence-based movement in fact gains consensus by displacing normative discourse with aggregate or statistically-derived empirical evidence as the "bottom line". Therefore, along with wavering on the fact/value distinction, evidence-based ethics threatens bioethics' normative mandate. The appeal of the evidence-based approach is that it offers a means of negotiating the demands of moral pluralism. Rather than appealing to explicit values that are likely not shared by all, "the evidence" is proposed to adjudicate between competing claims. Quantified measures are notably more "neutral" and democratic than liberal markers like "species normal functioning". Yet the positivist notion that claims stand or fall in light of the evidence is untenable; furthermore, the legacy of positivism entails the quieting of empirically non-verifiable (or at least non-falsifiable) considerations like moral claims and judgments. As a result, evidence-based ethics proposes to operate with the implicit normativity that accompanies the production and presentation of all biomedical and scientific facts unchecked. Summary The "empirical turn" in bioethics signals a need for

  7. STEAM: a software tool based on empirical analysis for micro electro mechanical systems

    NASA Astrophysics Data System (ADS)

    Devasia, Archana; Pasupuleti, Ajay; Sahin, Ferat

    2006-03-01

    In this research a generalized software framework that enables accurate computer aided design of MEMS devices is developed. The proposed simulation engine utilizes a novel material property estimation technique that generates effective material properties at the microscopic level. The material property models were developed based on empirical analysis and the behavior extraction of standard test structures. A literature review is provided on the physical phenomena that govern the mechanical behavior of thin films materials. This survey indicates that the present day models operate under a wide range of assumptions that may not be applicable to the micro-world. Thus, this methodology is foreseen to be an essential tool for MEMS designers as it would develop empirical models that relate the loading parameters, material properties, and the geometry of the microstructures with its performance characteristics. This process involves learning the relationship between the above parameters using non-parametric learning algorithms such as radial basis function networks and genetic algorithms. The proposed simulation engine has a graphical user interface (GUI) which is very adaptable, flexible, and transparent. The GUI is able to encompass all parameters associated with the determination of the desired material property so as to create models that provide an accurate estimation of the desired property. This technique was verified by fabricating and simulating bilayer cantilevers consisting of aluminum and glass (TEOS oxide) in our previous work. The results obtained were found to be very encouraging.

  8. Empirical relations between sense of coherence and self-efficacy, National Danish Survey.

    PubMed

    Trap, Rete; Rejkjær, Lillan; Hansen, Ebba Holme

    2016-09-01

    Salutogenic orientation is a health promotion paradigm focusing on the resources of the individual. This study analyzed the relationship between sense of coherence (SOC) and self-efficacy (SE) based on population data. By conducting an empirical analysis of the two models, we wanted to see whether we could make a valid judgement as to whether both SOC and SE could be utilized in health promotion practice, or whether one is preferable to the other. The study population was randomly selected from the Danish Central Population Register and consisted of five birth-year cohorts (1920, 1930, 1940, 1965 and 1975). The study used the 13-item SOC scale and the general SE scale. The main findings were that SOC score increased by age cohort (p = 0.0004), and there is a positive and graded correlation between SOC and SE (r = 0.39; p < 0.0001) and adjusted OR = 10.3 (CI = 6.7-15.4). We found the strongest association at the lowest level of SOC. For health promotion practice, this finding signifies the importance of focusing on improving SOC in people with a low SOC score, as they are most in need and most likely to increase their SOC level. The finding of higher SOC scores in the older age cohorts indicates that SOC changes over lifetime. Public health work focusing on lifestyle change by increasing SOC can be effective throughout life, however early intervention is important. The finding of a positive correlation between SOC and SE indicates that health promotion altering one of the constructs is paralleled in the other. PMID:26069296

  9. WATERSHED-BASED SURVEY DESIGNS

    EPA Science Inventory

    Water-based sampling design and assessment tools help serve the multiple goals for water quality monitoring required under the Clean Water Act, including assessment of regional conditions to meet Section 305(b), identification if impaired water bodies or watersheds to meet Sectio...

  10. EMT - Empirical-mode-decomposition-based Magneto-Telluric Processing

    NASA Astrophysics Data System (ADS)

    Neukirch, M.; Garcia, X.

    2012-04-01

    We present a new Magneto-Telluric (MT) data processing scheme based on an emerging non linear, non stationary time series analysis tool, called the Empirical Mode Decomposition (EMD) or Hilbert-Huang Transform (HHT), to transform data into a non-stationary frequency domain and a robust principal component regression to estimate the most likely MT transfer functions from the data with the 2-σ confidence intervals computed by a bootstrap algorithm. Optionally, data quality can be controlled by a physical coherence and a signal power filter. MT sources are assumed to be quasi stationary and therefore a (windowed) Fourier Transform is often applied to transform the time series into the frequency domain in which Transfer Functions (TF) are defined between the electromagnetic field components. This assumption can break down in the presence of noise or when the sources are non stationary, and then TF estimates can become unreliable when obtained through a stationary transform like the Fourier transform. Our TF estimation scheme naturally deals with non stationarity without introducing artifacts and, therefore, potentially can distinguish quasi-stationary sources and non-stationary noise. In contrast to previous works on using HHT for MT processing, we argue the necessity of a multivariate EMD to model the MT problem physically correctly and highlight the resulting possibility to use instantaneous parameters as independent and identically distributed variables. Furthermore, we define a homogenization between data channels of frequency discrepancies due to non stationarity and noise. The TF estimation in the frequency domain bases on a robust principal component analysis in order to find two source polarizations. These two principal components are used as predictor to regress robustly the data channels within a bootstrap algorithm to estimate the Earth's Transfer function with 2-σ confidence interval supplied by the measured data.The scheme can be used with and without

  11. A Survey of Graduate Training in Empirically Supported and Manualized Treatments: A Preliminary Report

    ERIC Educational Resources Information Center

    Karekla, Maria; Lundgren, Jennifer D.; Forsyth, John P.

    2004-01-01

    The promotion and dissemination of empirically supported (ESTs) and manualized therapies are important, albeit controversial, developments within clinical science and practice. To date, studies evaluating training opportunities and attitudes about such treatments at the graduate, predoctoral internship, and postdoctoral levels have focused on the…

  12. Do Master's in Social Work Programs Teach Empirically Supported Interventions? A Survey of Deans and Directors

    ERIC Educational Resources Information Center

    Woody, Jane D.; D'Souza, Henry J.; Dartman, Rebecca

    2006-01-01

    Objective: A questionnaire to examine efforts toward the teaching of empirically supported interventions (ESI) was mailed to the 165 deans and directors of Council on Social Work Education-accredited Master's in social work (MSW) programs; 66 (40%) responded. Method: Questions included program characteristics and items assessing both faculty and…

  13. Empirical Evaluation of Social, Psychological, & Educational Construct Variables Used in NCES Surveys. Working Paper Series.

    ERIC Educational Resources Information Center

    Freidlin, Boris; Salvucci, Sameena

    The purpose of this study was to provide an analysis and evaluation of composite variables in the National Education Longitudinal Study of 1988 (NELS:88) and School and Staffing Survey (SASS) in a way that provides guidance to the staff of the National Center for Education Statistics (NCES) in the more effective use of survey resources. The study…

  14. Lightning Detection Efficiency Analysis Process: Modeling Based on Empirical Data

    NASA Technical Reports Server (NTRS)

    Rompala, John T.

    2005-01-01

    A ground based lightning detection system employs a grid of sensors, which record and evaluate the electromagnetic signal produced by a lightning strike. Several detectors gather information on that signal s strength, time of arrival, and behavior over time. By coordinating the information from several detectors, an event solution can be generated. That solution includes the signal s point of origin, strength and polarity. Determination of the location of the lightning strike uses algorithms based on long used techniques of triangulation. Determination of the event s original signal strength relies on the behavior of the generated magnetic field over distance and time. In general the signal from the event undergoes geometric dispersion and environmental attenuation as it progresses. Our knowledge of that radial behavior together with the strength of the signal received by detecting sites permits an extrapolation and evaluation of the original strength of the lightning strike. It also limits the detection efficiency (DE) of the network. For expansive grids and with a sparse density of detectors, the DE varies widely over the area served. This limits the utility of the network in gathering information on regional lightning strike density and applying it to meteorological studies. A network of this type is a grid of four detectors in the Rondonian region of Brazil. The service area extends over a million square kilometers. Much of that area is covered by rain forests. Thus knowledge of lightning strike characteristics over the expanse is of particular value. I have been developing a process that determines the DE over the region [3]. In turn, this provides a way to produce lightning strike density maps, corrected for DE, over the entire region of interest. This report offers a survey of that development to date and a record of present activity.

  15. Comparison of empirical, semi-empirical and physically based models of soil hydraulic functions derived for bi-modal soils

    NASA Astrophysics Data System (ADS)

    Kutílek, M.; Jendele, L.; Krejča, M.

    2009-02-01

    The accelerated flow in soil pores is responsible for a rapid transport of pollutants from the soil surface to deeper layers up to groundwater. The term preferential flow is used for this type of transport. Our study was aimed at the preferential flow realized in the structural porous domain in bi-modal soils. We compared equations describing the soil water retention function h( θ) and unsaturated hydraulic conductivity K( h), eventually K( θ) modified for bi-modal soils, where θ is the soil water content and h is the pressure head. The analytical description of a curve passing experimental data sets of the soil hydraulic function is typical for the empirical equation characterized by fitting parameters only. If the measured data are described by the equation derived by the physical model without using fitting parameters, we speak about a physically based model. There exist several transitional subtypes between empirical and physically based models. They are denoted as semi-empirical, or semi-physical. We tested 3 models of soil water retention function and 3 models of unsaturated conductivity using experimental data sets of sand, silt, silt loam and loam. All used soils are typical by their bi-modality of the soil porous system. The model efficiency was estimated by RMSE (Root mean square error) and by RSE (Relative square error). The semi-empirical equation of the soil water retention function had the lowest values of RMSE and RSE and was qualified as "optimal" for the formal description of the shape of the water retention function. With this equation, the fit of the modelled data to experiments was the closest one. The fitting parameters smoothed the difference between the model and the physical reality of the soil porous media. The physical equation based upon the model of the pore size distribution did not allow exact fitting of the modelled data to the experimental data due to the rigidity and simplicity of the physical model when compared to the real soil

  16. Comparison of empirical, semi-empirical and physically based models of soil hydraulic functions derived for bi-modal soils.

    PubMed

    Kutílek, M; Jendele, L; Krejca, M

    2009-02-16

    The accelerated flow in soil pores is responsible for a rapid transport of pollutants from the soil surface to deeper layers up to groundwater. The term preferential flow is used for this type of transport. Our study was aimed at the preferential flow realized in the structural porous domain in bi-modal soils. We compared equations describing the soil water retention function h(theta) and unsaturated hydraulic conductivity K(h), eventually K(theta) modified for bi-modal soils, where theta is the soil water content and h is the pressure head. The analytical description of a curve passing experimental data sets of the soil hydraulic function is typical for the empirical equation characterized by fitting parameters only. If the measured data are described by the equation derived by the physical model without using fitting parameters, we speak about a physically based model. There exist several transitional subtypes between empirical and physically based models. They are denoted as semi-empirical, or semi-physical. We tested 3 models of soil water retention function and 3 models of unsaturated conductivity using experimental data sets of sand, silt, silt loam and loam. All used soils are typical by their bi-modality of the soil porous system. The model efficiency was estimated by RMSE (Root mean square error) and by RSE (Relative square error). The semi-empirical equation of the soil water retention function had the lowest values of RMSE and RSE and was qualified as "optimal" for the formal description of the shape of the water retention function. With this equation, the fit of the modelled data to experiments was the closest one. The fitting parameters smoothed the difference between the model and the physical reality of the soil porous media. The physical equation based upon the model of the pore size distribution did not allow exact fitting of the modelled data to the experimental data due to the rigidity and simplicity of the physical model when compared to the

  17. Accuracy of Population Validity and Cross-Validity Estimation: An Empirical Comparison of Formula-Based, Traditional Empirical, and Equal Weights Procedures.

    ERIC Educational Resources Information Center

    Raju, Nambury S.; Bilgic, Reyhan; Edwards, Jack E.; Fleer, Paul F.

    1999-01-01

    Performed an empirical Monte Carlo study using predictor and criterion data from 84,808 U.S. Air Force enlistees. Compared formula-based, traditional empirical, and equal-weights procedures. Discusses issues for basic research on validation and cross-validation. (SLD)

  18. A Comparison of Web-Based and Paper-Based Survey Methods: Testing Assumptions of Survey Mode and Response Cost

    ERIC Educational Resources Information Center

    Greenlaw, Corey; Brown-Welty, Sharon

    2009-01-01

    Web-based surveys have become more prevalent in areas such as evaluation, research, and marketing research to name a few. The proliferation of these online surveys raises the question, how do response rates compare with traditional surveys and at what cost? This research explored response rates and costs for Web-based surveys, paper surveys, and…

  19. National Stream Survey data-base guide

    SciTech Connect

    Mitch, M.E.; Kaufmann, P.R.; Herlihy, A.T.; Overton, W.S.; Sale, M.J.

    1990-07-01

    The National Stream Survey (NSS), conducted in the spring of 1985 and 1986, is one component of the U.S. Environmental Protection Agency's National Surface Water Survey. This effort is in support of the National Acid Precipitation Assessment Program. The NSS was a synoptic, spring survey of 500 streams in regions of the Southeastern and Mid-Atlantic United States expected to contain larger numbers of low alkalinity streams. The NSS is based on a probability sample from an explicitly defined population of surface waters. In the NSS, 500 streams were sampled, representing a regional population of 64,700 stream reaches. The NSS database includes stream and watershed physical characteristics, in situ measurements, and water chemistry data. Accompanying the database is a comprehensive user's guide that provides an overview of the NSS design, database structure, and transfer media.

  20. Landfill modelling in LCA - a contribution based on empirical data.

    PubMed

    Obersteiner, Gudrun; Binner, Erwin; Mostbauer, Peter; Salhofer, Stefan

    2007-01-01

    Landfills at various stages of development, depending on their age and location, can be found throughout Europe. The type of facilities goes from uncontrolled dumpsites to highly engineered facilities with leachate and gas management. In addition, some landfills are designed to receive untreated waste, while others can receive incineration residues (MSWI) or residues after mechanical biological treatment (MBT). Dimension, type and duration of the emissions from landfills depend on the quality of the disposed waste, the technical design, and the location of the landfill. Environmental impacts are produced by the leachate (heavy metals, organic loading), emissions into the air (CH(4), hydrocarbons, halogenated hydrocarbons) and from the energy or fuel requirements for the operation of the landfill (SO(2) and NO(x) from the production of electricity from fossil fuels). To include landfilling in an life-cycle assessment (LCA) approach entails several methodological questions (multi-input process, site-specific influence, time dependency). Additionally, no experiences are available with regard to mid-term behaviour (decades) for the relatively new types of landfill (MBT landfill, landfill for residues from MSWI). The present paper focuses on two main issues concerning modelling of landfills in LCA: Firstly, it is an acknowledged fact that emissions from landfills may prevail for a very long time, often thousands of years or longer. The choice of time frame in the LCA of landfilling may therefore clearly affect the results. Secondly, the reliability of results obtained through a life-cycle assessment depends on the availability and quality of Life Cycle Inventory (LCI) data. Therefore the choice of the general approach, using multi-input inventory tool versus empirical results, may also influence the results. In this paper the different approaches concerning time horizon and LCI will be introduced and discussed. In the application of empirical results, the presence of

  1. An Empirical Taxonomy of Youths' Fears: Cluster Analysis of the American Fear Survey Schedule

    ERIC Educational Resources Information Center

    Burnham, Joy J.; Schaefer, Barbara A.; Giesen, Judy

    2006-01-01

    Fears profiles among children and adolescents were explored using the Fear Survey Schedule for Children-American version (FSSC-AM; J.J. Burnham, 1995, 2005). Eight cluster profiles were identified via multistage Euclidean grouping and supported by homogeneity coefficients and replication. Four clusters reflected overall level of fears (i.e., very…

  2. [Psyche - soma - delinquency. Empirical and clinical datas of psychosomatic disturbed delinquents: discussion and survey].

    PubMed

    Kobbé, Ulrich

    2002-06-01

    Psychosomatic disorders of delinquents are current clinical phenomena, but they are rarely examined and discussed. The contribution presents empirical-statistical results of a field-study with n = 202 unselected forensic patients in disciplinary measures. Psychosomatic diseases have been diagnosed in n = 48 cases. The interpretation of multi- and bivariate statistics hints at an iatrogene influence on detainees: Their psychosomatic disorders can be described as forms of autoplastic defence which must be considered in the context of specific basic personality structures and corresponding disturbed ego functions (control of impulses and anxiety, competence of communication and sublimation). This psychosomatic structure can be differentiated from perverse structured personalities. Characterised as an autoplastic and auto-aggressive reaction, the psychosomatic symptom can be distinguished from alloplastic and hetero-aggressive acting. Nevertheless these psychodynamics seem to be forensic-psychological predicators for anticipated delinquent relapses.

  3. Attachment-Based Family Therapy: A Review of the Empirical Support.

    PubMed

    Diamond, Guy; Russon, Jody; Levy, Suzanne

    2016-09-01

    Attachment-based family therapy (ABFT) is an empirically supported treatment designed to capitalize on the innate, biological desire for meaningful and secure relationships. The therapy is grounded in attachment theory and provides an interpersonal, process-oriented, trauma-focused approach to treating adolescent depression, suicidality, and trauma. Although a process-oriented therapy, ABFT offers a clear structure and road map to help therapists quickly address attachment ruptures that lie at the core of family conflict. Several clinical trials and process studies have demonstrated empirical support for the model and its proposed mechanism of change. This article provides an overview of the clinical model and the existing empirical support for ABFT.

  4. An empirical formula based on Monte Carlo simulation for diffuse reflectance from turbid media

    NASA Astrophysics Data System (ADS)

    Gnanatheepam, Einstein; Aruna, Prakasa Rao; Ganesan, Singaravelu

    2016-03-01

    Diffuse reflectance spectroscopy has been widely used in diagnostic oncology and characterization of laser irradiated tissue. However, still accurate and simple analytical equation does not exist for estimation of diffuse reflectance from turbid media. In this work, a diffuse reflectance lookup table for a range of tissue optical properties was generated using Monte Carlo simulation. Based on the generated Monte Carlo lookup table, an empirical formula for diffuse reflectance was developed using surface fitting method. The variance between the Monte Carlo lookup table surface and the surface obtained from the proposed empirical formula is less than 1%. The proposed empirical formula may be used for modeling of diffuse reflectance from tissue.

  5. Empirical Analysis and Refinement of Expert System Knowledge Bases

    PubMed Central

    Weiss, Sholom M.; Politakis, Peter; Ginsberg, Allen

    1986-01-01

    Recent progress in knowledge base refinement for expert systems is reviewed. Knowledge base refinement is characterized by the constrained modification of rule-components in an existing knowledge base. The goals are to localize specific weaknesses in a knowledge base and to improve an expert system's performance. Systems that automate some aspects of knowledge base refinement can have a significant impact on the related problems of knowledge base acquisition, maintenance, verification, and learning from experience. The SEEK empiricial analysis and refinement system is reviewed and its successor system, SEEK2, is introduced. Important areas for future research in knowledge base refinement are described.

  6. Deep in Data. Empirical Data Based Software Accuracy Testing Using the Building America Field Data Repository

    SciTech Connect

    Neymark, J.; Roberts, D.

    2013-06-01

    This paper describes progress toward developing a usable, standardized, empirical data-based software accuracy test suite using home energy consumption and building description data. Empirical data collected from around the United States have been translated into a uniform Home Performance Extensible Markup Language format that may enable software developers to create translators to their input schemes for efficient access to the data. This could allow for modeling many homes expediently, and thus implementing software accuracy test cases by applying the translated data.

  7. An Empirical Analysis of Knowledge Based Hypertext Navigation

    PubMed Central

    Snell, J.R.; Boyle, C.

    1990-01-01

    Our purpose is to investigate the effectiveness of knowledge-based navigation in a dermatology hypertext network. The chosen domain is a set of dermatology class notes implemented in Hypercard and SINS. The study measured time, number of moves, and success rates for subjects to find solutions to ten questions. The subjects were required to navigate within a dermatology hypertext network in order to find the solutions to a question. Our results indicate that knowledge-based navigation can assist the user in finding information of interest in a fewer number of node visits (moves) than with traditional button-based browsing or keyword searching. The time necessary to find an item of interest was lower for traditional-based methods. There was no difference in success rates for the two test groups.

  8. An empirical determination of the dust mass absorption coefficient, κd, using the Herschel Reference Survey

    NASA Astrophysics Data System (ADS)

    Clark, Christopher J. R.; Schofield, Simon P.; Gomez, Haley L.; Davies, Jonathan I.

    2016-06-01

    We use the published photometry and spectroscopy of 22 galaxies in the Herschel Reference Survey to determine that the value of the dust mass absorption coefficient κd at a wavelength of 500 μm is kappa _{500} = 0.051^{+0.070}_{-0.026} m^{2 kg^{-1}}. We do so by taking advantage of the fact that the dust-to-metals ratio in the interstellar medium of galaxies appears to be constant. We argue that our value for κd supersedes that of James et al. - who pioneered this approach for determining κd - because we take advantage of superior data, and account for a number of significant systematic effects that they did not consider. We comprehensively incorporate all methodological and observational contributions to establish the uncertainty on our value, which represents a marked improvement on the oft-quoted `order-of-magnitude' uncertainty on κd. We find no evidence that the value of κd differs significantly between galaxies, or that it correlates with any other measured or derived galaxy properties. We note, however, that the availability of data limits our sample to relatively massive (109.7 < M⋆ < 1011.0 M⊙), high metallicity (8.61 < [ 12 + log_{10} fracOH ] < 8.86) galaxies; future work will allow us to investigate a wider range of systems.

  9. Performance-Based Service Quality Model: An Empirical Study on Japanese Universities

    ERIC Educational Resources Information Center

    Sultan, Parves; Wong, Ho

    2010-01-01

    Purpose: This paper aims to develop and empirically test the performance-based higher education service quality model. Design/methodology/approach: The study develops 67-item instrument for measuring performance-based service quality with a particular focus on the higher education sector. Scale reliability is confirmed using the Cronbach's alpha.…

  10. Empirically Based School Interventions Targeted at Academic and Mental Health Functioning

    ERIC Educational Resources Information Center

    Hoagwood, Kimberly E.; Olin, S. Serene; Kerker, Bonnie D.; Kratochwill, Thomas R.; Crowe, Maura; Saka, Noa

    2007-01-01

    This review examines empirically based studies of school-based mental health interventions. The review identified 64 out of more than 2,000 articles published between 1990 and 2006 that met methodologically rigorous criteria for inclusion. Of these 64 articles, only 24 examined both mental health "and" educational outcomes. The majority of…

  11. Semantically-Based Child Grammars: Some Empirical Inadequacies.

    ERIC Educational Resources Information Center

    Hyams, Nina

    It is argued that the general consensus of researchers of child language that the grammatical system underlying the child's earliest multiword utterances is semantically-based, fails to provide an adequate description of even the earliest multiword utterances, and that the most sparing account of the acquisition data must include reference to…

  12. Empirical comparison of structure-based pathway methods

    PubMed Central

    Jaakkola, Maria K.

    2016-01-01

    Multiple methods have been proposed to estimate pathway activities from expression profiles, and yet, there is not enough information available about the performance of those methods. This makes selection of a suitable tool for pathway analysis difficult. Although methods based on simple gene lists have remained the most common approach, various methods that also consider pathway structure have emerged. To provide practical insight about the performance of both list-based and structure-based methods, we tested six different approaches to estimate pathway activities in two different case study settings of different characteristics. The first case study setting involved six renal cell cancer data sets, and the differences between expression profiles of case and control samples were relatively big. The second case study setting involved four type 1 diabetes data sets, and the profiles of case and control samples were more similar to each other. In general, there were marked differences in the outcomes of the different pathway tools even with the same input data. In the cancer studies, the results of a tested method were typically consistent across the different data sets, yet different between the methods. In the more challenging diabetes studies, almost all the tested methods detected as significant only few pathways if any. PMID:26197809

  13. Polymer electrolyte membrane fuel cell fault diagnosis based on empirical mode decomposition

    NASA Astrophysics Data System (ADS)

    Damour, Cédric; Benne, Michel; Grondin-Perez, Brigitte; Bessafi, Miloud; Hissel, Daniel; Chabriat, Jean-Pierre

    2015-12-01

    Diagnosis tool for water management is relevant to improve the reliability and lifetime of polymer electrolyte membrane fuel cells (PEMFCs). This paper presents a novel signal-based diagnosis approach, based on Empirical Mode Decomposition (EMD), dedicated to PEMFCs. EMD is an empirical, intuitive, direct and adaptive signal processing method, without pre-determined basis functions. The proposed diagnosis approach relies on the decomposition of FC output voltage to detect and isolate flooding and drying faults. The low computational cost of EMD, the reduced number of required measurements, and the high diagnosis accuracy of flooding and drying faults diagnosis make this approach a promising online diagnosis tool for PEMFC degraded modes management.

  14. The Empirical Investigation of Perspective-Based Reading

    NASA Technical Reports Server (NTRS)

    Basili, Victor R.; Green, Scott; Laitenberger, Oliver; Shull, Forrest; Sorumgard, Sivert; Zelkowitz, Marvin V.

    1996-01-01

    We consider reading techniques a fundamental means of achieving high quality software. Due to the lack of research in this area, we are experimenting with the application and comparison of various reading techniques. This paper deals with our experiences with Perspective-Based Reading (PBR), a particular reading technique for requirements documents. The goal of PBR is to provide operational scenarios where members of a review team read a document from a particular perspective (e.g., tester, developer, user). Our assumption is that the combination of different perspectives provides better coverage of the document than the same number of readers using their usual technique.

  15. Behavioral Modeling Based on Probabilistic Finite Automata: An Empirical Study.

    PubMed

    Tîrnăucă, Cristina; Montaña, José L; Ontañón, Santiago; González, Avelino J; Pardo, Luis M

    2016-01-01

    Imagine an agent that performs tasks according to different strategies. The goal of Behavioral Recognition (BR) is to identify which of the available strategies is the one being used by the agent, by simply observing the agent's actions and the environmental conditions during a certain period of time. The goal of Behavioral Cloning (BC) is more ambitious. In this last case, the learner must be able to build a model of the behavior of the agent. In both settings, the only assumption is that the learner has access to a training set that contains instances of observed behavioral traces for each available strategy. This paper studies a machine learning approach based on Probabilistic Finite Automata (PFAs), capable of achieving both the recognition and cloning tasks. We evaluate the performance of PFAs in the context of a simulated learning environment (in this case, a virtual Roomba vacuum cleaner robot), and compare it with a collection of other machine learning approaches. PMID:27347956

  16. Navigating pitfalls of web-based survey development and administration.

    PubMed

    Littleton, Dawn

    2007-01-01

    With multiple low-cost Web-based delivery methods increasingly available, the number of surveys being distributed to gather data on library users and potential users is dramatically increasing. This paper provides an overview of survey administration including history, guidelines on decision-making in survey development, and best practices for survey research when delivered via the Web. Questions on survey development issues such as probability and non-probability sampling and privacy concerns are clarified.

  17. The Empirical Investigation of Perspective-Based Reading

    NASA Technical Reports Server (NTRS)

    Basili, Victor R.; Green, Scott; Laitenberger, Oliver; Shull, Forrest; Sorumgard, Sivert; Zelkowitz, Marvin V.

    1995-01-01

    We consider reading techniques a fundamental means of achieving high quality software. Due to lack of research in this area, we are experimenting with the application and comparison of various reading techniques. This paper deals with our experiences with Perspective Based Reading (PBR) a particular reading technique for requirement documents. The goal of PBR is to provide operation scenarios where members of a review team read a document from a particular perspective (eg., tester, developer, user). Our assumption is that the combination of different perspective provides better coverage of the document than the same number of readers using their usual technique. To test the efficacy of PBR, we conducted two runs of a controlled experiment in the environment of NASA GSFC Software Engineering Laboratory (SEL), using developers from the environment. The subjects read two types of documents, one generic in nature and the other from the NASA Domain, using two reading techniques, PBR and their usual technique. The results from these experiment as well as the experimental design, are presented and analyzed. When there is a statistically significant distinction, PBR performs better than the subjects' usual technique. However, PBR appears to be more effective on the generic documents than on the NASA documents.

  18. Towards an Empirically Based Parametric Explosion Spectral Model

    SciTech Connect

    Ford, S R; Walter, W R; Ruppert, S; Matzel, E; Hauk, T; Gok, R

    2009-08-31

    Small underground nuclear explosions need to be confidently detected, identified, and characterized in regions of the world where they have never before been tested. The focus of our work is on the local and regional distances (< 2000 km) and phases (Pn, Pg, Sn, Lg) necessary to see small explosions. We are developing a parametric model of the nuclear explosion seismic source spectrum that is compatible with the earthquake-based geometrical spreading and attenuation models developed using the Magnitude Distance Amplitude Correction (MDAC) techniques (Walter and Taylor, 2002). The explosion parametric model will be particularly important in regions without any prior explosion data for calibration. The model is being developed using the available body of seismic data at local and regional distances for past nuclear explosions at foreign and domestic test sites. Parametric modeling is a simple and practical approach for widespread monitoring applications, prior to the capability to carry out fully deterministic modeling. The achievable goal of our parametric model development is to be able to predict observed local and regional distance seismic amplitudes for event identification and yield determination in regions with incomplete or no prior history of underground nuclear testing. The relationship between the parametric equations and the geologic and containment conditions will assist in our physical understanding of the nuclear explosion source.

  19. Introduction to the Application of Web-Based Surveys.

    ERIC Educational Resources Information Center

    Timmerman, Annemarie

    This paper discusses some basic assumptions and issues concerning web-based surveys. Discussion includes: assumptions regarding cost and ease of use; disadvantages of web-based surveys, concerning the inability to compensate for four common errors of survey research: coverage error, sampling error, measurement error and nonresponse error; and…

  20. Understanding why users tag: A survey of tagging motivation literature and results from an empirical study

    PubMed Central

    Strohmaier, Markus; Körner, Christian; Kern, Roman

    2012-01-01

    While recent progress has been achieved in understanding the structure and dynamics of social tagging systems, we know little about the underlying user motivations for tagging, and how they influence resulting folksonomies and tags. This paper addresses three issues related to this question. (1) What distinctions of user motivations are identified by previous research, and in what ways are the motivations of users amenable to quantitative analysis? (2) To what extent does tagging motivation vary across different social tagging systems? (3) How does variability in user motivation influence resulting tags and folksonomies? In this paper, we present measures to detect whether a tagger is primarily motivated by categorizing or describing resources, and apply these measures to datasets from seven different tagging systems. Our results show that (a) users’ motivation for tagging varies not only across, but also within tagging systems, and that (b) tag agreement among users who are motivated by categorizing resources is significantly lower than among users who are motivated by describing resources. Our findings are relevant for (1) the development of tag-based user interfaces, (2) the analysis of tag semantics and (3) the design of search algorithms for social tagging systems. PMID:23471473

  1. Untangling the Evidence: Introducing an Empirical Model for Evidence-Based Library and Information Practice

    ERIC Educational Resources Information Center

    Gillespie, Ann

    2014-01-01

    Introduction: This research is the first to investigate the experiences of teacher-librarians as evidence-based practice. An empirically derived model is presented in this paper. Method: This qualitative study utilised the expanded critical incident approach, and investigated the real-life experiences of fifteen Australian teacher-librarians,…

  2. Satellite-based empirical models linking river plume dynamics with hypoxic area andvolume

    EPA Science Inventory

    Satellite-based empirical models explaining hypoxic area and volume variation were developed for the seasonally hypoxic (O2 < 2 mg L−1) northern Gulf of Mexico adjacent to the Mississippi River. Annual variations in midsummer hypoxic area and ...

  3. An Empirically Based Method of Q-Matrix Validation for the DINA Model: Development and Applications

    ERIC Educational Resources Information Center

    de la Torre, Jimmy

    2008-01-01

    Most model fit analyses in cognitive diagnosis assume that a Q matrix is correct after it has been constructed, without verifying its appropriateness. Consequently, any model misfit attributable to the Q matrix cannot be addressed and remedied. To address this concern, this paper proposes an empirically based method of validating a Q matrix used…

  4. An Empirically-Based Statewide System for Identifying Quality Pre-Kindergarten Programs

    ERIC Educational Resources Information Center

    Williams, Jeffrey M.; Landry, Susan H.; Anthony, Jason L.; Swank, Paul R.; Crawford, April D.

    2012-01-01

    This study presents an empirically-based statewide system that links information about pre-kindergarten programs with children's school readiness scores to certify pre-kindergarten classrooms as promoting school readiness. Over 8,000 children from 1,255 pre-kindergarten classrooms were followed longitudinally for one year. Pre-kindergarten quality…

  5. Empirically Derived Consequences: A Data-Based Method for Prescribing Treatments for Destructive Behavior.

    ERIC Educational Resources Information Center

    Fisher, Wayne; And Others

    1994-01-01

    This study used a data-based assessment to identify reinforcers and punishers for successful treatment of two children with severe destructive behaviors. Results suggested that empirically derived consequences may be useful in decreasing destructive behavior when a functional assessment is inconclusive or suggests that internal stimuli are…

  6. Feasibility of an Empirically Based Program for Parents of Preschoolers with Autism Spectrum Disorder

    ERIC Educational Resources Information Center

    Dababnah, Sarah; Parish, Susan L.

    2016-01-01

    This article reports on the feasibility of implementing an existing empirically based program, "The Incredible Years," tailored to parents of young children with autism spectrum disorder. Parents raising preschool-aged children (aged 3-6?years) with autism spectrum disorder (N?=?17) participated in a 15-week pilot trial of the…

  7. Multisystemic Therapy: An Empirically Supported, Home-Based Family Therapy Approach.

    ERIC Educational Resources Information Center

    Sheidow, Ashli J.; Woodford, Mark S.

    2003-01-01

    Multisystemic Therapy (MST) is a well-validated, evidenced-based treatment for serious clinical problems presented by adolescents and their families. This article is an introduction to the MST approach and outlines key clinical features, describes the theoretical underpinnings, and discusses the empirical support for MST's effectiveness with a…

  8. Empirical vs. Expected IRT-Based Reliability Estimation in Computerized Multistage Testing (MST)

    ERIC Educational Resources Information Center

    Zhang, Yanwei; Breithaupt, Krista; Tessema, Aster; Chuah, David

    2006-01-01

    Two IRT-based procedures to estimate test reliability for a certification exam that used both adaptive (via a MST model) and non-adaptive design were considered in this study. Both procedures rely on calibrated item parameters to estimate error variance. In terms of score variance, one procedure (Method 1) uses the empirical ability distribution…

  9. Development of an Empirically Based Questionnaire to Investigate Young Students' Ideas about Nature of Science

    ERIC Educational Resources Information Center

    Chen, Sufen; Chang, Wen-Hua; Lieu, Sang-Chong; Kao, Huey-Lien; Huang, Mao-Tsai; Lin, Shu-Fen

    2013-01-01

    This study developed an empirically based questionnaire to monitor young learners' conceptions of nature of science (NOS). The questionnaire, entitled Students' Ideas about Nature of Science (SINOS), measured views on theory-ladenness, use of creativity and imagination, tentativeness of scientific knowledge, durability of scientific knowledge,…

  10. Implementing Evidence-Based Practice: A Review of the Empirical Research Literature

    ERIC Educational Resources Information Center

    Gray, Mel; Joy, Elyssa; Plath, Debbie; Webb, Stephen A.

    2013-01-01

    The article reports on the findings of a review of empirical studies examining the implementation of evidence-based practice (EBP) in the human services. Eleven studies were located that defined EBP as a research-informed, clinical decision-making process and identified barriers and facilitators to EBP implementation. A thematic analysis of the…

  11. Training Community-Based Professionals to Implement an Empirically Supported Parenting Program

    ERIC Educational Resources Information Center

    Fox, Robert A.; Duffy, Kathleen M.; Keller, Kathryn M.

    2006-01-01

    Professionals representing 14 community-based organizations were trained at three different sites serving urban and rural families to implement an empirically supported parenting program for families of young children with challenging behaviors. Of the 44 practitioners trained, 23 successfully completed the program, which involved passing a…

  12. Empirically-Based Predictions and Their Generalizability for Decision-Making in Occupational Education

    ERIC Educational Resources Information Center

    Passmore, David L.; Irvin, Donald E.

    1974-01-01

    The generation of empirically-based equations is within the grasp of most researchers in occupational education. This paper was designed to develop a sensitivity to the need for cross-validation in these circumstances. Program CROSVAL is a means of bridging the gap between the need for, and mechanization of, cross validation. (Author/DS)

  13. Attachment-Based Family Therapy: A Review of the Empirical Support.

    PubMed

    Diamond, Guy; Russon, Jody; Levy, Suzanne

    2016-09-01

    Attachment-based family therapy (ABFT) is an empirically supported treatment designed to capitalize on the innate, biological desire for meaningful and secure relationships. The therapy is grounded in attachment theory and provides an interpersonal, process-oriented, trauma-focused approach to treating adolescent depression, suicidality, and trauma. Although a process-oriented therapy, ABFT offers a clear structure and road map to help therapists quickly address attachment ruptures that lie at the core of family conflict. Several clinical trials and process studies have demonstrated empirical support for the model and its proposed mechanism of change. This article provides an overview of the clinical model and the existing empirical support for ABFT. PMID:27541199

  14. Empirical metallicity-dependent calibrations of effective temperature against colours for dwarfs and giants based on interferometric data

    NASA Astrophysics Data System (ADS)

    Huang, Y.; Liu, X.-W.; Yuan, H.-B.; Xiang, M.-S.; Chen, B.-Q.; Zhang, H.-W.

    2015-12-01

    We present empirical metallicity-dependent calibrations of effective temperature against colours for dwarfs of luminosity classes IV and V and for giants of luminosity classes II and III, based on a collection from the literature of about two hundred nearby stars with direct effective temperature measurements of better than 2.5 per cent. The calibrations are valid for an effective temperature range 3100-10 000 K for dwarfs of spectral types M5 to A0 and 3100-5700 K for giants of spectral types K5 to G5. A total of 21 colours for dwarfs and 18 colours for giants of bands of four photometric systems, i.e. the Johnson (UBVRJIJJHK), the Cousins (RCIC), the Sloan Digital Sky Survey (gr) and the Two Micron All Sky Survey (JHKs), have been calibrated. Restricted by the metallicity range of the current sample, the calibrations are mainly applicable for disc stars ([Fe/H] ≳ - 1.0). The normalized percentage residuals of the calibrations are typically 2.0 and 1.5 per cent for dwarfs and giants, respectively. Some systematic discrepancies at various levels are found between the current scales and those available in the literature (e.g. those based on the infrared flux method or spectroscopy). Based on the current calibrations, we have re-determined the colours of the Sun. We have also investigated the systematic errors in effective temperatures yielded by the current on-going large-scale low- to intermediate-resolution stellar spectroscopic surveys. We show that the calibration of colour (g - Ks) presented in this work provides an invaluable tool for the estimation of stellar effective temperature for those on-going or upcoming surveys.

  15. [Empirically based early intervention programs for children with autistic disorders - a selective literature review].

    PubMed

    Freitag, Christine M

    2010-07-01

    Autistic Disorders (AD) are characterized by impairments in social interaction and communication, as well as by stereotyped behaviors and interests. Early intervention programs in AD aim to improve several aspects of the child's abilities: joint attention, play abilities, language development, and especially social interaction and communication. In this review article based on a selective literature search, the relatively best empirically based early intervention programs will be discussed with a focus on the proven efficacy of these interventions.

  16. Increasing Response Rates to Web-Based Surveys

    ERIC Educational Resources Information Center

    Monroe, Martha C.; Adams, Damian C.

    2012-01-01

    We review a popular method for collecing data--Web-based surveys. Although Web surveys are popular, one major concern is their typically low response rates. Using the Dillman et al. (2009) approach, we designed, pre-tested, and implemented a survey on climate change with Extension professionals in the Southeast. The Dillman approach worked well,…

  17. Increasing Your Productivity with Web-Based Surveys

    ERIC Educational Resources Information Center

    Wissmann, Mary; Stone, Brittney; Schuster, Ellen

    2012-01-01

    Web-based survey tools such as Survey Monkey can be used in many ways to increase the efficiency and effectiveness of Extension professionals. This article describes how Survey Monkey has been used at the state and county levels to collect community and internal staff information for the purposes of program planning, administration, evaluation and…

  18. Triangulation-based 3D surveying borescope

    NASA Astrophysics Data System (ADS)

    Pulwer, S.; Steglich, P.; Villringer, C.; Bauer, J.; Burger, M.; Franz, M.; Grieshober, K.; Wirth, F.; Blondeau, J.; Rautenberg, J.; Mouti, S.; Schrader, S.

    2016-04-01

    In this work, a measurement concept based on triangulation was developed for borescopic 3D-surveying of surface defects. The integration of such measurement system into a borescope environment requires excellent space utilization. The triangulation angle, the projected pattern, the numerical apertures of the optical system, and the viewing angle were calculated using partial coherence imaging and geometric optical raytracing methods. Additionally, optical aberrations and defocus were considered by the integration of Zernike polynomial coefficients. The measurement system is able to measure objects with a size of 50 μm in all dimensions with an accuracy of +/- 5 μm. To manage the issue of a low depth of field while using an optical high resolution system, a wavelength dependent aperture was integrated. Thereby, we are able to control depth of field and resolution of the optical system and can use the borescope in measurement mode with high resolution and low depth of field or in inspection mode with low resolution and higher depth of field. First measurements of a demonstrator system are in good agreement with our simulations.

  19. Scaling up explanation generation: Large-scale knowledge bases and empirical studies

    SciTech Connect

    Lester, J.C.; Porter, B.W.

    1996-12-31

    To explain complex phenomena, an explanation system must be able to select information from a formal representation of domain knowledge, organize the selected information into multisentential discourse plans, and realize the discourse plans in text. Although recent years have witnessed significant progress in the development of sophisticated computational mechanisms for explanation, empirical results have been limited. This paper reports on a seven year effort to empirically study explanation generation from semantically rich, large-scale knowledge bases. We first describe Knight, a robust explanation system that constructs multi-sentential and multi-paragraph explanations from the Biology Knowledge Base, a large-scale knowledge base in the domain of botanical anatomy, physiology, and development. We then introduce the Two Panel evaluation methodology and describe how Knight`s performance was assessed with this methodology in the most extensive empirical evaluation conducted on an explanation system. In this evaluation, Knight scored within {open_quotes}half a grade{close_quote} of domain experts, and its performance exceeded that of one of the domain experts.

  20. Survey Says? A Primer on Web-based Survey Design and Distribution

    PubMed Central

    Oppenheimer, Adam J.; Pannucci, Christopher J.; Kasten, Steven J.; Haase, Steven C.

    2011-01-01

    The internet has changed the way in which we gather and interpret information. While books were once the exclusive bearers of data, knowledge is now only a keystroke away. The internet has also facilitated the synthesis of new knowledge. Specifically, it has become a tool through which medical research is conducted. A review of the literature reveals that in the past year, over one-hundred medical publications have been based on web-based survey data alone. Due to emerging internet technologies, web-based surveys can now be launched with little computer knowledge. They may also be self-administered, eliminating personnel requirements. Ultimately, an investigator may build, implement, and analyze survey results with speed and efficiency, obviating the need for mass mailings and data processing. All of these qualities have rendered telephone and mail-based surveys virtually obsolete. Despite these capabilities, web-based survey techniques are not without their limitations, namely recall and response biases. When used properly, however, web-based surveys can greatly simplify the research process. This article discusses the implications of web-based surveys and provides guidelines for their effective design and distribution. PMID:21701347

  1. An empirical mass-loss law for Population II giants from the Spitzer-IRAC survey of Galactic globular clusters

    NASA Astrophysics Data System (ADS)

    Origlia, L.; Ferraro, F. R.; Fabbri, S.; Fusi Pecci, F.; Dalessandro, E.; Rich, R. M.; Valenti, E.

    2014-04-01

    Aims: The main aim of the present work is to derive an empirical mass-loss (ML) law for Population II stars in first and second ascent red giant branches. Methods: We used the Spitzer InfraRed Array Camera (IRAC) photometry obtained in the 3.6-8 μm range of a carefully chosen sample of 15 Galactic globular clusters spanning the entire metallicity range and sampling the vast zoology of horizontal branch (HB) morphologies. We complemented the IRAC photometry with near-infrared data to build suitable color-magnitude and color-color diagrams and identify mass-losing giant stars. Results: We find that while the majority of stars show colors typical of cool giants, some stars show an excess of mid-infrared light that is larger than expected from their photospheric emission and that is plausibly due to dust formation in mass flowing from them. For these stars, we estimate dust and total (gas + dust) ML rates and timescales. We finally calibrate an empirical ML law for Population II red and asymptotic giant branch stars with varying metallicity. We find that at a given red giant branch luminosity only a fraction of the stars are losing mass. From this, we conclude that ML is episodic and is active only a fraction of the time, which we define as the duty cycle. The fraction of mass-losing stars increases by increasing the stellar luminosity and metallicity. The ML rate, as estimated from reasonable assumptions for the gas-to-dust ratio and expansion velocity, depends on metallicity and slowly increases with decreasing metallicity. In contrast, the duty cycle increases with increasing metallicity, with the net result that total ML increases moderately with increasing metallicity, about 0.1 M⊙ every dex in [Fe/H]. For Population II asymptotic giant branch stars, we estimate a total ML of ≤0.1 M⊙, nearly constant with varying metallicity. This work is based on observations made with the Spitzer Space Telescope, which is operated by the Jet Propulsion Laboratory

  2. Web-Based Surveys Facilitate Undergraduate Research and Knowledge

    ERIC Educational Resources Information Center

    Grimes, Paul, Ed.; Steele, Scott R.

    2008-01-01

    The author presents Web-based surveying as a valuable tool for achieving quality undergraduate research in upper-level economics courses. Web-based surveys can be employed in efforts to integrate undergraduate research into the curriculum without overburdening students or faculty. The author discusses the value of undergraduate research, notes…

  3. School-Based Health Care State Policy Survey. Executive Summary

    ERIC Educational Resources Information Center

    National Assembly on School-Based Health Care, 2012

    2012-01-01

    The National Assembly on School-Based Health Care (NASBHC) surveys state public health and Medicaid offices every three years to assess state-level public policies and activities that promote the growth and sustainability of school-based health services. The FY2011 survey found 18 states (see map below) reporting investments explicitly dedicated…

  4. Outcome (competency) based education: an exploration of its origins, theoretical basis, and empirical evidence.

    PubMed

    Morcke, Anne Mette; Dornan, Tim; Eika, Berit

    2013-10-01

    Outcome based or competency based education (OBE) is so firmly established in undergraduate medical education that it might not seem necessary to ask why it was included in recommendations for the future, like the Flexner centenary report. Uncritical acceptance may not, however, deliver its greatest benefits. Our aim was to explore the underpinnings of OBE: its historical origins, theoretical basis, and empirical evidence of its effects in order to answer the question: How can predetermined learning outcomes influence undergraduate medical education? This literature review had three components: A review of historical landmarks in the evolution of OBE; a review of conceptual frameworks and theories; and a systematic review of empirical publications from 1999 to 2010 that reported data concerning the effects of learning outcomes on undergraduate medical education. OBE had its origins in behaviourist theories of learning. It is tightly linked to the assessment and regulation of proficiency, but less clearly linked to teaching and learning activities. Over time, there have been cycles of advocacy for, then criticism of, OBE. A recurring critique concerns the place of complex personal and professional attributes as "competencies". OBE has been adopted by consensus in the face of weak empirical evidence. OBE, which has been advocated for over 50 years, can contribute usefully to defining requisite knowledge and skills, and blueprinting assessments. Its applicability to more complex aspects of clinical performance is not clear. OBE, we conclude, provides a valuable approach to some, but not all, important aspects of undergraduate medical education.

  5. Band structure calculation of GaSe-based nanostructures using empirical pseudopotential method

    NASA Astrophysics Data System (ADS)

    Osadchy, A. V.; Volotovskiy, S. G.; Obraztsova, E. D.; Savin, V. V.; Golovashkin, D. L.

    2016-08-01

    In this paper we present the results of band structure computer simulation of GaSe- based nanostructures using the empirical pseudopotential method. Calculations were performed using a specially developed software that allows performing simulations using cluster computing. Application of this method significantly reduces the demands on computing resources compared to traditional approaches based on ab-initio techniques and provides receiving the adequate comparable results. The use of cluster computing allows to obtain information for structures that require an explicit account of a significant number of atoms, such as quantum dots and quantum pillars.

  6. Methodologies for Crawler Based Web Surveys.

    ERIC Educational Resources Information Center

    Thelwall, Mike

    2002-01-01

    Describes Web survey methodologies used to study the content of the Web, and discusses search engines and the concept of crawling the Web. Highlights include Web page selection methodologies; obstacles to reliable automatic indexing of Web sites; publicly indexable pages; crawling parameters; and tests for file duplication. (Contains 62…

  7. Scan MDCs for GPS-Based Gamma Radiation Surveys.

    PubMed

    Alecksen, Tyler; Whicker, Randy

    2016-08-01

    A method for estimating the minimum detectable concentration of a contaminant radionuclide in soil when scanning with gamma radiation detectors (known as the "scan MDC") is described in the Multi-Agency Radiation Survey and Site Investigation Manual (MARSSIM). This paper presents an alternate method for estimating scan MDCs for GPS-based gamma surveys based on detector efficiencies modeled with the probabilistic Monte Carlo N-Particle Extended (MCNPX) Transport simulation code. Results are compared to those provided in MARSSIM. An extensive database of MCNPX-based detection efficiencies has been developed to represent a variety of gamma survey applications and potential scanning configurations (detector size, scan height, size of contaminated soil volume, etc.), and an associated web-based user interface has been developed to provide survey designers and regulators with access to a reasonably wide range of calculated scan MDC values for survey planning purposes. PMID:27356162

  8. Scan MDCs for GPS-Based Gamma Radiation Surveys.

    PubMed

    Alecksen, Tyler; Whicker, Randy

    2016-08-01

    A method for estimating the minimum detectable concentration of a contaminant radionuclide in soil when scanning with gamma radiation detectors (known as the "scan MDC") is described in the Multi-Agency Radiation Survey and Site Investigation Manual (MARSSIM). This paper presents an alternate method for estimating scan MDCs for GPS-based gamma surveys based on detector efficiencies modeled with the probabilistic Monte Carlo N-Particle Extended (MCNPX) Transport simulation code. Results are compared to those provided in MARSSIM. An extensive database of MCNPX-based detection efficiencies has been developed to represent a variety of gamma survey applications and potential scanning configurations (detector size, scan height, size of contaminated soil volume, etc.), and an associated web-based user interface has been developed to provide survey designers and regulators with access to a reasonably wide range of calculated scan MDC values for survey planning purposes.

  9. Simulating the Performance of Ground-Based Optical Asteroid Surveys

    NASA Astrophysics Data System (ADS)

    Christensen, Eric J.; Shelly, Frank C.; Gibbs, Alex R.; Grauer, Albert D.; Hill, Richard E.; Johnson, Jess A.; Kowalski, Richard A.; Larson, Stephen M.

    2014-11-01

    We are developing a set of asteroid survey simulation tools in order to estimate the capability of existing and planned ground-based optical surveys, and to test a variety of possible survey cadences and strategies. The survey simulator is composed of several layers, including a model population of solar system objects and an orbital integrator, a site-specific atmospheric model (including inputs for seeing, haze and seasonal cloud cover), a model telescope (with a complete optical path to estimate throughput), a model camera (including FOV, pixel scale, and focal plane fill factor) and model source extraction and moving object detection layers with tunable detection requirements. We have also developed a flexible survey cadence planning tool to automatically generate nightly survey plans. Inputs to the cadence planner include camera properties (FOV, readout time), telescope limits (horizon, declination, hour angle, lunar and zenithal avoidance), preferred and restricted survey regions in RA/Dec, ecliptic, and Galactic coordinate systems, and recent coverage by other asteroid surveys. Simulated surveys are created for a subset of current and previous NEO surveys (LINEAR, Pan-STARRS and the three Catalina Sky Survey telescopes), and compared against the actual performance of these surveys in order to validate the model’s performance. The simulator tracks objects within the FOV of any pointing that were not discovered (e.g. too few observations, too trailed, focal plane array gaps, too fast or slow), thus dividing the population into “discoverable” and “discovered” subsets, to inform possible survey design changes. Ongoing and future work includes generating a realistic “known” subset of the model NEO population, running multiple independent simulated surveys in coordinated and uncoordinated modes, and testing various cadences to find optimal strategies for detecting NEO sub-populations. These tools can also assist in quantifying the efficiency of novel

  10. Fault Diagnosis of Rotating Machinery Based on an Adaptive Ensemble Empirical Mode Decomposition

    PubMed Central

    Lei, Yaguo; Li, Naipeng; Lin, Jing; Wang, Sizhe

    2013-01-01

    The vibration based signal processing technique is one of the principal tools for diagnosing faults of rotating machinery. Empirical mode decomposition (EMD), as a time-frequency analysis technique, has been widely used to process vibration signals of rotating machinery. But it has the shortcoming of mode mixing in decomposing signals. To overcome this shortcoming, ensemble empirical mode decomposition (EEMD) was proposed accordingly. EEMD is able to reduce the mode mixing to some extent. The performance of EEMD, however, depends on the parameters adopted in the EEMD algorithms. In most of the studies on EEMD, the parameters were selected artificially and subjectively. To solve the problem, a new adaptive ensemble empirical mode decomposition method is proposed in this paper. In the method, the sifting number is adaptively selected, and the amplitude of the added noise changes with the signal frequency components during the decomposition process. The simulation, the experimental and the application results demonstrate that the adaptive EEMD provides the improved results compared with the original EEMD in diagnosing rotating machinery. PMID:24351666

  11. Empirical and physics based mathematical models of uranium hydride decomposition kinetics with quantified uncertainties.

    SciTech Connect

    Salloum, Maher N.; Gharagozloo, Patricia E.

    2013-10-01

    Metal particle beds have recently become a major technique for hydrogen storage. In order to extract hydrogen from such beds, it is crucial to understand the decomposition kinetics of the metal hydride. We are interested in obtaining a a better understanding of the uranium hydride (UH3) decomposition kinetics. We first developed an empirical model by fitting data compiled from different experimental studies in the literature and quantified the uncertainty resulting from the scattered data. We found that the decomposition time range predicted by the obtained kinetics was in a good agreement with published experimental results. Secondly, we developed a physics based mathematical model to simulate the rate of hydrogen diffusion in a hydride particle during the decomposition. We used this model to simulate the decomposition of the particles for temperatures ranging from 300K to 1000K while propagating parametric uncertainty and evaluated the kinetics from the results. We compared the kinetics parameters derived from the empirical and physics based models and found that the uncertainty in the kinetics predicted by the physics based model covers the scattered experimental data. Finally, we used the physics-based kinetics parameters to simulate the effects of boundary resistances and powder morphological changes during decomposition in a continuum level model. We found that the species change within the bed occurring during the decomposition accelerates the hydrogen flow by increasing the bed permeability, while the pressure buildup and the thermal barrier forming at the wall significantly impede the hydrogen extraction.

  12. Survey of Commercially Available Computer-Readable Bibliographic Data Bases.

    ERIC Educational Resources Information Center

    Schneider, John H., Ed.; And Others

    This document contains the results of a survey of 94 U. S. organizations, and 36 organizations in other countries that were thought to prepare machine-readable data bases. Of those surveyed, 55 organizations (40 in U. S., 15 in other countries) provided completed camera-ready forms describing 81 commercially available, machine-readable data bases…

  13. Empirical metallicity-dependent calibrations of effective temperature against colours for dwarfs and giants based on interferometric data

    NASA Astrophysics Data System (ADS)

    Huang, Yang

    2015-08-01

    We present empirical metallicity-dependent calibrations of effective temperature against colours for dwarfs of luminosity classes IV and V and giants of luminosity classes II and III, based on a collection of about two hundred nearby stars with direct effective temperature measurements of better than 2.5 per cent from the literature. The calibrations are valid for an effective temperature range 3,100 - 10,000 K for dwarfs of spectral types M5 to A0 and 3,100 - 5,700 K for giants of spectral types K5 to G5. A total of twenty-one colours for dwarfs and eighteen colours for giants of bands between four photometric systems, i.e. the Johnson (UBV RJIJJHK), the Cousins (RCIC), the Sloan Digital Sky Survey (SDSS, gr) and the Two Micron All Sky Survey (2MASS, JHKs), have been calibrated, respectively. Restricted by the metallicity range of the current sample, the calibrations are mainly applicable for disk stars ([Fe/H] ~ -1.0). The normalized percentage residuals of the calibrations are typically 2.0 and 1.5 per cent for dwarf and giant stars, respectively. Some systematic discrepancies of different levels are found between the current scales and those available in the literature (e.g. those based on the infrared flux method IRFM or spectroscopy). Based on the current calibrations, we have re-determined the colours of the Sun. We have also investigated the potential systematic errors in the estimates of effective temperature delivered by the current on-going large scale low- to intermediate-resolution stellar spectroscopic surveys. We propose that colour (g - Ks) calibrated in the current work can be an invaluable tool for the estimation of stellar effective temperature for the current/upcoming large scale stellar spectroscopic surveys. With the temperature yielded by this method , surface gravity (e.g. mass, density) determined by newly asteroseismology knowledge and metallicity ([Fe/H]) from high-resolution spectroscopy, we can give more constrains to the stellar

  14. A survey of machine readable data bases

    NASA Technical Reports Server (NTRS)

    Matlock, P.

    1981-01-01

    Forty-two of the machine readable data bases available to the technologist and researcher in the natural sciences and engineering are described and compared with the data bases and date base services offered by NASA.

  15. Comparing results from two continental geochemical surveys to world soil composition and deriving Predicted Empirical Global Soil (PEGS2) reference values

    NASA Astrophysics Data System (ADS)

    de Caritat, Patrice; Reimann, Clemens; Bastrakov, E.; Bowbridge, D.; Boyle, P.; Briggs, S.; Brown, D.; Brown, M.; Brownlie, K.; Burrows, P.; Burton, G.; Byass, J.; de Caritat, P.; Chanthapanya, N.; Cooper, M.; Cranfield, L.; Curtis, S.; Denaro, T.; Dhnaram, C.; Dhu, T.; Diprose, G.; Fabris, A.; Fairclough, M.; Fanning, S.; Fidler, R.; Fitzell, M.; Flitcroft, P.; Fricke, C.; Fulton, D.; Furlonger, J.; Gordon, G.; Green, A.; Green, G.; Greenfield, J.; Harley, J.; Heawood, S.; Hegvold, T.; Henderson, K.; House, E.; Husain, Z.; Krsteska, B.; Lam, J.; Langford, R.; Lavigne, T.; Linehan, B.; Livingstone, M.; Lukss, A.; Maier, R.; Makuei, A.; McCabe, L.; McDonald, P.; McIlroy, D.; McIntyre, D.; Morris, P.; O'Connell, G.; Pappas, B.; Parsons, J.; Petrick, C.; Poignand, B.; Roberts, R.; Ryle, J.; Seymon, A.; Sherry, K.; Skinner, J.; Smith, M.; Strickland, C.; Sutton, S.; Swindell, R.; Tait, H.; Tang, J.; Thomson, A.; Thun, C.; Uppill, B.; Wall, K.; Watkins, J.; Watson, T.; Webber, L.; Whiting, A.; Wilford, J.; Wilson, T.; Wygralak, A.; Albanese, S.; Andersson, M.; Arnoldussen, A.; Baritz, R.; Batista, M. J.; Bel-lan, A.; Birke, M.; Cicchella, C.; Demetriades, A.; Dinelli, E.; De Vivo, B.; De Vos, W.; Duris, M.; Dusza-Dobek, A.; Eggen, O. A.; Eklund, M.; Ernstsen, V.; Filzmoser, P.; Finne, T. E.; Flight, D.; Forrester, S.; Fuchs, M.; Fugedi, U.; Gilucis, A.; Gosar, M.; Gregorauskiene, V.; Gulan, A.; Halamić, J.; Haslinger, E.; Hayoz, P.; Hobiger, G.; Hoffmann, R.; Hoogewerff, J.; Hrvatovic, H.; Husnjak, S.; Janik, L.; Johnson, C. C.; Jordan, G.; Kirby, J.; Kivisilla, J.; Klos, V.; Krone, F.; Kwecko, P.; Kuti, L.; Ladenberger, A.; Lima, A.; Locutura, J.; Lucivjansky, P.; Mackovych, D.; Malyuk, B. I.; Maquil, R.; McLaughlin, M.; Meuli, R. G.; Miosic, N.; Mol, G.; Négrel, P.; O'Connor, P.; Oorts, K.; Ottesen, R. T.; Pasieczna, A.; Petersell, V.; Pfleiderer, S.; Poňavič, M.; Prazeres, C.; Rauch, U.; Reimann, C.; Salpeteur, I.; Schedl, A.; Scheib, A.; Schoeters, I.; Sefcik, P.; Sellersjö, E.; Skopljak, F.; Slaninka, I.; Šorša, A.; Srvkota, R.; Stafilov, T.; Tarvainen, T.; Trendavilov, V.; Valera, P.; Verougstraete, V.; Vidojević, D.; Zissimos, A. M.; Zomeni, Z.

    2012-02-01

    Analytical data for 10 major oxides (Al2O3, CaO, Fe2O3, K2O, MgO, MnO, Na2O, P2O5, SiO2 and TiO2), 16 total trace elements (As, Ba, Ce, Co, Cr, Ga, Nb, Ni, Pb, Rb, Sr, Th, V, Y, Zn and Zr), 14 aqua regia extracted elements (Ag, As, Bi, Cd, Ce, Co, Cs, Cu, Fe, La, Li, Mn, Mo and Pb), Loss On Ignition (LOI) and pH from 3526 soil samples from two continents (Australia and Europe) are presented and compared to (1) the composition of the upper continental crust, (2) published world soil average values, and (3) data from other continental-scale soil surveys. It can be demonstrated that average upper continental crust values do not provide reliable estimates for natural concentrations of elements in soils. For many elements there exist substantial differences between published world soil averages and the median concentrations observed on two continents. Direct comparison with other continental datasets is hampered by the fact that often mean, instead of the statistically more robust median, is reported. Using a database of the worldwide distribution of lithological units, it can be demonstrated that lithology is a poor predictor of soil chemistry. Climate-related processes such as glaciation and weathering are strong modifiers of the geochemical signature inherited from bedrock during pedogenesis. To overcome existing shortcomings of predicted global or world soil geochemical reference values, we propose Preliminary Empirical Global Soil reference values based on analytical results of a representative number of soil samples from two continents (PEGS2).

  16. An ISAR imaging algorithm for the space satellite based on empirical mode decomposition theory

    NASA Astrophysics Data System (ADS)

    Zhao, Tao; Dong, Chun-zhu

    2014-11-01

    Currently, high resolution imaging of the space satellite is a popular topic in the field of radar technology. In contrast with regular targets, the satellite target often moves along with its trajectory and simultaneously its solar panel substrate changes the direction toward the sun to obtain energy. Aiming at the imaging problem, a signal separating and imaging approach based on the empirical mode decomposition (EMD) theory is proposed, and the approach can realize separating the signal of two parts in the satellite target, the main body and the solar panel substrate and imaging for the target. The simulation experimentation can demonstrate the validity of the proposed method.

  17. Inequality of Higher Education in China: An Empirical Test Based on the Perspective of Relative Deprivation

    ERIC Educational Resources Information Center

    Hou, Liming

    2014-01-01

    The primary goal of this paper is to examine what makes Chinese college students dissatisfied with entrance opportunities for higher education. Based on the author's survey data, we test two parameters which could be a potential cause of this dissatisfaction: 1) distributive inequality, which emphasizes the individual's dissatisfaction caused by…

  18. Advances on Empirical Mode Decomposition-based Time-Frequency Analysis Methods in Hydrocarbon Detection

    NASA Astrophysics Data System (ADS)

    Chen, H. X.; Xue, Y. J.; Cao, J.

    2015-12-01

    Empirical mode decomposition (EMD), which is a data-driven adaptive decomposition method and is not limited by time-frequency uncertainty spreading, is proved to be more suitable for seismic signals which are nonlinear and non-stationary. Compared with other Fourier-based and wavelet-based time-frequency methods, EMD-based time-frequency methods have higher temporal and spatial resolution and yield hydrocarbon interpretations with more statistical significance. Empirical mode decomposition algorithm has now evolved from EMD to Ensemble EMD (EEMD) to Complete Ensemble EMD (CEEMD). Even though EMD-based time-frequency methods offer many promising features for analyzing and processing geophysical data, there are some limitations or defects in EMD-based time-frequency methods. This presentation will present a comparative study on hydrocarbon detection using seven EMD-based time-frequency analysis methods, which include: (1) first, EMD combined with Hilbert transform (HT) as a time-frequency analysis method is used for hydrocarbon detection; and (2) second, Normalized Hilbert transform (NHT) and HU Methods respectively combined with HT as improved time-frequency analysis methods are applied for hydrocarbon detection; and (3) three, EMD combined with Teager-Kaiser energy (EMD/TK) is investigated for hydrocarbon detection; and (4) four, EMD combined with wavelet transform (EMDWave) as a seismic attenuation estimation method is comparatively studied; and (5) EEMD- and CEEMD- based time-frequency analysis methods used as highlight volumes technology are studied. The differences between these methods in hydrocarbon detection will be discussed. The question of getting a meaningful instantaneous frequency by HT and mode-mixing issues in EMD will be analysed. The work was supported by NSFC under grant Nos. 41430323, 41404102 and 41274128.

  19. Dip-separated structural filtering using seislet transform and adaptive empirical mode decomposition based dip filter

    NASA Astrophysics Data System (ADS)

    Chen, Yangkang

    2016-07-01

    The seislet transform has been demonstrated to have a better compression performance for seismic data compared with other well-known sparsity promoting transforms, thus it can be used to remove random noise by simply applying a thresholding operator in the seislet domain. Since the seislet transform compresses the seismic data along the local structures, the seislet thresholding can be viewed as a simple structural filtering approach. Because of the dependence on a precise local slope estimation, the seislet transform usually suffers from low compression ratio and high reconstruction error for seismic profiles that have dip conflicts. In order to remove the limitation of seislet thresholding in dealing with conflicting-dip data, I propose a dip-separated filtering strategy. In this method, I first use an adaptive empirical mode decomposition based dip filter to separate the seismic data into several dip bands (5 or 6). Next, I apply seislet thresholding to each separated dip component to remove random noise. Then I combine all the denoised components to form the final denoised data. Compared with other dip filters, the empirical mode decomposition based dip filter is data-adaptive. One only needs to specify the number of dip components to be separated. Both complicated synthetic and field data examples show superior performance of my proposed approach than the traditional alternatives. The dip-separated structural filtering is not limited to seislet thresholding, and can also be extended to all those methods that require slope information.

  20. Determination of knock characteristics in spark ignition engines: an approach based on ensemble empirical mode decomposition

    NASA Astrophysics Data System (ADS)

    Li, Ning; Yang, Jianguo; Zhou, Rui; Liang, Caiping

    2016-04-01

    Knock is one of the major constraints to improve the performance and thermal efficiency of spark ignition (SI) engines. It can also result in severe permanent engine damage under certain operating conditions. Based on the ensemble empirical mode decomposition (EEMD), this paper proposes a new approach to determine the knock characteristics in SI engines. By adding a uniformly distributed and finite white Gaussian noise, the EEMD can preserve signal continuity in different scales and therefore alleviates the mode-mixing problem occurring in the classic empirical mode decomposition (EMD). The feasibilities of applying the EEMD to detect the knock signatures of a test SI engine via the pressure signal measured from combustion chamber and the vibration signal measured from cylinder head are investigated. Experimental results show that the EEMD-based method is able to detect the knock signatures from both the pressure signal and vibration signal, even in initial stage of knock. Finally, by comparing the application results with those obtained by short-time Fourier transform (STFT), Wigner-Ville distribution (WVD) and discrete wavelet transform (DWT), the superiority of the EEMD method in determining knock characteristics is demonstrated.

  1. Dynamics of bloggers’ communities: Bipartite networks from empirical data and agent-based modeling

    NASA Astrophysics Data System (ADS)

    Mitrović, Marija; Tadić, Bosiljka

    2012-11-01

    We present an analysis of the empirical data and the agent-based modeling of the emotional behavior of users on the Web portals where the user interaction is mediated by posted comments, like Blogs and Diggs. We consider the dataset of discussion-driven popular Diggs, in which all comments are screened by machine-learning emotion detection in the text, to determine positive and negative valence (attractiveness and aversiveness) of each comment. By mapping the data onto a suitable bipartite network, we perform an analysis of the network topology and the related time-series of the emotional comments. The agent-based model is then introduced to simulate the dynamics and to capture the emergence of the emotional behaviors and communities. The agents are linked to posts on a bipartite network, whose structure evolves through their actions on the posts. The emotional states (arousal and valence) of each agent fluctuate in time, subject to the current contents of the posts to which the agent is exposed. By an agent’s action on a post its current emotions are transferred to the post. The model rules and the key parameters are inferred from the considered empirical data to ensure their realistic values and mutual consistency. The model assumes that the emotional arousal over posts drives the agent’s action. The simulations are preformed for the case of constant flux of agents and the results are analyzed in full analogy with the empirical data. The main conclusions are that the emotion-driven dynamics leads to long-range temporal correlations and emergent networks with community structure, that are comparable with the ones in the empirical system of popular posts. In view of pure emotion-driven agents actions, this type of comparisons provide a quantitative measure for the role of emotions in the dynamics on real blogs. Furthermore, the model reveals the underlying mechanisms which relate the post popularity with the emotion dynamics and the prevalence of negative

  2. Polarizable Empirical Force Field for Hexopyranose Monosaccharides Based on the Classical Drude Oscillator

    PubMed Central

    2015-01-01

    A polarizable empirical force field based on the classical Drude oscillator is presented for the hexopyranose form of selected monosaccharides. Parameter optimization targeted quantum mechanical (QM) dipole moments, solute–water interaction energies, vibrational frequencies, and conformational energies. Validation of the model was based on experimental data on crystals, densities of aqueous-sugar solutions, diffusion constants of glucose, and rotational preferences of the exocylic hydroxymethyl of d-glucose and d-galactose in aqueous solution as well as additional QM data. Notably, the final model involves a single electrostatic model for all sixteen diastereomers of the monosaccharides, indicating the transferability of the polarizable model. The presented parameters are anticipated to lay the foundation for a comprehensive polarizable force field for saccharides that will be compatible with the polarizable Drude parameters for lipids and proteins, allowing for simulations of glycolipids and glycoproteins. PMID:24564643

  3. Multi-Scale Pixel-Based Image Fusion Using Multivariate Empirical Mode Decomposition

    PubMed Central

    Rehman, Naveed ur; Ehsan, Shoaib; Abdullah, Syed Muhammad Umer; Akhtar, Muhammad Jehanzaib; Mandic, Danilo P.; McDonald-Maier, Klaus D.

    2015-01-01

    A novel scheme to perform the fusion of multiple images using the multivariate empirical mode decomposition (MEMD) algorithm is proposed. Standard multi-scale fusion techniques make a priori assumptions regarding input data, whereas standard univariate empirical mode decomposition (EMD)-based fusion techniques suffer from inherent mode mixing and mode misalignment issues, characterized respectively by either a single intrinsic mode function (IMF) containing multiple scales or the same indexed IMFs corresponding to multiple input images carrying different frequency information. We show that MEMD overcomes these problems by being fully data adaptive and by aligning common frequency scales from multiple channels, thus enabling their comparison at a pixel level and subsequent fusion at multiple data scales. We then demonstrate the potential of the proposed scheme on a large dataset of real-world multi-exposure and multi-focus images and compare the results against those obtained from standard fusion algorithms, including the principal component analysis (PCA), discrete wavelet transform (DWT) and non-subsampled contourlet transform (NCT). A variety of image fusion quality measures are employed for the objective evaluation of the proposed method. We also report the results of a hypothesis testing approach on our large image dataset to identify statistically-significant performance differences. PMID:26007714

  4. An empirical Bayesian approach for model-based inference of cellular signaling networks

    PubMed Central

    2009-01-01

    Background A common challenge in systems biology is to infer mechanistic descriptions of biological process given limited observations of a biological system. Mathematical models are frequently used to represent a belief about the causal relationships among proteins within a signaling network. Bayesian methods provide an attractive framework for inferring the validity of those beliefs in the context of the available data. However, efficient sampling of high-dimensional parameter space and appropriate convergence criteria provide barriers for implementing an empirical Bayesian approach. The objective of this study was to apply an Adaptive Markov chain Monte Carlo technique to a typical study of cellular signaling pathways. Results As an illustrative example, a kinetic model for the early signaling events associated with the epidermal growth factor (EGF) signaling network was calibrated against dynamic measurements observed in primary rat hepatocytes. A convergence criterion, based upon the Gelman-Rubin potential scale reduction factor, was applied to the model predictions. The posterior distributions of the parameters exhibited complicated structure, including significant covariance between specific parameters and a broad range of variance among the parameters. The model predictions, in contrast, were narrowly distributed and were used to identify areas of agreement among a collection of experimental studies. Conclusion In summary, an empirical Bayesian approach was developed for inferring the confidence that one can place in a particular model that describes signal transduction mechanisms and for inferring inconsistencies in experimental measurements. PMID:19900289

  5. Determinants of Obesity and Associated Population Attributability, South Africa: Empirical Evidence from a National Panel Survey, 2008-2012

    PubMed Central

    Sartorius, Benn; Veerman, Lennert J.; Manyema, Mercy; Chola, Lumbwe; Hofman, Karen

    2015-01-01

    Background Obesity is a major risk factor for emerging non-communicable diseases (NCDS) in middle income countries including South Africa (SA). Understanding the multiple and complex determinants of obesity and their true population attributable impact is critical for informing and developing effective prevention efforts using scientific based evidence. This study identified contextualised high impact factors associated with obesity in South Africa. Methods Analysis of three national cross sectional (repeated panel) surveys, using a multilevel logistic regression and population attributable fraction estimation allowed for identification of contextualised high impact factors associated with obesity (BMI>30 kg/m2) among adults (15years+). Results Obesity prevalence increased significantly from 23.5% in 2008 to 27.2% in 2012, with a significantly (p-value<0.001) higher prevalence among females (37.9% in 2012) compared to males (13.3% in 2012). Living in formal urban areas, white ethnicity, being married, not exercising and/or in higher socio-economic category were significantly associated with male obesity. Females living in formal or informal urban areas, higher crime areas, African/White ethnicity, married, not exercising, in a higher socio-economic category and/or living in households with proportionate higher spending on food (and unhealthy food options) were significantly more likely to be obese. The identified determinants appeared to account for 75% and 43% of male and female obesity respectively. White males had the highest relative gain in obesity from 2008 to 2012. Conclusions The rising prevalence of obesity in South Africa is significant and over the past 5 years the rising prevalence of Type-2 diabetes has mirrored this pattern, especially among females. Targeting young adolescent girls should be a priority. Addressing determinants of obesity will involve a multifaceted strategy and requires at individual and population levels. With rising costs in the

  6. Using Web-Based Surveys to Conduct Counseling Research.

    ERIC Educational Resources Information Center

    Granello, Darcy Haag; Wheaton, Joe E.

    In spite of the increased use of the Internet for data collection, there is little published research about the process of data collection online. That is, discipline specific studies publish the results of their web-based surveys in discipline-specific journals, but little information is available on the process of Internet-based data collection.…

  7. A Human ECG Identification System Based on Ensemble Empirical Mode Decomposition

    PubMed Central

    Zhao, Zhidong; Yang, Lei; Chen, Diandian; Luo, Yi

    2013-01-01

    In this paper, a human electrocardiogram (ECG) identification system based on ensemble empirical mode decomposition (EEMD) is designed. A robust preprocessing method comprising noise elimination, heartbeat normalization and quality measurement is proposed to eliminate the effects of noise and heart rate variability. The system is independent of the heart rate. The ECG signal is decomposed into a number of intrinsic mode functions (IMFs) and Welch spectral analysis is used to extract the significant heartbeat signal features. Principal component analysis is used reduce the dimensionality of the feature space, and the K-nearest neighbors (K-NN) method is applied as the classifier tool. The proposed human ECG identification system was tested on standard MIT-BIH ECG databases: the ST change database, the long-term ST database, and the PTB database. The system achieved an identification accuracy of 95% for 90 subjects, demonstrating the effectiveness of the proposed method in terms of accuracy and robustness. PMID:23698274

  8. Distributed optical fiber-based theoretical and empirical methods monitoring hydraulic engineering subjected to seepage velocity

    NASA Astrophysics Data System (ADS)

    Su, Huaizhi; Tian, Shiguang; Cui, Shusheng; Yang, Meng; Wen, Zhiping; Xie, Wei

    2016-09-01

    In order to systematically investigate the general principle and method of monitoring seepage velocity in the hydraulic engineering, the theoretical analysis and physical experiment were implemented based on distributed fiber-optic temperature sensing (DTS) technology. During the coupling influence analyses between seepage field and temperature field in the embankment dam or dike engineering, a simplified model was constructed to describe the coupling relationship of two fields. Different arrangement schemes of optical fiber and measuring approaches of temperature were applied on the model. The inversion analysis idea was further used. The theoretical method of monitoring seepage velocity in the hydraulic engineering was finally proposed. A new concept, namely the effective thermal conductivity, was proposed referring to the thermal conductivity coefficient in the transient hot-wire method. The influence of heat conduction and seepage could be well reflected by this new concept, which was proved to be a potential approach to develop an empirical method monitoring seepage velocity in the hydraulic engineering.

  9. Orientation-Independent Empirical Mode Decomposition for Images Based on Unconstrained Optimization.

    PubMed

    Colominas, Marcelo A; Humeau-Heurtier, Anne; Schlotthauer, Gastón

    2016-05-01

    This paper introduces a 2D extension of the empirical mode decomposition (EMD), through a novel approach based on unconstrained optimization. EMD is a fully data-driven method that locally separates, in a completely data-driven and unsupervised manner, signals into fast and slow oscillations. The present proposal implements the method in a very simple and fast way, and it is compared with the state-of-the-art methods evidencing the advantages of being computationally efficient, orientation-independent, and leads to better performances for the decomposition of amplitude modulated-frequency modulated (AM-FM) images. The resulting genuine 2D method is successfully tested on artificial AM-FM images and its capabilities are illustrated on a biomedical example. The proposed framework leaves room for an nD extension (n > 2 ). PMID:26992022

  10. Empirical likelihood based detection procedure for change point in mean residual life functions under random censorship.

    PubMed

    Chen, Ying-Ju; Ning, Wei; Gupta, Arjun K

    2016-05-01

    The mean residual life (MRL) function is one of the basic parameters of interest in survival analysis that describes the expected remaining time of an individual after a certain age. The study of changes in the MRL function is practical and interesting because it may help us to identify some factors such as age and gender that may influence the remaining lifetimes of patients after receiving a certain surgery. In this paper, we propose a detection procedure based on the empirical likelihood for the changes in MRL functions with right censored data. Two real examples are also given: Veterans' administration lung cancer study and Stanford heart transplant to illustrate the detecting procedure. Copyright © 2016 John Wiley & Sons, Ltd.

  11. Empirical likelihood based detection procedure for change point in mean residual life functions under random censorship.

    PubMed

    Chen, Ying-Ju; Ning, Wei; Gupta, Arjun K

    2016-05-01

    The mean residual life (MRL) function is one of the basic parameters of interest in survival analysis that describes the expected remaining time of an individual after a certain age. The study of changes in the MRL function is practical and interesting because it may help us to identify some factors such as age and gender that may influence the remaining lifetimes of patients after receiving a certain surgery. In this paper, we propose a detection procedure based on the empirical likelihood for the changes in MRL functions with right censored data. Two real examples are also given: Veterans' administration lung cancer study and Stanford heart transplant to illustrate the detecting procedure. Copyright © 2016 John Wiley & Sons, Ltd. PMID:26936529

  12. The application of Kriging and empirical Kriging based on the variables selected by SCAD.

    PubMed

    Peng, Xiao-Ling; Yin, Hong; Li, Runze; Fang, Kai-Tai

    2006-09-25

    The commonly used approach for building a structure-activity/property relationship consists of three steps. First, one determines the descriptors for the molecular structure, then builds a metamodel by using some proper mathematical methods, and finally evaluates the meta-model. Some existing methods only can select important variables from the candidates, while most metamodels just explore linear relationships between inputs and outputs. Some techniques are useful to build more complicated relationship, but they may not be able to select important variables from a large number of variables. In this paper, we propose to screen important variables by the smoothly clipped absolute deviation (SCAD) variable selection procedure, and then apply Kriging model and empirical Kriging model for quantitative structure-activity/property relationship (QSAR/QSPR) research based on the selected important variables. We demonstrate the proposed procedure retains the virtues of both variable selection and Kriging model. PMID:17723710

  13. The Utility of Teacher and Student Surveys in Principal Evaluations: An Empirical Investigation. REL 2015-047

    ERIC Educational Resources Information Center

    Liu, Keke; Stuit, David; Springer, Jeff; Lindsay, Jim; Wan, Yinmei

    2014-01-01

    This study examined whether adding student and teacher survey measures to existing principal evaluation measures increases the overall power of the principal evaluation model to explain variation in student achievement across schools. The study was conducted using data from 2011-12 on 39 elementary and secondary schools within a midsize urban…

  14. The "Public Opinion Survey of Human Attributes-Stuttering" (POSHA-S): Summary Framework and Empirical Comparisons

    ERIC Educational Resources Information Center

    St. Louis, Kenneth O.

    2011-01-01

    Purpose: The "Public Opinion Survey of Human Attributes-Stuttering" ("POSHA-S") was developed to make available worldwide a standard measure of public attitudes toward stuttering that is practical, reliable, valid, and translatable. Mean data from past field studies as comparisons for interpretation of "POSHA-S" results are reported. Method: Means…

  15. Gender Wage Gaps by College Major in Taiwan: Empirical Evidence from the 1997-2003 Manpower Utilization Survey

    ERIC Educational Resources Information Center

    Lin, Eric S.

    2010-01-01

    In this article, we examine the effect of incorporating the fields of study on the explained and unexplained components of the standard Oaxaca decomposition for the gender wage gaps in Taiwan using 1997-2003 Manpower Utilization Survey data. Using several existing and lately developed measures, we inspect the gender wage gap by college major to…

  16. Web-based application for Data INterpolation Empirical Orthogonal Functions (DINEOF) analysis

    NASA Astrophysics Data System (ADS)

    Tomazic, Igor; Alvera-Azcarate, Aida; Barth, Alexander; Beckers, Jean-Marie

    2014-05-01

    DINEOF (Data INterpolating Empirical Orthogonal Functions) is a powerful tool based on EOF decomposition developed at the University of Liege/GHER for the reconstruction of missing data in satellite datasets, as well as for the reduction of noise and detection of outliers. DINEOF is openly available as a series of Fortran routines to be compiled by the user, and as binaries (that can be run directly without any compilation) both for Windows and Linux platforms. In order to facilitate the use of DINEOF and increase the number of interested users, we developed a web-based interface for DINEOF with the necessary parameters available to run high-quality DINEOF analysis. This includes choosing variable within selected dataset, defining a domain, time range, filtering criteria based on available variables in the dataset (e.g. quality flag, satellite zenith angle …) and defining necessary DINEOF parameters. Results, including reconstructed data and calculated EOF modes will be disseminated in NetCDF format using OpenDAP and WMS server allowing easy visualisation and analysis. First, we will include several satellite datasets of sea surface temperature and chlorophyll concentration obtained from MyOcean data centre and already remapped to the regular grid (L3C). Later, based on user's request, we plan to extend number of datasets available for reconstruction.

  17. An Empirical Study on Washback Effects of the Internet-Based College English Test Band 4 in China

    ERIC Educational Resources Information Center

    Wang, Chao; Yan, Jiaolan; Liu, Bao

    2014-01-01

    Based on Bailey's washback model, in respect of participants, process and products, the present empirical study was conducted to find the actual washback effects of the internet-based College English Test Band 4 (IB CET-4). The methods adopted are questionnaires, class observation, interview and the analysis of both the CET-4 teaching and testing…

  18. How "Does" the Comforting Process Work? An Empirical Test of an Appraisal-Based Model of Comforting

    ERIC Educational Resources Information Center

    Jones, Susanne M.; Wirtz, John G.

    2006-01-01

    Burleson and Goldsmith's (1998) comforting model suggests an appraisal-based mechanism through which comforting messages can bring about a positive change in emotional states. This study is a first empirical test of three causal linkages implied by the appraisal-based comforting model. Participants (N=258) talked about an upsetting event with a…

  19. Empirical source strength correlations for rans-based acoustic analogy methods

    NASA Astrophysics Data System (ADS)

    Kube-McDowell, Matthew Tyndall

    JeNo is a jet noise prediction code based on an acoustic analogy method developed by Mani, Gliebe, Balsa, and Khavaran. Using the flow predictions from a standard Reynolds-averaged Navier-Stokes computational fluid dynamics solver, JeNo predicts the overall sound pressure level and angular spectra for high-speed hot jets over a range of observer angles, with a processing time suitable for rapid design purposes. JeNo models the noise from hot jets as a combination of two types of noise sources; quadrupole sources dependent on velocity fluctuations, which represent the major noise of turbulent mixing, and dipole sources dependent on enthalpy fluctuations, which represent the effects of thermal variation. These two sources are modeled by JeNo as propagating independently into the far-field, with no cross-correlation at the observer location. However, high-fidelity computational fluid dynamics solutions demonstrate that this assumption is false. In this thesis, the theory, assumptions, and limitations of the JeNo code are briefly discussed, and a modification to the acoustic analogy method is proposed in which the cross-correlation of the two primary noise sources is allowed to vary with the speed of the jet and the observer location. As a proof-of-concept implementation, an empirical correlation correction function is derived from comparisons between JeNo's noise predictions and a set of experimental measurements taken for the Air Force Aero-Propulsion Laboratory. The empirical correlation correction is then applied to JeNo's predictions of a separate data set of hot jets tested at NASA's Glenn Research Center. Metrics are derived to measure the qualitative and quantitative performance of JeNo's acoustic predictions, and the empirical correction is shown to provide a quantitative improvement in the noise prediction at low observer angles with no freestream flow, and a qualitative improvement in the presence of freestream flow. However, the results also demonstrate

  20. Written institutional ethics policies on euthanasia: an empirical-based organizational-ethical framework.

    PubMed

    Lemiengre, Joke; Dierckx de Casterlé, Bernadette; Schotsmans, Paul; Gastmans, Chris

    2014-05-01

    As euthanasia has become a widely debated issue in many Western countries, hospitals and nursing homes especially are increasingly being confronted with this ethically sensitive societal issue. The focus of this paper is how healthcare institutions can deal with euthanasia requests on an organizational level by means of a written institutional ethics policy. The general aim is to make a critical analysis whether these policies can be considered as organizational-ethical instruments that support healthcare institutions to take their institutional responsibility for dealing with euthanasia requests. By means of an interpretative analysis, we conducted a process of reinterpretation of results of former Belgian empirical studies on written institutional ethics policies on euthanasia in dialogue with the existing international literature. The study findings revealed that legal regulations, ethical and care-oriented aspects strongly affected the development, the content, and the impact of written institutional ethics policies on euthanasia. Hence, these three cornerstones-law, care and ethics-constituted the basis for the empirical-based organizational-ethical framework for written institutional ethics policies on euthanasia that is presented in this paper. However, having a euthanasia policy does not automatically lead to more legal transparency, or to a more professional and ethical care practice. The study findings suggest that the development and implementation of an ethics policy on euthanasia as an organizational-ethical instrument should be considered as a dynamic process. Administrators and ethics committees must take responsibility to actively create an ethical climate supporting care providers who have to deal with ethical dilemmas in their practice.

  1. Open-circuit sensitivity model based on empirical parameters for a capacitive-type MEMS acoustic sensor

    NASA Astrophysics Data System (ADS)

    Lee, Jaewoo; Jeon, J. H.; Je, C. H.; Lee, S. Q.; Yang, W. S.; Lee, S.-G.

    2016-03-01

    An empirical-based open-circuit sensitivity model for a capacitive-type MEMS acoustic sensor is presented. To intuitively evaluate the characteristic of the open-circuit sensitivity, the empirical-based model is proposed and analysed by using a lumped spring-mass model and a pad test sample without a parallel plate capacitor for the parasitic capacitance. The model is composed of three different parameter groups: empirical, theoretical, and mixed data. The empirical residual stress from the measured pull-in voltage of 16.7 V and the measured surface topology of the diaphragm were extracted as +13 MPa, resulting in the effective spring constant of 110.9 N/m. The parasitic capacitance for two probing pads including the substrate part was 0.25 pF. Furthermore, to verify the proposed model, the modelled open-circuit sensitivity was compared with the measured value. The MEMS acoustic sensor had an open- circuit sensitivity of -43.0 dBV/Pa at 1 kHz with a bias of 10 V, while the modelled open- circuit sensitivity was -42.9 dBV/Pa, which showed good agreement in the range from 100 Hz to 18 kHz. This validates the empirical-based open-circuit sensitivity model for designing capacitive-type MEMS acoustic sensors.

  2. Empirical Study on Designing of Gaze Tracking Camera Based on the Information of User's Head Movement.

    PubMed

    Pan, Weiyuan; Jung, Dongwook; Yoon, Hyo Sik; Lee, Dong Eun; Naqvi, Rizwan Ali; Lee, Kwan Woo; Park, Kang Ryoung

    2016-08-31

    Gaze tracking is the technology that identifies a region in space that a user is looking at. Most previous non-wearable gaze tracking systems use a near-infrared (NIR) light camera with an NIR illuminator. Based on the kind of camera lens used, the viewing angle and depth-of-field (DOF) of a gaze tracking camera can be different, which affects the performance of the gaze tracking system. Nevertheless, to our best knowledge, most previous researches implemented gaze tracking cameras without ground truth information for determining the optimal viewing angle and DOF of the camera lens. Eye-tracker manufacturers might also use ground truth information, but they do not provide this in public. Therefore, researchers and developers of gaze tracking systems cannot refer to such information for implementing gaze tracking system. We address this problem providing an empirical study in which we design an optimal gaze tracking camera based on experimental measurements of the amount and velocity of user's head movements. Based on our results and analyses, researchers and developers might be able to more easily implement an optimal gaze tracking system. Experimental results show that our gaze tracking system shows high performance in terms of accuracy, user convenience and interest.

  3. Investigating properties of the cardiovascular system using innovative analysis algorithms based on ensemble empirical mode decomposition.

    PubMed

    Yeh, Jia-Rong; Lin, Tzu-Yu; Chen, Yun; Sun, Wei-Zen; Abbod, Maysam F; Shieh, Jiann-Shing

    2012-01-01

    Cardiovascular system is known to be nonlinear and nonstationary. Traditional linear assessments algorithms of arterial stiffness and systemic resistance of cardiac system accompany the problem of nonstationary or inconvenience in practical applications. In this pilot study, two new assessment methods were developed: the first is ensemble empirical mode decomposition based reflection index (EEMD-RI) while the second is based on the phase shift between ECG and BP on cardiac oscillation. Both methods utilise the EEMD algorithm which is suitable for nonlinear and nonstationary systems. These methods were used to investigate the properties of arterial stiffness and systemic resistance for a pig's cardiovascular system via ECG and blood pressure (BP). This experiment simulated a sequence of continuous changes of blood pressure arising from steady condition to high blood pressure by clamping the artery and an inverse by relaxing the artery. As a hypothesis, the arterial stiffness and systemic resistance should vary with the blood pressure due to clamping and relaxing the artery. The results show statistically significant correlations between BP, EEMD-based RI, and the phase shift between ECG and BP on cardiac oscillation. The two assessments results demonstrate the merits of the EEMD for signal analysis.

  4. Empirical Study on Designing of Gaze Tracking Camera Based on the Information of User's Head Movement.

    PubMed

    Pan, Weiyuan; Jung, Dongwook; Yoon, Hyo Sik; Lee, Dong Eun; Naqvi, Rizwan Ali; Lee, Kwan Woo; Park, Kang Ryoung

    2016-01-01

    Gaze tracking is the technology that identifies a region in space that a user is looking at. Most previous non-wearable gaze tracking systems use a near-infrared (NIR) light camera with an NIR illuminator. Based on the kind of camera lens used, the viewing angle and depth-of-field (DOF) of a gaze tracking camera can be different, which affects the performance of the gaze tracking system. Nevertheless, to our best knowledge, most previous researches implemented gaze tracking cameras without ground truth information for determining the optimal viewing angle and DOF of the camera lens. Eye-tracker manufacturers might also use ground truth information, but they do not provide this in public. Therefore, researchers and developers of gaze tracking systems cannot refer to such information for implementing gaze tracking system. We address this problem providing an empirical study in which we design an optimal gaze tracking camera based on experimental measurements of the amount and velocity of user's head movements. Based on our results and analyses, researchers and developers might be able to more easily implement an optimal gaze tracking system. Experimental results show that our gaze tracking system shows high performance in terms of accuracy, user convenience and interest. PMID:27589768

  5. Empirical Evaluation Indicators in Thai Higher Education: Theory-Based Multidimensional Learners' Assessment

    ERIC Educational Resources Information Center

    Sritanyarat, Dawisa; Russ-Eft, Darlene

    2016-01-01

    This study proposed empirical indicators which can be validated and adopted in higher education institutions to evaluate quality of teaching and learning, and to serve as an evaluation criteria for human resource management and development of higher institutions in Thailand. The main purpose of this study was to develop empirical indicators of a…

  6. Impact of Inadequate Empirical Therapy on the Mortality of Patients with Bloodstream Infections: a Propensity Score-Based Analysis

    PubMed Central

    Retamar, Pilar; Portillo, María M.; López-Prieto, María Dolores; Rodríguez-López, Fernando; de Cueto, Marina; García, María V.; Gómez, María J.; del Arco, Alfonso; Muñoz, Angel; Sánchez-Porto, Antonio; Torres-Tortosa, Manuel; Martín-Aspas, Andrés; Arroyo, Ascensión; García-Figueras, Carolina; Acosta, Federico; Corzo, Juan E.; León-Ruiz, Laura; Escobar-Lara, Trinidad

    2012-01-01

    The impact of the adequacy of empirical therapy on outcome for patients with bloodstream infections (BSI) is key for determining whether adequate empirical coverage should be prioritized over other, more conservative approaches. Recent systematic reviews outlined the need for new studies in the field, using improved methodologies. We assessed the impact of inadequate empirical treatment on the mortality of patients with BSI in the present-day context, incorporating recent methodological recommendations. A prospective multicenter cohort including all BSI episodes in adult patients was performed in 15 hospitals in Andalucía, Spain, over a 2-month period in 2006 to 2007. The main outcome variables were 14- and 30-day mortality. Adjusted analyses were performed by multivariate analysis and propensity score-based matching. Eight hundred one episodes were included. Inadequate empirical therapy was administered in 199 (24.8%) episodes; mortality at days 14 and 30 was 18.55% and 22.6%, respectively. After controlling for age, Charlson index, Pitt score, neutropenia, source, etiology, and presentation with severe sepsis or shock, inadequate empirical treatment was associated with increased mortality at days 14 and 30 (odds ratios [ORs], 2.12 and 1.56; 95% confidence intervals [95% CI], 1.34 to 3.34 and 1.01 to 2.40, respectively). The adjusted ORs after a propensity score-based matched analysis were 3.03 and 1.70 (95% CI, 1.60 to 5.74 and 0.98 to 2.98, respectively). In conclusion, inadequate empirical therapy is independently associated with increased mortality in patients with BSI. Programs to improve the quality of empirical therapy in patients with suspicion of BSI and optimization of definitive therapy should be implemented. PMID:22005999

  7. Is Project Based Learning More Effective than Direct Instruction in School Science Classrooms? An Analysis of the Empirical Research Evidence

    NASA Astrophysics Data System (ADS)

    Dann, Clifford

    An increasingly loud call by parents, school administrators, teachers, and even business leaders for "authentic learning", emphasizing both group-work and problem solving, has led to growing enthusiasm for inquiry-based learning over the past decade. Although "inquiry" can be defined in many ways, a curriculum called "project-based learning" has recently emerged as the inquiry practice-of-choice with roots in the educational constructivism that emerged in the mid-twentieth century. Often, project-based learning is framed as an alternative instructional strategy to direct instruction for maximizing student content knowledge. This study investigates the empirical evidence for such a comparison while also evaluating the overall quality of the available studies in the light of accepted standards for educational research. Specifically, this thesis investigates what the body of quantitative research says about the efficacy of project-based learning vs. direct instruction when considering student acquisition of content knowledge in science classrooms. Further, existing limitations of the research pertaining to project based learning and secondary school education are explored. The thesis concludes with a discussion of where and how we should focus our empirical efforts in the future. The research revealed that the available empirical research contains flaws in both design and instrumentation. In particular, randomization is poor amongst all the studies considered. The empirical evidence indicates that project-based learning curricula improved student content knowledge but that, while the results were statistically significant, increases in raw test scores were marginal.

  8. Evaluation of Physically and Empirically Based Models for the Estimation of Green Roof Evapotranspiration

    NASA Astrophysics Data System (ADS)

    Digiovanni, K. A.; Montalto, F. A.; Gaffin, S.; Rosenzweig, C.

    2010-12-01

    Green roofs and other urban green spaces can provide a variety of valuable benefits including reduction of the urban heat island effect, reduction of stormwater runoff, carbon sequestration, oxygen generation, air pollution mitigation etc. As many of these benefits are directly linked to the processes of evaporation and transpiration, accurate and representative estimation of urban evapotranspiration (ET) is a necessary tool for predicting and quantifying such benefits. However, many common ET estimation procedures were developed for agricultural applications, and thus carry inherent assumptions that may only be rarely applicable to urban green spaces. Various researchers have identified the estimation of expected urban ET rates as critical, yet poorly studied components of urban green space performance prediction and cite that further evaluation is needed to reconcile differences in predictions from varying ET modeling approaches. A small scale green roof lysimeter setup situated on the green roof of the Ethical Culture Fieldston School in the Bronx, NY has been the focus of ongoing monitoring initiated in June 2009. The experimental setup includes a 0.6 m by 1.2 m Lysimeter replicating the anatomy of the 500 m2 green roof of the building, with a roof membrane, drainage layer, 10 cm media depth, and planted with a variety of Sedum species. Soil moisture sensors and qualitative runoff measurements are also recorded in the Lysimeter, while a weather station situated on the rooftop records climatologic data. Direct quantification of actual evapotranspiration (AET) from the green roof weighing lysimeter was achieved through a mass balance approaches during periods absent of precipitation and drainage. A comparison of AET to estimates of potential evapotranspiration (PET) calculated from empirically and physically based ET models was performed in order to evaluate the applicability of conventional ET equations for the estimation of ET from green roofs. Results have

  9. Validating a Theory-Based Survey to Evaluate Teaching Effectiveness in Higher Education

    ERIC Educational Resources Information Center

    Amrein-Beardsley, A.; Haladyna, T.

    2012-01-01

    Surveys to evaluate instructor effectiveness are commonly used in higher education. Yet the survey items included are often drawn from other surveys without reference to a theory of adult learning. The authors present the results from a validation study of such a theory-based survey. They evidence that an evaluation survey based on a theory that…

  10. Place Based Assistance Tools: Networking and Resident Surveys.

    ERIC Educational Resources Information Center

    Department of Housing and Urban Development, Washington, DC.

    "Place-based assistance" is not a new concept. Asking what people want and finding ways to give it to them sounds simplistic, but it can result in "win-win" solutions in which everyone involved benefits. This document is a guide to using networking and surveys of residents to determine community needs. Some case studies show networking and surveys…

  11. Lake Superior Phytoplankton Characterization from the 2006 Probability Based Survey

    EPA Science Inventory

    We conducted a late summer probability based survey of Lake Superior in 2006 which consisted of 52 sites stratified across 3 depth zones. As part of this effort, we collected composite phytoplankton samples from the epilimnion and the fluorescence maxima (Fmax) at 29 of the site...

  12. School-Based Budgeting Survey Study of Pilot Schools.

    ERIC Educational Resources Information Center

    Robinson, Carol

    This report describes results of a survey on school-based budgeting (SBB) in the Albuquerque, New Mexico, public schools (APS). SBB began in the 1986-87 school year at 33 of the 116 APS schools and alternative schools, with 16 elementary, 11 middle, and 6 high schools, participating in the first year. A total of 131 responses were received from…

  13. Feasibility of an empirically based program for parents of preschoolers with autism spectrum disorder.

    PubMed

    Dababnah, Sarah; Parish, Susan L

    2016-01-01

    This article reports on the feasibility of implementing an existing empirically based program, The Incredible Years, tailored to parents of young children with autism spectrum disorder. Parents raising preschool-aged children (aged 3-6 years) with autism spectrum disorder (N = 17) participated in a 15-week pilot trial of the intervention. Quantitative assessments of the program revealed fidelity was generally maintained, with the exception of program-specific videos. Qualitative data from individual post-intervention interviews reported parents benefited most from child emotion regulation strategies, play-based child behavior skills, parent stress management, social support, and visual resources. More work is needed to further refine the program to address parent self-care, partner relationships, and the diverse behavioral and communication challenges of children across the autism spectrum. Furthermore, parent access and retention could potentially be increased by providing in-home childcare vouchers and a range of times and locations in which to offer the program. The findings suggest The Incredible Years is a feasible intervention for parents seeking additional support for child- and family-related challenges and offers guidance to those communities currently using The Incredible Years or other related parenting programs with families of children with autism spectrum disorder.

  14. An empirically based steady state friction law and implications for fault stability

    NASA Astrophysics Data System (ADS)

    Spagnuolo, E.; Nielsen, S.; Violay, M.; Di Toro, G.

    2016-04-01

    Empirically based rate-and-state friction laws (RSFLs) have been proposed to model the dependence of friction forces with slip and time. The relevance of the RSFL for earthquake mechanics is that few constitutive parameters define critical conditions for fault stability (i.e., critical stiffness and frictional fault behavior). However, the RSFLs were determined from experiments conducted at subseismic slip rates (V < 1 cm/s), and their extrapolation to earthquake deformation conditions (V > 0.1 m/s) remains questionable on the basis of the experimental evidence of (1) large dynamic weakening and (2) activation of particular fault lubrication processes at seismic slip rates. Here we propose a modified RSFL (MFL) based on the review of a large published and unpublished data set of rock friction experiments performed with different testing machines. The MFL, valid at steady state conditions from subseismic to seismic slip rates (0.1 µm/s < V < 3 m/s), describes the initiation of a substantial velocity weakening in the 1-20 cm/s range resulting in a critical stiffness increase that creates a peak of potential instability in that velocity regime. The MFL leads to a new definition of fault frictional stability with implications for slip event styles and relevance for models of seismic rupture nucleation, propagation, and arrest.

  15. Phospholipid-based nonlamellar mesophases for delivery systems: bridging the gap between empirical and rational design.

    PubMed

    Martiel, Isabelle; Sagalowicz, Laurent; Mezzenga, Raffaele

    2014-07-01

    Phospholipids are ubiquitous cell membrane components and relatively well-accepted ingredients due to their natural origin. Phosphatidylcholine (PC) in particular offers a promising alternative to monoglycerides for lyotropic liquid crystalline (LLC) delivery system applications in the food, cosmetics and pharmaceutical industries, provided its strong tendency to form zero-mean curvature lamellar mesophases in water can be overcome. Higher negative curvatures are usually reached through the addition of a third lipid component, forming a ternary diagram phospholipid/water/oil. The initial part of this work summarizes the potential advantages and the challenges of phospholipid-based delivery system applications. In the next part, various ternary PC/water/oil systems are discussed, with a special emphasis on the PC/water/cyclohexane and PC/water/α-tocopherol systems. We report that R-(+)-limonene has a quantitatively similar effect as cyclohexane. The last part is devoted to the theoretical interpretation of the observed phase behaviors. A fruitful parallel is drawn with PC polymer-like reverse micelles, leading to a thermodynamic description in terms of interfacial bending energy. Investigations at the molecular level are reviewed to help in bridging the empirical and theoretical approaches. Predictive rules are finally derived from this wide-ranging overview, thereby opening the way to a future rational design of PC-based LLC delivery systems. PMID:24685272

  16. Time Domain Strain/Stress Reconstruction Based on Empirical Mode Decomposition: Numerical Study and Experimental Validation

    PubMed Central

    He, Jingjing; Zhou, Yibin; Guan, Xuefei; Zhang, Wei; Zhang, Weifang; Liu, Yongming

    2016-01-01

    Structural health monitoring has been studied by a number of researchers as well as various industries to keep up with the increasing demand for preventive maintenance routines. This work presents a novel method for reconstruct prompt, informed strain/stress responses at the hot spots of the structures based on strain measurements at remote locations. The structural responses measured from usage monitoring system at available locations are decomposed into modal responses using empirical mode decomposition. Transformation equations based on finite element modeling are derived to extrapolate the modal responses from the measured locations to critical locations where direct sensor measurements are not available. Then, two numerical examples (a two-span beam and a 19956-degree of freedom simplified airfoil) are used to demonstrate the overall reconstruction method. Finally, the present work investigates the effectiveness and accuracy of the method through a set of experiments conducted on an aluminium alloy cantilever beam commonly used in air vehicle and spacecraft. The experiments collect the vibration strain signals of the beam via optical fiber sensors. Reconstruction results are compared with theoretical solutions and a detailed error analysis is also provided. PMID:27537889

  17. A hybrid filtering method based on a novel empirical mode decomposition for friction signals

    NASA Astrophysics Data System (ADS)

    Li, Chengwei; Zhan, Liwei

    2015-12-01

    During a measurement, the measured signal usually contains noise. To remove the noise and preserve the important feature of the signal, we introduce a hybrid filtering method that uses a new intrinsic mode function (NIMF) and a modified Hausdorff distance. The NIMF is defined as the difference between the noisy signal and each intrinsic mode function (IMF), which is obtained by empirical mode decomposition (EMD), ensemble EMD, complementary ensemble EMD, or complete ensemble EMD with adaptive noise (CEEMDAN). The relevant mode selecting is based on the similarity between the first NIMF and the rest of the NIMFs. With this filtering method, the EMD and improved versions are used to filter the simulation and friction signals. The friction signal between an airplane tire and the runaway is recorded during a simulated airplane touchdown and features spikes of various amplitudes and noise. The filtering effectiveness of the four hybrid filtering methods are compared and discussed. The results show that the filtering method based on CEEMDAN outperforms other signal filtering methods.

  18. An empirically based steady state friction law and implications for fault stability

    PubMed Central

    Nielsen, S.; Violay, M.; Di Toro, G.

    2016-01-01

    Abstract Empirically based rate‐and‐state friction laws (RSFLs) have been proposed to model the dependence of friction forces with slip and time. The relevance of the RSFL for earthquake mechanics is that few constitutive parameters define critical conditions for fault stability (i.e., critical stiffness and frictional fault behavior). However, the RSFLs were determined from experiments conducted at subseismic slip rates (V < 1 cm/s), and their extrapolation to earthquake deformation conditions (V > 0.1 m/s) remains questionable on the basis of the experimental evidence of (1) large dynamic weakening and (2) activation of particular fault lubrication processes at seismic slip rates. Here we propose a modified RSFL (MFL) based on the review of a large published and unpublished data set of rock friction experiments performed with different testing machines. The MFL, valid at steady state conditions from subseismic to seismic slip rates (0.1 µm/s < V < 3 m/s), describes the initiation of a substantial velocity weakening in the 1–20 cm/s range resulting in a critical stiffness increase that creates a peak of potential instability in that velocity regime. The MFL leads to a new definition of fault frictional stability with implications for slip event styles and relevance for models of seismic rupture nucleation, propagation, and arrest.

  19. Going Global: A Model for Evaluating Empirically Supported Family-Based Interventions in New Contexts.

    PubMed

    Sundell, Knut; Ferrer-Wreder, Laura; Fraser, Mark W

    2014-06-01

    The spread of evidence-based practice throughout the world has resulted in the wide adoption of empirically supported interventions (ESIs) and a growing number of controlled trials of imported and culturally adapted ESIs. This article is informed by outcome research on family-based interventions including programs listed in the American Blueprints Model and Promising Programs. Evidence from these controlled trials is mixed and, because it is comprised of both successful and unsuccessful replications of ESIs, it provides clues for the translation of promising programs in the future. At least four explanations appear plausible for the mixed results in replication trials. One has to do with methodological differences across trials. A second deals with ambiguities in the cultural adaptation process. A third explanation is that ESIs in failed replications have not been adequately implemented. A fourth source of variation derives from unanticipated contextual influences that might affect the effects of ESIs when transported to other cultures and countries. This article describes a model that allows for the differential examination of adaptations of interventions in new cultural contexts.

  20. Time Domain Strain/Stress Reconstruction Based on Empirical Mode Decomposition: Numerical Study and Experimental Validation.

    PubMed

    He, Jingjing; Zhou, Yibin; Guan, Xuefei; Zhang, Wei; Zhang, Weifang; Liu, Yongming

    2016-08-16

    Structural health monitoring has been studied by a number of researchers as well as various industries to keep up with the increasing demand for preventive maintenance routines. This work presents a novel method for reconstruct prompt, informed strain/stress responses at the hot spots of the structures based on strain measurements at remote locations. The structural responses measured from usage monitoring system at available locations are decomposed into modal responses using empirical mode decomposition. Transformation equations based on finite element modeling are derived to extrapolate the modal responses from the measured locations to critical locations where direct sensor measurements are not available. Then, two numerical examples (a two-span beam and a 19956-degree of freedom simplified airfoil) are used to demonstrate the overall reconstruction method. Finally, the present work investigates the effectiveness and accuracy of the method through a set of experiments conducted on an aluminium alloy cantilever beam commonly used in air vehicle and spacecraft. The experiments collect the vibration strain signals of the beam via optical fiber sensors. Reconstruction results are compared with theoretical solutions and a detailed error analysis is also provided.

  1. Time Domain Strain/Stress Reconstruction Based on Empirical Mode Decomposition: Numerical Study and Experimental Validation.

    PubMed

    He, Jingjing; Zhou, Yibin; Guan, Xuefei; Zhang, Wei; Zhang, Weifang; Liu, Yongming

    2016-01-01

    Structural health monitoring has been studied by a number of researchers as well as various industries to keep up with the increasing demand for preventive maintenance routines. This work presents a novel method for reconstruct prompt, informed strain/stress responses at the hot spots of the structures based on strain measurements at remote locations. The structural responses measured from usage monitoring system at available locations are decomposed into modal responses using empirical mode decomposition. Transformation equations based on finite element modeling are derived to extrapolate the modal responses from the measured locations to critical locations where direct sensor measurements are not available. Then, two numerical examples (a two-span beam and a 19956-degree of freedom simplified airfoil) are used to demonstrate the overall reconstruction method. Finally, the present work investigates the effectiveness and accuracy of the method through a set of experiments conducted on an aluminium alloy cantilever beam commonly used in air vehicle and spacecraft. The experiments collect the vibration strain signals of the beam via optical fiber sensors. Reconstruction results are compared with theoretical solutions and a detailed error analysis is also provided. PMID:27537889

  2. An empirically based steady state friction law and implications for fault stability

    PubMed Central

    Nielsen, S.; Violay, M.; Di Toro, G.

    2016-01-01

    Abstract Empirically based rate‐and‐state friction laws (RSFLs) have been proposed to model the dependence of friction forces with slip and time. The relevance of the RSFL for earthquake mechanics is that few constitutive parameters define critical conditions for fault stability (i.e., critical stiffness and frictional fault behavior). However, the RSFLs were determined from experiments conducted at subseismic slip rates (V < 1 cm/s), and their extrapolation to earthquake deformation conditions (V > 0.1 m/s) remains questionable on the basis of the experimental evidence of (1) large dynamic weakening and (2) activation of particular fault lubrication processes at seismic slip rates. Here we propose a modified RSFL (MFL) based on the review of a large published and unpublished data set of rock friction experiments performed with different testing machines. The MFL, valid at steady state conditions from subseismic to seismic slip rates (0.1 µm/s < V < 3 m/s), describes the initiation of a substantial velocity weakening in the 1–20 cm/s range resulting in a critical stiffness increase that creates a peak of potential instability in that velocity regime. The MFL leads to a new definition of fault frictional stability with implications for slip event styles and relevance for models of seismic rupture nucleation, propagation, and arrest. PMID:27667875

  3. Space-based infrared surveys of small bodies

    NASA Astrophysics Data System (ADS)

    Mommert, M.

    2014-07-01

    Most small bodies in the Solar System are too small and too distant to be spatially resolved, precluding a direct diameter derivation. Furthermore, measurements of the optical brightness alone only allow a rough estimate of the diameter, since the surface albedo is usually unknown and can have values between about 3 % and 60 % or more. The degeneracy can be resolved by considering the thermal emission of these objects, which is less prone to albedo effects and mainly a function of the diameter. Hence, the combination of optical and thermal-infrared observational data provides a means to independently derive an object's diameter and albedo. This technique is used in asteroid thermal models or more sophisticated thermophysical models (see, e.g., [1]). Infrared observations require cryogenic detectors and/or telescopes, depending on the actual wavelength range observed. Observations from the ground are additionally compromised by the variable transparency of Earth's atmosphere in major portions of the infrared wavelength ranges. Hence, space-based infrared telescopes, providing stable conditions and significantly better sensitivities than ground-based telescopes, are now used routinely to exploit this wavelength range. Two observation strategies are used with space-based infrared observatories: Space-based Infrared All-Sky Surveys. Asteroid surveys in the thermal infrared are less prone to albedo-related discovery bias compared to surveys with optical telescopes, providing a more complete picture of small body populations. The first space-based infrared survey of Solar System small bodies was performed with the Infrared Astronomical Satellite (IRAS) for 10 months in 1983. In the course of the 'IRAS Minor Planet Survey' [2], 2228 asteroids (3 new discoveries) and more than 25 comets (6 new discoveries) were observed. More recent space-based infrared all-sky asteroid surveys were performed by Akari (launched 2006) and the Wide-field Infrared Survey Explorer (WISE

  4. Survey-based Indices for Nursing Home Quality Incentive Reimbursement

    PubMed Central

    Willemain, Thomas R.

    1983-01-01

    Incentive payments are a theoretically appealing complement to nursing home quality assurance systems that rely on regulatory enforcement. However, the practical aspects of incentive program design are not yet well understood. After reviewing the rationale for incentive approaches and recent State and. Federal initiatives, the article considers a basic program design issue: creating an index of nursing home quality. It focuses on indices constructed from routine licensure and certification survey results because State initiatives have relied heavily on these readily accessible data. It also suggests a procedure for creating a survey-based index and discusses a sampling of Implementation issues. PMID:10309858

  5. Selecting Great Lakes streams for lampricide treatment based on larval sea lamprey surveys

    USGS Publications Warehouse

    Christie, Gavin C.; Adams, Jean V.; Steeves, Todd B.; Slade, Jeffrey W.; Cuddy, Douglas W.; Fodale, Michael F.; Young, Robert J.; Kuc, Miroslaw; Jones, Michael L.

    2003-01-01

    The Empiric Stream Treatment Ranking (ESTR) system is a data-driven, model-based, decision tool for selecting Great Lakes streams for treatment with lampricide, based on estimates from larval sea lamprey (Petromyzon marinus) surveys conducted throughout the basin. The 2000 ESTR system was described and applied to larval assessment surveys conducted from 1996 to 1999. A comparative analysis of stream survey and selection data was conducted and improvements to the stream selection process were recommended. Streams were selected for treatment based on treatment cost, predicted treatment effectiveness, and the projected number of juvenile sea lampreys produced. On average, lampricide treatments were applied annually to 49 streams with 1,075 ha of larval habitat, killing 15 million larval and 514,000 juvenile sea lampreys at a total cost of $5.3 million, and marginal and mean costs of $85 and $10 per juvenile killed. The numbers of juvenile sea lampreys killed for given treatment costs showed a pattern of diminishing returns with increasing investment. Of the streams selected for treatment, those with > 14 ha of larval habitat targeted 73% of the juvenile sea lampreys for 60% of the treatment cost. Suggested improvements to the ESTR system were to improve accuracy and precision of model estimates, account for uncertainty in estimates, include all potentially productive streams in the process (not just those surveyed in the current year), consider the value of all larvae killed during treatment (not just those predicted to metamorphose the following year), use lake-specific estimates of damage, and establish formal suppression targets.

  6. Survey on Existing Science Gateways - Based on DEGREE

    NASA Astrophysics Data System (ADS)

    Schwichtenberg, Horst; Claus, Steffen

    2010-05-01

    Science Gateways gather community-developed specific tools, applications, and data that is usually combined and presented in a graphical user interface which is customized to the needs of the target user community. Science Gateways serve as a single point of entry for the users and are usually represented by fat clients or web portals. Part of the DEGREE project (Dissemination and Exploitation of Grids in Earth Science) was a state-of-the-art survey of portal usage in Earth Science (ES) applications. This survey considered a list of 29 portals, including 17 ES portals and 12 generic developments coming from outside of the ES domain. The survey identified three common usage types of ES portals, including data dissemination (e.g. observational data), collaboration as well as usage of Grid-based resources (e.g. for processing of ES datasets). Based on these three usage types, key requirements could be extracted. These requirements were furthermore used for a feature comparison with existing portal developments coming from outside of the ES domain. This presentation gives an overview of the results of the survey (including a feature comparison of ES and non-ES portals). Furthermore, three portals are discussed in detail, one for each usage type (data dissemination, collaboration, Grid-based).

  7. An Empirical Introduction to the Concept of Chemical Element Based on Van Hiele's Theory of Level Transitions

    ERIC Educational Resources Information Center

    Vogelezang, Michiel; Van Berkel, Berry; Verdonk, Adri

    2015-01-01

    Between 1970 and 1990, the Dutch working group "Empirical Introduction to Chemistry" developed a secondary school chemistry education curriculum based on the educational vision of the mathematicians van Hiele and van Hiele-Geldof. This approach viewed learning as a process in which students must go through discontinuous level transitions…

  8. Empirical Differences in Omission Tendency and Reading Ability in PISA: An Application of Tree-Based Item Response Models

    ERIC Educational Resources Information Center

    Okumura, Taichi

    2014-01-01

    This study examined the empirical differences between the tendency to omit items and reading ability by applying tree-based item response (IRTree) models to the Japanese data of the Programme for International Student Assessment (PISA) held in 2009. For this purpose, existing IRTree models were expanded to contain predictors and to handle…

  9. Design-based and model-based inference in surveys of freshwater mollusks

    USGS Publications Warehouse

    Dorazio, R.M.

    1999-01-01

    Well-known concepts in statistical inference and sampling theory are used to develop recommendations for planning and analyzing the results of quantitative surveys of freshwater mollusks. Two methods of inference commonly used in survey sampling (design-based and model-based) are described and illustrated using examples relevant in surveys of freshwater mollusks. The particular objectives of a survey and the type of information observed in each unit of sampling can be used to help select the sampling design and the method of inference. For example, the mean density of a sparsely distributed population of mollusks can be estimated with higher precision by using model-based inference or by using design-based inference with adaptive cluster sampling than by using design-based inference with conventional sampling. More experience with quantitative surveys of natural assemblages of freshwater mollusks is needed to determine the actual benefits of different sampling designs and inferential procedures.

  10. Empirical likelihood-based confidence intervals for length-biased data

    PubMed Central

    Ning, J.; Qin, J.; Asgharian, M.; Shen, Y.

    2013-01-01

    Logistic or other constraints often preclude the possibility of conducting incident cohort studies. A feasible alternative in such cases is to conduct a cross-sectional prevalent cohort study for which we recruit prevalent cases, i.e. subjects who have already experienced the initiating event, say the onset of a disease. When the interest lies in estimating the lifespan between the initiating event and a terminating event, say death for instance, such subjects may be followed prospectively until the terminating event or loss to follow-up, whichever happens first. It is well known that prevalent cases have, on average, longer lifespans. As such they do not constitute a representative random sample from the target population; they comprise a biased sample. If the initiating events are generated from a stationary Poisson process, the so-called stationarity assumption, this bias is called length bias. The current literature on length-biased sampling lacks a simple method for estimating the margin of errors of commonly used summary statistics. We fill this gap using the empirical likelihood-based confidence intervals by adapting this method to right-censored length-biased survival data. Both large and small sample behaviors of these confidence intervals are studied. We illustrate our method using a set of data on survival with dementia, collected as part of the Canadian Study of Health and Aging. PMID:23027662

  11. Robust multitask learning with three-dimensional empirical mode decomposition-based features for hyperspectral classification

    NASA Astrophysics Data System (ADS)

    He, Zhi; Liu, Lin

    2016-11-01

    Empirical mode decomposition (EMD) and its variants have recently been applied for hyperspectral image (HSI) classification due to their ability to extract useful features from the original HSI. However, it remains a challenging task to effectively exploit the spectral-spatial information by the traditional vector or image-based methods. In this paper, a three-dimensional (3D) extension of EMD (3D-EMD) is proposed to naturally treat the HSI as a cube and decompose the HSI into varying oscillations (i.e. 3D intrinsic mode functions (3D-IMFs)). To achieve fast 3D-EMD implementation, 3D Delaunay triangulation (3D-DT) is utilized to determine the distances of extrema, while separable filters are adopted to generate the envelopes. Taking the extracted 3D-IMFs as features of different tasks, robust multitask learning (RMTL) is further proposed for HSI classification. In RMTL, pairs of low-rank and sparse structures are formulated by trace-norm and l1,2 -norm to capture task relatedness and specificity, respectively. Moreover, the optimization problems of RMTL can be efficiently solved by the inexact augmented Lagrangian method (IALM). Compared with several state-of-the-art feature extraction and classification methods, the experimental results conducted on three benchmark data sets demonstrate the superiority of the proposed methods.

  12. Satellite-based empirical models linking river plume dynamics with hypoxic area and volume

    NASA Astrophysics Data System (ADS)

    Le, Chengfeng; Lehrter, John C.; Hu, Chuanmin; Obenour, Daniel R.

    2016-03-01

    Satellite-based empirical models explaining hypoxic area and volume variation were developed for the seasonally hypoxic (O2 < 2 mg L-1) northern Gulf of Mexico adjacent to the Mississippi River. Annual variations in midsummer hypoxic area and volume were related to Moderate Resolution Imaging Spectroradiometer-derived monthly estimates of river plume area (km2) and average, inner shelf chlorophyll a concentration (Chl a, mg m-3). River plume area in June was negatively related with midsummer hypoxic area (km2) and volume (km3), while July inner shelf Chl a was positively related to hypoxic area and volume. Multiple regression models using river plume area and Chl a as independent variables accounted for most of the variability in hypoxic area (R2 = 0.92) or volume (R2 = 0.89). These models explain more variation in hypoxic area than models using Mississippi River nutrient loads as independent variables. The results here also support a hypothesis that confinement of the river plume to the inner shelf is an important mechanism controlling hypoxia area and volume in this region.

  13. Empirical Study of User Preferences Based on Rating Data of Movies.

    PubMed

    Zhao, YingSi; Shen, Bo

    2016-01-01

    User preference plays a prominent role in many fields, including electronic commerce, social opinion, and Internet search engines. Particularly in recommender systems, it directly influences the accuracy of the recommendation. Though many methods have been presented, most of these have only focused on how to improve the recommendation results. In this paper, we introduce an empirical study of user preferences based on a set of rating data about movies. We develop a simple statistical method to investigate the characteristics of user preferences. We find that the movies have potential characteristics of closure, which results in the formation of numerous cliques with a power-law size distribution. We also find that a user related to a small clique always has similar opinions on the movies in this clique. Then, we suggest a user preference model, which can eliminate the predictions that are considered to be impracticable. Numerical results show that the model can reflect user preference with remarkable accuracy when data elimination is allowed, and random factors in the rating data make prediction error inevitable. In further research, we will investigate many other rating data sets to examine the universality of our findings. PMID:26735847

  14. Empirical prediction of Indian summer monsoon rainfall with different lead periods based on global SST anomalies

    NASA Astrophysics Data System (ADS)

    Pai, D. S.; Rajeevan, M.

    2006-02-01

    The main objective of this study was to develop empirical models with different seasonal lead time periods for the long range prediction of seasonal (June to September) Indian summer monsoon rainfall (ISMR). For this purpose, 13 predictors having significant and stable relationships with ISMR were derived by the correlation analysis of global grid point seasonal Sea-Surface Temperature (SST) anomalies and the tendency in the SST anomalies. The time lags of the seasonal SST anomalies were varied from 1 season to 4 years behind the reference monsoon season. The basic SST data set used was the monthly NOAA Extended Reconstructed Global SST (ERSST) data at 2° × 2° spatial grid for the period 1951 2003. The time lags of the 13 predictors derived from various areas of all three tropical ocean basins (Indian, Pacific and Atlantic Oceans) varied from 1 season to 3 years. Based on these inter-correlated predictors, 3 predictor sub sets A, B and C were formed with prediction lead time periods of 0, 1 and 2 seasons, respectively, from the beginning of the monsoon season. The selected principal components (PCs) of these predictor sets were used as the input parameters for the models A, B and C, respectively. The model development period was 1955 1984. The correct model size was derived using all-possible regressions procedure and Mallow’s “Cp” statistics.

  15. Ship classification using nonlinear features of radiated sound: an approach based on empirical mode decomposition.

    PubMed

    Bao, Fei; Li, Chen; Wang, Xinlong; Wang, Qingfu; Du, Shuanping

    2010-07-01

    Classification for ship-radiated underwater sound is one of the most important and challenging subjects in underwater acoustical signal processing. An approach to ship classification is proposed in this work based on analysis of ship-radiated acoustical noise in subspaces of intrinsic mode functions attained via the ensemble empirical mode decomposition. It is shown that detection and acquisition of stable and reliable nonlinear features become practically feasible by nonlinear analysis of the time series of individual decomposed components, each of which is simple enough and well represents an oscillatory mode of ship dynamics. Surrogate and nonlinear predictability analysis are conducted to probe and measure the nonlinearity and regularity. The results of both methods, which verify each other, substantiate that ship-radiated noises contain components with deterministic nonlinear features well serving for efficient classification of ships. The approach perhaps opens an alternative avenue in the direction toward object classification and identification. It may also import a new view of signals as complex as ship-radiated sound.

  16. Empirical Study of User Preferences Based on Rating Data of Movies

    PubMed Central

    Zhao, YingSi; Shen, Bo

    2016-01-01

    User preference plays a prominent role in many fields, including electronic commerce, social opinion, and Internet search engines. Particularly in recommender systems, it directly influences the accuracy of the recommendation. Though many methods have been presented, most of these have only focused on how to improve the recommendation results. In this paper, we introduce an empirical study of user preferences based on a set of rating data about movies. We develop a simple statistical method to investigate the characteristics of user preferences. We find that the movies have potential characteristics of closure, which results in the formation of numerous cliques with a power-law size distribution. We also find that a user related to a small clique always has similar opinions on the movies in this clique. Then, we suggest a user preference model, which can eliminate the predictions that are considered to be impracticable. Numerical results show that the model can reflect user preference with remarkable accuracy when data elimination is allowed, and random factors in the rating data make prediction error inevitable. In further research, we will investigate many other rating data sets to examine the universality of our findings. PMID:26735847

  17. Seismic facies analysis based on self-organizing map and empirical mode decomposition

    NASA Astrophysics Data System (ADS)

    Du, Hao-kun; Cao, Jun-xing; Xue, Ya-juan; Wang, Xing-jian

    2015-01-01

    Seismic facies analysis plays an important role in seismic interpretation and reservoir model building by offering an effective way to identify the changes in geofacies inter wells. The selections of input seismic attributes and their time window have an obvious effect on the validity of classification and require iterative experimentation and prior knowledge. In general, it is sensitive to noise when waveform serves as the input data to cluster analysis, especially with a narrow window. To conquer this limitation, the Empirical Mode Decomposition (EMD) method is introduced into waveform classification based on SOM. We first de-noise the seismic data using EMD and then cluster the data using 1D grid SOM. The main advantages of this method are resolution enhancement and noise reduction. 3D seismic data from the western Sichuan basin, China, are collected for validation. The application results show that seismic facies analysis can be improved and better help the interpretation. The powerful tolerance for noise makes the proposed method to be a better seismic facies analysis tool than classical 1D grid SOM method, especially for waveform cluster with a narrow window.

  18. Pseudo-empirical Likelihood-Based Method Using Calibration for Longitudinal Data with Drop-Out

    PubMed Central

    Chen, Baojiang; Zhou, Xiao-Hua; Chan, Kwun Chuen Gary

    2014-01-01

    Summary In observational studies, interest mainly lies in estimation of the population-level relationship between the explanatory variables and dependent variables, and the estimation is often undertaken using a sample of longitudinal data. In some situations, the longitudinal data sample features biases and loss of estimation efficiency due to non-random drop-out. However, inclusion of population-level information can increase estimation efficiency. In this paper we propose an empirical likelihood-based method to incorporate population-level information in a longitudinal study with drop-out. The population-level information is incorporated via constraints on functions of the parameters, and non-random drop-out bias is corrected by using a weighted generalized estimating equations method. We provide a three-step estimation procedure that makes computation easier. Some commonly used methods are compared in simulation studies, which demonstrate that our proposed method can correct the non-random drop-out bias and increase the estimation efficiency, especially for small sample size or when the missing proportion is high. In some situations, the efficiency improvement is substantial. Finally, we apply this method to an Alzheimer’s disease study. PMID:25587200

  19. Interdigitated silver-polymer-based antibacterial surface system activated by oligodynamic iontophoresis - an empirical characterization study.

    PubMed

    Shirwaiker, Rohan A; Wysk, Richard A; Kariyawasam, Subhashinie; Voigt, Robert C; Carrion, Hector; Nembhard, Harriet Black

    2014-02-01

    There is a pressing need to control the occurrences of nosocomial infections due to their detrimental effects on patient well-being and the rising treatment costs. To prevent the contact transmission of such infections via health-critical surfaces, a prophylactic surface system that consists of an interdigitated array of oppositely charged silver electrodes with polymer separations and utilizes oligodynamic iontophoresis has been recently developed. This paper presents a systematic study that empirically characterizes the effects of the surface system parameters on its antibacterial efficacy, and validates the system's effectiveness. In the first part of the study, a fractional factorial design of experiments (DOE) was conducted to identify the statistically significant system parameters. The data were used to develop a first-order response surface model to predict the system's antibacterial efficacy based on the input parameters. In the second part of the study, the effectiveness of the surface system was validated by evaluating it against four bacterial species responsible for several nosocomial infections - Staphylococcus aureus, Escherichia coli, Pseudomonas aeruginosa, and Enterococcus faecalis - alongside non-antibacterial polymer (acrylic) control surfaces. The system demonstrated statistically significant efficacy against all four bacteria. The results indicate that given a constant total effective surface area, the system designed with micro-scale features (minimum feature width: 20 μm) and activated by 15 μA direct current will provide the most effective antibacterial prophylaxis.

  20. Literature review of theory-based empirical studies examining adolescent tanning practices.

    PubMed

    Reynolds, Diane

    2007-10-01

    Lifetime exposure to ultraviolet radiation is a major risk factor for all types of skin cancer. The purpose of this manuscript is to examine theory-guided empirical studies examining adolescent tanning practices.

  1. Implementing community-based provider participation in research: an empirical study

    PubMed Central

    2012-01-01

    Background Since 2003, the United States National Institutes of Health (NIH) has sought to restructure the clinical research enterprise in the United States by promoting collaborative research partnerships between academically-based investigators and community-based physicians. By increasing community-based provider participation in research (CBPPR), the NIH seeks to advance the science of discovery by conducting research in clinical settings where most people get their care, and accelerate the translation of research results into everyday clinical practice. Although CBPPR is seen as a promising strategy for promoting the use of evidence-based clinical services in community practice settings, few empirical studies have examined the organizational factors that facilitate or hinder the implementation of CBPPR. The purpose of this study is to explore the organizational start-up and early implementation of CBPPR in community-based practice. Methods We used longitudinal, case study research methods and an organizational model of innovation implementation to theoretically guide our study. Our sample consisted of three community practice settings that recently joined the National Cancer Institute’s (NCI) Community Clinical Oncology Program (CCOP) in the United States. Data were gathered through site visits, telephone interviews, and archival documents from January 2008 to May 2011. Results The organizational model for innovation implementation was useful in identifying and investigating the organizational factors influencing start-up and early implementation of CBPPR in CCOP organizations. In general, the three CCOP organizations varied in the extent to which they achieved consistency in CBPPR over time and across physicians. All three CCOP organizations demonstrated mixed levels of organizational readiness for change. Hospital management support and resource availability were limited across CCOP organizations early on, although they improved in one CCOP organization

  2. How much does participatory flood management contribute to stakeholders' social capacity building? Empirical findings based on a triangulation of three evaluation approaches

    NASA Astrophysics Data System (ADS)

    Buchecker, M.; Menzel, S.; Home, R.

    2013-06-01

    Recent literature suggests that dialogic forms of risk communication are more effective to build stakeholders' hazard-related social capacities. In spite of the high theoretical expectations, there is a lack of univocal empirical evidence on the relevance of these effects. This is mainly due to the methodological limitations of the existing evaluation approaches. In our paper we aim at eliciting the contribution of participatory river revitalisation projects on stakeholders' social capacity building by triangulating the findings of three evaluation studies that were based on different approaches: a field-experimental, a qualitative long-term ex-post and a cross-sectional household survey approach. The results revealed that social learning and avoiding the loss of trust were more relevant benefits of participatory flood management than acceptance building. The results suggest that stakeholder involvements should be more explicitly designed as tools for long-term social learning.

  3. Development of a Web-Based Survey for Monitoring Daily Health and its Application in an Epidemiological Survey

    PubMed Central

    Ohkusa, Yasushi; Akahane, Manabu; Sano, Tomomi; Okabe, Nobuhiko; Imamura, Tomoaki

    2011-01-01

    Background Early detection of symptoms arising from exposure to pathogens, harmful substances, or environmental changes is required for timely intervention. The administration of Web-based questionnaires is a potential method for collecting information from a sample population. Objective The objective of our study was to develop a Web-based daily questionnaire for health (WDQH) for symptomatic surveillance. Methods We adopted two different survey methods to develop the WDQH: an Internet panel survey, which included participants already registered with an Internet survey company, and the Tokyo Consumers’ Co-operative Union (TCCU) Internet survey, in cooperation with the Japanese Consumers’ Co-operative Union, which recruited participants by website advertising. The Internet panel survey participants were given a fee every day for providing answers, and the survey was repeated twice with modified surveys and collection methods: Internet Panel Survey I was conducted every day, and Internet Panel Survey II was conducted every 3 days to reduce costs. We examined whether the survey remained valid by reporting health conditions on day 1 over a 3-day period, and whether the response rate would vary among groups with different incentives. In the TCCU survey, participants were given a fee only for initially registering, and health information was provided in return for survey completion. The WDQH included the demographic details of participants and prompted them to answer questions about the presence of various symptoms by email. Health information collected by the WDQH was then used for the syndromic surveillance of infection. Results Response rates averaged 47.3% for Internet Panel Survey I, 42.7% for Internet Panel Survey II, and 40.1% for the TCCU survey. During a seasonal influenza epidemic, the WDQH detected a rapid increase in the number of participants with fever through the early aberration reporting system. Conclusions We developed a health observation method

  4. The Effect of Survey Mode on High School Risk Behavior Data: A Comparison between Web and Paper-Based Surveys

    ERIC Educational Resources Information Center

    Raghupathy, Shobana; Hahn-Smith, Stephen

    2013-01-01

    There has been increasing interest in using of web-based surveys--rather than paper based surveys--for collecting data on alcohol and other drug use in middle and high schools in the US. However, prior research has indicated that respondent confidentiality is an underlying concern with online data collection especially when computer-assisted…

  5. The development of a fast radiative transfer model based on an empirical orthogonal functions (EOF) technique

    NASA Astrophysics Data System (ADS)

    Havemann, Stephan

    2006-12-01

    Remote sensing with the new generation of highly spectrally resolving instruments like the Atmospheric Research Interferometer Evaluation System (ARIES) or the assimilation of highly resolved spectra from satellites into Numerical Weather Prediction (NWP) systems requires radiative transfer computations that deliver results essentially instantaneous. This paper reports on the development of such a new fast radiative transfer model. The model is based on an Empirical Orthogonal Functions (EOF) technique. The model can be used for the simulation of sensors with different characteristics and in different spectral ranges from the solar to the infrared. For the purpose of airborne remote sensing, the fast model has been designed to work on any altitude and for slant paths whilst looking down or up. The fast model works for situations with diverse temperature and humidity profiles to an accuracy of better than 0.01K for most of the instrument channels. The EOF fast model works for clear-sky atmospheres and is applicable to atmospheres with scattering layers of aerosols or clouds. The fast model is trained with a large set of diverse atmospheric training profiles. In forward calculations corresponding high resolution spectra are obtained. An EOF analysis is performed on these spectra and only the leading EOF are retained (data compression). When the fast model is applied to a new independent profile, only the weights of the EOF need to be calculated (=predicted). Monochromatic radiances at suitable frequencies are used as predictors. The frequency selection is done by a cluster algorithm, which sorts frequencies with similar characteristics into clusters.

  6. HIV testing in national population-based surveys: experience from the Demographic and Health Surveys.

    PubMed Central

    Mishra, Vinod; Vaessen, Martin; Boerma, J. Ties; Arnold, Fred; Way, Ann; Barrere, Bernard; Cross, Anne; Hong, Rathavuth; Sangha, Jasbir

    2006-01-01

    OBJECTIVES: To describe the methods used in the Demographic and Health Surveys (DHS) to collect nationally representative data on the prevalence of human immunodeficiency virus (HIV) and assess the value of such data to country HIV surveillance systems. METHODS: During 2001-04, national samples of adult women and men in Burkina Faso, Cameroon, Dominican Republic, Ghana, Mali, Kenya, United Republic of Tanzania and Zambia were tested for HIV. Dried blood spot samples were collected for HIV testing, following internationally accepted ethical standards. The results for each country are presented by age, sex, and urban versus rural residence. To estimate the effects of non-response, HIV prevalence among non-responding males and females was predicted using multivariate statistical models for those who were tested, with a common set of predictor variables. RESULTS: Rates of HIV testing varied from 70% among Kenyan men to 92% among women in Burkina Faso and Cameroon. Despite large differences in HIV prevalence between the surveys (1-16%), fairly consistent patterns of HIV infection were observed by age, sex and urban versus rural residence, with considerably higher rates in urban areas and in women, especially at younger ages. Analysis of non-response bias indicates that although predicted HIV prevalence tended to be higher in non-tested males and females than in those tested, the overall effects of non-response on the observed national estimates of HIV prevalence are insignificant. CONCLUSIONS: Population-based surveys can provide reliable, direct estimates of national and regional HIV seroprevalence among men and women irrespective of pregnancy status. Survey data greatly enhance surveillance systems and the accuracy of national estimates in generalized epidemics. PMID:16878227

  7. Assessment of diffuse trace metal inputs into surface waters - Combining empirical estimates with process based simulations

    NASA Astrophysics Data System (ADS)

    Schindewolf, Marcus; Steinz, André; Schmidt, Jürgen

    2015-04-01

    As a result of mining activities since the 13th century, surface waters of the German Mulde catchment suffer from deleterious dissolved and sediment attached lead (Pb) and zinc (Zn) inputs. The leaching rate of trace metals with drainage water is a significant criterion for assessing trace metal concentrations of soils and associated risks of ground water pollution. However, the vertical transport rates of trace metals in soils are difficult to quantify. Monitoring is restricted to small lysimeter plots, which limits the transferability of results. Additionally the solid-liquid-transfer conditions in soils are highly variable, primarily due to the fluctuating retention time of percolating soil water. In contrast, lateral sediment attached trace metal inputs are mostly associated with soil erosion and resulting sediment inputs into surface waters. Since soil erosion by water is related to rare single events, monitoring and empirical estimates reveal visible shortcomings. This gap in knowledge can only be closed by process based model calculations. Concerning these calculations it has to be considered, that Pb and Zn are predominantly attached to the fine-grained soil particles (<0.063 mm). The selective nature of soil erosion causes a preferential transport of these fine particles, while less contaminated larger particles remain on site. Consequently trace metals are enriched in the eroded sediment compared to the origin soil. This paper aims to introduce both, a new method that allows the assessment of trace metal leaching rates from contaminated top soils for standardised transfer conditions and a process based modelling approach for sediment attached trace metal inputs into surface waters. Pb and Zn leaching rates amounts to 20 Mg ha-1 yr-1 resp. 114 Mg ha-1 yr-1. Deviations to observed dissolved trace metal yields at the Bad Düben gauging station are caused by plant uptake and subsoil retention. Sediment attached Pb and Zn input rates amounts to 114 Mg ha-1 yr

  8. Mental Health Functioning in the Human Rights Field: Findings from an International Internet-Based Survey

    PubMed Central

    Joscelyne, Amy; Knuckey, Sarah; Satterthwaite, Margaret L.; Bryant, Richard A.; Li, Meng; Qian, Meng; Brown, Adam D.

    2015-01-01

    Human rights advocates play a critical role in promoting respect for human rights world-wide, and engage in a broad range of strategies, including documentation of rights violations, monitoring, press work and report-writing, advocacy, and litigation. However, little is known about the impact of human rights work on the mental health of human rights advocates. This study examined the mental health profile of human rights advocates and risk factors associated with their psychological functioning. 346 individuals currently or previously working in the field of human rights completed an internet-based survey regarding trauma exposure, depression, posttraumatic stress disorder (PTSD), resilience and occupational burnout. PTSD was measured with the Posttraumatic Stress Disorder Checklist-Civilian Version (PCL-C) and depression was measured with the Patient History Questionnaire-9 (PHQ-9). These findings revealed that among human rights advocates that completed the survey, 19.4% met criteria for PTSD, 18.8% met criteria for subthreshold PTSD, and 14.7% met criteria for depression. Multiple linear regressions revealed that after controlling for symptoms of depression, PTSD symptom severity was predicted by human rights-related trauma exposure, perfectionism and negative self-appraisals about human rights work. In addition, after controlling for symptoms of PTSD, depressive symptoms were predicted by perfectionism and lower levels of self-efficacy. Survey responses also suggested high levels of resilience: 43% of responders reported minimal symptoms of PTSD. Although survey responses suggest that many human rights workers are resilient, they also suggest that human rights work is associated with elevated rates of PTSD and depression. The field of human rights would benefit from further empirical research, as well as additional education and training programs in the workplace about enhancing resilience in the context of human rights work. PMID:26700305

  9. Mental Health Functioning in the Human Rights Field: Findings from an International Internet-Based Survey.

    PubMed

    Joscelyne, Amy; Knuckey, Sarah; Satterthwaite, Margaret L; Bryant, Richard A; Li, Meng; Qian, Meng; Brown, Adam D

    2015-01-01

    Human rights advocates play a critical role in promoting respect for human rights world-wide, and engage in a broad range of strategies, including documentation of rights violations, monitoring, press work and report-writing, advocacy, and litigation. However, little is known about the impact of human rights work on the mental health of human rights advocates. This study examined the mental health profile of human rights advocates and risk factors associated with their psychological functioning. 346 individuals currently or previously working in the field of human rights completed an internet-based survey regarding trauma exposure, depression, posttraumatic stress disorder (PTSD), resilience and occupational burnout. PTSD was measured with the Posttraumatic Stress Disorder Checklist-Civilian Version (PCL-C) and depression was measured with the Patient History Questionnaire-9 (PHQ-9). These findings revealed that among human rights advocates that completed the survey, 19.4% met criteria for PTSD, 18.8% met criteria for subthreshold PTSD, and 14.7% met criteria for depression. Multiple linear regressions revealed that after controlling for symptoms of depression, PTSD symptom severity was predicted by human rights-related trauma exposure, perfectionism and negative self-appraisals about human rights work. In addition, after controlling for symptoms of PTSD, depressive symptoms were predicted by perfectionism and lower levels of self-efficacy. Survey responses also suggested high levels of resilience: 43% of responders reported minimal symptoms of PTSD. Although survey responses suggest that many human rights workers are resilient, they also suggest that human rights work is associated with elevated rates of PTSD and depression. The field of human rights would benefit from further empirical research, as well as additional education and training programs in the workplace about enhancing resilience in the context of human rights work. PMID:26700305

  10. Mental Health Functioning in the Human Rights Field: Findings from an International Internet-Based Survey.

    PubMed

    Joscelyne, Amy; Knuckey, Sarah; Satterthwaite, Margaret L; Bryant, Richard A; Li, Meng; Qian, Meng; Brown, Adam D

    2015-01-01

    Human rights advocates play a critical role in promoting respect for human rights world-wide, and engage in a broad range of strategies, including documentation of rights violations, monitoring, press work and report-writing, advocacy, and litigation. However, little is known about the impact of human rights work on the mental health of human rights advocates. This study examined the mental health profile of human rights advocates and risk factors associated with their psychological functioning. 346 individuals currently or previously working in the field of human rights completed an internet-based survey regarding trauma exposure, depression, posttraumatic stress disorder (PTSD), resilience and occupational burnout. PTSD was measured with the Posttraumatic Stress Disorder Checklist-Civilian Version (PCL-C) and depression was measured with the Patient History Questionnaire-9 (PHQ-9). These findings revealed that among human rights advocates that completed the survey, 19.4% met criteria for PTSD, 18.8% met criteria for subthreshold PTSD, and 14.7% met criteria for depression. Multiple linear regressions revealed that after controlling for symptoms of depression, PTSD symptom severity was predicted by human rights-related trauma exposure, perfectionism and negative self-appraisals about human rights work. In addition, after controlling for symptoms of PTSD, depressive symptoms were predicted by perfectionism and lower levels of self-efficacy. Survey responses also suggested high levels of resilience: 43% of responders reported minimal symptoms of PTSD. Although survey responses suggest that many human rights workers are resilient, they also suggest that human rights work is associated with elevated rates of PTSD and depression. The field of human rights would benefit from further empirical research, as well as additional education and training programs in the workplace about enhancing resilience in the context of human rights work.

  11. Avian survey and field guide for Osan Air Base, Korea.

    SciTech Connect

    Levenson, J.

    2006-12-05

    This report summarizes the results of the avian surveys conducted at Osan Air Base (AB). This ongoing survey is conducted to comply with requirements of the Environmental Governing Standards (EGS) for the Republic of Korea, the Integrated Natural Resources Management Plan (INRMP) for Osan AB, and the 51st Fighter Wing's Bird Aircraft Strike Hazard (BASH) Plan. One hundred ten bird species representing 35 families were identified and recorded. Seven species are designated as Natural Monuments, and their protection is accorded by the Korean Ministry of Culture and Tourism. Three species appear on the Korean Association for Conservation of Nature's (KACN's) list of Reserved Wild Species and are protected by the Korean Ministry of Environment. Combined, ten different species are Republic of Korea (ROK)-protected. The primary objective of the avian survey at Osan AB was to determine what species of birds are present on the airfield and their respective habitat requirements during the critical seasons of the year. This requirement is specified in Annex J.14.c of the 51st Fighter BASH Plan 91-212 (51 FW OPLAN 91-212). The second objective was to initiate surveys to determine what bird species are present on Osan AB throughout the year and from the survey results, determine if threatened, endangered, or other Korean-listed bird species are present on Osan AB. This overall census satisfies Criterion 13-3.e of the EGS for Korea. The final objective was to formulate management strategies within Osan AB's operational requirements to protect and enhance habitats of known threatened, endangered, and ROK-protected species in accordance with EGS Criterion 13-3.a that are also favorable for the reproduction of indigenous species in accordance with the EGS Criterion 13-3.h.

  12. Competence-based demands made of senior physicians: an empirical study to evaluate leadership competencies.

    PubMed

    Lehr, Bosco; Ostermann, Herwig; Schubert, Harald

    2011-01-01

    As a result of more economising in German hospitals, changes evolve in organising the deployment of senior medical staff. New demands are made of senior hospital management. Leadership competencies in the training and development of physicians are of prime importance to the successful perception of managerial responsibilities. The present study investigates the actual and targeted demands of leadership made of senior medical staff in terms of how these demands are perceived. To this end, the demands of leadership were surveyed using a competence-based questionnaire and investigated with a view to potentials in professional development by way of example of the senior management of psychiatric hospitals in Germany. In all, the results show high ratings in personal performance, the greatest significance being attributed to value-oriented competence in the actual assessment of demands on leadership. Besides gender-specific differences in the actual assessments of single fields of competence, the greatest differences between the targeted and the actual demands are, in all, shown to be in the competencies of self-management and communication. Competence-based core areas in leadership can be demonstrated for the professional development of physicians and an adaptive mode of procedure deduced.

  13. The EMPIRE Survey: Systematic Variations in the Dense Gas Fraction and Star Formation Efficiency from Full-disk Mapping of M51

    NASA Astrophysics Data System (ADS)

    Bigiel, Frank; Leroy, Adam K.; Jiménez-Donaire, Maria J.; Pety, Jérôme; Usero, Antonio; Cormier, Diane; Bolatto, Alberto; Garcia-Burillo, Santiago; Colombo, Dario; González-García, Manuel; Hughes, Annie; Kepley, Amanda A.; Kramer, Carsten; Sandstrom, Karin; Schinnerer, Eva; Schruba, Andreas; Schuster, Karl; Tomicic, Neven; Zschaechner, Laura

    2016-05-01

    We present the first results from the EMPIRE survey, an IRAM large program that is mapping tracers of high-density molecular gas across the disks of nine nearby star-forming galaxies. Here, we present new maps of the 3 mm transitions of HCN, HCO+, and HNC across the whole disk of our pilot target, M51. As expected, dense gas correlates with tracers of recent star formation, filling the “luminosity gap” between Galactic cores and whole galaxies. In detail, we show that both the fraction of gas that is dense, {f}{dense} traced by HCN/CO, and the rate at which dense gas forms stars, {{SFE}}{dense} traced by IR/HCN, depend on environment in the galaxy. The sense of the dependence is that high-surface-density, high molecular gas fraction regions of the galaxy show high dense gas fractions and low dense gas star formation efficiencies. This agrees with recent results for individual pointings by Usero et al. but using unbiased whole-galaxy maps. It also agrees qualitatively with the behavior observed contrasting our own Solar Neighborhood with the central regions of the Milky Way. The sense of the trends can be explained if the dense gas fraction tracks interstellar pressure but star formation occurs only in regions of high density contrast.

  14. Coalitional affiliation as a missing link between ethnic polarization and well-being: An empirical test from the European Social Survey.

    PubMed

    Firat, Rengin B; Boyer, Pascal

    2015-09-01

    Many studies converge in suggesting (a) that ethnic and racial minorities fare worse than host populations in reported well-being and objective measures of health and (b) that ethnic/racial diversity has a negative impact on various measures of social trust and well-being, including in the host or majority population. However, there is much uncertainty about the processes that connect diversity variables with personal outcomes. In this paper, we are particularly interested in different levels of coalitional affiliation, which refers to people's social allegiances that guide their expectations of social support, in-group strength and cohesion. We operationalize coalitional affiliation as the extent to which people rely on a homogeneous social network, and we measure it with indicators of friendships across ethnic boundaries and frequency of contact with friends. Using multi-level models and data from the European Social Survey (Round 1, 2002-2003) for 19 countries, we demonstrate that coalitional affiliation provides an empirically reliable, as well as theoretically coherent, explanation for various effects of ethnic/racial diversity.

  15. An Empirically-based Steady-state Friction Law and its Implications for Fault Stability

    NASA Astrophysics Data System (ADS)

    Spagnuolo, E.; Nielsen, S. B.; Di Toro, G.; Violay, M.

    2015-12-01

    Empirically-based rate-and-state friction laws (RSFL) have been proposed to model the dependence of friction forces with slip and time. The relevance of the RSFL for earthquakes mechanics is that few constitutive parameters (e.g. A-B= dτ/dlog(V) with τ and V the shear stress and slip rate respectively, allow us to define the stability conditions of a fault. According to RSFL if A-B> 0, τ increases with V (rate-hardening behavior) resulting in an unconditionally stable behavior; if A-B< 0, τ decreases with V (rate-weakening behavior) potentially resulting in an unstable behavior leading to dynamic runaway. Given that τ at steady state conditions allows us also to define a critical fault stiffness, the RSFL determine a condition of stability for faults as their stiffness approaches the critical conditions. However, the conditions of fault stability, determined by the critical stiffness under the assumption of either a rate-weakening or a rate-hardening behavior, might be restrictive given that frictional properties sensibly change as a function of slip or slip rate. Moreover, the RSFL were determined from experiments conducted at sub-seismic slip rates (< 1 cm/s) and their extrapolation to earthquake deformation conditions remains questionable on the basis of the experimental evidence of large dynamic weakening at seismic slip rates and the plethora of slip events which characterize the seismic cycle. Here, we propose a modified RSFL based on the review of a large published and unpublished dataset of rock-friction experiments performed with different testing machines (rotary shear, bi-axial, tri-axial). The modified RSFL is valid at steady-state conditions from sub-seismic to seismic slip rates (0.1 μm/s

  16. Simulation of Long Lived Tracers Using an Improved Empirically Based Two-Dimensional Model Transport Algorithm

    NASA Technical Reports Server (NTRS)

    Fleming, E. L.; Jackman, C. H.; Stolarski, R. S.; Considine, D. B.

    1998-01-01

    We have developed a new empirically-based transport algorithm for use in our GSFC two-dimensional transport and chemistry model. The new algorithm contains planetary wave statistics, and parameterizations to account for the effects due to gravity waves and equatorial Kelvin waves. As such, this scheme utilizes significantly more information compared to our previous algorithm which was based only on zonal mean temperatures and heating rates. The new model transport captures much of the qualitative structure and seasonal variability observed in long lived tracers, such as: isolation of the tropics and the southern hemisphere winter polar vortex; the well mixed surf-zone region of the winter sub-tropics and mid-latitudes; the latitudinal and seasonal variations of total ozone; and the seasonal variations of mesospheric H2O. The model also indicates a double peaked structure in methane associated with the semiannual oscillation in the tropical upper stratosphere. This feature is similar in phase but is significantly weaker in amplitude compared to the observations. The model simulations of carbon-14 and strontium-90 are in good agreement with observations, both in simulating the peak in mixing ratio at 20-25 km, and the decrease with altitude in mixing ratio above 25 km. We also find mostly good agreement between modeled and observed age of air determined from SF6 outside of the northern hemisphere polar vortex. However, observations inside the vortex reveal significantly older air compared to the model. This is consistent with the model deficiencies in simulating CH4 in the northern hemisphere winter high latitudes and illustrates the limitations of the current climatological zonal mean model formulation. The propagation of seasonal signals in water vapor and CO2 in the lower stratosphere showed general agreement in phase, and the model qualitatively captured the observed amplitude decrease in CO2 from the tropics to midlatitudes. However, the simulated seasonal

  17. Universal Design for Instruction in Postsecondary Education: A Systematic Review of Empirically Based Articles

    ERIC Educational Resources Information Center

    Roberts, Kelly D.; Park, Hye Jin; Brown, Steven; Cook, Bryan

    2011-01-01

    Universal Design for Instruction (UDI) in postsecondary education is a relatively new concept/framework that has generated significant support. The purpose of this literature review was to examine existing empirical research, including qualitative, quantitative, and mixed methods, on the use of UDI (and related terms) in postsecondary education.…

  18. Comparisons of experiment with cellulose models based on electronic structure and empirical force field theories

    Technology Transfer Automated Retrieval System (TEKTRAN)

    Studies of cellobiose conformations with HF/6-31G* and B3LYP/6-31+G*quantum theory [1] gave a reference for studies with the much faster empirical methods such as MM3, MM4, CHARMM and AMBER. The quantum studies also enable a substantial reduction in the number of exo-cyclic group orientations that...

  19. Model Selection for Equating Testlet-Based Tests in the NEAT Design: An Empirical Study

    ERIC Educational Resources Information Center

    He, Wei; Li, Feifei; Wolfe, Edward W.; Mao, Xia

    2012-01-01

    For those tests solely composed of testlets, local item independency assumption tends to be violated. This study, by using empirical data from a large-scale state assessment program, was interested in investigates the effects of using different models on equating results under the non-equivalent group anchor-test (NEAT) design. Specifically, the…

  20. Understanding Transactional Distance in Web-Based Learning Environments: An Empirical Study

    ERIC Educational Resources Information Center

    Huang, Xiaoxia; Chandra, Aruna; DePaolo, Concetta A.; Simmons, Lakisha L.

    2016-01-01

    Transactional distance is an important pedagogical theory in distance education that calls for more empirical support. The purpose of this study was to verify the theory by operationalizing and examining the relationship of (1) dialogue, structure and learner autonomy to transactional distance, and (2) environmental factors and learner demographic…

  1. A Survey of Artificial Immune System Based Intrusion Detection

    PubMed Central

    Li, Tao; Hu, Xinlei; Wang, Feng; Zou, Yang

    2014-01-01

    In the area of computer security, Intrusion Detection (ID) is a mechanism that attempts to discover abnormal access to computers by analyzing various interactions. There is a lot of literature about ID, but this study only surveys the approaches based on Artificial Immune System (AIS). The use of AIS in ID is an appealing concept in current techniques. This paper summarizes AIS based ID methods from a new view point; moreover, a framework is proposed for the design of AIS based ID Systems (IDSs). This framework is analyzed and discussed based on three core aspects: antibody/antigen encoding, generation algorithm, and evolution mode. Then we collate the commonly used algorithms, their implementation characteristics, and the development of IDSs into this framework. Finally, some of the future challenges in this area are also highlighted. PMID:24790549

  2. A survey of artificial immune system based intrusion detection.

    PubMed

    Yang, Hua; Li, Tao; Hu, Xinlei; Wang, Feng; Zou, Yang

    2014-01-01

    In the area of computer security, Intrusion Detection (ID) is a mechanism that attempts to discover abnormal access to computers by analyzing various interactions. There is a lot of literature about ID, but this study only surveys the approaches based on Artificial Immune System (AIS). The use of AIS in ID is an appealing concept in current techniques. This paper summarizes AIS based ID methods from a new view point; moreover, a framework is proposed for the design of AIS based ID Systems (IDSs). This framework is analyzed and discussed based on three core aspects: antibody/antigen encoding, generation algorithm, and evolution mode. Then we collate the commonly used algorithms, their implementation characteristics, and the development of IDSs into this framework. Finally, some of the future challenges in this area are also highlighted. PMID:24790549

  3. Methodology of the National School-based Health Survey in Malaysia, 2012.

    PubMed

    Yusoff, Fadhli; Saari, Riyanti; Naidu, Balkish M; Ahmad, Noor Ani; Omar, Azahadi; Aris, Tahir

    2014-09-01

    The National School-Based Health Survey 2012 was a nationwide school health survey of students in Standard 4 to Form 5 (10-17 years of age), who were schooling in government schools in Malaysia during the period of data collection. The survey comprised 3 subsurveys: the Global School Health Survey (GSHS), the Mental Health Survey, and the National School-Based Nutrition Survey. The aim of the survey was to provide data on the health status of adolescents in Malaysia toward strengthening the adolescent health program in the country. The design of the survey was created to fulfill the requirements of the 3 subsurveys. A 2-stage stratified sampling method was adopted in the sampling. The methods for data collection were via questionnaire and physical examination. The National School-Based Health Survey 2012 adopted an appropriate methodology for a school-based survey to ensure valid and reliable findings.

  4. Empirical model of equatorial electrojet based on ground-based magnetometer data during solar minimum in fall

    NASA Astrophysics Data System (ADS)

    Hamid, Nurul Shazana Abdul; Liu, Huixin; Uozumi, Teiji; Yoshikawa, Akimasa

    2015-12-01

    In this study, we constructed an empirical model of the equatorial electrojet (EEJ), including local time and longitudinal dependence, based on simultaneous data from 12 magnetometer stations located in six longitude sectors. An analysis was carried out using the equatorial electrojet index, EUEL, calculated from the geomagnetic northward H component. The magnetic EEJ strength is calculated as the difference between the normalized EUEL index of the magnetic dip equator station and the normalized EUEL index of the off-dip equator station located beyond the EEJ band. Analysis showed that this current is always strongest in the South American sector, regardless of local time (LT), and weakest in the Indian sector during 0900 and 1000 LT, but shifted to the African sector during 1100 to 1400 LT. These longitude variations of EEJ roughly follow variations of the inversed main field strength along the dip equator, except for the Indian and Southeast Asian sectors. The result showed that the EEJ component derived from the model exhibits a similar pattern with measured EEJ from ground data during noontime, mainly before 1300 LT.

  5. Developing Empirically-Based Ground Truth Criteria for Varying Geological Complexity using Regional Seismic Networks

    NASA Astrophysics Data System (ADS)

    Brazier, R. A.; O'Donnell, J.; Boomer, K.; Nyblade, A.; Kokoska, J.; Liu, S.

    2011-12-01

    We have extended the approaches of Bondár et al. (2004), Bondár and McLaughlin (2009), and Boomer et al. (2010) by developing new empirically based ground truth (EBGT) local criteria for a variety of geologic settings for which data sets containing GT0 events (explosions and mine tremors) are available, local crustal structure is well known, and hand-picked arrival times have been obtained. Boomer et al. (2010) describes the development of local criteria for the simple structure of the Archean Kaapvaal Craton in southern Africa. Continuing the development of local criteria in regions of varying geologic complexity, we now have criteria for the Main Ethiopian Rift and for the Tibetan plateau. In the geologically complex region of the Main Ethiopian Rift, we use the 2003 Ethiopia-Afar Geoscientific Lithosphere Experiment (EAGLE; Maguire et al., 2003) data to obtain EBGT595% criteria. Four of the 25 large refraction line shots were used as reference events to develop the criteria; the remainder of the shots are used for verification. We require an event to be recorded on at least 8 stations within the Pg/Pn crossover distance and a network quality metric (Bondár and McLaughlin, 2009) less than 0.43 for an event to be classified as EBGT595%. Using these criteria to identify GT events within the Ethiopian Broadband Seismic Experiment, we have identified ten events to add to the NNSA knowledge database. There are also 196 potential events from the EAGLE dataset from which we expect to yield an additional 20 GT events. The crust and upper mantle structure of the Tibetan plateau is arguably more complicated than for the Kaapvaal Craton and yet less complicated than the Main Ethiopian Rift, and includes a number of prominent suture zones. Five of the 11 larger shots from the International Deep Profiling of Tibet and the Himalaya (INDEPTH III) refraction line were used to develop the criteria. The remaining 6 shots are used to validate the criteria. The criteria for Tibet

  6. Comparing and combining physically-based and empirically-based approaches for estimating the hydrology of ungauged catchments

    NASA Astrophysics Data System (ADS)

    Booker, D. J.; Woods, R. A.

    2014-01-01

    Methods for estimating various hydrological indices at ungauged sites were compared.Methods included a TopNet rainfall-runoff model and a Random Forest empirical model.TopNet estimates were improved through correction using Random Forest estimates.Random Forests provided the best estimates of all indices except mean flow.Mean flow was best estimated using an already published empirical method.

  7. Automatic emotion recognition based on body movement analysis: a survey.

    PubMed

    Zacharatos, Haris; Gatzoulis, Christos; Chrysanthou, Yiorgos L

    2014-01-01

    Humans are emotional beings, and their feelings influence how they perform and interact with computers. One of the most expressive modalities for humans is body posture and movement, which researchers have recently started exploiting for emotion recognition. This survey describes emerging techniques and modalities related to emotion recognition based on body movement, as well as recent advances in automatic emotion recognition. It also describes application areas and notation systems and explains the importance of movement segmentation. It then discusses unsolved problems and provides promising directions for future research. The Web extra (a PDF file) contains tables with additional information related to the article. PMID:25216477

  8. Social-Emotional Well-Being and Resilience of Children in Early Childhood Settings--PERIK: An Empirically Based Observation Scale for Practitioners

    ERIC Educational Resources Information Center

    Mayr, Toni; Ulich, Michaela

    2009-01-01

    Compared with the traditional focus on developmental problems, research on positive development is relatively new. Empirical research in children's well-being has been scarce. The aim of this study was to develop a theoretically and empirically based instrument for practitioners to observe and assess preschool children's well-being in early…

  9. A Reliability Test of a Complex System Based on Empirical Likelihood

    PubMed Central

    Zhang, Jun; Hui, Yongchang

    2016-01-01

    To analyze the reliability of a complex system described by minimal paths, an empirical likelihood method is proposed to solve the reliability test problem when the subsystem distributions are unknown. Furthermore, we provide a reliability test statistic of the complex system and extract the limit distribution of the test statistic. Therefore, we can obtain the confidence interval for reliability and make statistical inferences. The simulation studies also demonstrate the theorem results. PMID:27760130

  10. Organizational Learning, Strategic Flexibility and Business Model Innovation: An Empirical Research Based on Logistics Enterprises

    NASA Astrophysics Data System (ADS)

    Bao, Yaodong; Cheng, Lin; Zhang, Jian

    Using the data of 237 Jiangsu logistics firms, this paper empirically studies the relationship among organizational learning capability, business model innovation, strategic flexibility. The results show as follows; organizational learning capability has positive impacts on business model innovation performance; strategic flexibility plays mediating roles on the relationship between organizational learning capability and business model innovation; interaction among strategic flexibility, explorative learning and exploitative learning play significant roles in radical business model innovation and incremental business model innovation.

  11. Deep in Data: Empirical Data Based Software Accuracy Testing Using the Building America Field Data Repository: Preprint

    SciTech Connect

    Neymark, J.; Roberts, D.

    2013-06-01

    An opportunity is available for using home energy consumption and building description data to develop a standardized accuracy test for residential energy analysis tools. That is, to test the ability of uncalibrated simulations to match real utility bills. Empirical data collected from around the United States have been translated into a uniform Home Performance Extensible Markup Language format that may enable software developers to create translators to their input schemes for efficient access to the data. This may facilitate the possibility of modeling many homes expediently, and thus implementing software accuracy test cases by applying the translated data. This paper describes progress toward, and issues related to, developing a usable, standardized, empirical data-based software accuracy test suite.

  12. 77 FR 73432 - Proposed Information Collection; Comment Request; Survey of Shore-Based and Boat-Based Non...

    Federal Register 2010, 2011, 2012, 2013, 2014

    2012-12-10

    ... Shore-Based and Boat-Based Non-Commercial Fishing on St. Croix, U.S. Virgin Islands AGENCY: National... consist of a survey of non-commercial, shore-based fishers. This survey will ascertain the catch, ] effort... place over a 12-month period, will be comprised of two data collection activities: (1) A survey of...

  13. Smartphone-Based, Self-Administered Intervention System for Alcohol Use Disorders: Theory and Empirical Evidence Basis

    PubMed Central

    Dulin, Patrick L.; Gonzalez, Vivian M.; King, Diane K.; Giroux, Danielle; Bacon, Samantha

    2013-01-01

    Advances in mobile technology provide an opportunity to deliver in-the-moment interventions to individuals with alcohol use disorders, yet availability of effective “apps” that deliver evidence-based interventions is scarce. We developed an immediately available, portable, smartphone-based intervention system whose purpose is to provide stand-alone, self-administered assessment and intervention. In this paper, we describe how theory and empirical evidence, combined with smartphone functionality contributed to the construction of a user-friendly, engaging alcohol intervention. With translation in mind, we discuss how we selected appropriate intervention components including assessments, feedback and tools, that work together to produce the hypothesized outcomes. PMID:24347811

  14. Smartphone-Based, Self-Administered Intervention System for Alcohol Use Disorders: Theory and Empirical Evidence Basis.

    PubMed

    Dulin, Patrick L; Gonzalez, Vivian M; King, Diane K; Giroux, Danielle; Bacon, Samantha

    2013-01-01

    Advances in mobile technology provide an opportunity to deliver in-the-moment interventions to individuals with alcohol use disorders, yet availability of effective "apps" that deliver evidence-based interventions is scarce. We developed an immediately available, portable, smartphone-based intervention system whose purpose is to provide stand-alone, self-administered assessment and intervention. In this paper, we describe how theory and empirical evidence, combined with smartphone functionality contributed to the construction of a user-friendly, engaging alcohol intervention. With translation in mind, we discuss how we selected appropriate intervention components including assessments, feedback and tools, that work together to produce the hypothesized outcomes. PMID:24347811

  15. Sensor Systems Based on FPGAs and Their Applications: A Survey

    PubMed Central

    de la Piedra, Antonio; Braeken, An; Touhafi, Abdellah

    2012-01-01

    In this manuscript, we present a survey of designs and implementations of research sensor nodes that rely on FPGAs, either based upon standalone platforms or as a combination of microcontroller and FPGA. Several current challenges in sensor networks are distinguished and linked to the features of modern FPGAs. As it turns out, low-power optimized FPGAs are able to enhance the computation of several types of algorithms in terms of speed and power consumption in comparison to microcontrollers of commercial sensor nodes. We show that architectures based on the combination of microcontrollers and FPGA can play a key role in the future of sensor networks, in fields where processing capabilities such as strong cryptography, self-testing and data compression, among others, are paramount.

  16. Fault Location Based on Synchronized Measurements: A Comprehensive Survey

    PubMed Central

    Al-Mohammed, A. H.; Abido, M. A.

    2014-01-01

    This paper presents a comprehensive survey on transmission and distribution fault location algorithms that utilize synchronized measurements. Algorithms based on two-end synchronized measurements and fault location algorithms on three-terminal and multiterminal lines are reviewed. Series capacitors equipped with metal oxide varistors (MOVs), when set on a transmission line, create certain problems for line fault locators and, therefore, fault location on series-compensated lines is discussed. The paper reports the work carried out on adaptive fault location algorithms aiming at achieving better fault location accuracy. Work associated with fault location on power system networks, although limited, is also summarized. Additionally, the nonstandard high-frequency-related fault location techniques based on wavelet transform are discussed. Finally, the paper highlights the area for future research. PMID:24701191

  17. Fault location based on synchronized measurements: a comprehensive survey.

    PubMed

    Al-Mohammed, A H; Abido, M A

    2014-01-01

    This paper presents a comprehensive survey on transmission and distribution fault location algorithms that utilize synchronized measurements. Algorithms based on two-end synchronized measurements and fault location algorithms on three-terminal and multiterminal lines are reviewed. Series capacitors equipped with metal oxide varistors (MOVs), when set on a transmission line, create certain problems for line fault locators and, therefore, fault location on series-compensated lines is discussed. The paper reports the work carried out on adaptive fault location algorithms aiming at achieving better fault location accuracy. Work associated with fault location on power system networks, although limited, is also summarized. Additionally, the nonstandard high-frequency-related fault location techniques based on wavelet transform are discussed. Finally, the paper highlights the area for future research.

  18. Comparison of ensemble post-processing approaches, based on empirical and dynamical error modelisation of rainfall-runoff model forecasts

    NASA Astrophysics Data System (ADS)

    Chardon, J.; Mathevet, T.; Le Lay, M.; Gailhard, J.

    2012-04-01

    In the context of a national energy company (EDF : Electricité de France), hydro-meteorological forecasts are necessary to ensure safety and security of installations, meet environmental standards and improve water ressources management and decision making. Hydrological ensemble forecasts allow a better representation of meteorological and hydrological forecasts uncertainties and improve human expertise of hydrological forecasts, which is essential to synthesize available informations, coming from different meteorological and hydrological models and human experience. An operational hydrological ensemble forecasting chain has been developed at EDF since 2008 and is being used since 2010 on more than 30 watersheds in France. This ensemble forecasting chain is characterized ensemble pre-processing (rainfall and temperature) and post-processing (streamflow), where a large human expertise is solicited. The aim of this paper is to compare 2 hydrological ensemble post-processing methods developed at EDF in order improve ensemble forecasts reliability (similar to Monatanari &Brath, 2004; Schaefli et al., 2007). The aim of the post-processing methods is to dress hydrological ensemble forecasts with hydrological model uncertainties, based on perfect forecasts. The first method (called empirical approach) is based on a statistical modelisation of empirical error of perfect forecasts, by streamflow sub-samples of quantile class and lead-time. The second method (called dynamical approach) is based on streamflow sub-samples of quantile class and streamflow variation, and lead-time. On a set of 20 watersheds used for operational forecasts, results show that both approaches are necessary to ensure a good post-processing of hydrological ensemble, allowing a good improvement of reliability, skill and sharpness of ensemble forecasts. The comparison of the empirical and dynamical approaches shows the limits of the empirical approach which is not able to take into account hydrological

  19. A new high frequency Earth rotation model based on an empirical ocean tide model from satellite altimetry

    NASA Astrophysics Data System (ADS)

    Madzak, Matthias; Böhm, Sigrid; Böhm, Johannes; Bosch, Wolfgang; Hagedoorn, Jan; Schuh, Harald

    2014-05-01

    A new model for Earth rotation variations based on ocean tide models is highly desirable in order to close the gap between geophysical Earth rotation models and geodetic observations. The current high frequency Earth rotation model mentioned in the IERS Conventions 2010 and thus used by most analysis institutions was developed in 1994. Since then several satellite missions have collected lots of altimetry data and were used to obtain new ocean tide models. Due to the increase of accuracy and resolution of these models, we will develop an improved Earth rotation model for (sub-) daily periods. In order to reduce (hydrodynamic) modeling effects, we use the empirical ocean tide model EOT11a, provided by DGFI, Munich. Global oceanic currents, which are required for ocean tidal angular momentum but not included in empirical models, are obtained using a linearized and simplified Navier-Stokes equation (Ray, 2001). We compare the new model with the model from the IERS Conventions 2010 as well as with an empirical Earth rotation model (Artz et al., 2011) and show the expected differences in the analysis of VLBI observations. For this purpose we use the Vienna VLBI Software (VieVS).

  20. [DGRW-update: neurology--from empirical strategies towards evidence based interventions].

    PubMed

    Schupp, W

    2011-12-01

    Stroke, Multiple Sclerosis (MS), traumatic brain injuries (TBI) and neuropathies are the most important diseases in neurological rehabilitation financed by the German Pension Insurance. The primary goal is vocational (re)integration. Driven by multiple findings of neuroscience research the traditional holistic approach with mainly empirically derived strategies was developed further and improved by new evidence-based interventions. This process had been, and continues to be, necessary to meet the health-economic pressures for ever shorter and more efficient rehab measures. Evidence-based interventions refer to symptom-oriented measures, to team-management concepts, as well as to education and psychosocial interventions. Drug therapy and/or neurophysiological measures can be added to increase neuroregeneration and neuroplasticity. Evidence-based aftercare concepts support sustainability and steadiness of rehab results.Mirror therapy, robot-assisted training, mental training, task-specific training, and above all constraint-induced movement therapy (CIMT) can restore motor arm and hand functions. Treadmill training and robot-assisted training improve stance and gait. Botulinum toxine injections in combination with physical and redressing methods are superior in managing spasticity. Guideline-oriented management of associated pain syndromes (myofascial, neuropathic, complex-regional=dystrophic) improve primary outcome and quality of life. Drug therapy with so-called co-analgetics and physical therapy play an important role in pain management. Swallowing disorders lead to higher mortality and morbidity in the acute phase; stepwise diagnostics (screening, endoscopy, radiology) and specific swallowing therapy can reduce these risks and frequently can restore normal eating und drinking.In our modern industrial societies communicative and cognitive disturbances are more impairing than the above mentioned disorders. Speech and language therapy (SLT) is dominant in

  1. [DGRW-update: neurology--from empirical strategies towards evidence based interventions].

    PubMed

    Schupp, W

    2011-12-01

    Stroke, Multiple Sclerosis (MS), traumatic brain injuries (TBI) and neuropathies are the most important diseases in neurological rehabilitation financed by the German Pension Insurance. The primary goal is vocational (re)integration. Driven by multiple findings of neuroscience research the traditional holistic approach with mainly empirically derived strategies was developed further and improved by new evidence-based interventions. This process had been, and continues to be, necessary to meet the health-economic pressures for ever shorter and more efficient rehab measures. Evidence-based interventions refer to symptom-oriented measures, to team-management concepts, as well as to education and psychosocial interventions. Drug therapy and/or neurophysiological measures can be added to increase neuroregeneration and neuroplasticity. Evidence-based aftercare concepts support sustainability and steadiness of rehab results.Mirror therapy, robot-assisted training, mental training, task-specific training, and above all constraint-induced movement therapy (CIMT) can restore motor arm and hand functions. Treadmill training and robot-assisted training improve stance and gait. Botulinum toxine injections in combination with physical and redressing methods are superior in managing spasticity. Guideline-oriented management of associated pain syndromes (myofascial, neuropathic, complex-regional=dystrophic) improve primary outcome and quality of life. Drug therapy with so-called co-analgetics and physical therapy play an important role in pain management. Swallowing disorders lead to higher mortality and morbidity in the acute phase; stepwise diagnostics (screening, endoscopy, radiology) and specific swallowing therapy can reduce these risks and frequently can restore normal eating und drinking.In our modern industrial societies communicative and cognitive disturbances are more impairing than the above mentioned disorders. Speech and language therapy (SLT) is dominant in

  2. Accounting protesting and warm glow bidding in Contingent Valuation surveys considering the management of environmental goods--an empirical case study assessing the value of protecting a Natura 2000 wetland area in Greece.

    PubMed

    Grammatikopoulou, Ioanna; Olsen, Søren Bøye

    2013-11-30

    Based on a Contingent Valuation survey aiming to reveal the willingness to pay (WTP) for conservation of a wetland area in Greece, we show how protest and warm glow motives can be taken into account when modeling WTP. In a sample of more than 300 respondents, we find that 54% of the positive bids are rooted to some extent in warm glow reasoning while 29% of the zero bids can be classified as expressions of protest rather than preferences. In previous studies, warm glow bidders are only rarely identified while protesters are typically identified and excluded from further analysis. We test for selection bias associated with simple removal of both protesters and warm glow bidders in our data. Our findings show that removal of warm glow bidders does not significantly distort WTP whereas we find strong evidence of selection bias associated with removal of protesters. We show how to correct for such selection bias by using a sample selection model. In our empirical sample, using the typical approach of removing protesters from the analysis, the value of protecting the wetland is significantly underestimated by as much as 46% unless correcting for selection bias.

  3. Measuring microscopic evolution processes of complex networks based on empirical data

    NASA Astrophysics Data System (ADS)

    Chi, Liping

    2015-04-01

    Aiming at understanding the microscopic mechanism of complex systems in real world, we perform the measurement that characterizes the evolution properties on two empirical data sets. In the Autonomous Systems Internet data, the network size keeps growing although the system suffers a high rate of node deletion (r = 0.4) and link deletion (q = 0.81). However, the average degree keeps almost unchanged during the whole time range. At each time step the external links attached to a new node are about c = 1.1 and the internal links added between existing nodes are approximately m = 8. For the Scientific Collaboration data, it is a cumulated result of all the authors from 1893 up to the considered year. There is no deletion of nodes and links, r = q = 0. The external and internal links at each time step are c = 1.04 and m = 0, correspondingly. The exponents of degree distribution p(k) ∼ k-γ of these two empirical datasets γdata are in good agreement with that obtained theoretically γtheory. The results indicate that these evolution quantities may provide an insight into capturing the microscopic dynamical processes that govern the network topology.

  4. Empirical formula for rates of hot pixel defects based on pixel size, sensor area, and ISO

    NASA Astrophysics Data System (ADS)

    Chapman, Glenn H.; Thomas, Rohit; Koren, Zahava; Koren, Israel

    2013-02-01

    Experimentally, image sensors measurements show a continuous development of in-field permanent hot pixel defects increasing in numbers over time. In our tests we accumulated data on defects in cameras ranging from large area (<300 sq mm) DSLR's, medium sized (~40 sq mm) point and shoot, and small (20 sq mm) cell phone cameras. The results show that the rate of defects depends on the technology (APS or CCD), and on design parameters like imager area, pixel size (from 1.5 to 7 um), and gain (from ISO100 to 1600). Comparing different sensor sizes with similar pixel sizes has shown that defect rates scale linearly with sensor area, suggesting the metric of defects/year/sq mm, which we call defect density. A search was made to model this defect density as a function of the two parameters pixel size and ISO. The best empirical fit was obtained by a power law curve. For CCD imagers, the defect densities are proportional to the pixel size to the power of -2.25 times the ISO to the power of 0.69. For APS (CMOS) sensors the power law had the defect densities proportional to the pixel size to the power of -3.07 times the ISO raised to the power of 0.5. Extending our empirical formula to include ISO allows us to predict the expected defect development rate for a wide set of sensor parameters.

  5. Empirical evaluation of H.265/HEVC-based dynamic adaptive video streaming over HTTP (HEVC-DASH)

    NASA Astrophysics Data System (ADS)

    Irondi, Iheanyi; Wang, Qi; Grecos, Christos

    2014-05-01

    Real-time HTTP streaming has gained global popularity for delivering video content over Internet. In particular, the recent MPEG-DASH (Dynamic Adaptive Streaming over HTTP) standard enables on-demand, live, and adaptive Internet streaming in response to network bandwidth fluctuations. Meanwhile, emerging is the new-generation video coding standard, H.265/HEVC (High Efficiency Video Coding) promises to reduce the bandwidth requirement by 50% at the same video quality when compared with the current H.264/AVC standard. However, little existing work has addressed the integration of the DASH and HEVC standards, let alone empirical performance evaluation of such systems. This paper presents an experimental HEVC-DASH system, which is a pull-based adaptive streaming solution that delivers HEVC-coded video content through conventional HTTP servers where the client switches to its desired quality, resolution or bitrate based on the available network bandwidth. Previous studies in DASH have focused on H.264/AVC, whereas we present an empirical evaluation of the HEVC-DASH system by implementing a real-world test bed, which consists of an Apache HTTP Server with GPAC, an MP4Client (GPAC) with open HEVC-based DASH client and a NETEM box in the middle emulating different network conditions. We investigate and analyze the performance of HEVC-DASH by exploring the impact of various network conditions such as packet loss, bandwidth and delay on video quality. Furthermore, we compare the Intra and Random Access profiles of HEVC coding with the Intra profile of H.264/AVC when the correspondingly encoded video is streamed with DASH. Finally, we explore the correlation among the quality metrics and network conditions, and empirically establish under which conditions the different codecs can provide satisfactory performance.

  6. Comparisons of ground motions from the 1999 Chi-Chi, earthquake with empirical predictions largely based on data from California

    USGS Publications Warehouse

    Boore, D.M.

    2001-01-01

    This article has the modest goal of comparing the ground motions recorded during the 1999 Chi-Chi, Taiwan, mainshock with predictions from four empirical-based equations commonly used for western North America; these empirical predictions are largely based on data from California. Comparisons are made for peak acceleration and 5%-damped response spectra at periods between 0.1 and 4 sec. The general finding is that the Chi-Chi ground motions are smaller than those predicted from the empirically based equations for periods less than about 1 sec by factors averaging about 0.4 but as small as 0.26 (depending on period, on which equation is used, and on whether the sites are assumed to be rock or soil). There is a trend for the observed motions to approach or even exceed the predicted motions for longer periods. Motions at similar distances (30-60 km) to the east and to the west of the fault differ dramatically at periods between about 2 and 20 sec: Long-duration wave trains are present on the motions to the west, and when normalized to similar amplitudes at short periods, the response spectra of the motions at the western stations are as much as five times larger than those of motions from eastern stations. The explanation for the difference is probably related to site and propagation effects; the western stations are on the Coastal Plain, whereas the eastern stations are at the foot of young and steep mountains, either in the relatively narrow Longitudinal Valley or along the eastern coast-the sediments underlying the eastern stations are probably shallower and have higher velocity than those under the western stations.

  7. An Empirical Investigation of a Theoretically Based Measure of Perceived Wellness

    ERIC Educational Resources Information Center

    Harari, Marc J.; Waehler, Charles A.; Rogers, James R.

    2005-01-01

    The Perceived Wellness Survey (PWS; T. Adams, 1995; T. Adams, J. Bezner, & M. Steinhardt, 1997) is a recently developed instrument intended to operationalize the comprehensive Perceived Wellness Model (T. Adams, J. Bezner, & M. Steinhardt, 1997), an innovative model that attempts to include the balance of multiple life activities in its evaluation…

  8. University-Based Evaluation Training Programs in the United States 1980-2008: An Empirical Examination

    ERIC Educational Resources Information Center

    LaVelle, John M.; Donaldson, Stewart I.

    2010-01-01

    Evaluation practice has grown in leaps and bounds in recent years. In contrast, the most recent survey data suggest that there has been a sharp decline in the number and strength of preservice evaluation training programs in the United States. In an effort to further understand this curious trend, an alternative methodology was used to examine the…

  9. Fuzzy logic based clustering in wireless sensor networks: a survey

    NASA Astrophysics Data System (ADS)

    Singh, Ashutosh Kumar; Purohit, N.; Varma, S.

    2013-01-01

    Wireless sensor networks (WSNs) have limited resources, thus extending the lifetime has always been an issue of great interest. Recent developments in WSNs have led to various new fuzzy systems, specifically designed for WSNs where energy awareness is an essential consideration. In several applications, the clustered WSN are known to perform better than flat WSN, if the energy consumption in clustering operation itself could be minimised. Routing in clustered WSN is very efficient, especially when the challenge of finding the optimum number of intermediate cluster heads can be resolved. Fortunately, several fuzzy logic based solutions have been proposed for these jobs. Both single- and two-level fuzzy logic approaches are being used for cluster head election in which several distinguished features of WSN have been considered in making a decision. This article surveys the recent fuzzy applications for cluster head selection in WSNs and presents a comparative study for the various approaches pursued.

  10. A method for extracting human gait series from accelerometer signals based on the ensemble empirical mode decomposition

    NASA Astrophysics Data System (ADS)

    Fu, Mao-Jing; Zhuang, Jian-Jun; Hou, Feng-Zhen; Zhan, Qing-Bo; Shao, Yi; Ning, Xin-Bao

    2010-05-01

    In this paper, the ensemble empirical mode decomposition (EEMD) is applied to analyse accelerometer signals collected during normal human walking. First, the self-adaptive feature of EEMD is utilised to decompose the accelerometer signals, thus sifting out several intrinsic mode functions (IMFs) at disparate scales. Then, gait series can be extracted through peak detection from the eigen IMF that best represents gait rhythmicity. Compared with the method based on the empirical mode decomposition (EMD), the EEMD-based method has the following advantages: it remarkably improves the detection rate of peak values hidden in the original accelerometer signal, even when the signal is severely contaminated by the intermittent noises; this method effectively prevents the phenomenon of mode mixing found in the process of EMD. And a reasonable selection of parameters for the stop-filtering criteria can improve the calculation speed of the EEMD-based method. Meanwhile, the endpoint effect can be suppressed by using the auto regressive and moving average model to extend a short-time series in dual directions. The results suggest that EEMD is a powerful tool for extraction of gait rhythmicity and it also provides valuable clues for extracting eigen rhythm of other physiological signals.

  11. Inferring causal molecular networks: empirical assessment through a community-based effort.

    PubMed

    Hill, Steven M; Heiser, Laura M; Cokelaer, Thomas; Unger, Michael; Nesser, Nicole K; Carlin, Daniel E; Zhang, Yang; Sokolov, Artem; Paull, Evan O; Wong, Chris K; Graim, Kiley; Bivol, Adrian; Wang, Haizhou; Zhu, Fan; Afsari, Bahman; Danilova, Ludmila V; Favorov, Alexander V; Lee, Wai Shing; Taylor, Dane; Hu, Chenyue W; Long, Byron L; Noren, David P; Bisberg, Alexander J; Mills, Gordon B; Gray, Joe W; Kellen, Michael; Norman, Thea; Friend, Stephen; Qutub, Amina A; Fertig, Elana J; Guan, Yuanfang; Song, Mingzhou; Stuart, Joshua M; Spellman, Paul T; Koeppl, Heinz; Stolovitzky, Gustavo; Saez-Rodriguez, Julio; Mukherjee, Sach

    2016-04-01

    It remains unclear whether causal, rather than merely correlational, relationships in molecular networks can be inferred in complex biological settings. Here we describe the HPN-DREAM network inference challenge, which focused on learning causal influences in signaling networks. We used phosphoprotein data from cancer cell lines as well as in silico data from a nonlinear dynamical model. Using the phosphoprotein data, we scored more than 2,000 networks submitted by challenge participants. The networks spanned 32 biological contexts and were scored in terms of causal validity with respect to unseen interventional data. A number of approaches were effective, and incorporating known biology was generally advantageous. Additional sub-challenges considered time-course prediction and visualization. Our results suggest that learning causal relationships may be feasible in complex settings such as disease states. Furthermore, our scoring approach provides a practical way to empirically assess inferred molecular networks in a causal sense.

  12. Empirical mode decomposition-based facial pose estimation inside video sequences

    NASA Astrophysics Data System (ADS)

    Qing, Chunmei; Jiang, Jianmin; Yang, Zhijing

    2010-03-01

    We describe a new pose-estimation algorithm via integration of the strength in both empirical mode decomposition (EMD) and mutual information. While mutual information is exploited to measure the similarity between facial images to estimate poses, EMD is exploited to decompose input facial images into a number of intrinsic mode function (IMF) components, which redistribute the effect of noise, expression changes, and illumination variations as such that, when the input facial image is described by the selected IMF components, all the negative effects can be minimized. Extensive experiments were carried out in comparisons to existing representative techniques, and the results show that the proposed algorithm achieves better pose-estimation performances with robustness to noise corruption, illumination variation, and facial expressions.

  13. Inferring causal molecular networks: empirical assessment through a community-based effort.

    PubMed

    Hill, Steven M; Heiser, Laura M; Cokelaer, Thomas; Unger, Michael; Nesser, Nicole K; Carlin, Daniel E; Zhang, Yang; Sokolov, Artem; Paull, Evan O; Wong, Chris K; Graim, Kiley; Bivol, Adrian; Wang, Haizhou; Zhu, Fan; Afsari, Bahman; Danilova, Ludmila V; Favorov, Alexander V; Lee, Wai Shing; Taylor, Dane; Hu, Chenyue W; Long, Byron L; Noren, David P; Bisberg, Alexander J; Mills, Gordon B; Gray, Joe W; Kellen, Michael; Norman, Thea; Friend, Stephen; Qutub, Amina A; Fertig, Elana J; Guan, Yuanfang; Song, Mingzhou; Stuart, Joshua M; Spellman, Paul T; Koeppl, Heinz; Stolovitzky, Gustavo; Saez-Rodriguez, Julio; Mukherjee, Sach

    2016-04-01

    It remains unclear whether causal, rather than merely correlational, relationships in molecular networks can be inferred in complex biological settings. Here we describe the HPN-DREAM network inference challenge, which focused on learning causal influences in signaling networks. We used phosphoprotein data from cancer cell lines as well as in silico data from a nonlinear dynamical model. Using the phosphoprotein data, we scored more than 2,000 networks submitted by challenge participants. The networks spanned 32 biological contexts and were scored in terms of causal validity with respect to unseen interventional data. A number of approaches were effective, and incorporating known biology was generally advantageous. Additional sub-challenges considered time-course prediction and visualization. Our results suggest that learning causal relationships may be feasible in complex settings such as disease states. Furthermore, our scoring approach provides a practical way to empirically assess inferred molecular networks in a causal sense. PMID:26901648

  14. Gold price analysis based on ensemble empirical model decomposition and independent component analysis

    NASA Astrophysics Data System (ADS)

    Xian, Lu; He, Kaijian; Lai, Kin Keung

    2016-07-01

    In recent years, the increasing level of volatility of the gold price has received the increasing level of attention from the academia and industry alike. Due to the complexity and significant fluctuations observed in the gold market, however, most of current approaches have failed to produce robust and consistent modeling and forecasting results. Ensemble Empirical Model Decomposition (EEMD) and Independent Component Analysis (ICA) are novel data analysis methods that can deal with nonlinear and non-stationary time series. This study introduces a new methodology which combines the two methods and applies it to gold price analysis. This includes three steps: firstly, the original gold price series is decomposed into several Intrinsic Mode Functions (IMFs) by EEMD. Secondly, IMFs are further processed with unimportant ones re-grouped. Then a new set of data called Virtual Intrinsic Mode Functions (VIMFs) is reconstructed. Finally, ICA is used to decompose VIMFs into statistically Independent Components (ICs). The decomposition results reveal that the gold price series can be represented by the linear combination of ICs. Furthermore, the economic meanings of ICs are analyzed and discussed in detail, according to the change trend and ICs' transformation coefficients. The analyses not only explain the inner driving factors and their impacts but also conduct in-depth analysis on how these factors affect gold price. At the same time, regression analysis has been conducted to verify our analysis. Results from the empirical studies in the gold markets show that the EEMD-ICA serve as an effective technique for gold price analysis from a new perspective.

  15. A Survey of Early-Type Stars Based on the Two Micron All Sky Survey Database

    NASA Astrophysics Data System (ADS)

    Ortiz, R.; Malacarne, M.; Wilhelm, R.; Costa, R. D. D.; Rossi, S.; Maciel, W. J.; Costa, A. F. M.

    2007-09-01

    We report the results of a spectroscopic survey of blue stars near the south Galactic pole, selected according to their color indices in the 2MASS database. The main scope of this work is to determine the nature of the objects that comprise the sample. Some characteristics of the Hβ, Hγ, and Hδ lines which are sensitive to temperature and gravity were measured and eventually compared with a grid of model atmospheres to obtain Teff and log g. In some cases the data allowed a unique, reliable classification of the star. Among the 44 stars studied, 5 objects are reliably classified as subdwarfs, 12 as A-type stars, and 3 as field horizontal-branch (FHB) stars. The remaining objects are either A or FHB stars, but additional observations are necessary to distinguish between these two types. Neither degenerate nor low-gravity stars (post-AGB) have been identified in the sample. Based on observations obtained at the Pico dos Dias Observatory, operated by the National Laboratory for Astrophysics, Brazil.

  16. Racism, health status, and birth outcomes: results of a participatory community-based intervention and health survey.

    PubMed

    Carty, Denise C; Kruger, Daniel J; Turner, Tonya M; Campbell, Bettina; DeLoney, E Hill; Lewis, E Yvonne

    2011-02-01

    Many community-based participatory research (CBPR) partnerships address social determinants of health as a central consideration. However, research studies that explicitly address racism are scarce in the CBPR literature, and there is a dearth of available community-generated data to empirically examine how racism influences health disparities at the local level. In this paper, we provide results of a cross-sectional, population-based health survey conducted in the urban areas of Genesee and Saginaw Counties in Michigan to assess how a sustained community intervention to reduce racism and infant mortality influenced knowledge, beliefs, and experiences of racism and to explore how perceived racism is associated with self-rated health and birth outcomes. We used ANOVA and regression models to compare the responses of intervention participants and non-participants as well as African Americans and European Americans (N = 629). We found that intervention participants reported greater acknowledgment of the enduring and differential impact of racism in comparison to the non-intervention participants. Moreover, survey analyses revealed that racism was associated with health in the following ways: (1) experiences of racial discrimination predicted self-rated physical health, mental health, and smoking status; (2) perceived racism against one's racial group predicted lower self-rated physical health; and (3) emotional responses to racism-related experiences were marginally associated with lower birth-weight births in the study sample. Our study bolsters the published findings on perceived racism and health outcomes and highlights the usefulness of CBPR and community surveys to empirically investigate racism as a social determinant of health.

  17. Adolescent preventive health and team-games-tournaments: five decades of evidence for an empirically based paradigm.

    PubMed

    Wodarski, John S; Feit, Marvin D

    2011-01-01

    The problematic behaviors of teenagers and the subsequent negative consequences are extensive and well documented: unwanted pregnancy, substance abuse, violent behavior, depression, and social and psychological consequences of unemployment. In this article, the authors review an approach that uses a cooperative learning, empirically based intervention that employs peers as teachers. This intervention of choice is Teams-Games-Tournaments (TGT), a paradigm backed by five decades of empirical support. The application of TGT in preventive health programs incorporates elements in common with other prevention programs that are based on a public health orientation and constitute the essential components of health education, that is, skills training and practice in applying skills. The TGT intervention supports the idea that children and adolescents from various socioeconomic classes, between the ages of 8 and 18 and in classrooms or groups ranging in size from 4 to 17 members, can work together for one another. TGT has been applied successfully in such diverse areas as adolescent development, sexuality education, psychoactive substance abuse education, anger control, coping with depression and suicide, nutrition, comprehensive employment preparation, and family intervention. This article reviews the extensive research on TGT using examples of successful projects in substance abuse, violence, and nutrition. Issues are raised that relate to the implementation of preventive health strategies for adolescents, including cognitive aspects, social and family networks, and intervention components.

  18. Development of the Knowledge-based & Empirical Combined Scoring Algorithm (KECSA) to Score Protein-Ligand Interactions

    PubMed Central

    Zheng, Zheng

    2013-01-01

    We describe a novel knowledge-based protein-ligand scoring function that employs a new definition for the reference state, allowing us to relate a statistical potential to a Lennard-Jones (LJ) potential. In this way, the LJ potential parameters were generated from protein-ligand complex structural data contained in the PDB. Forty-nine types of atomic pairwise interactions were derived using this method, which we call the knowledge-based and empirical combined scoring algorithm (KECSA). Two validation benchmarks were introduced to test the performance of KECSA. The first validation benchmark included two test sets that address the training-set and enthalpy/entropy of KECSA The second validation benchmark suite included two large-scale and five small-scale test sets to compare the reproducibility of KECSA with respect to two empirical score functions previously developed in our laboratory (LISA and LISA+), as well as to other well-known scoring methods. Validation results illustrate that KECSA shows improved performance in all test sets when compared with other scoring methods especially in its ability to minimize the RMSE. LISA and LISA+ displayed similar performance using the correlation coefficient and Kendall τ as the metric of quality for some of the small test sets. Further pathways for improvement are discussed which would KECSA more sensitive to subtle changes in ligand structure. PMID:23560465

  19. A Compound fault diagnosis for rolling bearings method based on blind source separation and ensemble empirical mode decomposition.

    PubMed

    Wang, Huaqing; Li, Ruitong; Tang, Gang; Yuan, Hongfang; Zhao, Qingliang; Cao, Xi

    2014-01-01

    A Compound fault signal usually contains multiple characteristic signals and strong confusion noise, which makes it difficult to separate week fault signals from them through conventional ways, such as FFT-based envelope detection, wavelet transform or empirical mode decomposition individually. In order to improve the compound faults diagnose of rolling bearings via signals' separation, the present paper proposes a new method to identify compound faults from measured mixed-signals, which is based on ensemble empirical mode decomposition (EEMD) method and independent component analysis (ICA) technique. With the approach, a vibration signal is firstly decomposed into intrinsic mode functions (IMF) by EEMD method to obtain multichannel signals. Then, according to a cross correlation criterion, the corresponding IMF is selected as the input matrix of ICA. Finally, the compound faults can be separated effectively by executing ICA method, which makes the fault features more easily extracted and more clearly identified. Experimental results validate the effectiveness of the proposed method in compound fault separating, which works not only for the outer race defect, but also for the rollers defect and the unbalance fault of the experimental system. PMID:25289644

  20. Methods for the Design and Administration of Web-based Surveys

    PubMed Central

    Schleyer, Titus K. L.; Forrest, Jane L.

    2000-01-01

    This paper describes the design, development, and administration of a Web-based survey to determine the use of the Internet in clinical practice by 450 dental professionals. The survey blended principles of a controlled mail survey with data collection through a Web-based database application. The survey was implemented as a series of simple HTML pages and tested with a wide variety of operating environments. The response rate was 74.2 percent. Eighty-four percent of the participants completed the Web-based survey, and 16 percent used e-mail or fax. Problems identified during survey administration included incompatibilities/technical problems, usability problems, and a programming error. The cost of the Web-based survey was 38 percent less than that of an equivalent mail survey. A general formula for calculating breakeven points between electronic and hardcopy surveys is presented. Web-based surveys can significantly reduce turnaround time and cost compared with mail surveys and may enhance survey item completion rates. PMID:10887169

  1. Relevant modes selection method based on Spearman correlation coefficient for laser signal denoising using empirical mode decomposition

    NASA Astrophysics Data System (ADS)

    Duan, Yabo; Song, Chengtian

    2016-10-01

    Empirical mode decomposition (EMD) is a recently proposed nonlinear and nonstationary laser signal denoising method. A noisy signal is broken down using EMD into oscillatory components that are called intrinsic mode functions (IMFs). Thresholding-based denoising and correlation-based partial reconstruction of IMFs are the two main research directions for EMD-based denoising. Similar to other decomposition-based denoising approaches, EMD-based denoising methods require a reliable threshold to determine which IMFs are noise components and which IMFs are noise-free components. In this work, we propose a new approach in which each IMF is first denoised using EMD interval thresholding (EMD-IT), and then a robust thresholding process based on Spearman correlation coefficient is used for relevant modes selection. The proposed method tackles the problem using a thresholding-based denoising approach coupled with partial reconstruction of the relevant IMFs. Other traditional denoising methods, including correlation-based EMD partial reconstruction (EMD-Correlation), discrete Fourier transform and wavelet-based methods, are investigated to provide a comparison with the proposed technique. Simulation and test results demonstrate the superior performance of the proposed method when compared with the other methods.

  2. [Surveying a zoological facility through satellite-based geodesy].

    PubMed

    Böer, M; Thien, W; Tölke, D

    2000-06-01

    In the course of a thesis submitted for a diploma degree within the Fachhochschule Oldenburg the Serengeti Safaripark was surveyed in autumn and winter 1996/97 laying in the planning foundations for the application for licences from the controlling authorities. Taking into consideration the special way of keeping animals in the Serengeti Safaripark (game ranching, spacious walk-through-facilities) the intention was to employ the outstanding satellite based geodesy. This technology relies on special aerials receiving signals from 24 satellites which circle around the globe. These data are being gathered and examined. This examination produces the exact position of this aerial in a system of coordinates which allows depicting this point on a map. This procedure was used stationary (from a strictly defined point) as well as in the movement (in a moving car). Additionally conventional procedures were used when the satellite based geodesy came to its limits. Finally a detailed map of the Serengeti Safaripark was created which shows the position and size of stables and enclosures as well as wood and water areas and the sectors of the leisure park. Furthermore the established areas of the enclosures together with an already existing animal databank have flown into an information system with the help of which the stock of animals can be managed enclosure-orientated.

  3. ECG-based heartbeat classification for arrhythmia detection: A survey.

    PubMed

    Luz, Eduardo José da S; Schwartz, William Robson; Cámara-Chávez, Guillermo; Menotti, David

    2016-04-01

    An electrocardiogram (ECG) measures the electric activity of the heart and has been widely used for detecting heart diseases due to its simplicity and non-invasive nature. By analyzing the electrical signal of each heartbeat, i.e., the combination of action impulse waveforms produced by different specialized cardiac tissues found in the heart, it is possible to detect some of its abnormalities. In the last decades, several works were developed to produce automatic ECG-based heartbeat classification methods. In this work, we survey the current state-of-the-art methods of ECG-based automated abnormalities heartbeat classification by presenting the ECG signal preprocessing, the heartbeat segmentation techniques, the feature description methods and the learning algorithms used. In addition, we describe some of the databases used for evaluation of methods indicated by a well-known standard developed by the Association for the Advancement of Medical Instrumentation (AAMI) and described in ANSI/AAMI EC57:1998/(R)2008 (ANSI/AAMI, 2008). Finally, we discuss limitations and drawbacks of the methods in the literature presenting concluding remarks and future challenges, and also we propose an evaluation process workflow to guide authors in future works.

  4. Adolescents with attention-deficit/hyperactivity disorder: an overview of empirically based treatments.

    PubMed

    Barkley, Russell A

    2004-01-01

    The author first presents an overview of attention-deficit/hyperactivity disorder (ADHD) as it presents in adolescents. He reviews what is known about the predominantly inattentive subtype in adolescents, the persistence of symptoms into this developmental phase, and comorbid disorders in adolescent patients with ADHD. The author then reviews treatments for adolescents with ADHD for which there is some empirical support in the scientific literature. He first discusses common assumptions concerning the treatment of ADHD and evidence for or against these assumptions. Information on therapies that have been shown to be ineffective or the benefit of which is unproven are then described. These include cognitive-behavioral therapy and social skills training. The author then presents an overview of what is known about the medication treatment of ADHD and discusses how this information is applicable to adolescents with the disorder. Four main classes of drugs are discussed: stimulants, noradrenergic reuptake inhibitors, tricyclic antidepressants, and antihypertensive agents. The author then reviews the use of several psychosocial interventions, including contingency management strategies, parent training in behavior management methods, and teacher training in classroom management, and discusses how these strategies can best be used for adolescents with ADHD. The author then discusses the use of combined treatment with psychosocial interventions and medication. Finally, information on the use of physical exercise as therapy for adolescents with ADHD is discussed.

  5. Cardiopulmonary Resuscitation Pattern Evaluation Based on Ensemble Empirical Mode Decomposition Filter via Nonlinear Approaches

    PubMed Central

    Ma, Matthew Huei-Ming

    2016-01-01

    Good quality cardiopulmonary resuscitation (CPR) is the mainstay of treatment for managing patients with out-of-hospital cardiac arrest (OHCA). Assessment of the quality of the CPR delivered is now possible through the electrocardiography (ECG) signal that can be collected by an automated external defibrillator (AED). This study evaluates a nonlinear approximation of the CPR given to the asystole patients. The raw ECG signal is filtered using ensemble empirical mode decomposition (EEMD), and the CPR-related intrinsic mode functions (IMF) are chosen to be evaluated. In addition, sample entropy (SE), complexity index (CI), and detrended fluctuation algorithm (DFA) are collated and statistical analysis is performed using ANOVA. The primary outcome measure assessed is the patient survival rate after two hours. CPR pattern of 951 asystole patients was analyzed for quality of CPR delivered. There was no significant difference observed in the CPR-related IMFs peak-to-peak interval analysis for patients who are younger or older than 60 years of age, similarly to the amplitude difference evaluation for SE and DFA. However, there is a difference noted for the CI (p < 0.05). The results show that patients group younger than 60 years have higher survival rate with high complexity of the CPR-IMFs amplitude differences. PMID:27529068

  6. Development of Items for a Pedagogical Content Knowledge Test Based on Empirical Analysis of Pupils' Errors

    NASA Astrophysics Data System (ADS)

    Jüttner, Melanie; Neuhaus, Birgit J.

    2012-05-01

    In view of the lack of instruments for measuring biology teachers' pedagogical content knowledge (PCK), this article reports on a study about the development of PCK items for measuring teachers' knowledge of pupils' errors and ways for dealing with them. This study investigated 9th and 10th grade German pupils' (n = 461) drawings in an achievement test about the knee-jerk in biology, which were analysed by using the inductive qualitative analysis of their content. The empirical data were used for the development of the items in the PCK test. The validation of the items was determined with think-aloud interviews of German secondary school teachers (n = 5). If the item was determined, the reliability was tested by the results of German secondary school biology teachers (n = 65) who took the PCK test. The results indicated that these items are satisfactorily reliable (Cronbach's alpha values ranged from 0.60 to 0.65). We suggest a larger sample size and American biology teachers be used in our further studies. The findings of this study about teachers' professional knowledge from the PCK test could provide new information about the influence of teachers' knowledge on their pupils' understanding of biology and their possible errors in learning biology.

  7. Spectral analysis of Hall-effect thruster plasma oscillations based on the empirical mode decomposition

    NASA Astrophysics Data System (ADS)

    Kurzyna, J.; Mazouffre, S.; Lazurenko, A.; Albarède, L.; Bonhomme, G.; Makowski, K.; Dudeck, M.; Peradzyński, Z.

    2005-12-01

    Hall-effect thruster plasma oscillations recorded by means of probes located at the channel exit are analyzed using the empirical mode decomposition (EMD) method. This self-adaptive technique permits to decompose a nonstationary signal into a set of intrinsic modes, and acts as a very efficient filter allowing to separate contributions of different underlying physical mechanisms. Applying the Hilbert transform to the whole set of modes allows to identify peculiar events and to assign them a range of instantaneous frequency and power. In addition to 25kHz breathing-type oscillations which are unambiguously identified, the EMD approach confirms the existence of oscillations with instantaneous frequencies in the range of 100-500kHz typical for ion transit-time oscillations. Modeling of high-frequency modes (ν˜10MHz) resulting from EMD of measured wave forms supports the idea that high-frequency plasma oscillations originate from electron-density perturbations propagating azimuthally with the electron drift velocity.

  8. Cardiopulmonary Resuscitation Pattern Evaluation Based on Ensemble Empirical Mode Decomposition Filter via Nonlinear Approaches.

    PubMed

    Sadrawi, Muammar; Sun, Wei-Zen; Ma, Matthew Huei-Ming; Dai, Chun-Yi; Abbod, Maysam F; Shieh, Jiann-Shing

    2016-01-01

    Good quality cardiopulmonary resuscitation (CPR) is the mainstay of treatment for managing patients with out-of-hospital cardiac arrest (OHCA). Assessment of the quality of the CPR delivered is now possible through the electrocardiography (ECG) signal that can be collected by an automated external defibrillator (AED). This study evaluates a nonlinear approximation of the CPR given to the asystole patients. The raw ECG signal is filtered using ensemble empirical mode decomposition (EEMD), and the CPR-related intrinsic mode functions (IMF) are chosen to be evaluated. In addition, sample entropy (SE), complexity index (CI), and detrended fluctuation algorithm (DFA) are collated and statistical analysis is performed using ANOVA. The primary outcome measure assessed is the patient survival rate after two hours. CPR pattern of 951 asystole patients was analyzed for quality of CPR delivered. There was no significant difference observed in the CPR-related IMFs peak-to-peak interval analysis for patients who are younger or older than 60 years of age, similarly to the amplitude difference evaluation for SE and DFA. However, there is a difference noted for the CI (p < 0.05). The results show that patients group younger than 60 years have higher survival rate with high complexity of the CPR-IMFs amplitude differences. PMID:27529068

  9. Spectral analysis of Hall-effect thruster plasma oscillations based on the empirical mode decomposition

    SciTech Connect

    Kurzyna, J.; Mazouffre, S.; Lazurenko, A.; Albarede, L.; Bonhomme, G.; Makowski, K.; Dudeck, M.; Peradzynski, Z.

    2005-12-15

    Hall-effect thruster plasma oscillations recorded by means of probes located at the channel exit are analyzed using the empirical mode decomposition (EMD) method. This self-adaptive technique permits to decompose a nonstationary signal into a set of intrinsic modes, and acts as a very efficient filter allowing to separate contributions of different underlying physical mechanisms. Applying the Hilbert transform to the whole set of modes allows to identify peculiar events and to assign them a range of instantaneous frequency and power. In addition to 25 kHz breathing-type oscillations which are unambiguously identified, the EMD approach confirms the existence of oscillations with instantaneous frequencies in the range of 100-500 kHz typical for ion transit-time oscillations. Modeling of high-frequency modes ({nu}{approx}10 MHz) resulting from EMD of measured wave forms supports the idea that high-frequency plasma oscillations originate from electron-density perturbations propagating azimuthally with the electron drift velocity.

  10. Towards high performing hospital enterprise systems: an empirical and literature based design framework

    NASA Astrophysics Data System (ADS)

    dos Santos Fradinho, Jorge Miguel

    2014-05-01

    Our understanding of enterprise systems (ES) is gradually evolving towards a sense of design which leverages multidisciplinary bodies of knowledge that may bolster hybrid research designs and together further the characterisation of ES operation and performance. This article aims to contribute towards ES design theory with its hospital enterprise systems design (HESD) framework, which reflects a rich multidisciplinary literature and two in-depth hospital empirical cases from the US and UK. In doing so it leverages systems thinking principles and traditionally disparate bodies of knowledge to bolster the theoretical evolution and foundation of ES. A total of seven core ES design elements are identified and characterised with 24 main categories and 53 subcategories. In addition, it builds on recent work which suggests that hospital enterprises are comprised of multiple internal ES configurations which may generate different levels of performance. Multiple sources of evidence were collected including electronic medical records, 54 recorded interviews, observation, and internal documents. Both in-depth cases compare and contrast higher and lower performing ES configurations. Following literal replication across in-depth cases, this article concludes that hospital performance can be improved through an enriched understanding of hospital ES design.

  11. Inferring causal molecular networks: empirical assessment through a community-based effort

    PubMed Central

    Hill, Steven M.; Heiser, Laura M.; Cokelaer, Thomas; Unger, Michael; Nesser, Nicole K.; Carlin, Daniel E.; Zhang, Yang; Sokolov, Artem; Paull, Evan O.; Wong, Chris K.; Graim, Kiley; Bivol, Adrian; Wang, Haizhou; Zhu, Fan; Afsari, Bahman; Danilova, Ludmila V.; Favorov, Alexander V.; Lee, Wai Shing; Taylor, Dane; Hu, Chenyue W.; Long, Byron L.; Noren, David P.; Bisberg, Alexander J.; Mills, Gordon B.; Gray, Joe W.; Kellen, Michael; Norman, Thea; Friend, Stephen; Qutub, Amina A.; Fertig, Elana J.; Guan, Yuanfang; Song, Mingzhou; Stuart, Joshua M.; Spellman, Paul T.; Koeppl, Heinz; Stolovitzky, Gustavo; Saez-Rodriguez, Julio; Mukherjee, Sach

    2016-01-01

    Inferring molecular networks is a central challenge in computational biology. However, it has remained unclear whether causal, rather than merely correlational, relationships can be effectively inferred in complex biological settings. Here we describe the HPN-DREAM network inference challenge that focused on learning causal influences in signaling networks. We used phosphoprotein data from cancer cell lines as well as in silico data from a nonlinear dynamical model. Using the phosphoprotein data, we scored more than 2,000 networks submitted by challenge participants. The networks spanned 32 biological contexts and were scored in terms of causal validity with respect to unseen interventional data. A number of approaches were effective and incorporating known biology was generally advantageous. Additional sub-challenges considered time-course prediction and visualization. Our results constitute the most comprehensive assessment of causal network inference in a mammalian setting carried out to date and suggest that learning causal relationships may be feasible in complex settings such as disease states. Furthermore, our scoring approach provides a practical way to empirically assess the causal validity of inferred molecular networks. PMID:26901648

  12. Empirical application of empathy enhancing program based on movement concept for married couples in conflict

    PubMed Central

    Kim, Soo-Yeon; Kang, Hye-Won; Chung, Yong-Chul; Park, Seungha

    2013-01-01

    In the field of marital therapy, it is known that couple movement program helps married couples faced with conflict situation to rebuild the relationship and to maintain a family homeostasis. The purpose of this study was to configure and apply the kinesthetic empathy program and to assess the effectiveness for married couples in conflict. To achieve the research aims, qualitative research method has been conducted, subjecting three couples, 6 people, who are participating in expressive movement program for this study. The study used focus group interview method for collecting date and employed for the interview method by mixing the semi-structured and unstructured questionnaire. The results were followings. First, through kinesthetic empathy enhancing program, one could develop self-awareness and emotional attunement. Second, the result showed the relationship between intention and empathy. It shows that “knowing spouse’s hidden intention” is significant factors to understand others. Third, kinesthetic empathy program could complement general marriage counseling program. The results of this study provide empirical evidence that movement program functions as an empathy enhancer through the process of perceiving, feeling, thinking, and interacting with others. PMID:24278896

  13. Effects of Personalization and Invitation Email Length on Web-Based Survey Response Rates

    ERIC Educational Resources Information Center

    Trespalacios, Jesús H.; Perkins, Ross A.

    2016-01-01

    Individual strategies to increase response rate and survey completion have been extensively researched. Recently, efforts have been made to investigate a combination of interventions to yield better response rates for web-based surveys. This study examined the effects of four different survey invitation conditions on response rate. From a large…

  14. Empirical evidence for identical band gaps in substituted C{sub 60} and C{sub 70} based fullerenes

    SciTech Connect

    Mattias Andersson, L. Tanaka, Hideyuki

    2014-01-27

    Optical absorptance data, and a strong correlation between solar cell open circuit voltages and the ionization potentials of a wide range of differently substituted fullerene acceptors, are presented as empirical evidence for identical, or at least very similar, band gaps in all substituted C{sub 60} and C{sub 70} based fullerenes. Both the number and kind of substituents in this study are sufficiently varied to imply generality. While the band gaps of the fullerenes remain the same for all the different substitutions, their ionization potentials vary greatly in a span of more than 0.4 eV. The merits and drawbacks of using these results together with photoelectron based techniques to determine relative fullerene energy levels for, e.g., organic solar cell applications compared to more direct electrochemical methods are also discussed.

  15. [Removal Algorithm of Power Line Interference in Electrocardiogram Based on Morphological Component Analysis and Ensemble Empirical Mode Decomposition].

    PubMed

    Zhao, Wei; Xiao, Shixiao; Zhang, Baocan; Huang, Xiaojing; You, Rongyi

    2015-12-01

    Electrocardiogram (ECG) signals are susceptible to be disturbed by 50 Hz power line interference (PLI) in the process of acquisition and conversion. This paper, therefore, proposes a novel PLI removal algorithm based on morphological component analysis (MCA) and ensemble empirical mode decomposition (EEMD). Firstly, according to the morphological differences in ECG waveform characteristics, the noisy ECG signal was decomposed into the mutated component, the smooth component and the residual component by MCA. Secondly, intrinsic mode functions (IMF) of PLI was filtered. The noise suppression rate (NSR) and the signal distortion ratio (SDR) were used to evaluate the effect of de-noising algorithm. Finally, the ECG signals were re-constructed. Based on the experimental comparison, it was concluded that the proposed algorithm had better filtering functions than the improved Levkov algorithm, because it could not only effectively filter the PLI, but also have smaller SDR value. PMID:27079083

  16. An Empirical Study of Neural Network-Based Audience Response Technology in a Human Anatomy Course for Pharmacy Students.

    PubMed

    Fernández-Alemán, José Luis; López-González, Laura; González-Sequeros, Ofelia; Jayne, Chrisina; López-Jiménez, Juan José; Carrillo-de-Gea, Juan Manuel; Toval, Ambrosio

    2016-04-01

    This paper presents an empirical study of a formative neural network-based assessment approach by using mobile technology to provide pharmacy students with intelligent diagnostic feedback. An unsupervised learning algorithm was integrated with an audience response system called SIDRA in order to generate states that collect some commonality in responses to questions and add diagnostic feedback for guided learning. A total of 89 pharmacy students enrolled on a Human Anatomy course were taught using two different teaching methods. Forty-four students employed intelligent SIDRA (i-SIDRA), whereas 45 students received the same training but without using i-SIDRA. A statistically significant difference was found between the experimental group (i-SIDRA) and the control group (traditional learning methodology), with T (87) = 6.598, p < 0.001. In four MCQs tests, the difference between the number of correct answers in the first attempt and in the last attempt was also studied. A global effect size of 0.644 was achieved in the meta-analysis carried out. The students expressed satisfaction with the content provided by i-SIDRA and the methodology used during the process of learning anatomy (M = 4.59). The new empirical contribution presented in this paper allows instructors to perform post hoc analyses of each particular student's progress to ensure appropriate training. PMID:26815339

  17. An Empirical Study of Neural Network-Based Audience Response Technology in a Human Anatomy Course for Pharmacy Students.

    PubMed

    Fernández-Alemán, José Luis; López-González, Laura; González-Sequeros, Ofelia; Jayne, Chrisina; López-Jiménez, Juan José; Carrillo-de-Gea, Juan Manuel; Toval, Ambrosio

    2016-04-01

    This paper presents an empirical study of a formative neural network-based assessment approach by using mobile technology to provide pharmacy students with intelligent diagnostic feedback. An unsupervised learning algorithm was integrated with an audience response system called SIDRA in order to generate states that collect some commonality in responses to questions and add diagnostic feedback for guided learning. A total of 89 pharmacy students enrolled on a Human Anatomy course were taught using two different teaching methods. Forty-four students employed intelligent SIDRA (i-SIDRA), whereas 45 students received the same training but without using i-SIDRA. A statistically significant difference was found between the experimental group (i-SIDRA) and the control group (traditional learning methodology), with T (87) = 6.598, p < 0.001. In four MCQs tests, the difference between the number of correct answers in the first attempt and in the last attempt was also studied. A global effect size of 0.644 was achieved in the meta-analysis carried out. The students expressed satisfaction with the content provided by i-SIDRA and the methodology used during the process of learning anatomy (M = 4.59). The new empirical contribution presented in this paper allows instructors to perform post hoc analyses of each particular student's progress to ensure appropriate training.

  18. Combining Empirical Relationships with Data Based Mechanistic Modeling to Inform Solute Tracer Investigations across Stream Orders

    NASA Astrophysics Data System (ADS)

    Herrington, C.; Gonzalez-Pinzon, R.; Covino, T. P.; Mortensen, J.

    2015-12-01

    Solute transport studies in streams and rivers often begin with the introduction of conservative and reactive tracers into the water column. Information on the transport of these substances is then captured within tracer breakthrough curves (BTCs) and used to estimate, for instance, travel times and dissolved nutrient and carbon dynamics. Traditionally, these investigations have been limited to systems with small discharges (< 200 L/s) and with small reach lengths (< 500 m), partly due to the need for a priori information of the reach's hydraulic characteristics (e.g., channel geometry, resistance and dispersion coefficients) to predict arrival times, times to peak concentrations of the solute and mean travel times. Current techniques to acquire these channel characteristics through preliminary tracer injections become cost prohibitive at higher stream orders and the use of semi-continuous water quality sensors for collecting real-time information may be affected from erroneous readings that are masked by high turbidity (e.g., nitrate signals with SUNA instruments or fluorescence measures) and/or high total dissolved solids (e.g., making prohibitively expensive the use of salt tracers such as NaCl) in larger systems. Additionally, a successful time-of-travel study is valuable for only a single discharge and river stage. We have developed a method to predict tracer BTCs to inform sampling frequencies at small and large stream orders using empirical relationships developed from multiple tracer injections spanning several orders of magnitude in discharge and reach length. This method was successfully tested in 1st to 8th order systems along the Middle Rio Grande River Basin in New Mexico, USA.

  19. Ensemble Empirical Mode Decomposition based methodology for ultrasonic testing of coarse grain austenitic stainless steels.

    PubMed

    Sharma, Govind K; Kumar, Anish; Jayakumar, T; Purnachandra Rao, B; Mariyappa, N

    2015-03-01

    A signal processing methodology is proposed in this paper for effective reconstruction of ultrasonic signals in coarse grained high scattering austenitic stainless steel. The proposed methodology is comprised of the Ensemble Empirical Mode Decomposition (EEMD) processing of ultrasonic signals and application of signal minimisation algorithm on selected Intrinsic Mode Functions (IMFs) obtained by EEMD. The methodology is applied to ultrasonic signals obtained from austenitic stainless steel specimens of different grain size, with and without defects. The influence of probe frequency and data length of a signal on EEMD decomposition is also investigated. For a particular sampling rate and probe frequency, the same range of IMFs can be used to reconstruct the ultrasonic signal, irrespective of the grain size in the range of 30-210 μm investigated in this study. This methodology is successfully employed for detection of defects in a 50mm thick coarse grain austenitic stainless steel specimens. Signal to noise ratio improvement of better than 15 dB is observed for the ultrasonic signal obtained from a 25 mm deep flat bottom hole in 200 μm grain size specimen. For ultrasonic signals obtained from defects at different depths, a minimum of 7 dB extra enhancement in SNR is achieved as compared to the sum of selected IMF approach. The application of minimisation algorithm with EEMD processed signal in the proposed methodology proves to be effective for adaptive signal reconstruction with improved signal to noise ratio. This methodology was further employed for successful imaging of defects in a B-scan.

  20. Children's Experiences of Completing a Computer-Based Violence Survey: Finnish Child Victim Survey Revisited.

    PubMed

    Fagerlund, Monica; Ellonen, Noora

    2016-07-01

    The involvement of children as research subjects requires special considerations with regard to research practices and ethics. This is especially true concerning sensitive research topics such as sexual victimization. Prior research suggests that reflecting these experiences in a survey can cause negative feelings in child participants, although posing only a minimal to moderate risk. Analyzing only predefined, often negative feelings related to answering a sexual victimization survey has dominated the existing literature. In this article children's free-text comments about answering a victimization survey and experiences of sexual victimization are analyzed together to evaluate the effects of research participation in relation to this sensitive issue. Altogether 11,364 children, aged 11-12 and 15-16, participated in the Finnish Child Victim Survey in 2013. Of these, 69% (7,852) reflected on their feelings about answering the survey. Results indicate that both clearly negative and positive feelings are more prevalent among victimized children compared to their nonvictimized peers. Characteristics unique to sexual victimization as well as differences related to gender and age are also discussed. The study contributes to the important yet contradictory field of studying the effects of research participation on children. PMID:27472509

  1. Holding-based network of nations based on listed energy companies: An empirical study on two-mode affiliation network of two sets of actors

    NASA Astrophysics Data System (ADS)

    Li, Huajiao; Fang, Wei; An, Haizhong; Gao, Xiangyun; Yan, Lili

    2016-05-01

    Economic networks in the real world are not homogeneous; therefore, it is important to study economic networks with heterogeneous nodes and edges to simulate a real network more precisely. In this paper, we present an empirical study of the one-mode derivative holding-based network constructed by the two-mode affiliation network of two sets of actors using the data of worldwide listed energy companies and their shareholders. First, we identify the primitive relationship in the two-mode affiliation network of the two sets of actors. Then, we present the method used to construct the derivative network based on the shareholding relationship between two sets of actors and the affiliation relationship between actors and events. After constructing the derivative network, we analyze different topological features on the node level, edge level and entire network level and explain the meanings of the different values of the topological features combining the empirical data. This study is helpful for expanding the usage of complex networks to heterogeneous economic networks. For empirical research on the worldwide listed energy stock market, this study is useful for discovering the inner relationships between the nations and regions from a new perspective.

  2. The Empirical Interests of Developmental Psychology.

    ERIC Educational Resources Information Center

    McGraw, Kenneth O.

    1991-01-01

    Presents a classification system for developmental psychology that was constructed by means of a survey of articles published in the journals "Child Development" and "Developmental Psychology" in 1969 and 1987. The survey found that eight empirical questions encompassed the full range of empirical interests expressed by developmental…

  3. Perceptions and Experiences of Research Participants on Gender-Based Violence Community Based Survey: Implications for Ethical Guidelines

    PubMed Central

    Sikweyiya, Yandisa; Jewkes, Rachel

    2012-01-01

    Objective To explore how survey respondents perceived their experiences and the impact of participating in a survey, and to assess adverse consequences resulting from participation. Design Qualitative study involving purposefully selected participants who had participated in a household-based survey. Methods This qualitative study was nested within a survey that investigated the prevalence of gender-based violence perpetration and victimization with adult men and women in South Africa. 13 male- and 10 female-in-depth interviews were conducted with survey respondents. Results A majority of informants, without gender-differences, perceived the survey interview as a rare opportunity to share their adverse and or personal experiences in a 'safe' space. Gender-differences were noted in reporting perceptions of risks involved with survey participation. Some women remained fearful after completing the survey, that should breach of confidentiality or full survey content disclosure occur, they may be victimized by partners as a punishment for survey participation without men's approval. A number of informants generally discussed their survey participation with others. However, among women with interpersonal violence history or currently in abusive relationships, full survey content disclosure was done with fear; the partner responses were negative, and few women reported receiving threatening remarks but none reported being assaulted. In contrast no man reported adverse reaction by others. Informants with major life adversities reported that the survey had made them to relive the experiences causing them sadness and pain at the time. No informant perceived the survey as emotionally harmful or needed professional support because of survey questions. Rather the vast majority perceived benefit from survey participation. Conclusion Whilst no informant felt answering the survey questions had caused them emotional or physical harm, some were distressed and anxious, albeit

  4. Multiscale Detrended Cross-Correlation Analysis of Traffic Time Series Based on Empirical Mode Decomposition

    NASA Astrophysics Data System (ADS)

    Yin, Yi; Shang, Pengjian

    2015-04-01

    In this paper, we propose multiscale detrended cross-correlation analysis (MSDCCA) to detect the long-range power-law cross-correlation of considered signals in the presence of nonstationarity. For improving the performance and getting better robustness, we further introduce the empirical mode decomposition (EMD) to eliminate the noise effects and propose MSDCCA method combined with EMD, which is called MS-EDXA method, then systematically investigate the multiscale cross-correlation structure of the real traffic signals. We apply the MSDCCA and MS-EDXA methods to study the cross-correlations in three situations: velocity and volume on one lane, velocities on the present and the next moment and velocities on the adjacent lanes, and further compare their spectrums respectively. When the difference between the spectrums of MSDCCA and MS-EDXA becomes unobvious, there is a crossover which denotes the turning point of difference. The crossover results from the competition between the noise effects in the original signals and the intrinsic fluctuation of traffic signals and divides the plot of spectrums into two regions. In all the three case, MS-EDXA method makes the average of local scaling exponents increased and the standard deviation decreased and provides a relative stable persistent scaling cross-correlated behavior which gets the analysis more precise and more robust and improves the performance after noises being removed. Applying MS-EDXA method avoids the inaccurate characteristics of multiscale cross-correlation structure at the short scale including the spectrum minimum, the range for the spectrum fluctuation and general trend, which are caused by the noise in the original signals. We get the conclusions that the traffic velocity and volume are long-range cross-correlated, which is accordant to their actual evolution, while velocities on the present and the next moment and velocities on adjacent lanes reflect the strong cross-correlations both in temporal and

  5. An Empirical Study of Instructor Adoption of Web-Based Learning Systems

    ERIC Educational Resources Information Center

    Wang, Wei-Tsong; Wang, Chun-Chieh

    2009-01-01

    For years, web-based learning systems have been widely employed in both educational and non-educational institutions. Although web-based learning systems are emerging as a useful tool for facilitating teaching and learning activities, the number of users is not increasing as fast as expected. This study develops an integrated model of instructor…

  6. Meta-Analysis of Group Learning Activities: Empirically Based Teaching Recommendations

    ERIC Educational Resources Information Center

    Tomcho, Thomas J.; Foels, Rob

    2012-01-01

    Teaching researchers commonly employ group-based collaborative learning approaches in Teaching of Psychology teaching activities. However, the authors know relatively little about the effectiveness of group-based activities in relation to known psychological processes associated with group dynamics. Therefore, the authors conducted a meta-analytic…

  7. Outcome (Competency) Based Education: An Exploration of Its Origins, Theoretical Basis, and Empirical Evidence

    ERIC Educational Resources Information Center

    Morcke, Anne Mette; Dornan, Tim; Eika, Berit

    2013-01-01

    Outcome based or competency based education (OBE) is so firmly established in undergraduate medical education that it might not seem necessary to ask why it was included in recommendations for the future, like the Flexner centenary report. Uncritical acceptance may not, however, deliver its greatest benefits. Our aim was to explore the…

  8. Constructions, Semantic Compatibility, and Coercion: An Empirical Usage-Based Approach

    ERIC Educational Resources Information Center

    Yoon, Soyeon

    2012-01-01

    This study investigates the nature of semantic compatibility between constructions and lexical items that occur in them in relation with language use, and the related concept, coercion, based on a usage-based approach to language, in which linguistic knowledge (grammar) is grounded in language use. This study shows that semantic compatibility…

  9. Empirically Supported Family-Based Treatments for Conduct Disorder and Delinquency in Adolescents

    ERIC Educational Resources Information Center

    Henggeler, Scott W.; Sheidow, Ashli J.

    2012-01-01

    Several family-based treatments of conduct disorder and delinquency in adolescents have emerged as evidence-based and, in recent years, have been transported to more than 800 community practice settings. These models include multisystemic therapy, functional family therapy, multidimensional treatment foster care, and, to a lesser extent, brief…

  10. Formula-Based Public School Funding System in Victoria: An Empirical Analysis of Equity

    ERIC Educational Resources Information Center

    Bandaranayake, Bandara

    2013-01-01

    This article explores the formula-based school funding system in the state of Victoria, Australia, where state funds are directly allocated to schools based on a range of equity measures. The impact of Victoria' funding system for education in terms of alleviating inequality and disadvantage is contentious, to say the least. It is difficult…

  11. Lake Superior Zooplankton Biomass Predictions from LOPC Tow Surveys Compare Well with a Probability Based Net Survey

    EPA Science Inventory

    We conducted a probability-based sampling of Lake Superior in 2006 and compared the zooplankton biomass estimate with laser optical plankton counter (LOPC) predictions. The net survey consisted of 52 sites stratified across three depth zones (0-30, 30-150, >150 m). The LOPC tow...

  12. Cultivating mindfulness in health care professionals: a review of empirical studies of mindfulness-based stress reduction (MBSR).

    PubMed

    Irving, Julie Anne; Dobkin, Patricia L; Park, Jeeseon

    2009-05-01

    Demands faced by health care professionals include heavy caseloads, limited control over the work environment, long hours, as well as organizational structures and systems in transition. Such conditions have been directly linked to increased stress and symptoms of burnout, which in turn, have adverse consequences for clinicians and the quality of care that is provided to patients. Consequently, there exists an impetus for the development of curriculum aimed at fostering wellness and the necessary self-care skills for clinicians. This review will examine the potential benefits of mindfulness-based stress reduction (MBSR) programs aimed at enhancing well-being and coping with stress in this population. Empirical evidence indicates that participation in MBSR yields benefits for clinicians in the domains of physical and mental health. Conceptual and methodological limitations of the existing studies and suggestions for future research are discussed.

  13. Re-reading nursing and re-writing practice: towards an empirically based reformulation of the nursing mandate.

    PubMed

    Allen, Davina

    2004-12-01

    This article examines field studies of nursing work published in the English language between 1993 and 2003 as the first step towards an empirically based reformulation of the nursing mandate. A decade of ethnographic research reveals that, contrary to contemporary theories which promote an image of nursing work centred on individualised unmediated caring relationships, in real-life practice the core nursing contribution is that of the healthcare mediator. Eight bundles of activity that comprise this intermediary role are described utilising evidence from the literature. The mismatch between nursing's culture and ideals and the structure and constraints of the work setting is a chronic source of practitioner dissatisfaction. It is argued that the profession has little to gain by pursuing an agenda of holistic patient care centred on emotional intimacy and that an alternative occupational mandate focused on the healthcare mediator function might make for more humane health services and a more viable professional future.

  14. Gyroscope-driven mouse pointer with an EMOTIV® EEG headset and data analysis based on Empirical Mode Decomposition.

    PubMed

    Rosas-Cholula, Gerardo; Ramirez-Cortes, Juan Manuel; Alarcon-Aquino, Vicente; Gomez-Gil, Pilar; Rangel-Magdaleno, Jose de Jesus; Reyes-Garcia, Carlos

    2013-08-14

    This paper presents a project on the development of a cursor control emulating the typical operations of a computer-mouse, using gyroscope and eye-blinking electromyographic signals which are obtained through a commercial 16-electrode wireless headset, recently released by Emotiv. The cursor position is controlled using information from a gyroscope included in the headset. The clicks are generated through the user's blinking with an adequate detection procedure based on the spectral-like technique called Empirical Mode Decomposition (EMD). EMD is proposed as a simple and quick computational tool, yet effective, aimed to artifact reduction from head movements as well as a method to detect blinking signals for mouse control. Kalman filter is used as state estimator for mouse position control and jitter removal. The detection rate obtained in average was 94.9%. Experimental setup and some obtained results are presented.

  15. Analysis of Vibration and Noise of Construction Machinery Based on Ensemble Empirical Mode Decomposition and Spectral Correlation Analysis Method

    NASA Astrophysics Data System (ADS)

    Chen, Yuebiao; Zhou, Yiqi; Yu, Gang; Lu, Dan

    In order to analyze the effect of engine vibration on cab noise of construction machinery in multi-frequency bands, a new method based on ensemble empirical mode decomposition (EEMD) and spectral correlation analysis is proposed. Firstly, the intrinsic mode functions (IMFs) of vibration and noise signals were obtained by EEMD method, and then the IMFs which have the same frequency bands were selected. Secondly, we calculated the spectral correlation coefficients between the selected IMFs, getting the main frequency bands in which engine vibration has significant impact on cab noise. Thirdly, the dominated frequencies were picked out and analyzed by spectral analysis method. The study result shows that the main frequency bands and dominated frequencies in which engine vibration have serious impact on cab noise can be identified effectively by the proposed method, which provides effective guidance to noise reduction of construction machinery.

  16. Gyroscope-driven mouse pointer with an EMOTIV® EEG headset and data analysis based on Empirical Mode Decomposition.

    PubMed

    Rosas-Cholula, Gerardo; Ramirez-Cortes, Juan Manuel; Alarcon-Aquino, Vicente; Gomez-Gil, Pilar; Rangel-Magdaleno, Jose de Jesus; Reyes-Garcia, Carlos

    2013-01-01

    This paper presents a project on the development of a cursor control emulating the typical operations of a computer-mouse, using gyroscope and eye-blinking electromyographic signals which are obtained through a commercial 16-electrode wireless headset, recently released by Emotiv. The cursor position is controlled using information from a gyroscope included in the headset. The clicks are generated through the user's blinking with an adequate detection procedure based on the spectral-like technique called Empirical Mode Decomposition (EMD). EMD is proposed as a simple and quick computational tool, yet effective, aimed to artifact reduction from head movements as well as a method to detect blinking signals for mouse control. Kalman filter is used as state estimator for mouse position control and jitter removal. The detection rate obtained in average was 94.9%. Experimental setup and some obtained results are presented. PMID:23948873

  17. Empirically Supported Treatments in Psychotherapy: Towards an Evidence-Based or Evidence-Biased Psychology in Clinical Settings?

    PubMed Central

    Castelnuovo, Gianluca

    2010-01-01

    The field of research and practice in psychotherapy has been deeply influenced by two different approaches: the empirically supported treatments (ESTs) movement, linked with the evidence-based medicine (EBM) perspective and the “Common Factors” approach, typically connected with the “Dodo Bird Verdict”. About the first perspective, since 1998 a list of ESTs has been established in mental health field. Criterions for “well-established” and “probably efficacious” treatments have arisen. The development of these kinds of paradigms was motivated by the emergence of a “managerial” approach and related systems for remuneration also for mental health providers and for insurance companies. In this article ESTs will be presented underlining also some possible criticisms. Finally complementary approaches, that could add different evidence in the psychotherapy research in comparison with traditional EBM approach, are presented. PMID:21833197

  18. Using Information and Communications Technology in a National Population-Based Survey: The Kenya AIDS Indicator Survey 2012

    PubMed Central

    Ojwang’, James K.; Lee, Veronica C.; Waruru, Anthony; Ssempijja, Victor; Ng’ang’a, John G.; Wakhutu, Brian E.; Kandege, Nicholas O.; Koske, Danson K.; Kamiru, Samuel M.; Omondi, Kenneth O.; Kakinyi, Mutua; Kim, Andrea A.; Oluoch, Tom

    2016-01-01

    Background With improvements in technology, electronic data capture (EDC) for large surveys is feasible. EDC offers benefits over traditional paper-based data collection, including more accurate data, greater completeness of data, and decreased data cleaning burden. Methods The second Kenya AIDS Indicator Survey (KAIS 2012) was a population-based survey of persons aged 18 months to 64 years. A software application was designed to capture the interview, specimen collection, and home-based testing and counseling data. The application included: interview translations for local languages; options for single, multiple, and fill-in responses; and automated participant eligibility determination. Data quality checks were programmed to automate skip patterns and prohibit outlier responses. A data sharing architecture was developed to transmit the data in realtime from the field to a central server over a virtual private network. Results KAIS 2012 was conducted between October 2012 and February 2013. Overall, 68,202 records for the interviews, specimen collection, and home-based testing and counseling were entered into the application. Challenges arose during implementation, including poor connectivity and a systems malfunction that created duplicate records, which prevented timely data transmission to the central server. Data cleaning was minimal given the data quality control measures. Conclusions KAIS 2012 demonstrated the feasibility of using EDC in a population-based survey. The benefits of EDC were apparent in data quality and minimal time needed for data cleaning. Several important lessons were learned, such as the time and monetary investment required before survey implementation, the importance of continuous application testing, and contingency plans for data transmission due to connectivity challenges. PMID:24732816

  19. An Empirical Determination of the Intergalactic Background Light from UV to FIR Wavelengths Using FIR Deep Galaxy Surveys and the Gamma-Ray Opacity of the Universe

    NASA Astrophysics Data System (ADS)

    Stecker, Floyd W.; Scully, Sean T.; Malkan, Matthew A.

    2016-08-01

    We have previously calculated the intergalactic background light (IBL) as a function of redshift from the Lyman limit in the far-ultraviolet to a wavelength of 5 μm in the near-infrared range, based purely on data from deep galaxy surveys. Here, we use similar methods to determine the mid- and far-infrared IBL from 5 to 850 μm. Our approach enables us to constrain the range of photon densities by determining the uncertainties in observationally determined luminosity densities and spectral gradients. By also including the effect of the 2.7 K cosmic background photons, we determine upper and lower limits on the opacity of the universe to γ-rays up to PeV energies within a 68% confidence band. Our direct results on the IBL are consistent with those from complimentary γ-ray analyses using observations from the Fermi γ-ray space telescope and the H.E.S.S. air Čerenkov telescope. Thus, we find no evidence of previously suggested processes for the modification of γ-ray spectra other than that of absorption by pair production alone.

  20. An Empirical Study of Univariate and Genetic Algorithm-Based Feature Selection in Binary Classification with Microarray Data

    PubMed Central

    Lecocke, Michael; Hess, Kenneth

    2007-01-01

    Background We consider both univariate- and multivariate-based feature selection for the problem of binary classification with microarray data. The idea is to determine whether the more sophisticated multivariate approach leads to better misclassification error rates because of the potential to consider jointly significant subsets of genes (but without overfitting the data). Methods We present an empirical study in which 10-fold cross-validation is applied externally to both a univariate-based and two multivariate- (genetic algorithm (GA)-) based feature selection processes. These procedures are applied with respect to three supervised learning algorithms and six published two-class microarray datasets. Results Considering all datasets, and learning algorithms, the average 10-fold external cross-validation error rates for the univariate-, single-stage GA-, and two-stage GA-based processes are 14.2%, 14.6%, and 14.2%, respectively. We also find that the optimism bias estimates from the GA analyses were half that of the univariate approach, but the selection bias estimates from the GA analyses were 2.5 times that of the univariate results. Conclusions We find that the 10-fold external cross-validation misclassification error rates were very comparable. Further, we find that a two-stage GA approach did not demonstrate a significant advantage over a 1-stage approach. We also find that the univariate approach had higher optimism bias and lower selection bias compared to both GA approaches. PMID:19458774

  1. Process-based models not always better than empirical models for simulating budburst of Norway spruce and birch in Europe.

    PubMed

    Olsson, Cecilia; Jönsson, Anna Maria

    2014-11-01

    Budburst models have mainly been developed to capture the processes of individual trees, and vary in their complexity and plant physiological realism. We evaluated how well eleven models capture the variation in budburst of birch and Norway spruce in Germany, Austria, the United Kingdom and Finland. The comparison was based on the models performance in relation to their underlying physiological assumptions with four different calibration schemes. The models were not able to accurately simulate the timing of budburst. In general the models overestimated the temperature effect, thereby the timing of budburst was simulated too early in the United Kingdom and too late in Finland. Among the better performing models were three models based on the growing degree day concept, with or without day length or chilling, and an empirical model based on spring temperatures. These models were also the models least influenced by the calibration data. For birch the best calibration scheme was based on multiple sites in either Germany or Europe, and for Norway spruce the best scheme included multiple sites in Germany or cold years of all sites. Most model and calibration combinations indicated greater bias with higher spring temperatures, mostly simulating earlier than observed budburst.

  2. Determining Survey Satisficing of Online Longitudinal Survey Data in the Multicenter AIDS Cohort Study: A Group-Based Trajectory Analysis

    PubMed Central

    Di, Junrui; Li, Ying; Friedman, M Reuel; Reddy, Susheel; Surkan, Pamela J; Shoptaw, Steven

    2016-01-01

    Background Survey satisficing occurs when participants respond to survey questions rapidly without carefully reading or comprehending them. Studies have demonstrated the occurrence of survey satisficing, which can degrade survey quality, particularly in longitudinal studies. Objective The aim of this study is to use a group-based trajectory analysis method to identify satisficers when similar survey questions were asked periodically in a long-standing cohort, and to examine factors associated with satisficing in the surveys having sensitive human immunodeficiency virus (HIV)-related behavioral questions. Methods Behavioral data were collected semiannually online at all four sites of the Multicenter AIDS Cohort Study (MACS) from October 2008 through March 2013. Based on the start and end times, and the word counts per variable, response speed (word counts per second) for each participant visit was calculated. Two-step group-based trajectory analyses of the response speed across 9 study visits were performed to identify potential survey satisficing. Generalized linear models with repeated measures were used to investigate the factors associated with satisficing on HIV-related behavioral surveys. Results Among the total 2138 male participants, the median baseline age was 51 years (interquartile range, 45-58); most of the participants were non-Hispanic white (62.72%, 1341/2138) and college graduates (46.59%, 996/2138), and half were HIV seropositive (50.00%, 1069/2138). A total of 543 men (25.40%, 543/2138) were considered potential satisficers with respect to their increased trajectory tendency of response speed. In the multivariate analysis, being 10 years older at the baseline visit increased the odds of satisficing by 44% (OR 1.44, 95% CI 1.27-1.62, P<.001). Compared with the non-Hispanic white participants, non-Hispanic black participants were 122% more likely to satisfice the HIV-related behavioral survey (OR 2.22, 95% CI 1.69-2.91, P<.001), and 99% more likely

  3. Performance-based management and quality of work: an empirical assessment.

    PubMed

    Falzon, Pierre; Nascimento, Adelaide; Gaudart, Corinne; Piney, Cécile; Dujarier, Marie-Anne; Germe, Jean-François

    2012-01-01

    In France, in the private sector as in the public sector, performance-based management tends to become a norm. Performance-based management is supposed to improve service quality, productivity and efficiency, transparency of allotted means and achieved results, and to better focus the activity of employees and of the whole organization. This text reports a study conducted for the French Ministry of Budget by a team of researchers in ergonomics, sociology and management science, in order to assess the impact of performance-based management on employees, on teams and on work organization. About 100 interviews were conducted with employees of all categories and 6 working groups were set up in order to discuss and validate or amend our first analyses. Results concern several aspects: workload and work intensification, indicators and performance management and the transformation of jobs induced by performance management. PMID:22317310

  4. Accurate Young's modulus measurement based on Rayleigh wave velocity and empirical Poisson's ratio

    NASA Astrophysics Data System (ADS)

    Li, Mingxia; Feng, Zhihua

    2016-07-01

    This paper presents a method for Young's modulus measurement based on Rayleigh wave speed. The error in Poisson's ratio has weak influence on the measurement of Young's modulus based on Rayleigh wave speed, and Poisson's ratio minimally varies in a certain material; thus, we can accurately estimate Young's modulus with surface wave speed and a rough Poisson's ratio. We numerically analysed three methods using Rayleigh, longitudinal, and transversal wave speed, respectively, and the error in Poisson's ratio shows the least influence on the result in the method involving Rayleigh wave speed. An experiment was performed and has proved the feasibility of this method. Device for speed measuring could be small, and no sample pretreatment is needed. Hence, developing a portable instrument based on this method is possible. This method makes a good compromise between usability and precision.

  5. Voice Disorder Management Competencies: A Survey of School-Based Speech-Language Pathologists in Nebraska

    ERIC Educational Resources Information Center

    Teten, Amy F.; DeVeney, Shari L.; Friehe, Mary J.

    2016-01-01

    Purpose: The purpose of this survey was to determine the self-perceived competence levels in voice disorders of practicing school-based speech-language pathologists (SLPs) and identify correlated variables. Method: Participants were 153 master's level, school-based SLPs with a Nebraska teaching certificate and/or licensure who completed a survey,…

  6. Empirical Investigation into Motives for Choosing Web-Based Distance Learning Programs

    ERIC Educational Resources Information Center

    Alkhattabi, Mona

    2016-01-01

    Today, in association with rapid social and economic changes, there is an increasing level of demand for distance and online learning programs. This study will focus on identifying the main motivational factors for choosing a web-based distance-learning program. Moreover, it will investigate how these factors relate to age, gender, marital status…

  7. Young Readers' Narratives Based on a Picture Book: Model Readers and Empirical Readers

    ERIC Educational Resources Information Center

    Hoel, Trude

    2015-01-01

    The article present parts of a research project where the aim is to investigate six- to seven-year-old children's language use in storytelling. The children's oral texts are based on the wordless picture book "Frog, Where Are You?" Which has been, and still remains, a frequent tool for collecting narratives from children. The Frog story…

  8. Teaching Standards-Based Group Work Competencies to Social Work Students: An Empirical Examination

    ERIC Educational Resources Information Center

    Macgowan, Mark J.; Vakharia, Sheila P.

    2012-01-01

    Objectives: Accreditation standards and challenges in group work education require competency-based approaches in teaching social work with groups. The Association for the Advancement of Social Work with Groups developed Standards for Social Work Practice with Groups, which serve as foundation competencies for professional practice. However, there…

  9. Perceptions of the Effectiveness of System Dynamics-Based Interactive Learning Environments: An Empirical Study

    ERIC Educational Resources Information Center

    Qudrat-Ullah, Hassan

    2010-01-01

    The use of simulations in general and of system dynamics simulation based interactive learning environments (SDILEs) in particular is well recognized as an effective way of improving users' decision making and learning in complex, dynamic tasks. However, the effectiveness of SDILEs in classrooms has rarely been evaluated. This article describes…

  10. An Adaptive E-Learning System Based on Students' Learning Styles: An Empirical Study

    ERIC Educational Resources Information Center

    Drissi, Samia; Amirat, Abdelkrim

    2016-01-01

    Personalized e-learning implementation is recognized as one of the most interesting research areas in the distance web-based education. Since the learning style of each learner is different one must fit e-learning with the different needs of learners. This paper presents an approach to integrate learning styles into adaptive e-learning hypermedia.…

  11. An Empirical Typology of Residential Care/Assisted Living Based on a Four-State Study

    ERIC Educational Resources Information Center

    Park, Nan Sook; Zimmerman, Sheryl; Sloane, Philip D.; Gruber-Baldini, Ann L.; Eckert, J. Kevin

    2006-01-01

    Purpose: Residential care/assisted living describes diverse facilities providing non-nursing home care to a heterogeneous group of primarily elderly residents. This article derives typologies of assisted living based on theoretically and practically grounded evidence. Design and Methods: We obtained data from the Collaborative Studies of Long-Term…

  12. Introducing Evidence-Based Principles to Guide Collaborative Approaches to Evaluation: Results of an Empirical Process

    ERIC Educational Resources Information Center

    Shulha, Lyn M.; Whitmore, Elizabeth; Cousins, J. Bradley; Gilbert, Nathalie; al Hudib, Hind

    2016-01-01

    This article introduces a set of evidence-based principles to guide evaluation practice in contexts where evaluation knowledge is collaboratively produced by evaluators and stakeholders. The data from this study evolved in four phases: two pilot phases exploring the desirability of developing a set of principles; an online questionnaire survey…

  13. Journeys into Inquiry-Based Elementary Science: Literacy Practices, Questioning, and Empirical Study

    ERIC Educational Resources Information Center

    Howes, Elaine V.; Lim, Miyoun; Campos, Jaclyn

    2009-01-01

    Teaching literacy in inquiry-based science-teaching settings has recently become a focus of research in science education. Because professional scientists' uses of reading, writing, and speaking are foundational to their work, as well as to nonscientists' comprehension of it , it follows that literacy practices should also be central to science…

  14. Specification-based software sizing: An empirical investigation of function metrics

    NASA Technical Reports Server (NTRS)

    Jeffery, Ross; Stathis, John

    1993-01-01

    For some time the software industry has espoused the need for improved specification-based software size metrics. This paper reports on a study of nineteen recently developed systems in a variety of application domains. The systems were developed by a single software services corporation using a variety of languages. The study investigated several metric characteristics. It shows that: earlier research into inter-item correlation within the overall function count is partially supported; a priori function counts, in themself, do not explain the majority of the effort variation in software development in the organization studied; documentation quality is critical to accurate function identification; and rater error is substantial in manual function counting. The implication of these findings for organizations using function based metrics are explored.

  15. Conventional empirical law reverses in the phase transitions of 122-type iron-based superconductors

    PubMed Central

    Yu, Zhenhai; Wang, Lin; Wang, Luhong; Liu, Haozhe; Zhao, Jinggeng; Li, Chunyu; Sinogeikin, Stanislav; Wu, Wei; Luo, Jianlin; Wang, Nanlin; Yang, Ke; Zhao, Yusheng; Mao, Ho-kwang

    2014-01-01

    Phase transition of solid-state materials is a fundamental research topic in condensed matter physics, materials science and geophysics. It has been well accepted and widely proven that isostructural compounds containing different cations undergo same pressure-induced phase transitions but at progressively lower pressures as the cation radii increases. However, we discovered that this conventional law reverses in the structural transitions in 122-type iron-based superconductors. In this report, a combined low temperature and high pressure X-ray diffraction (XRD) measurement has identified the phase transition curves among the tetragonal (T), orthorhombic (O) and the collapsed-tetragonal (cT) phases in the structural phase diagram of the iron-based superconductor AFe2As2 (A = Ca, Sr, Eu, and Ba). The cation radii dependence of the phase transition pressure (T → cT) shows an opposite trend in which the compounds with larger ambient radii cations have a higher transition pressure. PMID:25417655

  16. Conventional empirical law reverses in the phase transitions of 122-type iron-based superconductors.

    PubMed

    Yu, Zhenhai; Wang, Lin; Wang, Luhong; Liu, Haozhe; Zhao, Jinggeng; Li, Chunyu; Sinogeikin, Stanislav; Wu, Wei; Luo, Jianlin; Wang, Nanlin; Yang, Ke; Zhao, Yusheng; Mao, Ho-kwang

    2014-01-01

    Phase transition of solid-state materials is a fundamental research topic in condensed matter physics, materials science and geophysics. It has been well accepted and widely proven that isostructural compounds containing different cations undergo same pressure-induced phase transitions but at progressively lower pressures as the cation radii increases. However, we discovered that this conventional law reverses in the structural transitions in 122-type iron-based superconductors. In this report, a combined low temperature and high pressure X-ray diffraction (XRD) measurement has identified the phase transition curves among the tetragonal (T), orthorhombic (O) and the collapsed-tetragonal (cT) phases in the structural phase diagram of the iron-based superconductor AFe2As2 (A = Ca, Sr, Eu, and Ba). The cation radii dependence of the phase transition pressure (T → cT) shows an opposite trend in which the compounds with larger ambient radii cations have a higher transition pressure. PMID:25417655

  17. Evidence-based architectural and space design supports Magnet® empirical outcomes.

    PubMed

    Ecoff, Laurie; Brown, Caroline E

    2010-12-01

    This department expands nursing leaders' knowledge and competencies in health facility design. The editor of this department, Dr Jaynelle Stichler, asked guest authors, Drs Ecoff and Brown, to describe the process of using the conceptual models of a nursing evidence-based practice model and the Magnet Recognition Program® as a structured process to lead decision making in the planning and design processes and to achieve desired outcomes in hospital design.

  18. Patients’ Acceptance towards a Web-Based Personal Health Record System: An Empirical Study in Taiwan

    PubMed Central

    Liu, Chung-Feng; Tsai, Yung-Chieh; Jang, Fong-Lin

    2013-01-01

    The health care sector has become increasingly interested in developing personal health record (PHR) systems as an Internet-based telehealthcare implementation to improve the quality and decrease the cost of care. However, the factors that influence patients’ intention to use PHR systems remain unclear. Based on physicians’ therapeutic expertise, we implemented a web-based infertile PHR system and proposed an extended Technology Acceptance Model (TAM) that integrates the physician-patient relationship (PPR) construct into TAM’s original perceived ease of use (PEOU) and perceived usefulness (PU) constructs to explore which factors will influence the behavioral intentions (BI) of infertile patients to use the PHR. From ninety participants from a medical center, 50 valid responses to a self-rating questionnaire were collected, yielding a response rate of 55.56%. The partial least squares (PLS) technique was used to assess the causal relationships that were hypothesized in the extended model. The results indicate that infertile patients expressed a moderately high intention to use the PHR system. The PPR and PU of patients had significant effects on their BI to use PHR, whereas the PEOU indirectly affected the patients’ BI through the PU. This investigation confirms that PPR can have a critical role in shaping patients’ perceptions of the use of healthcare information technologies. Hence, we suggest that hospitals should promote the potential usefulness of PHR and improve the quality of the physician-patient relationship to increase patients’ intention of using PHR. PMID:24142185

  19. Empirically Supported Family-Based Treatments for Conduct Disorder and Delinquency in Adolescents

    PubMed Central

    Henggeler, Scott W.; Sheidow, Ashli J.

    2011-01-01

    Several family-based treatments of conduct disorder and delinquency in adolescents have emerged as evidence-based and, in recent years, have been transported to more than 800 community practice settings. These models include multisystemic therapy, functional family therapy, multidimensional treatment foster care, and, to a lesser extent, brief strategic family therapy. In addition to summarizing the theoretical and clinical bases of these treatments, their results in efficacy and effectiveness trials are examined with particular emphasis on any demonstrated capacity to achieve favorable outcomes when implemented by real world practitioners in community practice settings. Special attention is also devoted to research on purported mechanisms of change as well as the long-term sustainability of outcomes achieved by these treatment models. Importantly, we note that the developers of each of the models have developed quality assurance systems to support treatment fidelity and youth and family outcomes; and the developers have formed purveyor organizations to facilitate the large scale transport of their respective treatments to community settings nationally and internationally. PMID:22283380

  20. Empirical Study on Designing of Gaze Tracking Camera Based on the Information of User’s Head Movement

    PubMed Central

    Pan, Weiyuan; Jung, Dongwook; Yoon, Hyo Sik; Lee, Dong Eun; Naqvi, Rizwan Ali; Lee, Kwan Woo; Park, Kang Ryoung

    2016-01-01

    Gaze tracking is the technology that identifies a region in space that a user is looking at. Most previous non-wearable gaze tracking systems use a near-infrared (NIR) light camera with an NIR illuminator. Based on the kind of camera lens used, the viewing angle and depth-of-field (DOF) of a gaze tracking camera can be different, which affects the performance of the gaze tracking system. Nevertheless, to our best knowledge, most previous researches implemented gaze tracking cameras without ground truth information for determining the optimal viewing angle and DOF of the camera lens. Eye-tracker manufacturers might also use ground truth information, but they do not provide this in public. Therefore, researchers and developers of gaze tracking systems cannot refer to such information for implementing gaze tracking system. We address this problem providing an empirical study in which we design an optimal gaze tracking camera based on experimental measurements of the amount and velocity of user’s head movements. Based on our results and analyses, researchers and developers might be able to more easily implement an optimal gaze tracking system. Experimental results show that our gaze tracking system shows high performance in terms of accuracy, user convenience and interest. PMID:27589768

  1. Empirical force field for cisplatin based on quantum dynamics data: case study of new parameterization scheme for coordination compounds.

    PubMed

    Yesylevskyy, S; Cardey, Bruno; Kraszewski, S; Foley, Sarah; Enescu, Mironel; da Silva, Antônio M; Dos Santos, Hélio F; Ramseyer, Christophe

    2015-10-01

    Parameterization of molecular complexes containing a metallic compound, such as cisplatin, is challenging due to the unconventional coordination nature of the bonds which involve platinum atoms. In this work, we develop a new methodology of parameterization for such compounds based on quantum dynamics (QD) calculations. We show that the coordination bonds and angles are more flexible than in normal covalent compounds. The influence of explicit solvent is also shown to be crucial to determine the flexibility of cisplatin in quantum dynamics simulations. Two empirical topologies of cisplatin were produced by fitting its atomic fluctuations against QD in vacuum and QD with explicit first solvation shell of water molecules respectively. A third topology built in a standard way from the static optimized structure was used for comparison. The later one leads to an excessively rigid molecule and exhibits much smaller fluctuations of the bonds and angles than QD reveals. It is shown that accounting for the high flexibility of cisplatin molecule is needed for adequate description of its first hydration shell. MD simulations with flexible QD-based topology also reveal a significant decrease of the barrier of passive diffusion of cisplatin accross the model lipid bilayer. These results confirm that flexibility of organometallic compounds is an important feature to be considered in classical molecular dynamics topologies. Proposed methodology based on QD simulations provides a systematic way of building such topologies.

  2. Lessons Learned: Cultural and linguistic enhancement of surveys through community-based participatory research

    PubMed Central

    Formea, Christine M.; Mohamed, Ahmed A.; Hassan, Abdullahi; Osman, Ahmed; Weis, Jennifer A.; Sia, Irene G.; Wieland, Mark L.

    2014-01-01

    Background Surveys are frequently implemented in community-based participatory research (CBPR), but adaptation and translation of surveys can be logistically and methodologically challenging when working with immigrant and refugee populations. Objective To describe a process of participatory survey adaptation and translation. Methods Within an established CBPR partnership, a survey about diabetes was adapted for health literacy and local relevance and then translated through a process of forward translation, group deliberation, and back translation. Lessons Learned The group deliberation process was the most time-intensive and important component of the process. The process enhanced community ownership of the larger project while maximizing local applicability of the product. Conclusions A participatory process of survey adaptation and translation resulted in significant revisions to approximate semantic, cultural, and conceptual equivalence with the original surveys. This approach is likely to enhance community acceptance of the survey instrument during the implementation phase. PMID:25435559

  3. An empirical analysis of exposure-based regulation to abate toxic air pollution

    SciTech Connect

    Marakovits, D.M.; Considine, T.J.

    1996-11-01

    Title III of the 1990 Clean Air Act Amendments requires the Environmental Protection Agency to regulate 189 air toxics, including emissions from by-product coke ovens. Economists criticize the inefficiency of uniform standards, but Title III makes no provision for flexible regulatory instruments. Environmental health scientists suggest that population exposure, not necessarily ambient air quality, should motivate environmental air pollution policies. Using an engineering-economic model of the United States steel industry, we estimate that an exposure-based policy can achieve the same level of public health as coke oven emissions standards and can reduce compliance costs by up to 60.0%. 18 refs., 3 figs., 1 tab.

  4. Conventional empirical law reverses in the phase transitions of 122-type iron-based superconductors

    SciTech Connect

    Yu, Zhenhai; Wang, Lin; Wang, Luhong; Liu, Haozhe; Zhao, Jinggeng; Li, Chunyu; Sinogeikin, Stanislav; Wu, Wei; Luo, Jianlin; Wang, Nanlin; Yang, Ke; Zhao, Yusheng; Mao, Ho -kwang

    2014-11-24

    Phase transition of solid-state materials is a fundamental research topic in condensed matter physics, materials science and geophysics. It has been well accepted and widely proven that isostructural compounds containing different cations undergo same pressure-induced phase transitions but at progressively lower pressures as the cation radii increases. However, we discovered that this conventional law reverses in the structural transitions in 122-type iron-based superconductors. In this report, a combined low temperature and high pressure X-ray diffraction (XRD) measurement has identified the phase transition curves among the tetragonal (T), orthorhombic (O) and the collapsed-tetragonal (cT) phases in the structural phase diagram of the iron-based superconductor AFe2As2 (A = Ca, Sr, Eu, and Ba). As a result, the cation radii dependence of the phase transition pressure (T → cT) shows an opposite trend in which the compounds with larger ambient radii cations have a higher transition pressure.

  5. Conventional empirical law reverses in the phase transitions of 122-type iron-based superconductors

    DOE PAGES

    Yu, Zhenhai; Wang, Lin; Wang, Luhong; Liu, Haozhe; Zhao, Jinggeng; Li, Chunyu; Sinogeikin, Stanislav; Wu, Wei; Luo, Jianlin; Wang, Nanlin; et al

    2014-11-24

    Phase transition of solid-state materials is a fundamental research topic in condensed matter physics, materials science and geophysics. It has been well accepted and widely proven that isostructural compounds containing different cations undergo same pressure-induced phase transitions but at progressively lower pressures as the cation radii increases. However, we discovered that this conventional law reverses in the structural transitions in 122-type iron-based superconductors. In this report, a combined low temperature and high pressure X-ray diffraction (XRD) measurement has identified the phase transition curves among the tetragonal (T), orthorhombic (O) and the collapsed-tetragonal (cT) phases in the structural phase diagram ofmore » the iron-based superconductor AFe2As2 (A = Ca, Sr, Eu, and Ba). As a result, the cation radii dependence of the phase transition pressure (T → cT) shows an opposite trend in which the compounds with larger ambient radii cations have a higher transition pressure.« less

  6. Environmental surveys in Alaska based upon ERTS data

    NASA Technical Reports Server (NTRS)

    Miller, J. M.

    1974-01-01

    Alaskan applications of ERTS data are summarized. Areas discussed are: (1) land use; (2) archaeology; (3) vegetation mapping; (4) ice reporting and mapping; (5) permafrost; (6) mineral and oil exploration; (7) geological surveys; (8) seismology; (9) geological faults and structures; (10) hydrology and water resources; (11) glaciology; (12) water circulation in Cook Inlet; and (13) fish and mammal populations.

  7. Teaching Margery and Julian in Anthology-Based Survey Courses

    ERIC Educational Resources Information Center

    Petersen, Zina

    2006-01-01

    Recognizing that many of us teach the medieval English women mystics Margery Kempe and Julian of Norwich in survey courses, this essay attempts to put these writers in context for teachers who may have only a passing familiarity with the period. Focusing on passages of their writings found in the Longman and Norton anthologies of British…

  8. Fall 1994 wildlife and vegetation survey, Norton Air Force Base, California

    SciTech Connect

    Not Available

    1994-12-15

    The fall 1994 wildlife and vegetation surveys were completed October 3-7, 1994, at Norton Air Force Base (AFB), California. Two biologists from CDM Federal Programs, the U.S. Environmental Protection Agency (EPA) regional biologist and the Oak Ridge National Laboratory (ORNL) lead biologist conducted the surveys. A habitat assessment of three Installation Restoration Project (IRP) sites at Norton Air Force Base was also completed during the fall survey period. The IRP sites include: Landfill No. 2 (Site 2); the Industrial Wastewater Treatment Plant (IWTP) area; and Former Fire Training Area No. 1 (Site 5). The assessments were designed to qualitatively characterize the sites of concern, identify potential ecological receptors, and provide information for Remedial Design/Remedial Action activities. A Reference Area (Santa Ana River Wash) and the base urban areas were also characterized. The reference area assessment was performed to provide a baseline for comparison with the IRP site habitats. The fall 1994 survey is the second of up to four surveys that may be completed. In order to develop a complete understanding of all plant and animal species using the base, these surveys were planned to be conducted over four seasons. Species composition can vary widely during the course of a year in Southern California, and therefore, seasonal surveys will provide the most complete and reliable data to address changes in habitat structure and wildlife use of the site. Subsequent surveys will focus on seasonal wildlife observations and a spring vegetation survey.

  9. Psychological first aid: a consensus-derived, empirically supported, competency-based training model.

    PubMed

    McCabe, O Lee; Everly, George S; Brown, Lisa M; Wendelboe, Aaron M; Abd Hamid, Nor Hashidah; Tallchief, Vicki L; Links, Jonathan M

    2014-04-01

    Surges in demand for professional mental health services occasioned by disasters represent a major public health challenge. To build response capacity, numerous psychological first aid (PFA) training models for professional and lay audiences have been developed that, although often concurring on broad intervention aims, have not systematically addressed pedagogical elements necessary for optimal learning or teaching. We describe a competency-based model of PFA training developed under the auspices of the Centers for Disease Control and Prevention and the Association of Schools of Public Health. We explain the approach used for developing and refining the competency set and summarize the observable knowledge, skills, and attitudes underlying the 6 core competency domains. We discuss the strategies for model dissemination, validation, and adoption in professional and lay communities. PMID:23865656

  10. An empirical study of the mechanisms of mindfulness in a mindfulness-based stress reduction program.

    PubMed

    Carmody, James; Baer, Ruth A; L B Lykins, Emily; Olendzki, Nicholas

    2009-06-01

    S. L. Shapiro and colleagues (2006) have described a testable theory of the mechanisms of mindfulness and how it affects positive change. They describe a model in which mindfulness training leads to a fundamental change in relationship to experience (reperceiving), which leads to changes in self-regulation, values clarification, cognitive and behavioral flexibility, and exposure. These four variables, in turn, result in salutogenic outcomes. Analyses of responses from participants in a mindfulness-based stress-reduction program did not support the mediating effect of changes in reperceiving on the relationship of mindfulness with those four variables. However, when mindfulness and reperceiving scores were combined, partial support was found for the mediating effect of the four variables on measures of psychological distress. Issues arising in attempts to test the proposed theory are discussed, including the description of the model variables and the challenges to their assessment.

  11. Psychological First Aid: A Consensus-Derived, Empirically Supported, Competency-Based Training Model

    PubMed Central

    Everly, George S.; Brown, Lisa M.; Wendelboe, Aaron M.; Abd Hamid, Nor Hashidah; Tallchief, Vicki L.; Links, Jonathan M.

    2014-01-01

    Surges in demand for professional mental health services occasioned by disasters represent a major public health challenge. To build response capacity, numerous psychological first aid (PFA) training models for professional and lay audiences have been developed that, although often concurring on broad intervention aims, have not systematically addressed pedagogical elements necessary for optimal learning or teaching. We describe a competency-based model of PFA training developed under the auspices of the Centers for Disease Control and Prevention and the Association of Schools of Public Health. We explain the approach used for developing and refining the competency set and summarize the observable knowledge, skills, and attitudes underlying the 6 core competency domains. We discuss the strategies for model dissemination, validation, and adoption in professional and lay communities. PMID:23865656

  12. Behavioral Modeling Based on Probabilistic Finite Automata: An Empirical Study †

    PubMed Central

    Tîrnăucă, Cristina; Montaña, José L.; Ontañón, Santiago; González, Avelino J.; Pardo, Luis M.

    2016-01-01

    Imagine an agent that performs tasks according to different strategies. The goal of Behavioral Recognition (BR) is to identify which of the available strategies is the one being used by the agent, by simply observing the agent’s actions and the environmental conditions during a certain period of time. The goal of Behavioral Cloning (BC) is more ambitious. In this last case, the learner must be able to build a model of the behavior of the agent. In both settings, the only assumption is that the learner has access to a training set that contains instances of observed behavioral traces for each available strategy. This paper studies a machine learning approach based on Probabilistic Finite Automata (PFAs), capable of achieving both the recognition and cloning tasks. We evaluate the performance of PFAs in the context of a simulated learning environment (in this case, a virtual Roomba vacuum cleaner robot), and compare it with a collection of other machine learning approaches. PMID:27347956

  13. An Empirical Pixel-Based CTE Correction for ACS/WFC

    NASA Astrophysics Data System (ADS)

    Anderson, Jay

    2010-07-01

    This presentation summarizes a paper that has been recently published in PASP, Anderson & Bedin (2010). The paper describes our pixel-based approach to correcting ACS data for imperfect CTE (charge-transfer efficiency). We developed the approach by characterizing the size and profiles of trails behind warm pixels in dark exposures. We found an algorithm that simulates the way imperfect CTE impacts the readout process. To correct images for imperfect CTE, we use a forwardmodeling procedure to determine the likely original distribution of charge, given the distribution that was read out. We applied this CTE-reconstruction algorithm to science images and found that the fluxes, positions and shapes of stars were restored to high fidelity. The ACS team is currently working to make this correction available to the public; they are also running tests to determine whether and how best to implement it in the pipeline.

  14. Psychological first aid: a consensus-derived, empirically supported, competency-based training model.

    PubMed

    McCabe, O Lee; Everly, George S; Brown, Lisa M; Wendelboe, Aaron M; Abd Hamid, Nor Hashidah; Tallchief, Vicki L; Links, Jonathan M

    2014-04-01

    Surges in demand for professional mental health services occasioned by disasters represent a major public health challenge. To build response capacity, numerous psychological first aid (PFA) training models for professional and lay audiences have been developed that, although often concurring on broad intervention aims, have not systematically addressed pedagogical elements necessary for optimal learning or teaching. We describe a competency-based model of PFA training developed under the auspices of the Centers for Disease Control and Prevention and the Association of Schools of Public Health. We explain the approach used for developing and refining the competency set and summarize the observable knowledge, skills, and attitudes underlying the 6 core competency domains. We discuss the strategies for model dissemination, validation, and adoption in professional and lay communities.

  15. Empirical estimation of consistency parameter in intertemporal choice based on Tsallis’ statistics

    NASA Astrophysics Data System (ADS)

    Takahashi, Taiki; Oono, Hidemi; Radford, Mark H. B.

    2007-07-01

    Impulsivity and inconsistency in intertemporal choice have been attracting attention in econophysics and neuroeconomics. Although loss of self-control by substance abusers is strongly related to their inconsistency in intertemporal choice, researchers in neuroeconomics and psychopharmacology have usually studied impulsivity in intertemporal choice using a discount rate (e.g. hyperbolic k), with little effort being expended on parameterizing subject's inconsistency in intertemporal choice. Recent studies using Tsallis’ statistics-based econophysics have found a discount function (i.e. q-exponential discount function), which may continuously parameterize a subject's consistency in intertemporal choice. In order to examine the usefulness of the consistency parameter (0⩽q⩽1) in the q-exponential discounting function in behavioral studies, we experimentally estimated the consistency parameter q in Tsallis’ statistics-based discounting function by assessing the points of subjective equality (indifference points) at seven delays (1 week-25 years) in humans (N=24). We observed that most (N=19) subjects’ intertemporal choice was completely inconsistent ( q=0, i.e. hyperbolic discounting), the mean consistency (0⩽q⩽1) was smaller than 0.5, and only one subject had a completely consistent intertemporal choice ( q=1, i.e. exponential discounting). There was no significant correlation between impulsivity and inconsistency parameters. Our results indicate that individual differences in consistency in intertemporal choice can be parameterized by introducing a q-exponential discount function and most people discount delayed rewards hyperbolically, rather than exponentially (i.e. mean q is smaller than 0.5). Further, impulsivity and inconsistency in intertemporal choice can be considered as separate behavioral tendencies. The usefulness of the consistency parameter q in psychopharmacological studies of addictive behavior was demonstrated in the present study.

  16. [Research on ECG de-noising method based on ensemble empirical mode decomposition and wavelet transform using improved threshold function].

    PubMed

    Ye, Linlin; Yang, Dan; Wang, Xu

    2014-06-01

    A de-noising method for electrocardiogram (ECG) based on ensemble empirical mode decomposition (EEMD) and wavelet threshold de-noising theory is proposed in our school. We decomposed noised ECG signals with the proposed method using the EEMD and calculated a series of intrinsic mode functions (IMFs). Then we selected IMFs and reconstructed them to realize the de-noising for ECG. The processed ECG signals were filtered again with wavelet transform using improved threshold function. In the experiments, MIT-BIH ECG database was used for evaluating the performance of the proposed method, contrasting with de-noising method based on EEMD and wavelet transform with improved threshold function alone in parameters of signal to noise ratio (SNR) and mean square error (MSE). The results showed that the ECG waveforms de-noised with the proposed method were smooth and the amplitudes of ECG features did not attenuate. In conclusion, the method discussed in this paper can realize the ECG denoising and meanwhile keep the characteristics of original ECG signal. PMID:25219236

  17. A UNIFIED EMPIRICAL MODEL FOR INFRARED GALAXY COUNTS BASED ON THE OBSERVED PHYSICAL EVOLUTION OF DISTANT GALAXIES

    SciTech Connect

    Bethermin, Matthieu; Daddi, Emanuele; Sargent, Mark T.; Elbaz, David; Mullaney, James; Pannella, Maurilio; Hezaveh, Yashar; Le Borgne, Damien; Buat, Veronique; Charmandaris, Vassilis; Lagache, Guilaine; Scott, Douglas

    2012-10-01

    We reproduce the mid-infrared to radio galaxy counts with a new empirical model based on our current understanding of the evolution of main-sequence (MS) and starburst (SB) galaxies. We rely on a simple spectral energy distribution (SED) library based on Herschel observations: a single SED for the MS and another one for SB, getting warmer with redshift. Our model is able to reproduce recent measurements of galaxy counts performed with Herschel, including counts per redshift slice. This agreement demonstrates the power of our 2-Star-Formation Modes (2SFM) decomposition in describing the statistical properties of infrared sources and their evolution with cosmic time. We discuss the relative contribution of MS and SB galaxies to the number counts at various wavelengths and flux densities. We also show that MS galaxies are responsible for a bump in the 1.4 GHz radio counts around 50 {mu}Jy. Material of the model (predictions, SED library, mock catalogs, etc.) is available online.

  18. An Empirical Orthogonal Function-Based Algorithm for Estimating Terrestrial Latent Heat Flux from Eddy Covariance, Meteorological and Satellite Observations.

    PubMed

    Feng, Fei; Li, Xianglan; Yao, Yunjun; Liang, Shunlin; Chen, Jiquan; Zhao, Xiang; Jia, Kun; Pintér, Krisztina; McCaughey, J Harry

    2016-01-01

    Accurate estimation of latent heat flux (LE) based on remote sensing data is critical in characterizing terrestrial ecosystems and modeling land surface processes. Many LE products were released during the past few decades, but their quality might not meet the requirements in terms of data consistency and estimation accuracy. Merging multiple algorithms could be an effective way to improve the quality of existing LE products. In this paper, we present a data integration method based on modified empirical orthogonal function (EOF) analysis to integrate the Moderate Resolution Imaging Spectroradiometer (MODIS) LE product (MOD16) and the Priestley-Taylor LE algorithm of Jet Propulsion Laboratory (PT-JPL) estimate. Twenty-two eddy covariance (EC) sites with LE observation were chosen to evaluate our algorithm, showing that the proposed EOF fusion method was capable of integrating the two satellite data sets with improved consistency and reduced uncertainties. Further efforts were needed to evaluate and improve the proposed algorithm at larger spatial scales and time periods, and over different land cover types. PMID:27472383

  19. Chemical and physical influences on aerosol activation in liquid clouds: an empirical study based on observations from the Jungfraujoch, Switzerland

    NASA Astrophysics Data System (ADS)

    Hoyle, C. R.; Webster, C. S.; Rieder, H. E.; Hammer, E.; Gysel, M.; Bukowiecki, N.; Weingartner, E.; Steinbacher, M.; Baltensperger, U.

    2015-06-01

    A simple empirical model to predict the number of aerosols which activate to form cloud droplets in a warm, free tropospheric cloud has been established, based on data from four summertime Cloud and Aerosol Characterisation Experiments (CLACE) campaigns at the Jungfraujoch (JFJ). It is shown that 76% of the observed variance in droplet numbers can be represented by a model accounting only for the number of potential CCN (defined as number of particles larger than 90 nm in diameter), while the mean errors in the model representation may be reduced by the addition of further explanatory variables, such as the mixing ratios of O3, CO and the height of the measurements above cloud base. The model has similar ability to represent the observed droplet numbers in each of the individual years, as well as for the two predominant local wind directions at the JFJ (north west and south east). Given the central European location of the JFJ, with air masses in summer being representative of the free troposphere with regular boundary layer in-mixing via convection, we expect that this model is applicable to warm, free tropospheric clouds over the European continent.

  20. An Empirical Orthogonal Function-Based Algorithm for Estimating Terrestrial Latent Heat Flux from Eddy Covariance, Meteorological and Satellite Observations.

    PubMed

    Feng, Fei; Li, Xianglan; Yao, Yunjun; Liang, Shunlin; Chen, Jiquan; Zhao, Xiang; Jia, Kun; Pintér, Krisztina; McCaughey, J Harry

    2016-01-01

    Accurate estimation of latent heat flux (LE) based on remote sensing data is critical in characterizing terrestrial ecosystems and modeling land surface processes. Many LE products were released during the past few decades, but their quality might not meet the requirements in terms of data consistency and estimation accuracy. Merging multiple algorithms could be an effective way to improve the quality of existing LE products. In this paper, we present a data integration method based on modified empirical orthogonal function (EOF) analysis to integrate the Moderate Resolution Imaging Spectroradiometer (MODIS) LE product (MOD16) and the Priestley-Taylor LE algorithm of Jet Propulsion Laboratory (PT-JPL) estimate. Twenty-two eddy covariance (EC) sites with LE observation were chosen to evaluate our algorithm, showing that the proposed EOF fusion method was capable of integrating the two satellite data sets with improved consistency and reduced uncertainties. Further efforts were needed to evaluate and improve the proposed algorithm at larger spatial scales and time periods, and over different land cover types.

  1. [Research on ECG de-noising method based on ensemble empirical mode decomposition and wavelet transform using improved threshold function].

    PubMed

    Ye, Linlin; Yang, Dan; Wang, Xu

    2014-06-01

    A de-noising method for electrocardiogram (ECG) based on ensemble empirical mode decomposition (EEMD) and wavelet threshold de-noising theory is proposed in our school. We decomposed noised ECG signals with the proposed method using the EEMD and calculated a series of intrinsic mode functions (IMFs). Then we selected IMFs and reconstructed them to realize the de-noising for ECG. The processed ECG signals were filtered again with wavelet transform using improved threshold function. In the experiments, MIT-BIH ECG database was used for evaluating the performance of the proposed method, contrasting with de-noising method based on EEMD and wavelet transform with improved threshold function alone in parameters of signal to noise ratio (SNR) and mean square error (MSE). The results showed that the ECG waveforms de-noised with the proposed method were smooth and the amplitudes of ECG features did not attenuate. In conclusion, the method discussed in this paper can realize the ECG denoising and meanwhile keep the characteristics of original ECG signal.

  2. An Empirical Orthogonal Function-Based Algorithm for Estimating Terrestrial Latent Heat Flux from Eddy Covariance, Meteorological and Satellite Observations

    PubMed Central

    Feng, Fei; Li, Xianglan; Yao, Yunjun; Liang, Shunlin; Chen, Jiquan; Zhao, Xiang; Jia, Kun; Pintér, Krisztina; McCaughey, J. Harry

    2016-01-01

    Accurate estimation of latent heat flux (LE) based on remote sensing data is critical in characterizing terrestrial ecosystems and modeling land surface processes. Many LE products were released during the past few decades, but their quality might not meet the requirements in terms of data consistency and estimation accuracy. Merging multiple algorithms could be an effective way to improve the quality of existing LE products. In this paper, we present a data integration method based on modified empirical orthogonal function (EOF) analysis to integrate the Moderate Resolution Imaging Spectroradiometer (MODIS) LE product (MOD16) and the Priestley-Taylor LE algorithm of Jet Propulsion Laboratory (PT-JPL) estimate. Twenty-two eddy covariance (EC) sites with LE observation were chosen to evaluate our algorithm, showing that the proposed EOF fusion method was capable of integrating the two satellite data sets with improved consistency and reduced uncertainties. Further efforts were needed to evaluate and improve the proposed algorithm at larger spatial scales and time periods, and over different land cover types. PMID:27472383

  3. PDE-based Non-Linear Diffusion Techniques for Denoising Scientific and Industrial Images: An Empirical Study

    SciTech Connect

    Weeratunga, S K; Kamath, C

    2001-12-20

    Removing noise from data is often the first step in data analysis. Denoising techniques should not only reduce the noise, but do so without blurring or changing the location of the edges. Many approaches have been proposed to accomplish this; in this paper, they focus on one such approach, namely the use of non-linear diffusion operators. This approach has been studied extensively from a theoretical viewpoint ever since the 1987 work of Perona and Malik showed that non-linear filters outperformed the more traditional linear Canny edge detector. They complement this theoretical work by investigating the performance of several isotropic diffusion operators on test images from scientific domains. They explore the effects of various parameters such as the choice of diffusivity function, explicit and implicit methods for the discretization of the PDE, and approaches for the spatial discretization of the non-linear operator etc. They also compare these schemes with simple spatial filters and the more complex wavelet-based shrinkage techniques. The empirical results show that, with an appropriate choice of parameters, diffusion-based schemes can be as effective as competitive techniques.

  4. A Cutting Pattern Recognition Method for Shearers Based on Improved Ensemble Empirical Mode Decomposition and a Probabilistic Neural Network

    PubMed Central

    Xu, Jing; Wang, Zhongbin; Tan, Chao; Si, Lei; Liu, Xinhua

    2015-01-01

    In order to guarantee the stable operation of shearers and promote construction of an automatic coal mining working face, an online cutting pattern recognition method with high accuracy and speed based on Improved Ensemble Empirical Mode Decomposition (IEEMD) and Probabilistic Neural Network (PNN) is proposed. An industrial microphone is installed on the shearer and the cutting sound is collected as the recognition criterion to overcome the disadvantages of giant size, contact measurement and low identification rate of traditional detectors. To avoid end-point effects and get rid of undesirable intrinsic mode function (IMF) components in the initial signal, IEEMD is conducted on the sound. The end-point continuation based on the practical storage data is performed first to overcome the end-point effect. Next the average correlation coefficient, which is calculated by the correlation of the first IMF with others, is introduced to select essential IMFs. Then the energy and standard deviation of the reminder IMFs are extracted as features and PNN is applied to classify the cutting patterns. Finally, a simulation example, with an accuracy of 92.67%, and an industrial application prove the efficiency and correctness of the proposed method. PMID:26528985

  5. Evaluation of empirical rule of linearly correlated peptide selection (ERLPS) for proteotypic peptide-based quantitative proteomics.

    PubMed

    Liu, Kehui; Zhang, Jiyang; Fu, Bin; Xie, Hongwei; Wang, Yingchun; Qian, Xiaohong

    2014-07-01

    Precise protein quantification is essential in comparative proteomics. Currently, quantification bias is inevitable when using proteotypic peptide-based quantitative proteomics strategy for the differences in peptides measurability. To improve quantification accuracy, we proposed an "empirical rule for linearly correlated peptide selection (ERLPS)" in quantitative proteomics in our previous work. However, a systematic evaluation on general application of ERLPS in quantitative proteomics under diverse experimental conditions needs to be conducted. In this study, the practice workflow of ERLPS was explicitly illustrated; different experimental variables, such as, different MS systems, sample complexities, sample preparations, elution gradients, matrix effects, loading amounts, and other factors were comprehensively investigated to evaluate the applicability, reproducibility, and transferability of ERPLS. The results demonstrated that ERLPS was highly reproducible and transferable within appropriate loading amounts and linearly correlated response peptides should be selected for each specific experiment. ERLPS was used to proteome samples from yeast to mouse and human, and in quantitative methods from label-free to O18/O16-labeled and SILAC analysis, and enabled accurate measurements for all proteotypic peptide-based quantitative proteomics over a large dynamic range.

  6. An empirical comparison of different LDA methods in fMRI-based brain states decoding.

    PubMed

    Xia, Maogeng; Song, Sutao; Yao, Li; Long, Zhiying

    2015-01-01

    Decoding brain states from response patterns with multivariate pattern recognition techniques is a popular method for detecting multivoxel patterns of brain activation. These patterns are informative with respect to a subject's perceptual or cognitive states. Linear discriminant analysis (LDA) cannot be directly applied to fMRI data analysis because of the "few samples and large features" nature of functional magnetic resonance imaging (fMRI) data. Although several improved LDA methods have been used in fMRI-based decoding, little is known regarding the relative performance of different LDA classifiers on fMRI data. In this study, we compared five LDA classifiers using both simulated data with varied noise levels and real fMRI data. The compared LDA classifiers include LDA combined with PCA (LDA-PCA), LDA with three types of regularizations (identity matrix, diagonal matrix and scaled identity matrix) and LDA with optimal-shrinkage covariance estimator using Ledoit and Wolf lemma (LDA-LW). The results indicated that LDA-LW was the most robust to noises. Moreover, LDA-LW and LDA with scaled identity matrix showed better stability and classification accuracy than the other methods. LDA-LW demonstrated the best overall performance. PMID:26405876

  7. An empirical comparison of different LDA methods in fMRI-based brain states decoding.

    PubMed

    Xia, Maogeng; Song, Sutao; Yao, Li; Long, Zhiying

    2015-01-01

    Decoding brain states from response patterns with multivariate pattern recognition techniques is a popular method for detecting multivoxel patterns of brain activation. These patterns are informative with respect to a subject's perceptual or cognitive states. Linear discriminant analysis (LDA) cannot be directly applied to fMRI data analysis because of the "few samples and large features" nature of functional magnetic resonance imaging (fMRI) data. Although several improved LDA methods have been used in fMRI-based decoding, little is known regarding the relative performance of different LDA classifiers on fMRI data. In this study, we compared five LDA classifiers using both simulated data with varied noise levels and real fMRI data. The compared LDA classifiers include LDA combined with PCA (LDA-PCA), LDA with three types of regularizations (identity matrix, diagonal matrix and scaled identity matrix) and LDA with optimal-shrinkage covariance estimator using Ledoit and Wolf lemma (LDA-LW). The results indicated that LDA-LW was the most robust to noises. Moreover, LDA-LW and LDA with scaled identity matrix showed better stability and classification accuracy than the other methods. LDA-LW demonstrated the best overall performance.

  8. Temporal asymmetries in Interbank Market: an empirically grounded Agent-Based Model

    NASA Astrophysics Data System (ADS)

    Zlatic, Vinko; Popovic, Marko; Abraham, Hrvoje; Caldarelli, Guido; Iori, Giulia

    2014-03-01

    We analyse the changes in the topology of the structure of the E-mid interbank market in the period from September 1st 1999 to September 1st 2009. We uncover a type of temporal irreversibility in the growth of the largest component of the interbank trading network, which is not common to any of the usual network growth models. Such asymmetry, which is also detected on the growth of the clustering and reciprocity coefficient, reveals that the trading mechanism is driven by different dynamics at the beginning and at the end of the day. We are able to recover the complexity of the system by means of a simple Agent Based Model in which the probability of matching between counter parties depends on a time varying vertex fitness (or attractiveness) describing banks liquidity needs. We show that temporal irreversibility is associated with heterogeneity in the banking system and emerges when the distribution of liquidity shocks across banks is broad. We acknowledge support from FET project FOC-II.

  9. Empirical modeling of the fine particle fraction for carrier-based pulmonary delivery formulations.

    PubMed

    Pacławski, Adam; Szlęk, Jakub; Lau, Raymond; Jachowicz, Renata; Mendyk, Aleksander

    2015-01-01

    In vitro study of the deposition of drug particles is commonly used during development of formulations for pulmonary delivery. The assay is demanding, complex, and depends on: properties of the drug and carrier particles, including size, surface characteristics, and shape; interactions between the drug and carrier particles and assay conditions, including flow rate, type of inhaler, and impactor. The aerodynamic properties of an aerosol are measured in vitro using impactors and in most cases are presented as the fine particle fraction, which is a mass percentage of drug particles with an aerodynamic diameter below 5 μm. In the present study, a model in the form of a mathematical equation was developed for prediction of the fine particle fraction. The feature selection was performed using the R-environment package "fscaret". The input vector was reduced from a total of 135 independent variables to 28. During the modeling stage, techniques like artificial neural networks, genetic programming, rule-based systems, and fuzzy logic systems were used. The 10-fold cross-validation technique was used to assess the generalization ability of the models created. The model obtained had good predictive ability, which was confirmed by a root-mean-square error and normalized root-mean-square error of 4.9 and 11%, respectively. Moreover, validation of the model using external experimental data was performed, and resulted in a root-mean-square error and normalized root-mean-square error of 3.8 and 8.6%, respectively.

  10. Empirically based recommendations to support parents facing the dilemma of paediatric cadaver organ donation.

    PubMed

    Bellali, T; Papazoglou, I; Papadatou, D

    2007-08-01

    The aim of the study was to describe the challenges donor and non-donor parents encounter before, during, and after the organ donation decision, and to identify parents' needs and expectations from health care professionals. A further aim was to propose evidence-based recommendations for effectively introducing the option of donation, and supporting families through the grieving process. This study was undertaken as part of a larger research project investigating the experiences of Greek parents who consented or declined organ and tissue donation, using a qualitative methodology for data collection and analysis. The experiences of 22 Greek bereaved parents of 14 underage brain dead children were studied through semi-structured interviews. Parents' decision-making process was described as challenging and fraught with difficulties both before and after the donation period. Identified challenges were clustered into: (a) personal challenges, (b) conditions of organ request, and (c) interpersonal challenges. Parents' main concern following donation was the lack of information about transplantation outcomes. Findings led to a list of recommendations for nurses and other health professionals for approaching and supporting parents in making choices about paediatric organ donation that are appropriate to them, and for facilitating their adjustment to the sudden death of their underage child. PMID:17475498

  11. An empirical approach to selecting community-based alcohol interventions: combining research evidence, rural community views and professional opinion

    PubMed Central

    2012-01-01

    Background Given limited research evidence for community-based alcohol interventions, this study examines the intervention preferences of rural communities and alcohol professionals, and factors that influence their choices. Method Community preferences were identified by a survey of randomly selected individuals across 20 regional Australian communities. The preferences of alcohol professionals were identified by a survey of randomly selected members of the Australasian Professional Society on Alcohol and Other Drugs. To identify preferred interventions and the extent of support for them, a budget allocation exercise was embedded in both surveys, asking respondents to allocate a given budget to different interventions. Tobit regression models were estimated to identify the characteristics that explain differences in intervention preferences. Results Community respondents selected school programs most often (88.0%) and allocated it the largest proportion of funds, followed by promotion of safer drinking (71.3%), community programs (61.4%) and police enforcement of alcohol laws (60.4%). Professionals selected GP training most often (61.0%) and allocated it the largest proportion of funds, followed by school programs (36.6%), community programs (33.8%) and promotion of safer drinking (31.7%). Community views were susceptible to response bias. There were no significant predictors of professionals' preferences. Conclusions In the absence of sufficient research evidence for effective community-based alcohol interventions, rural communities and professionals both strongly support school programs, promotion of safer drinking and community programs. Rural communities also supported police enforcement of alcohol laws and professionals supported GP training. The impact of a combination of these strategies needs to be rigorously evaluated. PMID:22233608

  12. An empirical approach to predicting long term behavior of metal particle based recording media

    NASA Technical Reports Server (NTRS)

    Hadad, Allan S.

    1991-01-01

    Alpha iron particles used for magnetic recording are prepared through a series of dehydration and reduction steps of alpha-Fe2O3-H2O resulting in acicular, polycrystalline, body centered cubic (bcc) alpha-Fe particles that are single magnetic domains. Since fine iron particles are pyrophoric by nature, stabilization processes had to be developed in order for iron particles to be considered as a viable recording medium for long term archival (i.e., 25+ years) information storage. The primary means of establishing stability is through passivation or controlled oxidation of the iron particle's surface. Since iron particles used for magnetic recording are small, additional oxidation has a direct impact on performance especially where archival storage of recorded information for long periods of time is important. Further stabilization chemistry/processes had to be developed to guarantee that iron particles could be considered as a viable long term recording medium. In an effort to retard the diffusion of iron ions through the oxide layer, other elements such as silicon, aluminum, and chromium have been added to the base iron to promote more dense scale formation or to alleviate some of the non-stoichiometric behavior of the oxide or both. The presence of water vapor has been shown to disrupt the passive layer, subsequently increasing the oxidation rate of the iron. A study was undertaken to examine the degradation in magnetic properties as a function of both temperature and humidity on silicon-containing iron particles between 50-120 deg C and 3-89 percent relative humidity. The methodology to which experimental data was collected and analyzed leading to predictive capability is discussed.

  13. The ethics of feedback of HIV test results in population-based surveys of HIV infection.

    PubMed

    Maher, Dermot

    2013-12-01

    Population-based disease prevalence surveys raise ethical questions, including whether participants should be routinely told their test results. Ethical guidelines call for informing survey participants of any clinically relevant finding to enable appropriate management. However, in anonymous surveys of human immunodeficiency virus (HIV) infection, participants can "opt out" of being given their test results or are offered the chance to undergo voluntary HIV testing in local counselling and testing services. This is aimed at minimizing survey participation bias. Those who opt out of being given their HIV test results and who do not seek their results miss the opportunity to receive life-saving antiretroviral therapy. The justification for HIV surveys without routine feedback of results to participants is based on a public health utility argument: that the benefits of more rigorous survey methods - reduced participation bias - outweigh the benefits to individuals of knowing their HIV status. However, people with HIV infection have a strong immediate interest in knowing their HIV status. In consideration of the ethical value of showing respect for people and thereby alleviating suffering, an argument based on public health utility is not an appropriate justification. In anonymous HIV surveys as well as other prevalence surveys of treatable conditions in any setting, participation should be on the basis of routine individual feedback of results as an integral part of fully informed participation. Ensuring that surveys are ethically sound may stimulate participation, increase a broader uptake of HIV testing and reduce stigmatization of people who are HIV-positive.

  14. GIS-based analysis and modelling with empirical and remotely-sensed data on coastline advance and retreat

    NASA Astrophysics Data System (ADS)

    Ahmad, Sajid Rashid

    With the understanding that far more research remains to be done on the development and use of innovative and functional geospatial techniques and procedures to investigate coastline changes this thesis focussed on the integration of remote sensing, geographical information systems (GIS) and modelling techniques to provide meaningful insights on the spatial and temporal dynamics of coastline changes. One of the unique strengths of this research was the parameterization of the GIS with long-term empirical and remote sensing data. Annual empirical data from 1941--2007 were analyzed by the GIS, and then modelled with statistical techniques. Data were also extracted from Landsat TM and ETM+ images. The band ratio method was used to extract the coastlines. Topographic maps were also used to extract digital map data. All data incorporated into ArcGIS 9.2 were analyzed with various modules, including Spatial Analyst, 3D Analyst, and Triangulated Irregular Networks. The Digital Shoreline Analysis System was used to analyze and predict rates of coastline change. GIS results showed the spatial locations along the coast that will either advance or retreat over time. The linear regression results highlighted temporal changes which are likely to occur along the coastline. Box-Jenkins modelling procedures were utilized to determine statistical models which best described the time series (1941--2007) of coastline change data. After several iterations and goodness-of-fit tests, second-order spatial cyclic autoregressive models, first-order autoregressive models and autoregressive moving average models were identified as being appropriate for describing the deterministic and random processes operating in Guyana's coastal system. The models highlighted not only cyclical patterns in advance and retreat of the coastline, but also the existence of short and long-term memory processes. Long-term memory processes could be associated with mudshoal propagation and stabilization while short

  15. Lessons learned in the conduct, validation, and interpretation of national population based HIV surveys.

    PubMed

    Calleja, Jesús M García; Marum, Lawrence H; Cárcamo, César P; Kaetano, Lovemore; Muttunga, James; Way, Ann

    2005-05-01

    In the past few years several countries have conducted national population-based HIV surveys. Survey methods, levels of participation bias from absence or refusal and lessons learned conducting such surveys are compared in four national population surveys: Mali, Kenya, Peru and Zambia. In Mali, Zambia, and Kenya, HIV testing of adult women and men was included in the national-level demographic and health surveys carried out regularly in these countries, whereas in Peru the national HIV survey targeted young people in 24 cities with populations over 50 000.The household response rate was above 90% in all countries, but some individuals were absent for interviews. HIV testing rates were between 70 and 79% of those eligible, with higher test rates for women. Three critical questions in this type of survey need to be answered: who did the surveys miss; how much it matters that they were missed; and what can be done to increase the participation of respondents so the coverage rates are adequate. The level of representativeness of the populations tested was adequate in each survey to provide a reliable national estimate of HIV prevalence that complements other methods of HIV surveillance. Different lessons were learned from each survey. These population-based HIV seroprevalence surveys demonstrate that reliable and useful results can be obtained, although they require careful planning and increased financial and human resource investment to maximize responses at the household and individual level, which are key elements to validate survey results.This review was initiated through an international meeting on 'New strategies for HIV/AIDS Surveillance in Resource-constrained Countries' held in Addis Ababa on 26-30 January 2004 to share and develop recommendations to guide future surveys. PMID:15930844

  16. Intelligence in Bali--A Case Study on Estimating Mean IQ for a Population Using Various Corrections Based on Theory and Empirical Findings

    ERIC Educational Resources Information Center

    Rindermann, Heiner; te Nijenhuis, Jan

    2012-01-01

    A high-quality estimate of the mean IQ of a country requires giving a well-validated test to a nationally representative sample, which usually is not feasible in developing countries. So, we used a convenience sample and four corrections based on theory and empirical findings to arrive at a good-quality estimate of the mean IQ in Bali. Our study…

  17. Demonstrating the Potential for Web-Based Survey Methodology with a Case Study.

    ERIC Educational Resources Information Center

    Mertler, Craig

    2002-01-01

    Describes personal experience with using the Internet to administer a teacher-motivation and job-satisfaction survey to elementary and secondary teachers. Concludes that advantages of Web-base surveys, such as cost savings and efficiency of data collection, outweigh disadvantages, such as the limitations of listservs. (Contains 10 references.)…

  18. Measuring Model-Based High School Science Instruction: Development and Application of a Student Survey

    ERIC Educational Resources Information Center

    Fulmer, Gavin W.; Liang, Ling L.

    2013-01-01

    This study tested a student survey to detect differences in instruction between teachers in a modeling-based science program and comparison group teachers. The Instructional Activities Survey measured teachers' frequency of modeling, inquiry, and lecture instruction. Factor analysis and Rasch modeling identified three subscales, Modeling and…

  19. Libraries and Desktop Storage Options: Results of a Web-Based Survey.

    ERIC Educational Resources Information Center

    Hendricks, Arthur; Wang, Jian

    2002-01-01

    Reports the results of a Web-based survey that investigated what plans, if any, librarians have for dealing with the expected obsolescence of the floppy disk and still retain effective library service. Highlights include data storage options, including compact disks, zip disks, and networked storage products; and a copy of the Web survey.…

  20. The Mood of American Youth. Based on a 1983 Survey of American Youth.

    ERIC Educational Resources Information Center

    National Association of Secondary School Principals, Reston, VA.

    This 1984 report by the National Association of Secondary School Principals is based on a survey of 1,500 students from grades 7 through 12 selected by means of a probability sample by National Family Opinions, Inc. The report compares today's students with those surveyed in a similar study in 1974. Chapter 1, "Students and Their Schools,"…

  1. Motion Trajectories for Wide-area Surveying with a Rover-based Distributed Spectrometer

    NASA Technical Reports Server (NTRS)

    Tunstel, Edward; Anderson, Gary; Wilson, Edmond

    2006-01-01

    A mobile ground survey application that employs remote sensing as a primary means of area coverage is highlighted. It is distinguished from mobile robotic area coverage problems that employ contact or proximity-based sensing. The focus is on a specific concept for performing mobile surveys in search of biogenic gases on planetary surfaces using a distributed spectrometer -- a rover-based instrument designed for wide measurement coverage of promising search areas. Navigation algorithms for executing circular and spiral survey trajectories are presented for widearea distributed spectroscopy and evaluated based on area covered and distance traveled.

  2. Predicting Student Performance in Web-Based Distance Education Courses Based on Survey Instruments Measuring Personality Traits and Technical Skills

    ERIC Educational Resources Information Center

    Hall, Michael

    2008-01-01

    Two common web-based surveys, "Is Online Learning Right for Me?' and "What Technical Skills Do I Need?", were combined into a single survey instrument and given to 228 on-campus and 83 distance education students. The students were enrolled in four different classes (business, computer information services, criminal justice, and…

  3. The dappled nature of causes of psychiatric illness: replacing the organic–functional/hardware–software dichotomy with empirically based pluralism

    PubMed Central

    Kendler, KS

    2012-01-01

    Our tendency to see the world of psychiatric illness in dichotomous and opposing terms has three major sources: the philosophy of Descartes, the state of neuropathology in late nineteenth century Europe (when disorders were divided into those with and without demonstrable pathology and labeled, respectively, organic and functional), and the influential concept of computer functionalism wherein the computer is viewed as a model for the human mind–brain system (brain = hardware, mind = software). These mutually re-enforcing dichotomies, which have had a pernicious influence on our field, make a clear prediction about how ‘difference-makers’ (aka causal risk factors) for psychiatric disorders should be distributed in nature. In particular, are psychiatric disorders like our laptops, which when they dysfunction, can be cleanly divided into those with software versus hardware problems? I propose 11 categories of difference-makers for psychiatric illness from molecular genetics through culture and review their distribution in schizophrenia, major depression and alcohol dependence. In no case do these distributions resemble that predicted by the organic–functional/hardware–software dichotomy. Instead, the causes of psychiatric illness are dappled, distributed widely across multiple categories. We should abandon Cartesian and computer-functionalism-based dichotomies as scientifically inadequate and an impediment to our ability to integrate the diverse information about psychiatric illness our research has produced. Empirically based pluralism provides a rigorous but dappled view of the etiology of psychiatric illness. Critically, it is based not on how we wish the world to be but how the difference-makers for psychiatric illness are in fact distributed. PMID:22230881

  4. The dappled nature of causes of psychiatric illness: replacing the organic-functional/hardware-software dichotomy with empirically based pluralism.

    PubMed

    Kendler, K S

    2012-04-01

    Our tendency to see the world of psychiatric illness in dichotomous and opposing terms has three major sources: the philosophy of Descartes, the state of neuropathology in late nineteenth century Europe (when disorders were divided into those with and without demonstrable pathology and labeled, respectively, organic and functional), and the influential concept of computer functionalism wherein the computer is viewed as a model for the human mind-brain system (brain=hardware, mind=software). These mutually re-enforcing dichotomies, which have had a pernicious influence on our field, make a clear prediction about how 'difference-makers' (aka causal risk factors) for psychiatric disorders should be distributed in nature. In particular, are psychiatric disorders like our laptops, which when they dysfunction, can be cleanly divided into those with software versus hardware problems? I propose 11 categories of difference-makers for psychiatric illness from molecular genetics through culture and review their distribution in schizophrenia, major depression and alcohol dependence. In no case do these distributions resemble that predicted by the organic-functional/hardware-software dichotomy. Instead, the causes of psychiatric illness are dappled, distributed widely across multiple categories. We should abandon Cartesian and computer-functionalism-based dichotomies as scientifically inadequate and an impediment to our ability to integrate the diverse information about psychiatric illness our research has produced. Empirically based pluralism provides a rigorous but dappled view of the etiology of psychiatric illness. Critically, it is based not on how we wish the world to be but how the difference-makers for psychiatric illness are in fact distributed.

  5. In silico structure-based screening of versatile P-glycoprotein inhibitors using polynomial empirical scoring functions.

    PubMed

    Shityakov, Sergey; Förster, Carola

    2014-01-01

    P-glycoprotein (P-gp) is an ATP (adenosine triphosphate)-binding cassette transporter that causes multidrug resistance of various chemotherapeutic substances by active efflux from mammalian cells. P-gp plays a pivotal role in limiting drug absorption and distribution in different organs, including the intestines and brain. Thus, the prediction of P-gp-drug interactions is of vital importance in assessing drug pharmacokinetic and pharmacodynamic properties. To find the strongest P-gp blockers, we performed an in silico structure-based screening of P-gp inhibitor library (1,300 molecules) by the gradient optimization method, using polynomial empirical scoring (POLSCORE) functions. We report a strong correlation (r (2)=0.80, F=16.27, n=6, P<0.0157) of inhibition constants (Kiexp or pKiexp; experimental Ki or negative decimal logarithm of Kiexp) converted from experimental IC50 (half maximal inhibitory concentration) values with POLSCORE-predicted constants (KiPOLSCORE or pKiPOLSCORE), using a linear regression fitting technique. The hydrophobic interactions between P-gp and selected drug substances were detected as the main forces responsible for the inhibition effect. The results showed that this scoring technique might be useful in the virtual screening and filtering of databases of drug-like compounds at the early stage of drug development processes. PMID:24711707

  6. Combined magnetic and kinetic control of advanced tokamak steady state scenarios based on semi-empirical modelling

    NASA Astrophysics Data System (ADS)

    Moreau, D.; Artaud, J. F.; Ferron, J. R.; Holcomb, C. T.; Humphreys, D. A.; Liu, F.; Luce, T. C.; Park, J. M.; Prater, R.; Turco, F.; Walker, M. L.

    2015-06-01

    This paper shows that semi-empirical data-driven models based on a two-time-scale approximation for the magnetic and kinetic control of advanced tokamak (AT) scenarios can be advantageously identified from simulated rather than real data, and used for control design. The method is applied to the combined control of the safety factor profile, q(x), and normalized pressure parameter, βN, using DIII-D parameters and actuators (on-axis co-current neutral beam injection (NBI) power, off-axis co-current NBI power, electron cyclotron current drive power, and ohmic coil). The approximate plasma response model was identified from simulated open-loop data obtained using a rapidly converging plasma transport code, METIS, which includes an MHD equilibrium and current diffusion solver, and combines plasma transport nonlinearity with 0D scaling laws and 1.5D ordinary differential equations. The paper discusses the results of closed-loop METIS simulations, using the near-optimal ARTAEMIS control algorithm (Moreau D et al 2013 Nucl. Fusion 53 063020) for steady state AT operation. With feedforward plus feedback control, the steady state target q-profile and βN are satisfactorily tracked with a time scale of about 10 s, despite large disturbances applied to the feedforward powers and plasma parameters. The robustness of the control algorithm with respect to disturbances of the H&CD actuators and of plasma parameters such as the H-factor, plasma density and effective charge, is also shown.

  7. In silico structure-based screening of versatile P-glycoprotein inhibitors using polynomial empirical scoring functions.

    PubMed

    Shityakov, Sergey; Förster, Carola

    2014-01-01

    P-glycoprotein (P-gp) is an ATP (adenosine triphosphate)-binding cassette transporter that causes multidrug resistance of various chemotherapeutic substances by active efflux from mammalian cells. P-gp plays a pivotal role in limiting drug absorption and distribution in different organs, including the intestines and brain. Thus, the prediction of P-gp-drug interactions is of vital importance in assessing drug pharmacokinetic and pharmacodynamic properties. To find the strongest P-gp blockers, we performed an in silico structure-based screening of P-gp inhibitor library (1,300 molecules) by the gradient optimization method, using polynomial empirical scoring (POLSCORE) functions. We report a strong correlation (r (2)=0.80, F=16.27, n=6, P<0.0157) of inhibition constants (Kiexp or pKiexp; experimental Ki or negative decimal logarithm of Kiexp) converted from experimental IC50 (half maximal inhibitory concentration) values with POLSCORE-predicted constants (KiPOLSCORE or pKiPOLSCORE), using a linear regression fitting technique. The hydrophobic interactions between P-gp and selected drug substances were detected as the main forces responsible for the inhibition effect. The results showed that this scoring technique might be useful in the virtual screening and filtering of databases of drug-like compounds at the early stage of drug development processes.

  8. Multi-fault diagnosis for rolling element bearings based on ensemble empirical mode decomposition and optimized support vector machines

    NASA Astrophysics Data System (ADS)

    Zhang, Xiaoyuan; Zhou, Jianzhong

    2013-12-01

    This study presents a novel procedure based on ensemble empirical mode decomposition (EEMD) and optimized support vector machine (SVM) for multi-fault diagnosis of rolling element bearings. The vibration signal is adaptively decomposed into a number of intrinsic mode functions (IMFs) by EEMD. Two types of features, the EEMD energy entropy and singular values of the matrix whose rows are IMFs, are extracted. EEMD energy entropy is used to specify whether the bearing has faults or not. If the bearing has faults, singular values are input to multi-class SVM optimized by inter-cluster distance in the feature space (ICDSVM) to specify the fault type. The proposed method was tested on a system with an electric motor which has two rolling bearings with 8 normal working conditions and 48 fault working conditions. Five groups of experiments were done to evaluate the effectiveness of the proposed method. The results show that the proposed method outperforms other methods both mentioned in this paper and published in other literatures.

  9. Empirically-Based Crop Insurance for China: A Pilot Study in the Down-middle Yangtze River Area of China

    NASA Astrophysics Data System (ADS)

    Wang, Erda; Yu, Yang; Little, Bertis B.; Chen, Zhongxin; Ren, Jianqiang

    Factors that caused slow growth in crop insurance participation and its ultimate failure in China were multi-faceted including high agricultural production risk, low participation rate, inadequate public awareness, high loss ratio, insufficient and interrupted government financial support. Thus, a clear and present need for data driven analyses and empirically-based risk management exists in China. In the present investigation, agricultural production data for two crops (corn, rice) in five counties in Jiangxi Province and Hunan province for design of a pilot crop insurance program in China. A crop insurance program was designed which (1) provides 75% coverage, (2) a 55% premium rate reduction for the farmer compared to catastrophic coverage most recently offered, and uses the currently approved governmental premium subsidy level. Thus a safety net for Chinese farmers that help maintain agricultural production at a level of self-sufficiency that costs less than half the current plans requires one change to the program: ≥80% of producers must participate in an area.

  10. Empirically Based Profiles of the Early Literacy Skills of Children With Language Impairment in Early Childhood Special Education.

    PubMed

    Justice, Laura; Logan, Jessica; Kaderavek, Joan; Schmitt, Mary Beth; Tompkins, Virginia; Bartlett, Christopher

    2015-01-01

    The purpose of this study was to empirically determine whether specific profiles characterize preschool-aged children with language impairment (LI) with respect to their early literacy skills (print awareness, name-writing ability, phonological awareness, alphabet knowledge); the primary interest was to determine if one or more profiles suggested vulnerability for future reading problems. Participants were 218 children enrolled in early childhood special education classrooms, 95% of whom received speech-language services. Children were administered an assessment of early literacy skills in the fall of the academic year. Based on results of latent profile analysis, four distinct literacy profiles were identified, with the single largest profile (55% of children) representing children with generally poor literacy skills across all areas examined. Children in the two low-risk categories had higher oral language skills than those in the high-risk and moderate-risk profiles. Across three of the four early literacy measures, children with language as their primary disability had higher scores than those with LI concomitant with other disabilities. These findings indicate that there are specific profiles of early literacy skills among children with LI, with about one half of children exhibiting a profile indicating potential susceptibility for future reading problems.

  11. An improved empirical model of electron and ion fluxes at geosynchronous orbit based on upstream solar wind conditions

    NASA Astrophysics Data System (ADS)

    Denton, M. H.; Henderson, M. G.; Jordanova, V. K.; Thomsen, M. F.; Borovsky, J. E.; Woodroffe, J.; Hartley, D. P.; Pitchford, D.

    2016-07-01

    A new empirical model of the electron fluxes and ion fluxes at geosynchronous orbit (GEO) is introduced, based on observations by Los Alamos National Laboratory (LANL) satellites. The model provides flux predictions in the energy range ~1 eV to ~40 keV, as a function of local time, energy, and the strength of the solar wind electric field (the negative product of the solar wind speed and the z component of the magnetic field). Given appropriate upstream solar wind measurements, the model provides a forecast of the fluxes at GEO with a ~1 h lead time. Model predictions are tested against in-sample observations from LANL satellites and also against out-of-sample observations from the Compact Environmental Anomaly Sensor II detector on the AMC-12 satellite. The model does not reproduce all structure seen in the observations. However, for the intervals studied here (quiet and storm times) the normalized root-mean-square deviation < ~0.3. It is intended that the model will improve forecasting of the spacecraft environment at GEO and also provide improved boundary/input conditions for physical models of the magnetosphere.

  12. In silico structure-based screening of versatile P-glycoprotein inhibitors using polynomial empirical scoring functions

    PubMed Central

    Shityakov, Sergey; Förster, Carola

    2014-01-01

    P-glycoprotein (P-gp) is an ATP (adenosine triphosphate)-binding cassette transporter that causes multidrug resistance of various chemotherapeutic substances by active efflux from mammalian cells. P-gp plays a pivotal role in limiting drug absorption and distribution in different organs, including the intestines and brain. Thus, the prediction of P-gp–drug interactions is of vital importance in assessing drug pharmacokinetic and pharmacodynamic properties. To find the strongest P-gp blockers, we performed an in silico structure-based screening of P-gp inhibitor library (1,300 molecules) by the gradient optimization method, using polynomial empirical scoring (POLSCORE) functions. We report a strong correlation (r2=0.80, F=16.27, n=6, P<0.0157) of inhibition constants (Kiexp or pKiexp; experimental Ki or negative decimal logarithm of Kiexp) converted from experimental IC50 (half maximal inhibitory concentration) values with POLSCORE-predicted constants (KiPOLSCORE or pKiPOLSCORE), using a linear regression fitting technique. The hydrophobic interactions between P-gp and selected drug substances were detected as the main forces responsible for the inhibition effect. The results showed that this scoring technique might be useful in the virtual screening and filtering of databases of drug-like compounds at the early stage of drug development processes. PMID:24711707

  13. An improved empirical model of electron and ion fluxes at geosynchronous orbit based on upstream solar wind conditions

    DOE PAGES

    Denton, M. H.; Henderson, M. G.; Jordanova, V. K.; Thomsen, M. F.; Borovsky, J. E.; Woodroffe, J.; Hartley, D. P.; Pitchford, D.

    2016-07-27

    In this study, a new empirical model of the electron fluxes and ion fluxes at geosynchronous orbit (GEO) is introduced, based on observations by Los Alamos National Laboratory (LANL) satellites. The model provides flux predictions in the energy range ~1 eV to ~40 keV, as a function of local time, energy, and the strength of the solar wind electric field (the negative product of the solar wind speed and the z component of the magnetic field). Given appropriate upstream solar wind measurements, the model provides a forecast of the fluxes at GEO with a ~1 h lead time. Model predictionsmore » are tested against in-sample observations from LANL satellites and also against out-of-sample observations from the Compact Environmental Anomaly Sensor II detector on the AMC-12 satellite. The model does not reproduce all structure seen in the observations. However, for the intervals studied here (quiet and storm times) the normalized root-mean-square deviation < ~0.3. It is intended that the model will improve forecasting of the spacecraft environment at GEO and also provide improved boundary/input conditions for physical models of the magnetosphere.« less

  14. Empirically Based Profiles of the Early Literacy Skills of Children With Language Impairment in Early Childhood Special Education.

    PubMed

    Justice, Laura; Logan, Jessica; Kaderavek, Joan; Schmitt, Mary Beth; Tompkins, Virginia; Bartlett, Christopher

    2015-01-01

    The purpose of this study was to empirically determine whether specific profiles characterize preschool-aged children with language impairment (LI) with respect to their early literacy skills (print awareness, name-writing ability, phonological awareness, alphabet knowledge); the primary interest was to determine if one or more profiles suggested vulnerability for future reading problems. Participants were 218 children enrolled in early childhood special education classrooms, 95% of whom received speech-language services. Children were administered an assessment of early literacy skills in the fall of the academic year. Based on results of latent profile analysis, four distinct literacy profiles were identified, with the single largest profile (55% of children) representing children with generally poor literacy skills across all areas examined. Children in the two low-risk categories had higher oral language skills than those in the high-risk and moderate-risk profiles. Across three of the four early literacy measures, children with language as their primary disability had higher scores than those with LI concomitant with other disabilities. These findings indicate that there are specific profiles of early literacy skills among children with LI, with about one half of children exhibiting a profile indicating potential susceptibility for future reading problems. PMID:24232733

  15. A survey of surveys

    SciTech Connect

    Kent, S.M.

    1994-11-01

    A new era for the field of Galactic structure is about to be opened with the advent of wide-area digital sky surveys. In this article, the author reviews the status and prospects for research for 3 new ground-based surveys: the Sloan Digital Sky Survey (SDSS), the Deep Near-Infrared Survey of the Southern Sky (DENIS) and the Two Micron AU Sky Survey (2MASS). These surveys will permit detailed studies of Galactic structure and stellar populations in the Galaxy with unprecedented detail. Extracting the information, however, will be challenging.

  16. Empirical characteristics of family-based linkage to a complex trait: the ADIPOQ region and adiponectin levels.

    PubMed

    Hellwege, Jacklyn N; Palmer, Nicholette D; Mark Brown, W; Brown, Mark W; Ziegler, Julie T; Sandy An, S; An, Sandy S; Guo, Xiuqing; Ida Chen, Y-D; Chen, Ida Y-D; Taylor, Kent; Hawkins, Gregory A; Ng, Maggie C Y; Speliotes, Elizabeth K; Lorenzo, Carlos; Norris, Jill M; Rotter, Jerome I; Wagenknecht, Lynne E; Langefeld, Carl D; Bowden, Donald W

    2015-02-01

    We previously identified a low-frequency (1.1 %) coding variant (G45R; rs200573126) in the adiponectin gene (ADIPOQ) which was the basis for a multipoint microsatellite linkage signal (LOD = 8.2) for plasma adiponectin levels in Hispanic families. We have empirically evaluated the ability of data from targeted common variants, exome chip genotyping, and genome-wide association study data to detect linkage and association to adiponectin protein levels at this locus. Simple two-point linkage and association analyses were performed in 88 Hispanic families (1,150 individuals) using 10,958 SNPs on chromosome 3. Approaches were compared for their ability to map the functional variant, G45R, which was strongly linked (two-point LOD = 20.98) and powerfully associated (p value = 8.1 × 10(-50)). Over 450 SNPs within a broad 61 Mb interval around rs200573126 showed nominal evidence of linkage (LOD > 3) but only four other SNPs in this region were associated with p values < 1.0 × 10(-4). When G45R was accounted for, the maximum LOD score across the interval dropped to 4.39 and the best p value was 1.1 × 10(-5). Linked and/or associated variants ranged in frequency (0.0018-0.50) and type (coding, non-coding) and had little detectable linkage disequilibrium with rs200573126 (r (2) < 0.20). In addition, the two-point linkage approach empirically outperformed multipoint microsatellite and multipoint SNP analysis. In the absence of data for rs200573126, family-based linkage analysis using a moderately dense SNP dataset, including both common and low-frequency variants, resulted in stronger evidence for an adiponectin locus than association data alone. Thus, linkage analysis can be a useful tool to facilitate identification of high-impact genetic variants.

  17. Testing an astronomically based decadal-scale empirical harmonic climate model versus the IPCC (2007) general circulation climate models

    NASA Astrophysics Data System (ADS)

    Scafetta, Nicola

    2012-05-01

    We compare the performance of a recently proposed empirical climate model based on astronomical harmonics against all CMIP3 available general circulation climate models (GCM) used by the IPCC (2007) to interpret the 20th century global surface temperature. The proposed astronomical empirical climate model assumes that the climate is resonating with, or synchronized to a set of natural harmonics that, in previous works (Scafetta, 2010b, 2011b), have been associated to the solar system planetary motion, which is mostly determined by Jupiter and Saturn. We show that the GCMs fail to reproduce the major decadal and multidecadal oscillations found in the global surface temperature record from 1850 to 2011. On the contrary, the proposed harmonic model (which herein uses cycles with 9.1, 10-10.5, 20-21, 60-62 year periods) is found to well reconstruct the observed climate oscillations from 1850 to 2011, and it is shown to be able to forecast the climate oscillations from 1950 to 2011 using the data covering the period 1850-1950, and vice versa. The 9.1-year cycle is shown to be likely related to a decadal Soli/Lunar tidal oscillation, while the 10-10.5, 20-21 and 60-62 year cycles are synchronous to solar and heliospheric planetary oscillations. We show that the IPCC GCM's claim that all warming observed from 1970 to 2000 has been anthropogenically induced is erroneous because of the GCM failure in reconstructing the quasi 20-year and 60-year climatic cycles. Finally, we show how the presence of these large natural cycles can be used to correct the IPCC projected anthropogenic warming trend for the 21st century. By combining this corrected trend with the natural cycles, we show that the temperature may not significantly increase during the next 30 years mostly because of the negative phase of the 60-year cycle. If multisecular natural cycles (which according to some authors have significantly contributed to the observed 1700-2010 warming and may contribute to an

  18. Discussion on climate oscillations: CMIP5 general circulation models versus a semi-empirical harmonic model based on astronomical cycles

    NASA Astrophysics Data System (ADS)

    Scafetta, Nicola

    2013-11-01

    Power spectra of global surface temperature (GST) records (available since 1850) reveal major periodicities at about 9.1, 10-11, 19-22 and 59-62 years. Equivalent oscillations are found in numerous multisecular paleoclimatic records. The Coupled Model Intercomparison Project 5 (CMIP5) general circulation models (GCMs), to be used in the IPCC Fifth Assessment Report (AR5, 2013), are analyzed and found not able to reconstruct this variability. In particular, from 2000 to 2013.5 a GST plateau is observed while the GCMs predicted a warming rate of about 2 °C/century. In contrast, the hypothesis that the climate is regulated by specific natural oscillations more accurately fits the GST records at multiple time scales. For example, a quasi 60-year natural oscillation simultaneously explains the 1850-1880, 1910-1940 and 1970-2000 warming periods, the 1880-1910 and 1940-1970 cooling periods and the post 2000 GST plateau. This hypothesis implies that about 50% of the ~ 0.5 °C global surface warming observed from 1970 to 2000 was due to natural oscillations of the climate system, not to anthropogenic forcing as modeled by the CMIP3 and CMIP5 GCMs. Consequently, the climate sensitivity to CO2 doubling should be reduced by half, for example from the 2.0-4.5 °C range (as claimed by the IPCC, 2007) to 1.0-2.3 °C with a likely median of ~ 1.5 °C instead of ~ 3.0 °C. Also modern paleoclimatic temperature reconstructions showing a larger preindustrial variability than the hockey-stick shaped temperature reconstructions developed in early 2000 imply a weaker anthropogenic effect and a stronger solar contribution to climatic changes. The observed natural oscillations could be driven by astronomical forcings. The ~ 9.1 year oscillation appears to be a combination of long soli-lunar tidal oscillations, while quasi 10-11, 20 and 60 year oscillations are typically found among major solar and heliospheric oscillations driven mostly by Jupiter and Saturn movements. Solar models based

  19. Public Release of Estimated Impact-Based Earthquake Alerts - An Update to the U.S. Geological Survey PAGER System

    NASA Astrophysics Data System (ADS)

    Wald, D. J.; Jaiswal, K. S.; Marano, K.; Hearne, M.; Earle, P. S.; So, E.; Garcia, D.; Hayes, G. P.; Mathias, S.; Applegate, D.; Bausch, D.

    2010-12-01

    The U.S. Geological Survey (USGS) has begun publicly releasing earthquake alerts for significant earthquakes around the globe based on estimates of potential casualties and economic losses. These estimates should significantly enhance the utility of the USGS Prompt Assessment of Global Earthquakes for Response (PAGER) system that has been providing estimated ShakeMaps and computing population exposures to specific shaking intensities since 2007. Quantifying earthquake impacts and communicating loss estimates (and their uncertainties) to the public has been the culmination of several important new and evolving components of the system. First, the operational PAGER system now relies on empirically-based loss models that account for estimated shaking hazard, population exposure, and employ country-specific fatality and economic loss functions derived using analyses of losses due to recent and past earthquakes. In some countries, our empirical loss models are informed in part by PAGER’s semi-empirical and analytical loss models, and building exposure and vulnerability data sets, all of which are being developed in parallel to the empirical approach. Second, human and economic loss information is now portrayed as a supplement to existing intensity/exposure content on both PAGER summary alert (available via cell phone/email) messages and web pages. Loss calculations also include estimates of the economic impact with respect to the country’s gross domestic product. Third, in order to facilitate rapid and appropriate earthquake responses based on our probable loss estimates, in early 2010 we proposed a four-level Earthquake Impact Scale (EIS). Instead of simply issuing median estimates for losses—which can be easily misunderstood and misused—this scale provides ranges of losses from which potential responders can gauge expected overall impact from strong shaking. EIS is based on two complementary criteria: the estimated cost of damage, which is most suitable for U

  20. Metamodels for Computer-Based Engineering Design: Survey and Recommendations

    NASA Technical Reports Server (NTRS)

    Simpson, Timothy W.; Peplinski, Jesse; Koch, Patrick N.; Allen, Janet K.

    1997-01-01

    The use of statistical techniques to build approximations of expensive computer analysis codes pervades much of todays engineering design. These statistical approximations, or metamodels, are used to replace the actual expensive computer analyses, facilitating multidisciplinary, multiobjective optimization and concept exploration. In this paper we review several of these techniques including design of experiments, response surface methodology, Taguchi methods, neural networks, inductive learning, and kriging. We survey their existing application in engineering design and then address the dangers of applying traditional statistical techniques to approximate deterministic computer analysis codes. We conclude with recommendations for the appropriate use of statistical approximation techniques in given situations and how common pitfalls can be avoided.

  1. South African maize production scenarios for 2055 using a combined empirical and process-based model approach

    NASA Astrophysics Data System (ADS)

    Estes, L.; Bradley, B.; Oppenheimer, M.; Wilcove, D.; Beukes, H.; Schulze, R. E.; Tadross, M.

    2011-12-01

    In South Africa, a semi-arid country with a diverse agricultural sector, climate change is projected to negatively impact staple crop production. Our study examines future impacts to maize, South Africa's most widely grown staple crop. Working at finer spatial resolution than previous studies, we combine the process-based DSSAT4.5 and the empirical MAXENT models to study future maize suitability. Climate scenarios were based on 9 GCMs run under SRES A2 and B1 emissions scenarios down-scaled (using self-organizing maps) to 5838 locations. Soil properties were derived from textural and compositional data linked to 26422 landforms. DSSAT was run with typical dryland planting parameters and mean projected CO2 values. MAXENT was trained using aircraft-observed distributions and monthly climatologies data derived from downscaled daily records, with future rainfall increased by 10% to simulate CO2 related water-use efficiency gains. We assessed model accuracy based on correlations between model output and a satellite-derived yield proxy (integrated NDVI), and the overlap of modeled and observed maize field distributions. DSSAT yields were linearly correlated to mean integrated NDVI (R2 = 0.38), while MAXENT's relationship was logistic. Binary suitability maps based on thresholding model outputs were slightly more accurate for MAXENT (88%) than for DSSAT (87%) when compared to current maize field distribution. We created 18 suitability maps for each model (9 GCMs X 2 SRES) using projected changes relative to historical suitability thresholds. Future maps largely agreed in eastern South Africa, but disagreed strongly in the semi-arid west. Using a 95% confidence criterion (17 models agree), MAXENT showed a 241305 km2 suitability loss relative to its modeled historical suitability, while DSSAT showed a potential loss of only 112446 km2. Even the smaller potential loss highlighted by DSSAT is uncertain, given that DSSAT's mean (across all 18 climate scenarios) projected yield

  2. Monte Carlo study for physiological interference reduction in near-infrared spectroscopy based on empirical mode decomposition

    NASA Astrophysics Data System (ADS)

    Zhang, Yan; Sun, JinWei; Rolfe, Peter

    2010-12-01

    Near-infrared spectroscopy (NIRS) can be used as the basis of non-invasive neuroimaging that may allow the measurement of haemodynamic changes in the human brain evoked by applied stimuli. Since this technique is very sensitive, physiological interference arising from the cardiac cycle and breathing can significantly affect the signal quality. Such interference is difficult to remove by conventional techniques because it occurs not only in the extracerebral layer but also in the brain tissue itself. Previous work on this problem employing temporal filtering, spatial filtering, and adaptive filtering have exhibited good performance for recovering brain activity data in evoked response studies. However, in this study, we present a time-frequency adaptive method for physiological interference reduction based on the combination of empirical mode decomposition (EMD) and Hilbert spectral analysis (HSA). Monte Carlo simulations based on a five-layered slab model of a human adult head were implemented to evaluate our methodology. We applied an EMD algorithm to decompose the NIRS time series derived from Monte Carlo simulations into a series of intrinsic mode functions (IMFs). In order to identify the IMFs associated with symmetric interference, the extracted components were then Hilbert transformed from which the instantaneous frequencies could be acquired. By reconstructing the NIRS signal by properly selecting IMFs, we determined that the evoked brain response is effectively filtered out with even higher signal-to-noise ratio (SNR). The results obtained demonstrated that EMD, combined with HSA, can effectively separate, identify and remove the contamination from the evoked brain response obtained with NIRS using a simple single source-detector pair.

  3. Assessing changes to South African maize production areas in 2055 using empirical and process-based crop models

    NASA Astrophysics Data System (ADS)

    Estes, L.; Bradley, B.; Oppenheimer, M.; Beukes, H.; Schulze, R. E.; Tadross, M.

    2010-12-01

    Rising temperatures and altered precipitation patterns associated with climate change pose a significant threat to crop production, particularly in developing countries. In South Africa, a semi-arid country with a diverse agricultural sector, anthropogenic climate change is likely to affect staple crops and decrease food security. Here, we focus on maize production, South Africa’s most widely grown crop and one with high socio-economic value. We build on previous coarser-scaled studies by working at a finer spatial resolution and by employing two different modeling approaches: the process-based DSSAT Cropping System Model (CSM, version 4.5), and an empirical distribution model (Maxent). For climate projections, we use an ensemble of 10 general circulation models (GCMs) run under both high and low CO2 emissions scenarios (SRES A2 and B1). The models were down-scaled to historical climate records for 5838 quinary-scale catchments covering South Africa (mean area = 164.8 km2), using a technique based on self-organizing maps (SOMs) that generates precipitation patterns more consistent with observed gradients than those produced by the parent GCMs. Soil hydrological and mechanical properties were derived from textural and compositional data linked to a map of 26422 land forms (mean area = 46 km2), while organic carbon from 3377 soil profiles was mapped using regression kriging with 8 spatial predictors. CSM was run using typical management parameters for the several major dryland maize production regions, and with projected CO2 values. The Maxent distribution model was trained using maize locations identified using annual phenology derived from satellite images coupled with airborne crop sampling observations. Temperature and precipitation projections were based on GCM output, with an additional 10% increase in precipitation to simulate higher water-use efficiency under future CO2 concentrations. The two modeling approaches provide spatially explicit projections of

  4. Correlation analysis of lung cancer and urban spatial factor: based on survey in Shanghai

    PubMed Central

    Xu, Wangyue; Tang, Jian; Jiang, Xiji

    2016-01-01

    Background The density of particulate matter (PM) in mega-cities in China such as Beijing and Shanghai has exceeded basic standards for health in recent years. Human exposure to PMs has been identified as traceable and controllable factor among all complicated risk factors for lung cancer. While the improvement of air quality needs tremendous efforts and time, certain revision of PM’s density might happen associated with the adjustment of built environment. It is also proved that urban built environment is directly relevant to respiratory disease. Studies have respectively explored the indoor and outdoor factors on respiratory diseases. More comprehensive spatial factors need to be analyzed to understand the cumulative effect of built environment upon respiratory system. This interdisciplinary study examines the impact of both indoor (including age of housing, interval after decoration, indoor humidity etc.) and outdoor spatial factors (including density, parking, green spaces etc.) on lung cancer. Methods A survey of lung cancer patients and a control group has been conducted in 2014 and 2015. A total of 472 interviewees are randomly selected within a pool of local residents who have resided in Shanghai for more than 5 years. Data are collected including their socio-demographic factors, lifestyle factors, and external and internal residential area factors. Regression models are established based on collected data to analyze the associations between lung cancer and urban spatial factors. Results Regression models illustrate that lung cancer presents significantly associated with a number of spatial factors. Significant outdoor spatial factors include external traffic volume (P=0.003), main plant type (P=0.035 for trees) of internal green space, internal water body (P=0.027) and land use of surrounding blocks (P=0.005 for residential areas of 7-9 floors, P=0.000 for residential areas of 4-6 floors, P=0.006 for business/commercial areas over 10 floors, P=0.005 for

  5. Modeling magnetic fields from a DC power cable buried beneath San Francisco Bay based on empirical measurements

    DOE PAGES

    Kavet, Robert; Wyman, Megan T.; Klimley, A. Peter; Carretero, Luis

    2016-02-25

    Here, the Trans Bay Cable (TBC) is a ±200-kilovolt (kV), 400 MW 85-km long High Voltage Direct Current (DC) buried transmission line linking Pittsburg, CA with San Francisco, CA (SF) beneath the San Francisco Estuary. The TBC runs parallel to the migratory route of various marine species, including green sturgeon, Chinook salmon, and steelhead trout. In July and August 2014, an extensive series of magnetic field measurements were taken using a pair of submerged Geometrics magnetometers towed behind a survey vessel in four locations in the San Francisco estuary along profiles that cross the cable’s path; these included the Sanmore » Francisco-Oakland Bay Bridge (BB), the Richmond-San Rafael Bridge (RSR), the Benicia- Martinez Bridge (Ben) and an area in San Pablo Bay (SP) in which a bridge is not present. In this paper, we apply basic formulas that ideally describe the magnetic field from a DC cable summed vectorially with the background geomagnetic field (in the absence of other sources that would perturb the ambient field) to derive characteristics of the cable that are otherwise not immediately observable. Magnetic field profiles from measurements taken along 170 survey lines were inspected visually for evidence of a distinct pattern representing the presence of the cable. Many profiles were dominated by field distortions unrelated to the cable caused by bridge structures or other submerged objects, and the cable’s contribution to the field was not detectable. BB, with 40 of the survey lines, did not yield usable data for these reasons. The unrelated anomalies could be up to 100 times greater than those from the cable. In total, discernible magnetic field profiles measured from 76 survey lines were regressed against the equations, representing eight days of measurement. The modeled field anomalies due to the cable (the difference between the maximum and minimum field along the survey line at the cable crossing) were virtually identical to the measured

  6. Modeling Magnetic Fields from a DC Power Cable Buried Beneath San Francisco Bay Based on Empirical Measurements

    PubMed Central

    Kavet, Robert; Wyman, Megan T.; Klimley, A. Peter

    2016-01-01

    The Trans Bay Cable (TBC) is a ±200-kilovolt (kV), 400 MW 85-km long High Voltage Direct Current (DC) buried transmission line linking Pittsburg, CA with San Francisco, CA (SF) beneath the San Francisco Estuary. The TBC runs parallel to the migratory route of various marine species, including green sturgeon, Chinook salmon, and steelhead trout. In July and August 2014, an extensive series of magnetic field measurements were taken using a pair of submerged Geometrics magnetometers towed behind a survey vessel in four locations in the San Francisco estuary along profiles that cross the cable’s path; these included the San Francisco-Oakland Bay Bridge (BB), the Richmond-San Rafael Bridge (RSR), the Benicia-Martinez Bridge (Ben) and an area in San Pablo Bay (SP) in which a bridge is not present. In this paper, we apply basic formulas that ideally describe the magnetic field from a DC cable summed vectorially with the background geomagnetic field (in the absence of other sources that would perturb the ambient field) to derive characteristics of the cable that are otherwise not immediately observable. Magnetic field profiles from measurements taken along 170 survey lines were inspected visually for evidence of a distinct pattern representing the presence of the cable. Many profiles were dominated by field distortions unrelated to the cable caused by bridge structures or other submerged objects, and the cable’s contribution to the field was not detectable. BB, with 40 of the survey lines, did not yield usable data for these reasons. The unrelated anomalies could be up to 100 times greater than those from the cable. In total, discernible magnetic field profiles measured from 76 survey lines were regressed against the equations, representing eight days of measurement. The modeled field anomalies due to the cable (the difference between the maximum and minimum field along the survey line at the cable crossing) were virtually identical to the measured values. The

  7. Modeling Magnetic Fields from a DC Power Cable Buried Beneath San Francisco Bay Based on Empirical Measurements.

    PubMed

    Kavet, Robert; Wyman, Megan T; Klimley, A Peter

    2016-01-01

    The Trans Bay Cable (TBC) is a ±200-kilovolt (kV), 400 MW 85-km long High Voltage Direct Current (DC) buried transmission line linking Pittsburg, CA with San Francisco, CA (SF) beneath the San Francisco Estuary. The TBC runs parallel to the migratory route of various marine species, including green sturgeon, Chinook salmon, and steelhead trout. In July and August 2014, an extensive series of magnetic field measurements were taken using a pair of submerged Geometrics magnetometers towed behind a survey vessel in four locations in the San Francisco estuary along profiles that cross the cable's path; these included the San Francisco-Oakland Bay Bridge (BB), the Richmond-San Rafael Bridge (RSR), the Benicia-Martinez Bridge (Ben) and an area in San Pablo Bay (SP) in which a bridge is not present. In this paper, we apply basic formulas that ideally describe the magnetic field from a DC cable summed vectorially with the background geomagnetic field (in the absence of other sources that would perturb the ambient field) to derive characteristics of the cable that are otherwise not immediately observable. Magnetic field profiles from measurements taken along 170 survey lines were inspected visually for evidence of a distinct pattern representing the presence of the cable. Many profiles were dominated by field distortions unrelated to the cable caused by bridge structures or other submerged objects, and the cable's contribution to the field was not detectable. BB, with 40 of the survey lines, did not yield usable data for these reasons. The unrelated anomalies could be up to 100 times greater than those from the cable. In total, discernible magnetic field profiles measured from 76 survey lines were regressed against the equations, representing eight days of measurement. The modeled field anomalies due to the cable (the difference between the maximum and minimum field along the survey line at the cable crossing) were virtually identical to the measured values. The modeling

  8. Modeling Magnetic Fields from a DC Power Cable Buried Beneath San Francisco Bay Based on Empirical Measurements.

    PubMed

    Kavet, Robert; Wyman, Megan T; Klimley, A Peter

    2016-01-01

    The Trans Bay Cable (TBC) is a ±200-kilovolt (kV), 400 MW 85-km long High Voltage Direct Current (DC) buried transmission line linking Pittsburg, CA with San Francisco, CA (SF) beneath the San Francisco Estuary. The TBC runs parallel to the migratory route of various marine species, including green sturgeon, Chinook salmon, and steelhead trout. In July and August 2014, an extensive series of magnetic field measurements were taken using a pair of submerged Geometrics magnetometers towed behind a survey vessel in four locations in the San Francisco estuary along profiles that cross the cable's path; these included the San Francisco-Oakland Bay Bridge (BB), the Richmond-San Rafael Bridge (RSR), the Benicia-Martinez Bridge (Ben) and an area in San Pablo Bay (SP) in which a bridge is not present. In this paper, we apply basic formulas that ideally describe the magnetic field from a DC cable summed vectorially with the background geomagnetic field (in the absence of other sources that would perturb the ambient field) to derive characteristics of the cable that are otherwise not immediately observable. Magnetic field profiles from measurements taken along 170 survey lines were inspected visually for evidence of a distinct pattern representing the presence of the cable. Many profiles were dominated by field distortions unrelated to the cable caused by bridge structures or other submerged objects, and the cable's contribution to the field was not detectable. BB, with 40 of the survey lines, did not yield usable data for these reasons. The unrelated anomalies could be up to 100 times greater than those from the cable. In total, discernible magnetic field profiles measured from 76 survey lines were regressed against the equations, representing eight days of measurement. The modeled field anomalies due to the cable (the difference between the maximum and minimum field along the survey line at the cable crossing) were virtually identical to the measured values. The modeling

  9. A hybrid model for PM₂.₅ forecasting based on ensemble empirical mode decomposition and a general regression neural network.

    PubMed

    Zhou, Qingping; Jiang, Haiyan; Wang, Jianzhou; Zhou, Jianling

    2014-10-15

    Exposure to high concentrations of fine particulate matter (PM₂.₅) can cause serious health problems because PM₂.₅ contains microscopic solid or liquid droplets that are sufficiently small to be ingested deep into human lungs. Thus, daily prediction of PM₂.₅ levels is notably important for regulatory plans that inform the public and restrict social activities in advance when harmful episodes are foreseen. A hybrid EEMD-GRNN (ensemble empirical mode decomposition-general regression neural network) model based on data preprocessing and analysis is firstly proposed in this paper for one-day-ahead prediction of PM₂.₅ concentrations. The EEMD part is utilized to decompose original PM₂.₅ data into several intrinsic mode functions (IMFs), while the GRNN part is used for the prediction of each IMF. The hybrid EEMD-GRNN model is trained using input variables obtained from principal component regression (PCR) model to remove redundancy. These input variables accurately and succinctly reflect the relationships between PM₂.₅ and both air quality and meteorological data. The model is trained with data from January 1 to November 1, 2013 and is validated with data from November 2 to November 21, 2013 in Xi'an Province, China. The experimental results show that the developed hybrid EEMD-GRNN model outperforms a single GRNN model without EEMD, a multiple linear regression (MLR) model, a PCR model, and a traditional autoregressive integrated moving average (ARIMA) model. The hybrid model with fast and accurate results can be used to develop rapid air quality warning systems. PMID:25089688

  10. Development of An Empirical Water Quality Model for Stormwater Based on Watershed Land Use in Puget Sound

    SciTech Connect

    Cullinan, Valerie I.; May, Christopher W.; Brandenberger, Jill M.; Judd, Chaeli; Johnston, Robert K.

    2007-03-29

    The Sinclair and Dyes Inlet watershed is located on the west side of Puget Sound in Kitsap County, Washington, U.S.A. (Figure 1). The Puget Sound Naval Shipyard (PSNS), U.S Environmental Protection Agency (USEPA), the Washington State Department of Ecology (WA-DOE), Kitsap County, City of Bremerton, City of Bainbridge Island, City of Port Orchard, and the Suquamish Tribe have joined in a cooperative effort to evaluate water-quality conditions in the Sinclair-Dyes Inlet watershed and correct identified problems. A major focus of this project, known as Project ENVVEST, is to develop Water Clean-up (TMDL) Plans for constituents listed on the 303(d) list within the Sinclair and Dyes Inlet watershed. Segments within the Sinclair and Dyes Inlet watershed were listed on the State of Washington’s 1998 303(d) because of fecal coliform contamination in marine water, metals in sediment and fish tissue, and organics in sediment and fish tissue (WA-DOE 2003). Stormwater loading was identified by ENVVEST as one potential source of sediment contamination, which lacked sufficient data for a contaminant mass balance calculation for the watershed. This paper summarizes the development of an empirical model for estimating contaminant concentrations in all streams discharging into Sinclair and Dyes Inlets based on watershed land use, 18 storm events, and wet/dry season baseflow conditions between November 2002 and May 2005. Stream pollutant concentrations along with estimates for outfalls and surface runoff will be used in estimating the loading and ultimately in establishing a Water Cleanup Plan (TMDL) for the Sinclair-Dyes Inlet watershed.

  11. Tectonic map of Liberia based on geophysical and geological surveys

    USGS Publications Warehouse

    Behrendt, John Charles; Wotorson, Cletus S.

    1972-01-01

    Interpretation of the results of aeromagnetic, total-gamma radioactivity, and gravity surveys combined with geologic data for Western Liberia from White and Leo (1969) and other geologic information allows the construction of a tectonic map of Liberia. The map approximately delineates the boundaries between the Liberian (ca. 2700 m.y.) province in the northwestern two-thirds of the country, the Eburnean (ca. 2000 m.y.) province in the south-eastern one-third, and the Pan-African (ca. 550 m.y.) province in the coastal area of the northwestern two-thirds of the country. Rock follation and tectonic structural features trend northeastward in the Liberian province, east-northeastward to north-northeastward in the Eburnean province, and northwestward in the Pan-African age province. Linear residual magnetic anomailes 20-80 km wide and 200-600 gammas in amplitude and following the northeast structural trend typical of the Liberian age province cross the entire country and extend into Sierra Leone and Ivory Coast.

  12. Empirical and targeted therapy of candidemia with fluconazole versus echinocandins: a propensity score-derived analysis of a population-based, multicentre prospective cohort.

    PubMed

    López-Cortés, L E; Almirante, B; Cuenca-Estrella, M; Garnacho-Montero, J; Padilla, B; Puig-Asensio, M; Ruiz-Camps, I; Rodríguez-Baño, J

    2016-08-01

    We compared the clinical efficacy of fluconazole and echinocandins in the treatment of candidemia in real practice. The CANDIPOP study is a prospective, population-based cohort study on candidemia carried out between May 2010 and April 2011 in 29 Spanish hospitals. Using strict inclusion criteria, we separately compared the impact of empirical and targeted therapy with fluconazole or echinocandins on 30-day mortality. Cox regression, including a propensity score (PS) for receiving echinocandins, stratified analysis on the PS quartiles and PS-based matched analyses, were performed. The empirical and targeted therapy cohorts comprised 316 and 421 cases, respectively; 30-day mortality was 18.7% with fluconazole and 33.9% with echinocandins (p 0.02) in the empirical therapy group and 19.8% with fluconazole and 27.7% with echinocandins (p 0.06) in the targeted therapy group. Multivariate Cox regression analysis including PS showed that empirical therapy with fluconazole was associated with better prognosis (adjusted hazard ratio 0.38; 95% confidence interval 0.17-0.81; p 0.01); no differences were found within each PS quartile or in cases matched according to PS. Targeted therapy with fluconazole did not show a significant association with mortality in the Cox regression analysis (adjusted hazard ratio 0.77; 95% confidence interval 0.41-1.46; p 0.63), in the PS quartiles or in PS-matched cases. The results were similar among patients with severe sepsis and septic shock. Empirical or targeted treatment with fluconazole was not associated with increased 30-day mortality compared to echinocandins among adults with candidemia.

  13. Current indications for renal biopsy: a questionnaire-based survey.

    PubMed

    Fuiano, G; Mazza, G; Comi, N; Caglioti, A; De Nicola, L; Iodice, C; Andreucci, M; Andreucci, V E

    2000-03-01

    Indications for renal biopsy are still ill defined. We recently sent a detailed questionnaire to 360 nephrologists in different areas of the world with the aim of providing information on this critical issue by evaluating the replies. The questionnaire was organized in four sections that included questions on renal biopsy indications in patients with normal renal function, renal insufficiency, and a transplanted kidney. In addition, the questions included methods applied to each renal biopsy procedure and to specimen processing. We received 166 replies; North Europe (50 replies), South Europe (47 replies), North America (31 replies), Australia and New Zealand (24 replies), and other countries (14 replies). In patients with normal renal function, primary indications for renal biopsy were microhematuria associated with proteinuria, particularly greater than 1 g/d of protein. In chronic renal insufficiency, kidney dimension was the major parameter considered before renal biopsy, whereas the presence of diabetes or serological abnormalities was not considered critical. In the course of acute renal failure (ARF) of unknown origin, 20% of the respondents would perform renal biopsy in the early stages, 26% after 1 week of nonrecovery, and 40% after 4 weeks. In a transplanted kidney, the majority of nephrologists would perform a renal biopsy in the case of graft failure after surgery, ARF after initial good function, slow progressive deterioration of renal function, and onset of nephrotic proteinuria. The last section provided comprehensive information on the technical aspects of renal biopsy. This survey represents the first attempt to provide a reliable consensus that can be used in developing guidelines on the use of kidney biopsy. PMID:10692270

  14. Issues in nursing: strategies for an Internet-based, computer-assisted telephone survey.

    PubMed

    Piamjariyakul, Ubolrat; Bott, Marjorie J; Taunton, Roma Lee

    2006-08-01

    The study describes the design and implementation of an Internet-based, computed-assisted telephone survey about the care-planning process in 107 long-term care facilities in the Midwest. Two structured telephone surveys were developed to interview the care planning coordinators and their team members. Questionmark Perception Software Version 3 was used to develop the surveys in a wide range of formats. The responses were drawn into a database that was exported to a spreadsheet format and converted to a statistical format by the Information Technology team. Security of the database was protected. Training sessions were provided to project staff. The interviews were tape-recorded for the quality checks. The inter-rater reliabilities were above 95% to 100% agreement. Investigators should consider using Internet-based survey tools, especially for multisite studies that allow access to larger samples at less cost. Exploring multiple software systems for the best fit to the study requirements is essential.

  15. Empirically Based Profiles of the Early Literacy Skills of Children with Language Impairment in Early Childhood Special Education

    ERIC Educational Resources Information Center

    Justice, Laura; Logan, Jessica; Kaderavek, Joan; Schmitt, Mary Beth; Tompkins, Virginia; Bartlett, Christopher

    2015-01-01

    The purpose of this study was to empirically determine whether specific profiles characterize preschool-aged children with language impairment (LI) with respect to their early literacy skills (print awareness, name-writing ability, phonological awareness, alphabet knowledge); the primary interest was to determine if one or more profiles suggested…

  16. Use of Evidence-Based Practice Resources and Empirically Supported Treatments for Posttraumatic Stress Disorder among University Counseling Center Psychologists

    ERIC Educational Resources Information Center

    Juel, Morgen Joray

    2012-01-01

    In the present study, an attempt was made to determine the degree to which psychologists at college and university counseling centers (UCCs) utilized empirically supported treatments with their posttraumatic stress disorder (PTSD) clients. In addition, an attempt was made to determine how frequently UCC psychologists utilized a number of…

  17. Land-based lidar mapping: a new surveying technique to shed light on rapid topographic change

    USGS Publications Warehouse

    Collins, Brian D.; Kayen, Robert

    2006-01-01

    The rate of natural change in such dynamic environments as rivers and coastlines can sometimes overwhelm the monitoring capacity of conventional surveying methods. In response to this limitation, U.S. Geological Survey (USGS) scientists are pioneering new applications of light detection and ranging (lidar), a laser-based scanning technology that promises to greatly increase our ability to track rapid topographic changes and manage their impact on affected communities.

  18. A survey on evolutionary algorithm based hybrid intelligence in bioinformatics.

    PubMed

    Li, Shan; Kang, Liying; Zhao, Xing-Ming

    2014-01-01

    With the rapid advance in genomics, proteomics, metabolomics, and other types of omics technologies during the past decades, a tremendous amount of data related to molecular biology has been produced. It is becoming a big challenge for the bioinformatists to analyze and interpret these data with conventional intelligent techniques, for example, support vector machines. Recently, the hybrid intelligent methods, which integrate several standard intelligent approaches, are becoming more and more popular due to their robustness and efficiency. Specifically, the hybrid intelligent approaches based on evolutionary algorithms (EAs) are widely used in various fields due to the efficiency and robustness of EAs. In this review, we give an introduction about the applications of hybrid intelligent methods, in particular those based on evolutionary algorithm, in bioinformatics. In particular, we focus on their applications to three common problems that arise in bioinformatics, that is, feature selection, parameter estimation, and reconstruction of biological networks.

  19. A Survey on Evolutionary Algorithm Based Hybrid Intelligence in Bioinformatics

    PubMed Central

    Li, Shan; Zhao, Xing-Ming

    2014-01-01

    With the rapid advance in genomics, proteomics, metabolomics, and other types of omics technologies during the past decades, a tremendous amount of data related to molecular biology has been produced. It is becoming a big challenge for the bioinformatists to analyze and interpret these data with conventional intelligent techniques, for example, support vector machines. Recently, the hybrid intelligent methods, which integrate several standard intelligent approaches, are becoming more and more popular due to their robustness and efficiency. Specifically, the hybrid intelligent approaches based on evolutionary algorithms (EAs) are widely used in various fields due to the efficiency and robustness of EAs. In this review, we give an introduction about the applications of hybrid intelligent methods, in particular those based on evolutionary algorithm, in bioinformatics. In particular, we focus on their applications to three common problems that arise in bioinformatics, that is, feature selection, parameter estimation, and reconstruction of biological networks. PMID:24729969

  20. Content based Image Retrieval based on Different Global and Local Color Histogram Methods: A Survey

    NASA Astrophysics Data System (ADS)

    Suhasini, Pallikonda Sarah; Sri Rama Krishna, K.; Murali Krishna, I. V.

    2016-06-01

    Different global and local color histogram methods for content based image retrieval (CBIR) are investigated in this paper. Color histogram is a widely used descriptor for CBIR. Conventional method of extracting color histogram is global, which misses the spatial content, is less invariant to deformation and viewpoint changes, and results in a very large three dimensional histogram corresponding to the color space used. To address the above deficiencies, different global and local histogram methods are proposed in recent research. Different ways of extracting local histograms to have spatial correspondence, invariant colour histogram to add deformation and viewpoint invariance and fuzzy linking method to reduce the size of the histogram are found in recent papers. The color space and the distance metric used are vital in obtaining color histogram. In this paper the performance of CBIR based on different global and local color histograms in three different color spaces, namely, RGB, HSV, L*a*b* and also with three distance measures Euclidean, Quadratic and Histogram intersection are surveyed, to choose appropriate method for future research.

  1. SU-E-T-05: A 2D EPID Transit Dosimetry Model Based On An Empirical Quadratic Formalism

    SciTech Connect

    Tan, Y; Metwaly, M; Glegg, M; Baggarley, S; Elliott, A

    2014-06-01

    Purpose: To describe a 2D electronic portal imaging device (EPID) transit dosimetry model, based on an empirical quadratic formalism, that can predict either EPID or in-phantom dose distribution for comparisons with EPID captured image or treatment planning system (TPS) dose respectively. Methods: A quadratic equation can be used to relate the reduction in intensity of an exit beam to the equivalent path length of the attenuator. The calibration involved deriving coefficients from a set of dose planes measured for homogeneous phantoms with known thicknesses under reference conditions. In this study, calibration dose planes were measured with EPID and ionisation chamber (IC) in water for the same reference beam (6MV, 100mu, 20×20cm{sup 2}) and set of thicknesses (0–30cm). Since the same calibration conditions were used, the EPID and IC measurements can be related through the quadratic equation. Consequently, EPID transit dose can be predicted from TPS exported dose planes and in-phantom dose can be predicted using EPID distribution captured during treatment as an input. The model was tested with 4 open fields, 6 wedge fields, and 7 IMRT fields on homogeneous and heterogeneous phantoms. Comparisons were done using 2D absolute gamma (3%/3mm) and results were validated against measurements with a commercial 2D array device. Results: The gamma pass rates for comparisons between EPID measured and predicted ranged from 93.6% to 100.0% for all fields and phantoms tested. Results from this study agreed with 2D array measurements to within 3.1%. Meanwhile, comparisons in-phantom between TPS computed and predicted ranged from 91.6% to 100.0%. Validation with 2D array device was not possible for inphantom comparisons. Conclusion: A 2D EPID transit dosimetry model for treatment verification was described and proven to be accurate. The model has the advantage of being generic and allows comparisons at the EPID plane as well as multiple planes in-phantom.

  2. Acceptance and Use of Game-Based Learning in Vocational Education and Training: An International Survey

    ERIC Educational Resources Information Center

    Schmitz, Birgit; Felicia, Patrick; Bignami, Filippo

    2015-01-01

    This paper presents findings from a study carried out between May and October 2013. Based on a survey, which was developed by the MoGaBa VET project partners, the study aimed at understanding the factors that influence the way vocational instructors perceive and use game-based learning. A total of 267 trainers from eight European countries took…

  3. Developing a Computer Information Systems Curriculum Based on an Industry Needs Survey.

    ERIC Educational Resources Information Center

    Ghafarian, Ahmad; Sisk, Kathy A.

    This paper details experiences in developing an undergraduate Computer Information Systems (CIS) curriculum at a small liberal arts school. The development of the program was based on the study of needs assessment. Findings were based on the analysis of four sources of data: the results of an industry needs survey, data from a needs assessment…

  4. OBSERVATIONS OF BINARY STARS WITH THE DIFFERENTIAL SPECKLE SURVEY INSTRUMENT. V. TOWARD AN EMPIRICAL METAL-POOR MASS–LUMINOSITY RELATION

    SciTech Connect

    Horch, Elliott P.; Van Altena, William F.; Demarque, Pierre; Howell, Steve B.; Everett, Mark E.; Ciardi, David R.; Teske, Johanna K.; Henry, Todd J.; Winters, Jennifer G. E-mail: william.vanaltena@yale.edu E-mail: steve.b.howell@nasa.gov E-mail: ciardi@ipac.caltech.edu E-mail: thenry@astro.gsu.edu

    2015-05-15

    In an effort to better understand the details of the stellar structure and evolution of metal-poor stars, the Gemini North telescope was used on two occasions to take speckle imaging data of a sample of known spectroscopic binary stars and other nearby stars in order to search for and resolve close companions. The observations were obtained using the Differential Speckle Survey Instrument, which takes data in two filters simultaneously. The results presented here are of 90 observations of 23 systems in which one or more companions was detected, and six stars where no companion was detected to the limit of the camera capabilities at Gemini. In the case of the binary and multiple stars, these results are then further analyzed to make first orbit determinations in five cases, and orbit refinements in four other cases. The mass information is derived, and since the systems span a range in metallicity, a study is presented that compares our results with the expected trend in total mass as derived from the most recent Yale isochrones as a function of metal abundance. These data suggest that metal-poor main-sequence stars are less massive at a given color than their solar-metallicity analogues in a manner consistent with that predicted from the theory.

  5. Observations of Binary Stars with the Differential Speckle Survey Instrument. V. Toward an Empirical Metal-Poor Mass-Luminosity Relation

    NASA Astrophysics Data System (ADS)

    Horch, Elliott P.; van Altena, William F.; Demarque, Pierre; Howell, Steve B.; Everett, Mark E.; Ciardi, David R.; Teske, Johanna K.; Henry, Todd J.; Winters, Jennifer G.

    2015-05-01

    In an effort to better understand the details of the stellar structure and evolution of metal-poor stars, the Gemini North telescope was used on two occasions to take speckle imaging data of a sample of known spectroscopic binary stars and other nearby stars in order to search for and resolve close companions. The observations were obtained using the Differential Speckle Survey Instrument, which takes data in two filters simultaneously. The results presented here are of 90 observations of 23 systems in which one or more companions was detected, and six stars where no companion was detected to the limit of the camera capabilities at Gemini. In the case of the binary and multiple stars, these results are then further analyzed to make first orbit determinations in five cases, and orbit refinements in four other cases. The mass information is derived, and since the systems span a range in metallicity, a study is presented that compares our results with the expected trend in total mass as derived from the most recent Yale isochrones as a function of metal abundance. These data suggest that metal-poor main-sequence stars are less massive at a given color than their solar-metallicity analogues in a manner consistent with that predicted from the theory.

  6. Trajectory-based visual localization in underwater surveying missions.

    PubMed

    Burguera, Antoni; Bonin-Font, Francisco; Oliver, Gabriel

    2015-01-14

    We present a new vision-based localization system applied to an autonomous underwater vehicle (AUV) with limited sensing and computation capabilities. The traditional EKF-SLAM approaches are usually expensive in terms of execution time; the approach presented in this paper strengthens this method by adopting a trajectory-based schema that reduces the computational requirements. The pose of the vehicle is estimated using an extended Kalman filter (EKF), which predicts the vehicle motion by means of a visual odometer and corrects these predictions using the data associations (loop closures) between the current frame and the previous ones. One of the most important steps in this procedure is the image registration method, as it reinforces the data association and, thus, makes it possible to close loops reliably. Since the use of standard EKFs entail linearization errors that can distort the vehicle pose estimations, the approach has also been tested using an iterated Kalman filter (IEKF). Experiments have been conducted using a real underwater vehicle in controlled scenarios and in shallow sea waters, showing an excellent performance with very small errors, both in the vehicle pose and in the overall trajectory estimates.

  7. Trajectory-Based Visual Localization in Underwater Surveying Missions

    PubMed Central

    Burguera, Antoni; Bonin-Font, Francisco; Oliver, Gabriel

    2015-01-01

    We present a new vision-based localization system applied to an autonomous underwater vehicle (AUV) with limited sensing and computation capabilities. The traditional EKF-SLAM approaches are usually expensive in terms of execution time; the approach presented in this paper strengthens this method by adopting a trajectory-based schema that reduces the computational requirements. The pose of the vehicle is estimated using an extended Kalman filter (EKF), which predicts the vehicle motion by means of a visual odometer and corrects these predictions using the data associations (loop closures) between the current frame and the previous ones. One of the most important steps in this procedure is the image registration method, as it reinforces the data association and, thus, makes it possible to close loops reliably. Since the use of standard EKFs entail linearization errors that can distort the vehicle pose estimations, the approach has also been tested using an iterated Kalman filter (IEKF). Experiments have been conducted using a real underwater vehicle in controlled scenarios and in shallow sea waters, showing an excellent performance with very small errors, both in the vehicle pose and in the overall trajectory estimates. PMID:25594602

  8. Trajectory-based visual localization in underwater surveying missions.

    PubMed

    Burguera, Antoni; Bonin-Font, Francisco; Oliver, Gabriel

    2015-01-01

    We present a new vision-based localization system applied to an autonomous underwater vehicle (AUV) with limited sensing and computation capabilities. The traditional EKF-SLAM approaches are usually expensive in terms of execution time; the approach presented in this paper strengthens this method by adopting a trajectory-based schema that reduces the computational requirements. The pose of the vehicle is estimated using an extended Kalman filter (EKF), which predicts the vehicle motion by means of a visual odometer and corrects these predictions using the data associations (loop closures) between the current frame and the previous ones. One of the most important steps in this procedure is the image registration method, as it reinforces the data association and, thus, makes it possible to close loops reliably. Since the use of standard EKFs entail linearization errors that can distort the vehicle pose estimations, the approach has also been tested using an iterated Kalman filter (IEKF). Experiments have been conducted using a real underwater vehicle in controlled scenarios and in shallow sea waters, showing an excellent performance with very small errors, both in the vehicle pose and in the overall trajectory estimates. PMID:25594602

  9. Implementing Project Based Survey Research Skills to Grade Six ELP Students with "The Survey Toolkit" and "TinkerPlots"[R

    ERIC Educational Resources Information Center

    Walsh, Thomas, Jr.

    2011-01-01

    "Survey Toolkit Collecting Information, Analyzing Data and Writing Reports" (Walsh, 2009a) is discussed as a survey research curriculum used by the author's sixth grade students. The report describes the implementation of "The Survey Toolkit" curriculum and "TinkerPlots"[R] software to provide instruction to students learning a project based…

  10. NVO, surveys and archiving ground-based data

    NASA Astrophysics Data System (ADS)

    Huchra, John P.

    2002-12-01

    The era of extremely large, public databases in astronomy is upon us. Such databases will open (are opening!) the field to new research and new researchers. However it is important to be sure the resources are available to properly archive groundbased astronomical data, and include the necessary quality checks and calibrations. An NVO without a proper archive will have limited usefulness. This also implies that with limited resources, not all data can or should be archived. NASA already has a very good handle on US space-based astronomical data. Agencies and organizations that operate astronomical facilities, particularly groundbased observatories, need to plan and budget for these activities now. We should not underestimate the effort required to produce high quality data products that will be useful for the broader community.

  11. An internet-based survey method for college student drinking research.

    PubMed

    Kypri, Kypros; Gallagher, Stephen J; Cashell-Smith, Martine L

    2004-10-01

    The purpose of this study was to describe and assess the utility of an Internet-based survey method for characterizing the alcohol consumption of college students. After extensive pilot research, a random sample of 1910 students aged 16-29 years was invited to complete a questionnaire, consisting of a series of web-pages linked to a relational database on a secure web-site. A branch structure allowed for tailoring of survey items by age and gender. The students received up to nine contacts, including a pre-notice letter with a token gift and an e-mail invitation (Phase 1), a reminder letter and e-mail message (Phase 2), and then telephone reminders and replacement access codes (Phase 3). Non-computer-users were offered a pen-and-paper alternative, making this a mixed-mode survey. The overall response to the survey was 82% (n = 1564). The median completion time was 16.7 min. Participants' comments showed high levels of satisfaction with the survey. Comparison of web (n = 1501) versus pen-and-paper completions (n = 63) revealed no modality effects. Technical problems addressed during the course of implementation included web-browser-operating system incompatibilities, and periodic network errors, although these resulted in little lost participation. Internet-based surveys are feasible for college student research and with carefully managed recruitment, can yield a high response.

  12. Sleepwalking in Parkinson's disease: a questionnaire-based survey.

    PubMed

    Oberholzer, Michael; Poryazova, Rositsa; Bassetti, Claudio L

    2011-07-01

    Sleepwalking (SW) corresponds to a complex sleep-associated behavior that includes locomotion, mental confusion, and amnesia. SW is present in about 10% of children and 2-3% of adults. In a retrospective series of 165 patients with Parkinson's disease (PD), we found adult-onset ("de novo") SW "de novo" in six (4%) of them. The aim of this study was to assess prospectively and systematically the frequency and characteristics of SW in PD patients. A questionnaire including items on sleep quality, sleep disorders, and specifically also SW and REM sleep behavior disorder (RBD), PD characteristics and severity, was sent to the members of the national PD patients organization in Switzerland. In the study, 36/417 patients (9%) reported SW, of which 22 (5%) had adult-onset SW. Patients with SW had significantly longer disease duration (p = 0.035), they reported more often hallucinations (p = 0.004) and nightmares (p = 0.003), and they had higher scores, suggestive for RBD in a validated questionnaire (p = 0.001). Patients with SW were also sleepier (trend to a higher Epworth Sleepiness Scale score, p = 0.055). Our data suggest that SW in PD patients is (1) more common than in the general population, and (2) is associated with RBD, nightmares, and hallucinations. Further studies including polysomnographic recordings are needed to confirm the results of this questionnaire-based analysis, to understand the relationship between SW and other nighttime wandering behaviors in PD, and to clarify the underlying mechanisms.

  13. Questionnaire-based survey of parturition in the queen.

    PubMed

    Musters, J; de Gier, J; Kooistra, H S; Okkens, A C

    2011-06-01

    The lack of scientific data concerning whether parturition in the queen proceeds normally or not may prevent veterinarians and cat owners from recognizing parturition problems in time. A questionnaire-based study of parturition in 197 queens was performed to determine several parameters of parturition and their influence on its progress. The mean length of gestation was 65.3 days (range 57 to 72 days) and it decreased with increasing litter size (P = 0.02). The median litter size was 4.5 kittens (range 1 to 9), with more males (53%) than females (46%) (P = 0.05). Sixty-nine percent of the kittens were born in anterior presentation and 31% in posterior presentation, indicating that either can be considered normal in the cat. Males were born in posterior position (34%) more often than females (26%) (P = 0.03). The mean birth weight was 98 g (range of 35 to 167 g) and decreased with increasing litter size (P < 0.01). Mean birth weight was higher in males and kittens born in posterior presentation (P < 0.01). Forty-four (5%) of the 887 kittens were stillborn. This was not correlated with the presentation at expulsion but stillborn kittens were more often female (P = 0.02) and weighed less than those born alive (P = 0.04). The median interkitten time was 30 min (range 2 to 343 min) and 95% were born within 100 min after expulsion of the preceding kitten. The interkitten time as a measure of the progress of parturition was not influenced by the kitten's gender, presentation at expulsion, birth weight, or stillbirth, or by the parity of the queen. The results of this study can be used to develop reference values for parturition parameters in the queen, both to determine whether a given parturition is abnormal and as the basis for a parturition protocol.

  14. Sleepwalking in Parkinson's disease: a questionnaire-based survey.

    PubMed

    Oberholzer, Michael; Poryazova, Rositsa; Bassetti, Claudio L

    2011-07-01

    Sleepwalking (SW) corresponds to a complex sleep-associated behavior that includes locomotion, mental confusion, and amnesia. SW is present in about 10% of children and 2-3% of adults. In a retrospective series of 165 patients with Parkinson's disease (PD), we found adult-onset ("de novo") SW "de novo" in six (4%) of them. The aim of this study was to assess prospectively and systematically the frequency and characteristics of SW in PD patients. A questionnaire including items on sleep quality, sleep disorders, and specifically also SW and REM sleep behavior disorder (RBD), PD characteristics and severity, was sent to the members of the national PD patients organization in Switzerland. In the study, 36/417 patients (9%) reported SW, of which 22 (5%) had adult-onset SW. Patients with SW had significantly longer disease duration (p = 0.035), they reported more often hallucinations (p = 0.004) and nightmares (p = 0.003), and they had higher scores, suggestive for RBD in a validated questionnaire (p = 0.001). Patients with SW were also sleepier (trend to a higher Epworth Sleepiness Scale score, p = 0.055). Our data suggest that SW in PD patients is (1) more common than in the general population, and (2) is associated with RBD, nightmares, and hallucinations. Further studies including polysomnographic recordings are needed to confirm the results of this questionnaire-based analysis, to understand the relationship between SW and other nighttime wandering behaviors in PD, and to clarify the underlying mechanisms. PMID:21293874

  15. Exoplanets -New Results from Space and Ground-based Surveys

    NASA Astrophysics Data System (ADS)

    Udry, Stephane

    The exploration of the outer solar system and in particular of the giant planets and their environments is an on-going process with the Cassini spacecraft currently around Saturn, the Juno mission to Jupiter preparing to depart and two large future space missions planned to launch in the 2020-2025 time frame for the Jupiter system and its satellites (Europa and Ganymede) on the one hand, and the Saturnian system and Titan on the other hand [1,2]. Titan, Saturn's largest satellite, is the only other object in our Solar system to possess an extensive nitrogen atmosphere, host to an active organic chemistry, based on the interaction of N2 with methane (CH4). Following the Voyager flyby in 1980, Titan has been intensely studied from the ground-based large telescopes (such as the Keck or the VLT) and by artificial satellites (such as the Infrared Space Observatory and the Hubble Space Telescope) for the past three decades. Prior to Cassini-Huygens, Titan's atmospheric composition was thus known to us from the Voyager missions and also through the explorations by the ISO. Our perception of Titan had thus greatly been enhanced accordingly, but many questions remained as to the nature of the haze surrounding the satellite and the composition of the surface. The recent revelations by the Cassini-Huygens mission have managed to surprise us with many discoveries [3-8] and have yet to reveal more of the interesting aspects of the satellite. The Cassini-Huygens mission to the Saturnian system has been an extraordinary success for the planetary community since the Saturn-Orbit-Insertion (SOI) in July 2004 and again the very successful probe descent and landing of Huygens on January 14, 2005. One of its main targets was Titan. Titan was revealed to be a complex world more like the Earth than any other: it has a dense mostly nitrogen atmosphere and active climate and meteorological cycles where the working fluid, methane, behaves under Titan conditions the way that water does on

  16. Empirical water depth predictions in Dublin Bay based on satellite EO multispectral imagery and multibeam data using spatially weighted geographical analysis

    NASA Astrophysics Data System (ADS)

    Monteys, Xavier; Harris, Paul; Caloca, Silvia

    2014-05-01

    The coastal shallow water zone can be a challenging and expensive environment within which to acquire bathymetry and other oceanographic data using traditional survey methods. Dangers and limited swath coverage make some of these areas unfeasible to survey using ship borne systems, and turbidity can preclude marine LIDAR. As a result, an extensive part of the coastline worldwide remains completely unmapped. Satellite EO multispectral data, after processing, allows timely, cost efficient and quality controlled information to be used for planning, monitoring, and regulating coastal environments. It has the potential to deliver repetitive derivation of medium resolution bathymetry, coastal water properties and seafloor characteristics in shallow waters. Over the last 30 years satellite passive imaging methods for bathymetry extraction, implementing analytical or empirical methods, have had a limited success predicting water depths. Different wavelengths of the solar light penetrate the water column to varying depths. They can provide acceptable results up to 20 m but become less accurate in deeper waters. The study area is located in the inner part of Dublin Bay, on the East coast of Ireland. The region investigated is a C-shaped inlet covering an area of 10 km long and 5 km wide with water depths ranging from 0 to 10 m. The methodology employed on this research uses a ratio of reflectance from SPOT 5 satellite bands, differing to standard linear transform algorithms. High accuracy water depths were derived using multibeam data. The final empirical model uses spatially weighted geographical tools to retrieve predicted depths. The results of this paper confirm that SPOT satellite scenes are suitable to predict depths using empirical models in very shallow embayments. Spatial regression models show better adjustments in the predictions over non-spatial models. The spatial regression equation used provides realistic results down to 6 m below the water surface, with

  17. Empirically Driven Software Engineering Research

    NASA Astrophysics Data System (ADS)

    Rombach, Dieter

    Software engineering is a design discipline. As such, its engineering methods are based on cognitive instead of physical laws, and their effectiveness depends highly on context. Empirical methods can be used to observe the effects of software engineering methods in vivo and in vitro, to identify improvement potentials, and to validate new research results. This paper summarizes both the current body of knowledge and further challenges wrt. empirical methods in software engineering as well as empirically derived evidence regarding software typical engineering methods. Finally, future challenges wrt. education, research, and technology transfer will be outlined.

  18. An in-Depth Survey of Visible Light Communication Based Positioning Systems

    PubMed Central

    Do, Trong-Hop; Yoo, Myungsik

    2016-01-01

    While visible light communication (VLC) has become the candidate for the wireless technology of the 21st century due to its inherent advantages, VLC based positioning also has a great chance of becoming the standard approach to positioning. Within the last few years, many studies on VLC based positioning have been published, but there are not many survey works in this field. In this paper, an in-depth survey of VLC based positioning systems is provided. More than 100 papers ranging from pioneering papers to the state-of-the-art in the field were collected and classified based on the positioning algorithms, the types of receivers, and the multiplexing techniques. In addition, current issues and research trends in VLC based positioning are discussed. PMID:27187395

  19. Evaluation of Midwater Trawl Selectivity and its Influence on Acoustic-Based Fish Population Surveys

    NASA Astrophysics Data System (ADS)

    Williams, Kresimir

    Trawls are used extensively during fisheries abundance surveys to derive estimates of fish density and, in the case of acoustic-based surveys, to identify acoustically sampled fish populations. However, trawls are selective in what fish they retain, resulting in biased estimates of density, species, and size compositions. Selectivity of the midwater trawl used in acoustic-based surveys of walleye pollock (Theragra chalcogramma) was evaluated using multiple methods. The effects of trawl selectivity on the acoustic-based survey abundance estimates and the stock assessment were evaluated for the Gulf of Alaska walleye pollock population. Selectivity was quantified using recapture, or pocket, nets attached to the outside of the trawl. Pocket net catches were modeled using a hierarchical Bayesian model to provide uncertainty in selectivity parameter estimates. Significant under-sampling of juvenile pollock by the midwater trawl was found, with lengths at 50% retention ranging from 14--26 cm over three experiments. Escapement was found to be light dependent, with more fish escaping in dark conditions. Highest escapement rates were observed in the aft of the trawl near to the codend though the bottom panel of the trawl. The behavioral mechanisms involved in the process of herding and escapement were evaluated using stereo-cameras, a DIDSON high frequency imaging sonar, and pocket nets. Fish maintained greater distances from the trawl panel during daylight, suggesting trawl modifications such as increased visibility of netting materials may evoke stronger herding responses and increased retention of fish. Selectivity and catchability of pollock by the midwater trawl was also investigated using acoustic density as an independent estimate of fish abundance to compare with trawl catches. A modeling framework was developed to evaluate potential explanatory factors for selectivity and catchability. Selectivity estimates were dependent on which vessel was used for the survey

  20. Empirically based Suggested Insights into the Concept of False-Self Defense: Contributions From a Study on Normalization of Children With Disabilities.

    PubMed

    Eichengreen, Adva; Hoofien, Dan; Bachar, Eytan

    2016-02-01

    The concept of the false self has been used widely in psychoanalytic theory and practice but seldom in empirical research. In this empirically based study, elevated features of false-self defense were hypothetically associated with risk factors attendant on processes of rehabilitation and integration of children with disabilities, processes that encourage adaptation of the child to the able-bodied environment. Self-report questionnaires and in-depth interviews were conducted with 88 deaf and hard-of-hearing students and a comparison group of 88 hearing counterparts. Results demonstrate that despite the important contribution of rehabilitation and integration to the well-being of these children, these efforts may put the child at risk of increased use of the false-self defense. The empirical findings suggest two general theoretical conclusions: (1) The Winnicottian concept of the environment, usually confined to the parent-child relationship, can be understood more broadly as including cultural, social, and rehabilitational variables that both influence the parent-child relationship and operate independently of it. (2) The monolithic conceptualization of the false self may be more accurately unpacked to reveal two distinct subtypes: the compliant and the split false self.

  1. Algorithms for personalized therapy of type 2 diabetes: results of a web-based international survey

    PubMed Central

    Gallo, Marco; Mannucci, Edoardo; De Cosmo, Salvatore; Gentile, Sandro; Candido, Riccardo; De Micheli, Alberto; Di Benedetto, Antonino; Esposito, Katherine; Genovese, Stefano; Medea, Gerardo; Ceriello, Antonio

    2015-01-01

    Objective In recent years increasing interest in the issue of treatment personalization for type 2 diabetes (T2DM) has emerged. This international web-based survey aimed to evaluate opinions of physicians about tailored therapeutic algorithms developed by the Italian Association of Diabetologists (AMD) and available online, and to get suggestions for future developments. Another aim of this initiative was to assess whether the online advertising and the survey would have increased the global visibility of the AMD algorithms. Research design and methods The web-based survey, which comprised five questions, has been available from the homepage of the web-version of the journal Diabetes Care throughout the month of December 2013, and on the AMD website between December 2013 and September 2014. Participation was totally free and responders were anonymous. Results Overall, 452 physicians (M=58.4%) participated in the survey. Diabetologists accounted for 76.8% of responders. The results of the survey show wide agreement (>90%) by participants on the utility of the algorithms proposed, even if they do not cover all possible needs of patients with T2DM for a personalized therapeutic approach. In the online survey period and in the months after its conclusion, a relevant and durable increase in the number of unique users who visited the websites was registered, compared to the period preceding the survey. Conclusions Patients with T2DM are heterogeneous, and there is interest toward accessible and easy to use personalized therapeutic algorithms. Responders opinions probably reflect the peculiar organization of diabetes care in each country. PMID:26301097

  2. A Community-Based Activities Survey: Systematically Determining the Impact on and of Faculty

    ERIC Educational Resources Information Center

    Perry, Lane; Farmer, Betty; Onder, David; Tanner, Benjamin; Burton, Carol

    2015-01-01

    As a descriptive case study from Western Carolina University (WCU), this article describes the development of a measuring, monitoring, and tracking system (the WCU Community-based Activities Survey) for faculty engagement in, adoption of, and impact through community engagement practices both internal and external to their courses. This paper will…

  3. Sexually Transmitted Diseases and Risk Behaviors among California Farmworkers: Results from a Population-Based Survey

    ERIC Educational Resources Information Center

    Brammeier, Monique; Chow, Joan M.; Samuel, Michael C.; Organista, Kurt C.; Miller, Jamie; Bolan, Gail

    2008-01-01

    Context: The prevalence of sexually transmitted diseases and associated risk behaviors among California farmworkers is not well described. Purpose: To estimate the prevalence of sexually transmitted diseases (STDs) and associated risk behaviors among California farmworkers. Methods: Cross-sectional analysis of population-based survey data from 6…

  4. Special Educators' Guide to Exemplary Curricula: Results of a National Field-Based Survey (1983).

    ERIC Educational Resources Information Center

    Ash, Paul M., Comp.

    A listing of 178 curriculum guides for exemplary special education programs is presented, based on a national 1983 survey of over 500 programs. For each program, information is presented on the title of the guide, the source and source address, development date, price, number of pages, recommended exceptionality area(s), and recommended level(s).…

  5. Medical Students' Experiences with Addicted Patients: A Web-Based Survey

    ERIC Educational Resources Information Center

    Midmer, Deana; Kahan, Meldon; Wilson, Lynn

    2008-01-01

    Project CREATE was an initiative to strengthen undergraduate medical education in addictions. As part of a needs assessment, forty-six medical students at Ontario's five medical schools completed a bi-weekly, interactive web-based survey about addiction-related learning events. In all, 704 unique events were recorded, for an average of 16.7…

  6. Social Workers' Orientation toward the Evidence-Based Practice Process: A Dutch Survey

    ERIC Educational Resources Information Center

    van der Zwet, Renske J. M.; Kolmer, Deirdre M. Beneken genaamd; Schalk, René

    2016-01-01

    Objectives: This study assesses social workers' orientation toward the evidence-based practice (EBP) process and explores which specific variables (e.g. age) are associated. Methods: Data were collected from 341 Dutch social workers through an online survey which included a Dutch translation of the EBP Process Assessment Scale (EBPPAS), along with…

  7. School Nutrition Directors are Receptive to Web-Based Training Opportunities: A National Survey

    ERIC Educational Resources Information Center

    Zoellner, Jamie; Carr, Deborah H.

    2009-01-01

    Purpose/Objective: The purpose of this study was to investigate school nutrition directors' (SNDs) previous experience with web-based training (WBT), interest in utilizing WBT within 14 functional areas, and logistical issues (time, price, educational credits, etc.) of developing and delivering WBT learning modules. Methods: A survey was developed…

  8. What Attributes Determine Severity of Function in Autism? A Web-Based Survey of Stakeholders

    ERIC Educational Resources Information Center

    Di Rezze, Briano; Rosenbaum, Peter; Zwaigenbaum, Lonnie

    2012-01-01

    Service providers and researchers in autism spectrum disorders (ASD) are challenged to categorize clinical variation in function. Classification systems for children with cerebral palsy have enabled clinicians and families to describe levels of function. A web-based survey engaged international ASD stakeholders to advise on considerations of…

  9. HIV/AIDS Misconceptions among Latinos: Findings from a Population-Based Survey of California Adults

    ERIC Educational Resources Information Center

    Ritieni, Assunta; Moskowitz, Joel; Tholandi, Maya

    2008-01-01

    Misconceptions about HIV/AIDS among Latino adults (N=454) in California were examined using data from a population-based telephone survey conducted in 2000. Common misconceptions concerning modes of HIV transmission included transmission via mosquito or animal bite (64.1%), public facilities (48.3%), or kissing someone on the cheek (24.8%). A…

  10. Proficiency-Based Credit Assessment: A National and Statewide Survey of Use.

    ERIC Educational Resources Information Center

    Institute for Behavioral Research in Creativity, Salt Lake City, UT.

    In recent years a movement has developed at the high school level to address issues of cost and educational improvement through Proficiency-Based Credit Assessment (PBCA). With PBCA, students can receive credit by demonstrating their proficiency without actually taking a course. A national mail survey conducted by the Utah State Office of…

  11. Our Environment, Our Health: A Community-Based Participatory Environmental Health Survey in Richmond, California

    ERIC Educational Resources Information Center

    Cohen, Alison; Lopez, Andrea; Malloy, Nile; Morello-Frosch, Rachel

    2012-01-01

    This study presents a health survey conducted by a community-based participatory research partnership between academic researchers and community organizers to consider environmental health and environmental justice issues in four neighborhoods of Richmond, California, a low-income community of color living along the fence line of a major oil…

  12. Are Web-based Surveys the New Epidemiological Mode For Healthcare Research? - The Saudi Perspective.

    PubMed

    Bahkali, Salwa; Almaiman, Ahmad; Al-Nasser, Lubna; Elmetwally, Ashraf; Househ, Mowafa

    2014-01-01

    Web-based surveys (WBS) are gaining popularity as simple, cost-effective and rapid tools for data collection in healthcare research. The purpose of this exploratory study is to summarize the role of website-based survey (WBS) in gathering data for epidemiological research and review challenges facing WBSs in healthcare research. Electronic search in related literature for advantages, characteristics and performance of WBS was conducted. Special focus was placed on WBSs use in the Saudi context. The results indicated that WBSs are being employed increasingly in Healthcare research because of their accessibility, rapidity, reduced cost, less need for human resources and elimination of human errors. However, Paper-based surveys still possess higher response rates, accuracy and are more applicable for probability-sampling methods. Selection bias favoring young, educated and professional participants is associated with WBSs. Most Saudi studies sampled either healthcare professionals or students as the target population. Challenges specific to WBSs include: accessibility, representativeness of sampled population and multiple responses/unintended participants. Although WBSs do appear to be a contemporary alternative to traditional paper-based survey tools, more research is needed to optimize their applicability in different cultural contexts and for diverse pools of participants.

  13. Evidence-Based Speech-Language Pathology Practices in Schools: Findings from a National Survey

    ERIC Educational Resources Information Center

    Hoffman, LaVae M.; Ireland, Marie; Hall-Mills, Shannon; Flynn, Perry

    2013-01-01

    Purpose: This study documented evidence-based practice (EBP) patterns as reported by speech-language pathologists (SLPs) employed in public schools during 2010-2011. Method: Using an online survey, practioners reported their EBP training experiences, resources available in their workplaces, and the frequency with which they engage in specific EBP…

  14. Empirical agreement in model validation.

    PubMed

    Jebeile, Julie; Barberousse, Anouk

    2016-04-01

    Empirical agreement is often used as an important criterion when assessing the validity of scientific models. However, it is by no means a sufficient criterion as a model can be so adjusted as to fit available data even though it is based on hypotheses whose plausibility is known to be questionable. Our aim in this paper is to investigate into the uses of empirical agreement within the process of model validation.

  15. Short-term memory of TiO2-based electrochemical capacitors: empirical analysis with adoption of a sliding threshold

    NASA Astrophysics Data System (ADS)

    Lim, Hyungkwang; Kim, Inho; Kim, Jin-Sang; Hwang, Cheol Seong; Jeong, Doo Seok

    2013-09-01

    Chemical synapses are important components of the large-scaled neural network in the hippocampus of the mammalian brain, and a change in their weight is thought to be in charge of learning and memory. Thus, the realization of artificial chemical synapses is of crucial importance in achieving artificial neural networks emulating the brain’s functionalities to some extent. This kind of research is often referred to as neuromorphic engineering. In this study, we report short-term memory behaviours of electrochemical capacitors (ECs) utilizing TiO2 mixed ionic-electronic conductor and various reactive electrode materials e.g. Ti, Ni, and Cr. By experiments, it turned out that the potentiation behaviours did not represent unlimited growth of synaptic weight. Instead, the behaviours exhibited limited synaptic weight growth that can be understood by means of an empirical equation similar to the Bienenstock-Cooper-Munro rule, employing a sliding threshold. The observed potentiation behaviours were analysed using the empirical equation and the differences between the different ECs were parameterized.

  16. Experience base for Radioactive Waste Thermal Processing Systems: A preliminary survey

    SciTech Connect

    Mayberry, J.; Geimer, R.; Gillins, R.; Steverson, E.M.; Dalton, D.; Anderson, G.L.

    1992-04-01

    In the process of considering thermal technologies for potential treatment of the Idaho National Engineering Laboratory mixed transuranic contaminated wastes, a preliminary survey of the experience base available from Radioactive Waste Thermal Processing Systems is reported. A list of known commercial radioactive waste facilities in the United States and some international thermal treatment facilities are provided. Survey focus is upon the US Department of Energy thermal treatment facilities. A brief facility description and a preliminary summary of facility status, and problems experienced is provided for a selected subset of the DOE facilities.

  17. Synchrotron-based organics and mineralogical survey of threeStardust tracks

    SciTech Connect

    Westphal, A.J.; Bajt, S.; Butterworth, A.L.; Fakra, S.; Gainsforth, Z.; Marcus, M.A.; Martin, M.C.; Snead, C.J.; Tyliszczak, T.

    2007-07-01

    Wild2 particles captured in the Stardust cometary dustcollector exhibit remarkable diversity between and within individualparticle tracks in aerogel. Several particles studied during the StardustPreliminary Examination are, so far, unique in the collection[1]. Thereis an urgent need to survey particles in tracks and to target specificparticles for extraction and destructive analysis. Here we describe an inaerogel survey at the Advanced Light Source (LBNL) of three tracks usingsynchrotron-based Fourier Transform Infrared Spectroscopy (FTIR),microbeam X-ray fluorescence (microXRF), X-ray Absorption Near-EdgeSpectroscopy (microXANES), X-ray Diffraction (XRD), and ScanningTransmission X-ray Microscopy (STXM).

  18. Measuring Model-Based High School Science Instruction: Development and Application of a Student Survey

    NASA Astrophysics Data System (ADS)

    Fulmer, Gavin W.; Liang, Ling L.

    2013-02-01

    This study tested a student survey to detect differences in instruction between teachers in a modeling-based science program and comparison group teachers. The Instructional Activities Survey measured teachers' frequency of modeling, inquiry, and lecture instruction. Factor analysis and Rasch modeling identified three subscales, Modeling and Reflecting, Communicating and Relating, and Investigative Inquiry. As predicted, treatment group teachers engaged in modeling and inquiry instruction more than comparison teachers, with effect sizes between 0.55 and 1.25. This study demonstrates the utility of student report data in measuring teachers' classroom practices and in evaluating outcomes of a professional development program.

  19. A Household-Based Distribution-Sensitive Human Development Index: An Empirical Application to Mexico, Nicaragua and Peru

    ERIC Educational Resources Information Center

    Lopez-Calva, Luis F.; Ortiz-Juarez, Eduardo

    2012-01-01

    In measuring human development, one of the main concerns relates to the inclusion of a measure that penalizes inequalities in the distribution of achievements across the population. Using indicators from nationally representative household surveys and census data, this paper proposes a straightforward methodology to estimate a household-based…

  20. A Comparative Investigation of TPB and Altruism Frameworks for an Empirically Based Communication Approach to Enhance Paper Recycling

    ERIC Educational Resources Information Center

    Chaisamrej, Rungrat; Zimmerman, Rick S.

    2014-01-01

    This research compared the ability of the theory of planned behavior (TPB) and the altruism framework (AM) to predict paper-recycling behavior. It was comprised of formative research and a major survey. Data collected from 628 undergraduate students in Thailand were analyzed using structural equation modeling. Results showed that TPB was superior…

  1. TESTING GROUND BASED GEOPHYSICAL TECHNIQUES TO REFINE ELECTROMAGNETIC SURVEYS NORTH OF THE 300 AREA HANFORD WASHINGTON

    SciTech Connect

    PETERSEN SW

    2010-12-02

    Airborne electromagnetic (AEM) surveys were flown during fiscal year (FY) 2008 within the 600 Area in an attempt to characterize the underlying subsurface and to aid in the closure and remediation design study goals for the 200-PO-1 Groundwater Operable Unit (OU). The rationale for using the AEM surveys was that airborne surveys can cover large areas rapidly at relatively low costs with minimal cultural impact, and observed geo-electrical anomalies could be correlated with important subsurface geologic and hydrogeologic features. Initial interpretation of the AEM surveys indicated a tenuous correlation with the underlying geology, from which several anomalous zones likely associated with channels/erosional features incised into the Ringold units were identified near the River Corridor. Preliminary modeling resulted in a slightly improved correlation but revealed that more information was required to constrain the modeling (SGW-39674, Airborne Electromagnetic Survey Report, 200-PO-1 Groundwater Operable Unit, 600 Area, Hanford Site). Both time-and frequency domain AEM surveys were collected with the densest coverage occurring adjacent to the Columbia River Corridor. Time domain surveys targeted deeper subsurface features (e.g., top-of-basalt) and were acquired using the HeliGEOTEM{reg_sign} system along north-south flight lines with a nominal 400 m (1,312 ft) spacing. The frequency domain RESOLVE system acquired electromagnetic (EM) data along tighter spaced (100 m [328 ft] and 200 m [656 ft]) north-south profiles in the eastern fifth of the 200-PO-1 Groundwater OU (immediately adjacent to the River Corridor). The overall goal of this study is to provide further quantification of the AEM survey results, using ground based geophysical methods, and to link results to the underlying geology and/or hydrogeology. Specific goals of this project are as follows: (1) Test ground based geophysical techniques for the efficacy in delineating underlying geology; (2) Use ground

  2. HI4PI: A full-sky H I survey based on EBHIS and GASS

    NASA Astrophysics Data System (ADS)

    HI4PI Collaboration; Ben Bekhti, N.; Flöer, L.; Keller, R.; Kerp, J.; Lenz, D.; Winkel, B.; Bailin, J.; Calabretta, M. R.; Dedes, L.; Ford, H. A.; Gibson, B. K.; Haud, U.; Janowiecki, S.; Kalberla, P. M. W.; Lockman, F. J.; McClure-Griffiths, N. M.; Murphy, T.; Nakanishi, H.; Pisano, D. J.; Staveley-Smith, L.

    2016-10-01

    Context. Measurement of the Galactic neutral atomic hydrogen (H I) column density, NH I, and brightness temperatures, TB, is of high scientific value for a broad range of astrophysical disciplines. In the past two decades, one of the most-used legacy H I datasets has been the Leiden/Argentine/Bonn Survey (LAB). Aims: We release the H I 4π survey (HI4PI), an all-sky database of Galactic H I, which supersedes the LAB survey. Methods: The HI4PI survey is based on data from the recently completed first coverage of the Effelsberg-Bonn H I Survey (EBHIS) and from the third revision of the Galactic All-Sky Survey (GASS). EBHIS and GASS share similar angular resolution and match well in sensitivity. Combined, they are ideally suited to be a successor to LAB. Results: The new HI4PI survey outperforms the LAB in angular resolution (ϑFWHM = 16´´.2) and sensitivity (σrms = 43 mK). Moreover, it has full spatial sampling and thus overcomes a major drawback of LAB, which severely undersamples the sky. We publish all-sky column density maps of the neutral atomic hydrogen in the Milky Way, along with full spectroscopic data, in several map projections including HEALPix. HI4PI datasets are only available at the CDS via anonymous ftp to http://cdsarc.u-strasbg.fr (http://130.79.128.5) or via http://cdsarc.u-strasbg.fr/viz-bin/qcat?J/A+A/594/A116

  3. Our environment, our health: a community-based participatory environmental health survey in Richmond, California.

    PubMed

    Cohen, Alison; Lopez, Andrea; Malloy, Nile; Morello-Frosch, Rachel

    2012-04-01

    This study presents a health survey conducted by a community-based participatory research partnership between academic researchers and community organizers to consider environmental health and environmental justice issues in four neighborhoods of Richmond, California, a low-income community of color living along the fence line of a major oil refinery and near other industrial and mobile sources of pollution. The Richmond health survey aimed to assess local concerns and perceptions of neighborhood conditions, health problems, mobile and stationary hazards, access to health care, and other issues affecting residents of Richmond. Although respondents thought their neighborhoods were good places to live, they expressed concerns about neighborhood stressors and particular sources of pollution, and identified elevated asthma rates for children and long-time Richmond residents. The Richmond health survey offers a holistic, community-centered perspective to understanding local environmental health issues, and can inform future environmental health research and organizing efforts for community-university collaboratives.

  4. 1995 Area 1 bird survey/Zone 1, Operable Unit 2, Robins Air Force Base, Georgia

    SciTech Connect

    Wade, M.C.

    1995-08-01

    Robins Air Force Base is located in Warner Robins, Georgia, approximately 90 miles southeast of Atlanta, Georgia. As part of the Baseline Investigation (CDM Federal 1994) a two day bird survey was conducted by M. C. Wade (Oak Ridge National Laboratory) and B.A. Beatty (CDM Federal Programs) in May 1995. The subject area of investigation includes the sludge lagoon, Landfill No. 4, and the wetland area east of the landfill and west of Hannah Road (including two ponds). This is known as Area 1. The Area 1 wetlands include bottomland hardwood forest, stream, and pond habitats. The objectives of this survey were to document bird species using the Area I wetlands and to see if the change in hydrology (due to the installation of the Sewage Treatment Plant effluent diversion and stormwater runon control systems) has resulted in changes at Area 1 since the previous survey of May 1992 (CDM Federal 1994).

  5. Development, use, and availability of a job exposure matrix based on national occupational hazard survey data.

    PubMed

    Sieber, W K; Sundin, D S; Frazier, T M; Robinson, C F

    1991-01-01

    A job exposure matrix has been developed based on potential exposure data collected during the 1972-1974 National Occupational Hazard Survey (NOHS). The survey sample was representative of all U.S. non-agricultural businesses covered under the Occupational Safety and Health Act of 1970 and employing eight or more employees. Potential worker exposure to all chemical, physical, or biological agents was recorded during the field survey if certain minimum guidelines for exposure were met. The job exposure matrix (JEM) itself is a computerized database that assists the user in determining potential chemical or physical exposures in occupational settings. We describe the structure and possible uses of the job exposure matrix. In one example, potential occupational exposures to elemental lead were grouped by industry and occupation. In a second example, the matrix was used to determine exposure classifications in a hypothetical case-control study. Present availability as well as future enhancements of the job exposure matrix are described.

  6. Improving Inpatient Surveys: Web-Based Computer Adaptive Testing Accessed via Mobile Phone QR Codes

    PubMed Central

    2016-01-01

    Background The National Health Service (NHS) 70-item inpatient questionnaire surveys inpatients on their perceptions of their hospitalization experience. However, it imposes more burden on the patient than other similar surveys. The literature shows that computerized adaptive testing (CAT) based on item response theory can help shorten the item length of a questionnaire without compromising its precision. Objective Our aim was to investigate whether CAT can be (1) efficient with item reduction and (2) used with quick response (QR) codes scanned by mobile phones. Methods After downloading the 2008 inpatient survey data from the Picker Institute Europe website and analyzing the difficulties of this 70-item questionnaire, we used an author-made Excel program using the Rasch partial credit model to simulate 1000 patients’ true scores followed by a standard normal distribution. The CAT was compared to two other scenarios of answering all items (AAI) and the randomized selection method (RSM), as we investigated item length (efficiency) and measurement accuracy. The author-made Web-based CAT program for gathering patient feedback was effectively accessed from mobile phones by scanning the QR code. Results We found that the CAT can be more efficient for patients answering questions (ie, fewer items to respond to) than either AAI or RSM without compromising its measurement accuracy. A Web-based CAT inpatient survey accessed by scanning a QR code on a mobile phone was viable for gathering inpatient satisfaction responses. Conclusions With advances in technology, patients can now be offered alternatives for providing feedback about hospitalization satisfaction. This Web-based CAT is a possible option in health care settings for reducing the number of survey items, as well as offering an innovative QR code access. PMID:26935793

  7. A robust method for analyzing the instantaneous attributes of seismic data: The instantaneous frequency estimation based on ensemble empirical mode decomposition

    NASA Astrophysics Data System (ADS)

    Li, Xiangfang; Chen, Wenchao; Zhou, Yanhui

    2014-12-01

    The Hilbert-Huang transform (HHT) includes two procedures. First, empirical mode decomposition (EMD) is used to decompose signals into several intrinsic mode functions (IMFs) before the Hilbert transform (HT) of these IMFs are calculated. Compared to the conventional Hilbert transform (HT), HHT is more sensitive to thickness variations of seismic beds. However, random noise in seismic signal may cause the mixture of the modes from HHT. The recent ensemble empirical mode decomposition (EEMD) presents the advantages in decreasing mode mixture and has the promising potential in seismic signal analysis. Currently, EEMD is mainly used in seismic spectral decomposition and noise attenuation. We extend the application of EEMD based instantaneous frequency to the analysis of bed thickness. The tests on complex Marmousi2 model and a 2D field data show that EEMD is more effective in weakening mode mixtures contained in the IMFs, compared with that calculated by EMD. Furthermore, the EEMD based instantaneous frequency is more sensitive to the seismic thickness variation than that based on EMD and more consistent with the stratigraphic structure, which means that E-IFPs are more advantageous in characterizing reservoirs.

  8. A 16-year examination of domestic violence among Asians and Asian Americans in the empirical knowledge base: a content analysis.

    PubMed

    Yick, Alice G; Oomen-Early, Jody

    2008-08-01

    Until recently, research studies have implied that domestic violence does not affect Asian American and immigrant communities, or even Asians abroad, because ethnicity or culture has not been addressed. In this content analysis, the authors examined trends in publications in leading scholarly journals on violence relating to Asian women and domestic violence. A coding schema was developed, with two raters coding the data with high interrater reliability. Sixty articles were published over the 16 years studied, most atheoretical and focusing on individual levels of analysis. The terms used in discussing domestic violence reflected a feminist perspective. Three quarters of the studies were empirical, with most guided by logical positivism using quantitative designs. Most targeted specific Asian subgroups (almost a third focused on Asian Indians) rather than categorizing Asians as a general ethnic category. The concept of "Asian culture" was most often assessed by discussing Asian family structure. Future research is discussed in light of the findings.

  9. NIRS-based classification of clench force and speed motor imagery with the use of empirical mode decomposition for BCI.

    PubMed

    Yin, Xuxian; Xu, Baolei; Jiang, Changhao; Fu, Yunfa; Wang, Zhidong; Li, Hongyi; Shi, Gang

    2015-03-01

    Near-infrared spectroscopy (NIRS) is a non-invasive optical technique used for brain-computer interface (BCI). This study aims to investigate the brain hemodynamic responses of clench force and speed motor imagery and extract task-relevant features to obtain better classification performance. Given the non-stationary characteristics of real hemodynamic measurements, empirical mode decomposition (EMD) was applied to reduce the physiological noise overwhelmed in the task-relevant NIRS signals. Compared with continuous wavelet decomposition, EMD does not require a pre-determined basis function. EMD decomposes the original signals into a set of intrinsic mode functions (IMFs). In this study, joint mutual information was applied to select the optimal features, and support vector machine was used as a classifier. Offline and pseudo-online analyses showed that the most feasible classification accuracy can be obtained using IMFs as input features. Accordingly, an alternative feature is provided to develop the NIRS-BCI system.

  10. Empirical model of global thermospheric temperature and composition based on data from the OGO-6 quadrupole mass spectrometer

    NASA Technical Reports Server (NTRS)

    Hedin, A. E.; Mayr, H. G.; Reber, C. A.; Spencer, N. W.; Carignan, G. R.

    1972-01-01

    An empirical global model for magnetically quiet conditions has been derived from longitudinally averaged N2, O, and He densities by means of an expansion in spherical harmonics. The data were obtained by the OGO-6 neutral mass spectrometer and cover the altitude range 400 to 600 km for the period 27 June 1969 to 13 May 1971. The accuracy of the analytical description is of the order of the experimental error for He and O and about three times experimental error for N2, thus providing a reasonable overall representation of the satellite observations. Two model schemes are used: one representing densities extrapolated to 450 km and one representing densities extrapolated to 120 km with exospheric temperatures inferred from N2 densities. Using the best fit model parameters the global thermospheric structure is presented in the form of a number of contour plots.

  11. A 16-year examination of domestic violence among Asians and Asian Americans in the empirical knowledge base: a content analysis.

    PubMed

    Yick, Alice G; Oomen-Early, Jody

    2008-08-01

    Until recently, research studies have implied that domestic violence does not affect Asian American and immigrant communities, or even Asians abroad, because ethnicity or culture has not been addressed. In this content analysis, the authors examined trends in publications in leading scholarly journals on violence relating to Asian women and domestic violence. A coding schema was developed, with two raters coding the data with high interrater reliability. Sixty articles were published over the 16 years studied, most atheoretical and focusing on individual levels of analysis. The terms used in discussing domestic violence reflected a feminist perspective. Three quarters of the studies were empirical, with most guided by logical positivism using quantitative designs. Most targeted specific Asian subgroups (almost a third focused on Asian Indians) rather than categorizing Asians as a general ethnic category. The concept of "Asian culture" was most often assessed by discussing Asian family structure. Future research is discussed in light of the findings. PMID:18259048

  12. Men who have sex with men: a comparison of a probability sample survey and a community based study.

    PubMed

    Dodds, J P; Mercer, C H; Mercey, D E; Copas, A J; Johnson, A M

    2006-02-01

    We compared characteristics of men who have sex with men (MSM) in a probability sample survey with a community based study in London. The majority of men in both surveys reported male sex partner(s) in the last year but MSM recruited through the population based survey had lower levels of HIV risk behaviour, reported fewer sexually transmitted infections and HIV testing than those recruited from gay venues. Community samples are likely to overestimate levels of risk behaviour among all MSM.

  13. Sources of Error in Substance Use Prevalence Surveys

    PubMed Central

    Johnson, Timothy P.

    2014-01-01

    Population-based estimates of substance use patterns have been regularly reported now for several decades. Concerns with the quality of the survey methodologies employed to produce those estimates date back almost as far. Those concerns have led to a considerable body of research specifically focused on understanding the nature and consequences of survey-based errors in substance use epidemiology. This paper reviews and summarizes that empirical research by organizing it within a total survey error model framework that considers multiple types of representation and measurement errors. Gaps in our knowledge of error sources in substance use surveys and areas needing future research are also identified. PMID:27437511

  14. An HIV epidemic model based on viral load dynamics: value in assessing empirical trends in HIV virulence and community viral load.

    PubMed

    Herbeck, Joshua T; Mittler, John E; Gottlieb, Geoffrey S; Mullins, James I

    2014-06-01

    Trends in HIV virulence have been monitored since the start of the AIDS pandemic, as studying HIV virulence informs our understanding of HIV epidemiology and pathogenesis. Here, we model changes in HIV virulence as a strictly evolutionary process, using set point viral load (SPVL) as a proxy, to make inferences about empirical SPVL trends from longitudinal HIV cohorts. We develop an agent-based epidemic model based on HIV viral load dynamics. The model contains functions for viral load and transmission, SPVL and disease progression, viral load trajectories in multiple stages of infection, and the heritability of SPVL across transmissions. We find that HIV virulence evolves to an intermediate level that balances infectiousness with longer infected lifespans, resulting in an optimal SPVL∼4.75 log10 viral RNA copies/mL. Adaptive viral evolution may explain observed HIV virulence trends: our model produces SPVL trends with magnitudes that are broadly similar to empirical trends. With regard to variation among studies in empirical SPVL trends, results from our model suggest that variation may be explained by the specific epidemic context, e.g. the mean SPVL of the founding lineage or the age of the epidemic; or improvements in HIV screening and diagnosis that results in sampling biases. We also use our model to examine trends in community viral load, a population-level measure of HIV viral load that is thought to reflect a population's overall transmission potential. We find that community viral load evolves in association with SPVL, in the absence of prevention programs such as antiretroviral therapy, and that the mean community viral load is not necessarily a strong predictor of HIV incidence.

  15. Difficulties with telephone-based surveys on alcohol in high-income countries: the Canadian example

    PubMed Central

    Shield, Kevin D.; Rehm, Jürgen

    2012-01-01

    Accurate information concerning alcohol consumption level and patterns is vital to formulating public health policy. The objective of this paper is to critically assess the extent to which survey design, response rate and alcohol consumption coverage obtained in random digit dialing, telephone-based surveys impact on conclusions about alcohol consumption and its patterns in the general population. Our analysis will be based on the Canadian Alcohol and Drug Use Monitoring Survey (CADUMS) 2008, a national survey intended to be representative of the general population. The conclusions of this paper are as follows: 1) ignoring people who are homeless, institutionalized and/or do not have a home phone may lead to an underestimation of the prevalence of alcohol consumption and related problems; 2) weighting of observations to population demographics may lead to a increase in the design effect, does not necessarily address the underlying selection bias, and may lead to overly influential observations; and 3) the accurate characterization of alcohol consumption patterns obtained by triangulating the data with the adult per capita consumption estimate is essential for comparative analyses and intervention planning especially when the alcohol coverage rate is low like in the CADUMS with 34%. PMID:22337654

  16. Survey of Selected Installations Actively Searching the ERIC Magnetic Tape Data Base in Batch Mode. Final Report. Volume I.

    ERIC Educational Resources Information Center

    Humphrey, Allan J.

    Findings of a survey of 29 selected information centers that use computers to search the Educational Resources Information Center (ERIC) data base in batch mode are reported. The survey was conducted from December 1972 to May 1973. The report includes some general observations based on the information gathered, a brief description of two "typical"…

  17. Empirical and theoretical bacterial diversity in four Arizona soils.

    PubMed

    Dunbar, John; Barns, Susan M; Ticknor, Lawrence O; Kuske, Cheryl R

    2002-06-01

    Understanding patterns of biodiversity in microbial communities is severely constrained by the difficulty of adequately sampling these complex systems. We illustrate the problem with empirical data from small surveys (200-member 16S rRNA gene clone libraries) of four bacterial soil communities from two locations in Arizona. Among the four surveys, nearly 500 species-level groups ( Dunbar et al., Appl. Environ. Microbiol. 65:662-1669, 1999) and 21 bacterial divisions were documented, including four new candidate divisions provisionally designated SC1, SC2, SC3, and SC4. We devised a simple approach to constructing theoretical null models of bacterial species abundance. These null models provide, for the first time, detailed descriptions of soil bacterial community structure that can be used to guide experimental design. Models based on a lognormal distribution were consistent with the observed sizes of the four communities and the richness of the clone surveys. Predictions from the models showed that the species richness of small surveys from complex communities is reproducible, whereas the species composition is not. By using the models, we can now estimate the required survey scale to document specified fractions of community diversity. For example, documentation of half the species in each model community would require surveys of 16,284 to 44,000 individuals. However, quantitative comparisons of half the species in two communities would require surveys at least 10-fold larger for each community. PMID:12039765

  18. The Importance of Adhering to Details of the Total Design Method (TDM) for Mail Surveys.

    ERIC Educational Resources Information Center

    Dillman, Don A.; And Others

    1984-01-01

    The empirical effects of adherence of details of the Total Design Method (TDM) approach to the design of mail surveys is discussed, based on the implementation of a common survey in 11 different states. The results suggest that greater adherence results in higher response, especially in the later stages of the TDM. (BW)

  19. Twitter Strategies for Web-Based Surveying: Descriptive Analysis From the International Concussion Study

    PubMed Central

    Düking, Peter; Mellalieu, Stephen D

    2016-01-01

    Background Social media provides researchers with an efficient means to reach and engage with a large and diverse audience. Twitter allows for the virtual social interaction among a network of users that enables researchers to recruit and administer surveys using snowball sampling. Although using Twitter to administer surveys for research is not new, strategies to improve response rates are yet to be reported. Objective To compare the potential and actual reach of 2 Twitter accounts that administered a Web-based concussion survey to rugby players and trainers using 2 distinct Twitter-targeting strategies. Furthermore, the study sought to determine the likelihood of receiving a retweet based on the time of the day and day of the week of posting. Methods A survey based on previous concussion research was exported to a Web-based survey website Survey Monkey. The survey comprised 2 questionnaires, one for players, and one for those involved in the game (eg, coaches and athletic trainers). The Web-based survey was administered using 2 existing Twitter accounts, with each account executing a distinct targeting strategy. A list of potential Twitter accounts to target was drawn up, together with a list of predesigned tweets. The list of accounts to target was divided into ‘High-Profile’ and ‘Low-Profile’, based on each accounts’ position to attract publicity with a high social interaction potential. The potential reach (number of followers of the targeted account), and actual reach (number of retweets received by each post) between the 2 strategies were compared. The number of retweets received by each account was further analyzed to understand when the most likely time of day, and day of the week, a retweet would be received. Results The number of retweets received by a Twitter account decreased by 72% when using the ‘high-profile strategy’ compared with the ‘low-profile strategy’ (incidence rate ratio (IRR); 0.28, 95% confidence interval (CI) 0

  20. Theoretical modeling of stream potholes based upon empirical observations from the Orange River, Republic of South Africa

    NASA Astrophysics Data System (ADS)

    Springer, Gregory S.; Tooth, Stephen; Wohl, Ellen E.

    2006-12-01

    Potholes carved into streambeds can be important components of channel incision, but they have received little quantitative attention. Here empirical evidence is presented from three sites along the Orange River, Republic of South Africa that demonstrates that the pothole dimensions of radius and depth are strongly correlated using a simple power law. Where radius is the dependent variable, the exponent of the power law describes the rate of increase in radius with increasing depth. Erosion within potholes is complexly related to erosion on the adjacent bed. Erosion efficiencies within small, hemispherical potholes must be high if the potholes are to survive in the face of bed translation (incision). As potholes deepen, however, the necessary efficiencies decline rapidly. Increasing concavity associated with growth imposes stricter constraints; comparatively deep potholes must erode orders of magnitude larger volumes of substrate than shallower potholes in the face of bed retreat. Hemispherical potholes are eventually converted to cylindrical potholes, the geometries of which favor enlargement while they are small. Geometric models constructed using the power law show unambiguously that more substrate is eroded by volume from cylindrical pothole walls during growth than from cylindrical pothole floors. Grinders thus play a secondary role to suspended sediment entrained within the vortices that occur in potholes. Continued growth leads to coalescence with other potholes or destruction through block detachment depending on local geology. The combination of geology and erosion mechanisms may determine whether a strath or inner channel develops as a consequence of the process.

  1. Single-Channel EMG Classification With Ensemble-Empirical-Mode-Decomposition-Based ICA for Diagnosing Neuromuscular Disorders.

    PubMed

    Naik, Ganesh R; Selvan, S Easter; Nguyen, Hung T

    2016-07-01

    An accurate and computationally efficient quantitative analysis of electromyography (EMG) signals plays an inevitable role in the diagnosis of neuromuscular disorders, prosthesis, and several related applications. Since it is often the case that the measured signals are the mixtures of electric potentials that emanate from surrounding muscles (sources), many EMG signal processing approaches rely on linear source separation techniques such as the independent component analysis (ICA). Nevertheless, naive implementations of ICA algorithms do not comply with the task of extracting the underlying sources from a single-channel EMG measurement. In this respect, the present work focuses on a classification method for neuromuscular disorders that deals with the data recorded using a single-channel EMG sensor. The ensemble empirical mode decomposition algorithm decomposes the single-channel EMG signal into a set of noise-canceled intrinsic mode functions, which in turn are separated by the FastICA algorithm. A reduced set of five time domain features extracted from the separated components are classified using the linear discriminant analysis, and the classification results are fine-tuned with a majority voting scheme. The performance of the proposed method has been validated with a clinical EMG database, which reports a higher classification accuracy (98%). The outcome of this study encourages possible extension of this approach to real settings to assist the clinicians in making correct diagnosis of neuromuscular disorders. PMID:26173218

  2. Natural language processing-based COTS software and related technologies survey.

    SciTech Connect

    Stickland, Michael G.; Conrad, Gregory N.; Eaton, Shelley M.

    2003-09-01

    Natural language processing-based knowledge management software, traditionally developed for security organizations, is now becoming commercially available. An informal survey was conducted to discover and examine current NLP and related technologies and potential applications for information retrieval, information extraction, summarization, categorization, terminology management, link analysis, and visualization for possible implementation at Sandia National Laboratories. This report documents our current understanding of the technologies, lists software vendors and their products, and identifies potential applications of these technologies.

  3. A survey of partition-based techniques for copy-move forgery detection.

    PubMed

    Diane, Wandji Nanda Nathalie; Xingming, Sun; Moise, Fah Kue

    2014-01-01

    A copy-move forged image results from a specific type of image tampering procedure carried out by copying a part of an image and pasting it on one or more parts of the same image generally to maliciously hide unwanted objects/regions or clone an object. Therefore, detecting such forgeries mainly consists in devising ways of exposing identical or relatively similar areas in images. This survey attempts to cover existing partition-based copy-move forgery detection techniques.

  4. A Survey of Partition-Based Techniques for Copy-Move Forgery Detection

    PubMed Central

    Nathalie Diane, Wandji Nanda; Xingming, Sun; Moise, Fah Kue

    2014-01-01

    A copy-move forged image results from a specific type of image tampering procedure carried out by copying a part of an image and pasting it on one or more parts of the same image generally to maliciously hide unwanted objects/regions or clone an object. Therefore, detecting such forgeries mainly consists in devising ways of exposing identical or relatively similar areas in images. This survey attempts to cover existing partition-based copy-move forgery detection techniques. PMID:25152931

  5. A design of strategic alliance based on value chain of surveying and mapping enterprises in China

    NASA Astrophysics Data System (ADS)

    Duan, Hong; Huang, Xianfeng

    2007-06-01

    In this paper, we use value chain and strategic alliance theories to analyzing the surveying and mapping Industry and enterprises. The value chain of surveying and mapping enterprises is highly-contacted but split by administrative interference, the enterprises are common small scale. According to the above things, we consider that establishing a nonequity- Holding strategic alliance based on value chain is an available way, it can not only let the enterprises share the superior resources in different sectors of the whole value chain each other but avoid offending the interests of related administrative departments, by this way, the surveying and mapping enterprises gain development respectively and totally. Then, we give the method to building up the strategic alliance model through parting the value chain and the using advantage of companies in different value chain sectors. Finally, we analyze the internal rule of strategic alliance and prove it is a suitable way to realize the development of surveying and mapping enterprises through game theory.

  6. Transportation to clinic: findings from a pilot clinic-based survey of low-income suburbanites.

    PubMed

    Silver, Diana; Blustein, Jan; Weitzman, Beth C

    2012-04-01

    Health care policymakers have cited transportation barriers as key obstacles to providing health care to low-income suburbanites, particularly because suburbs have become home to a growing number of recent immigrants who are less likely to own cars than their neighbors. In a suburb of New York City, we conducted a pilot survey of low income, largely immigrant clients in four public clinics, to find out how much transportation difficulties limit their access to primary care. Clients were receptive to the opportunity to participate in the survey (response rate = 94%). Nearly one-quarter reported having transportation problems that had caused them to miss or reschedule a clinic appointment in the past. Difficulties included limited and unreliable local bus service, and a tenuous connection to a car. Our pilot work suggests that this population is willing to participate in a survey on this topic. Further, since even among those attending clinic there was significant evidence of past transportation problems, it suggests that a population based survey would yield information about substantial transportation barriers to health care.

  7. Empirical derivation of the reference region for computing diagnostic sensitive ¹⁸fluorodeoxyglucose ratios in Alzheimer's disease based on the ADNI sample.

    PubMed

    Rasmussen, Jerod M; Lakatos, Anita; van Erp, Theo G M; Kruggel, Frithjof; Keator, David B; Fallon, James T; Macciardi, Fabio; Potkin, Steven G

    2012-03-01

    Careful selection of the reference region for non-quantitative positron emission tomography (PET) analyses is critically important for Region of Interest (ROI) data analyses. We introduce an empirical method of deriving the most suitable reference region for computing neurodegeneration sensitive (18)fluorodeoxyglucose (FDG) PET ratios based on the dataset collected by the Alzheimer's Disease Neuroimaging Initiative (ADNI) study. Candidate reference regions are selected based on a heat map of the difference in coefficients of variation (COVs) of FDG ratios over time for each of the Automatic Anatomical Labeling (AAL) atlas regions normalized by all other AAL regions. Visual inspection of the heat map suggests that the portion of the cerebellum and vermis superior to the horizontal fissure is the most sensitive reference region. Analyses of FDG ratio data show increases in significance on the order of ten-fold when using the superior portion of the cerebellum as compared with the traditionally used full cerebellum. The approach to reference region selection in this paper can be generalized to other radiopharmaceuticals and radioligands as well as to other disorders where brain changes over time are hypothesized and longitudinal data is available. Based on the empirical evidence presented in this study, we demonstrate the usefulness of the COV heat map method and conclude that intensity normalization based on the superior portion of the cerebellum may be most sensitive to measuring change when performing longitudinal analyses of FDG-PET ratios as well as group comparisons in Alzheimer's disease. This article is part of a Special Issue entitled: Imaging Brain Aging and Neurodegenerative disease. PMID:21958592

  8. An Empirical Determination of the Intergalactic Background Light Using Near-Infrared Deep Galaxy Survey Data Out to 5 Micrometers and the Gamma-Ray Opacity of the Universe

    NASA Technical Reports Server (NTRS)

    Scully, Sean T.; Malkan, Matthew A.; Stecker, Floyd W.

    2014-01-01

    We extend our previous model-independent determination of the intergalactic background light, based purely on galaxy survey data, out to a wavelength of 5 micrometers. Our approach enables us to constrain the range of photon densities, based on the uncertainties from observationally determined luminosity densities and colors. We further determine a 68% confidence upper and lower limit on the opacity of the universe to gamma-rays up to energies of 1.6/(1 + z) terraelectron volts. A comparison of our lower limit redshift-dependent opacity curves to the opacity limits derived from the results of both ground-based air Cerenkov telescope and Fermi-LAT observations of PKS 1424+240 allows us to place a new upper limit on the redshift of this source, independent of IBL modeling.

  9. Economic burden of multidrug-resistant bacteria in nursing homes in Germany: a cost analysis based on empirical data

    PubMed Central

    Huebner, Claudia; Roggelin, Marcus; Flessa, Steffen

    2016-01-01

    Objectives Infections and colonisations with multidrug-resistant organisms (MDROs) increasingly affect different types of healthcare facilities worldwide. So far, little is known about additional costs attributable to MDROs outside hospitals. The aim of this study was to analysis the economic burden of multidrug-resistant bacteria in nursing homes in Germany. Setting The cost analysis is performed from a microeconomic perspective of the healthcare facilities. Study took place in six long-term care facilities in north-eastern Germany. Participants Data of 71 residents with a positive MDRO status were included. Primary and secondary outcome measures The study analysed MDRO surveillance data from 2011 to 2013. It was supplemented by an empirical analysis to determine the burden on staff capacity and materials consumption. Results 11 793 days with a positive multidrug-resistant pathogen diagnosis could be included in the analysis. On average, 11.8 (SD±6.3) MDRO cases occurred per nursing home. Mean duration per case was 163.3 days (SD±97.1). The annual MDRO-related costs varied in nursing homes between €2449.72 and €153 263.74 on an average €12 682.23 per case. Main cost drivers were staff capacity (€43.95 per day and €7177.04 per case) and isolation materials (€24.70 per day and €4033.51 per case). Conclusions The importance of MDROs in nursing homes could be confirmed. MDRO-related cost data in this specific healthcare sector were collected for the first time. Knowledge about the burden of MDROs will enable to assess the efficiency of hygiene intervention measures in nursing homes in the future. PMID:26908511

  10. Semivolatile organic compounds in homes: strategies for efficient and systematic exposure measurement based on empirical and theoretical factors.

    PubMed

    Dodson, Robin E; Camann, David E; Morello-Frosch, Rachel; Brody, Julia G; Rudel, Ruthann A

    2015-01-01

    Residential exposure can dominate total exposure for commercial chemicals of health concern; however, despite the importance of consumer exposures, methods for estimating household exposures remain limited. We collected house dust and indoor air samples in 49 California homes and analyzed for 76 semivolatile organic compounds (SVOCs)--phthalates, polybrominated diphenyl ethers (PBDEs), polychlorinated biphenyls (PCBs), polycyclic aromatic hydrocarbons (PAHs), and pesticides. Sixty chemicals were detected in either dust or air and here we report 58 SVOCs detected in dust for the first time. In dust, phthalates (bis(2-ethylhexyl) phthalate, benzyl butyl phthalate, di-n-butyl phthalate) and flame retardants (PBDE 99, PBDE 47) were detected at the highest concentrations relative to other chemicals at the 95th percentile, while phthalates were highest at the median. Because SVOCs are found in both gas and condensed phases and redistribute from their original source over time, partitioning models can clarify their fate indoors. We use empirical data to validate air-dust partitioning models and use these results, combined with experience in SVOC exposure assessment, to recommend residential exposure measurement strategies. We can predict dust concentrations reasonably well from measured air concentrations (R(2) = 0.80). Partitioning models and knowledge of chemical Koa elucidate exposure pathways and suggest priorities for chemical regulation. These findings also inform study design by allowing researchers to select sampling approaches optimized for their chemicals of interest and study goals. While surface wipes are commonly used in epidemiology studies because of ease of implementation, passive air sampling may be more standardized between homes and also relatively simple to deploy. Validation of passive air sampling methods for SVOCs is a priority.

  11. Empirical chemosensitivity testing in a spheroid model of ovarian cancer using a microfluidics-based multiplex platform

    PubMed Central

    Das, Tamal; Meunier, Liliane; Barbe, Laurent; Provencher, Diane; Guenat, Olivier; Gervais, Thomas; Mes-Masson, Anne-Marie

    2013-01-01

    The use of biomarkers to infer drug response in patients is being actively pursued, yet significant challenges with this approach, including the complicated interconnection of pathways, have limited its application. Direct empirical testing of tumor sensitivity would arguably provide a more reliable predictive value, although it has garnered little attention largely due to the technical difficulties associated with this approach. We hypothesize that the application of recently developed microtechnologies, coupled to more complex 3-dimensional cell cultures, could provide a model to address some of these issues. As a proof of concept, we developed a microfluidic device where spheroids of the serous epithelial ovarian cancer cell line TOV112D are entrapped and assayed for their chemoresponse to carboplatin and paclitaxel, two therapeutic agents routinely used for the treatment of ovarian cancer. In order to index the chemoresponse, we analyzed the spatiotemporal evolution of the mortality fraction, as judged by vital dyes and confocal microscopy, within spheroids subjected to different drug concentrations and treatment durations inside the microfluidic device. To reflect microenvironment effects, we tested the effect of exogenous extracellular matrix and serum supplementation during spheroid formation on their chemotherapeutic response. Spheroids displayed augmented chemoresistance in comparison to monolayer culturing. This resistance was further increased by the simultaneous presence of both extracellular matrix and high serum concentration during spheroid formation. Following exposure to chemotherapeutics, cell death profiles were not uniform throughout the spheroid. The highest cell death fraction was found at the center of the spheroid and the lowest at the periphery. Collectively, the results demonstrate the validity of the approach, and provide the basis for further investigation of chemotherapeutic responses in ovarian cancer using microfluidics technology. In

  12. How good is crude MDL for solving the bias-variance dilemma? An empirical investigation based on Bayesian networks.

    PubMed

    Cruz-Ramírez, Nicandro; Acosta-Mesa, Héctor Gabriel; Mezura-Montes, Efrén; Guerra-Hernández, Alejandro; Hoyos-Rivera, Guillermo de Jesús; Barrientos-Martínez, Rocío Erandi; Gutiérrez-Fragoso, Karina; Nava-Fernández, Luis Alonso; González-Gaspar, Patricia; Novoa-del-Toro, Elva María; Aguilera-Rueda, Vicente Josué; Ameca-Alducin, María Yaneli

    2014-01-01

    The bias-variance dilemma is a well-known and important problem in Machine Learning. It basically relates the generalization capability (goodness of fit) of a learning method to its corresponding complexity. When we have enough data at hand, it is possible to use these data in such a way so as to minimize overfitting (the risk of selecting a complex model that generalizes poorly). Unfortunately, there are many situations where we simply do not have this required amount of data. Thus, we need to find methods capable of efficiently exploiting the available data while avoiding overfitting. Different metrics have been proposed to achieve this goal: the Minimum Description Length principle (MDL), Akaike's Information Criterion (AIC) and Bayesian Information Criterion (BIC), among others. In this paper, we focus on crude MDL and empirically evaluate its performance in selecting models with a good balance between goodness of fit and complexity: the so-called bias-variance dilemma, decomposition or tradeoff. Although the graphical interaction between these dimensions (bias and variance) is ubiquitous in the Machine Learning literature, few works present experimental evidence to recover such interaction. In our experiments, we argue that the resulting graphs allow us to gain insights that are difficult to unveil otherwise: that crude MDL naturally selects balanced models in terms of bias-variance, which not necessarily need be the gold-standard ones. We carry out these experiments using a specific model: a Bayesian network. In spite of these motivating results, we also should not overlook three other components that may significantly affect the final model selection: the search procedure, the noise rate and the sample size.

  13. How good is crude MDL for solving the bias-variance dilemma? An empirical investigation based on Bayesian networks.

    PubMed

    Cruz-Ramírez, Nicandro; Acosta-Mesa, Héctor Gabriel; Mezura-Montes, Efrén; Guerra-Hernández, Alejandro; Hoyos-Rivera, Guillermo de Jesús; Barrientos-Martínez, Rocío Erandi; Gutiérrez-Fragoso, Karina; Nava-Fernández, Luis Alonso; González-Gaspar, Patricia; Novoa-del-Toro, Elva María; Aguilera-Rueda, Vicente Josué; Ameca-Alducin, María Yaneli

    2014-01-01

    The bias-variance dilemma is a well-known and important problem in Machine Learning. It basically relates the generalization capability (goodness of fit) of a learning method to its corresponding complexity. When we have enough data at hand, it is possible to use these data in such a way so as to minimize overfitting (the risk of selecting a complex model that generalizes poorly). Unfortunately, there are many situations where we simply do not have this required amount of data. Thus, we need to find methods capable of efficiently exploiting the available data while avoiding overfitting. Different metrics have been proposed to achieve this goal: the Minimum Description Length principle (MDL), Akaike's Information Criterion (AIC) and Bayesian Information Criterion (BIC), among others. In this paper, we focus on crude MDL and empirically evaluate its performance in selecting models with a good balance between goodness of fit and complexity: the so-called bias-variance dilemma, decomposition or tradeoff. Although the graphical interaction between these dimensions (bias and variance) is ubiquitous in the Machine Learning literature, few works present experimental evidence to recover such interaction. In our experiments, we argue that the resulting graphs allow us to gain insights that are difficult to unveil otherwise: that crude MDL naturally selects balanced models in terms of bias-variance, which not necessarily need be the gold-standard ones. We carry out these experiments using a specific model: a Bayesian network. In spite of these motivating results, we also should not overlook three other components that may significantly affect the final model selection: the search procedure, the noise rate and the sample size. PMID:24671204

  14. How Good Is Crude MDL for Solving the Bias-Variance Dilemma? An Empirical Investigation Based on Bayesian Networks

    PubMed Central

    Cruz-Ramírez, Nicandro; Acosta-Mesa, Héctor Gabriel; Mezura-Montes, Efrén; Guerra-Hernández, Alejandro; Hoyos-Rivera, Guillermo de Jesús; Barrientos-Martínez, Rocío Erandi; Gutiérrez-Fragoso, Karina; Nava-Fernández, Luis Alonso; González-Gaspar, Patricia; Novoa-del-Toro, Elva María; Aguilera-Rueda, Vicente Josué; Ameca-Alducin, María Yaneli

    2014-01-01

    The bias-variance dilemma is a well-known and important problem in Machine Learning. It basically relates the generalization capability (goodness of fit) of a learning method to its corresponding complexity. When we have enough data at hand, it is possible to use these data in such a way so as to minimize overfitting (the risk of selecting a complex model that generalizes poorly). Unfortunately, there are many situations where we simply do not have this required amount of data. Thus, we need to find methods capable of efficiently exploiting the available data while avoiding overfitting. Different metrics have been proposed to achieve this goal: the Minimum Description Length principle (MDL), Akaike’s Information Criterion (AIC) and Bayesian Information Criterion (BIC), among others. In this paper, we focus on crude MDL and empirically evaluate its performance in selecting models with a good balance between goodness of fit and complexity: the so-called bias-variance dilemma, decomposition or tradeoff. Although the graphical interaction between these dimensions (bias and variance) is ubiquitous in the Machine Learning literature, few works present experimental evidence to recover such interaction. In our experiments, we argue that the resulting graphs allow us to gain insights that are difficult to unveil otherwise: that crude MDL naturally selects balanced models in terms of bias-variance, which not necessarily need be the gold-standard ones. We carry out these experiments using a specific model: a Bayesian network. In spite of these motivating results, we also should not overlook three other components that may significantly affect the final model selection: the search procedure, the noise rate and the sample size. PMID:24671204

  15. Semivolatile organic compounds in homes: strategies for efficient and systematic exposure measurement based on empirical and theoretical factors.

    PubMed

    Dodson, Robin E; Camann, David E; Morello-Frosch, Rachel; Brody, Julia G; Rudel, Ruthann A

    2015-01-01

    Residential exposure can dominate total exposure for commercial chemicals of health concern; however, despite the importance of consumer exposures, methods for estimating household exposures remain limited. We collected house dust and indoor air samples in 49 California homes and analyzed for 76 semivolatile organic compounds (SVOCs)--phthalates, polybrominated diphenyl ethers (PBDEs), polychlorinated biphenyls (PCBs), polycyclic aromatic hydrocarbons (PAHs), and pesticides. Sixty chemicals were detected in either dust or air and here we report 58 SVOCs detected in dust for the first time. In dust, phthalates (bis(2-ethylhexyl) phthalate, benzyl butyl phthalate, di-n-butyl phthalate) and flame retardants (PBDE 99, PBDE 47) were detected at the highest concentrations relative to other chemicals at the 95th percentile, while phthalates were highest at the median. Because SVOCs are found in both gas and condensed phases and redistribute from their original source over time, partitioning models can clarify their fate indoors. We use empirical data to validate air-dust partitioning models and use these results, combined with experience in SVOC exposure assessment, to recommend residential exposure measurement strategies. We can predict dust concentrations reasonably well from measured air concentrations (R(2) = 0.80). Partitioning models and knowledge of chemical Koa elucidate exposure pathways and suggest priorities for chemical regulation. These findings also inform study design by allowing researchers to select sampling approaches optimized for their chemicals of interest and study goals. While surface wipes are commonly used in epidemiology studies because of ease of implementation, passive air sampling may be more standardized between homes and also relatively simple to deploy. Validation of passive air sampling methods for SVOCs is a priority. PMID:25488487

  16. Semivolatile Organic Compounds in Homes: Strategies for Efficient and Systematic Exposure Measurement Based on Empirical and Theoretical Factors

    PubMed Central

    2014-01-01

    Residential exposure can dominate total exposure for commercial chemicals of health concern; however, despite the importance of consumer exposures, methods for estimating household exposures remain limited. We collected house dust and indoor air samples in 49 California homes and analyzed for 76 semivolatile organic compounds (SVOCs)—phthalates, polybrominated diphenyl ethers (PBDEs), polychlorinated biphenyls (PCBs), polycyclic aromatic hydrocarbons (PAHs), and pesticides. Sixty chemicals were detected in either dust or air and here we report 58 SVOCs detected in dust for the first time. In dust, phthalates (bis(2-ethylhexyl) phthalate, benzyl butyl phthalate, di-n-butyl phthalate) and flame retardants (PBDE 99, PBDE 47) were detected at the highest concentrations relative to other chemicals at the 95th percentile, while phthalates were highest at the median. Because SVOCs are found in both gas and condensed phases and redistribute from their original source over time, partitioning models can clarify their fate indoors. We use empirical data to validate air-dust partitioning models and use these results, combined with experience in SVOC exposure assessment, to recommend residential exposure measurement strategies. We can predict dust concentrations reasonably well from measured air concentrations (R2 = 0.80). Partitioning models and knowledge of chemical Koa elucidate exposure pathways and suggest priorities for chemical regulation. These findings also inform study design by allowing researchers to select sampling approaches optimized for their chemicals of interest and study goals. While surface wipes are commonly used in epidemiology studies because of ease of implementation, passive air sampling may be more standardized between homes and also relatively simple to deploy. Validation of passive air sampling methods for SVOCs is a priority. PMID:25488487

  17. Survey Says

    ERIC Educational Resources Information Center

    McCarthy, Susan K.

    2005-01-01

    Survey Says is a lesson plan designed to teach college students how to access Internet resources for valid data related to the sexual health of young people. Discussion questions based on the most recent available data from two national surveys, the Youth Risk Behavior Surveillance-United States, 2003 (CDC, 2004) and the National Survey of…

  18. Using SEM to Analyze Complex Survey Data: A Comparison between Design-Based Single-Level and Model-Based Multilevel Approaches

    ERIC Educational Resources Information Center

    Wu, Jiun-Yu; Kwok, Oi-man

    2012-01-01

    Both ad-hoc robust sandwich standard error estimators (design-based approach) and multilevel analysis (model-based approach) are commonly used for analyzing complex survey data with nonindependent observations. Although these 2 approaches perform equally well on analyzing complex survey data with equal between- and within-level model structures…

  19. Tests of HPGe- and scintillation-based backpack γ-radiation survey systems.

    PubMed

    Nilsson, Jonas M C; Östlund, Karl; Söderberg, Joakim; Mattsson, Sören; Rääf, Christopher

    2014-09-01

    The performance of three different backpack-mounted γ-radiation survey systems has been investigated. The systems are based on a LaBr3:Ce detector and a NaI(Tl) detector both with active volume dimensions of 76.2 mm in diameter and 76.2 mm length and a 123% relative efficiency HPGe detector. The detection limits of the systems were tested in a controlled outdoor environment in Sweden, followed by field tests of the HPGe- and LaBr3:Ce-based systems at the site of a radioactive waste repository in Georgia (in the Caucasus region of Eurasia). The results showed that the high efficiency HPGe detector performed significantly better than similar sized LaBr3:Ce and NaI(Tl) detectors, however, the HPGe detector was significantly heavier than the other systems. The use of different analysis methods revealed that creating maps of the survey area was the best method for offline analysis of survey data collected from a large area. Using off-site personnel for analysis of the data proved to be beneficial.

  20. A web-based survey of odontologist's opinions concerning bitemark analyses.

    PubMed

    Pretty, I A

    2003-09-01

    Within the field of forensic dentistry, the detection, collection, and analysis of bitemarks remains one of the most contentious areas. Attempts at the production of consensus documents have produced documents such as the ABFO's Guidelines for Bitemark Evidence Collection. Despite this, the range of differing analysis techniques, allied with a varied opinion base on the robustness of bitemark conclusions has led to polarized views within the profession. The purpose of this study was to survey forensic dentists to obtain their views on a number of crucial components of bitemark theory and contentious areas within the discipline. Using a web-based survey, 14 questions were asked of respondents. Seventy-two odontologists completed the survey, with 38% being of Diplomate status, 10% had completed 20 or more bitemark cases, and 20% between 10 and 20 cases, 91% of respondents believed that the human dentition was unique, with 78% believing that this uniqueness could be represented on human skin during the biting process. Seventy percent believed that they could positively identify an individual from a bitemark, and 22% stated that the statistical tool, the product rule, should be applied to bitemark conclusions. Over half of the odontologists used overlays for bitemark analysis. with a digital method of production the most popular. The implications of these and other findings are discussed. PMID:14535678

  1. Simulation of Strong Ground Motion Based on Conventional Empirical Green s Functions In the Michoacán State, Mexico

    NASA Astrophysics Data System (ADS)

    Vazquez Rosas, R.; Aguirre Gonzalez, J. J.; Mijares Arellano, H. H.

    2012-12-01

    In the present work, we study the state of Michoacán, one of the most important seimogenic zones in Mexico. Three kinds of sources exist in the state, producing tectonic earthquakes, volcanic earthquakes, and events due to local faults in the region. For this reason, it is of vital importance the study of source parameters in the Michoacán state. In this work in particular we applied the simulation of strong ground motions by the conventional empirical Green s functions proposed by Irikura (1986). We installed a temporary network consisting of 6 accelerograph stations across the state, at the following locations: Faro de Brucerías, Aguililla, Apatzingán, Pátzcuaro, Morelia, and Maravatío. The stations form a line that is perpendicular to the coastline and has a total length of 366 km, while the distance between neighboring stations varies from 60 to 80 km. Among all the seismic events recorded at this temporary network, we select 2 events originated along the coastline of Michoacán (May the 2007), with moment magnitudes of 4.3 and 5.1 Mw. In order to calibrate the model, the earthquake of May 31, 2007 (M 5.1) was simulated using the aftershock of May 27 of that year (M 4.3) with satisfactory results, following the same method and considering the ω2 spectral model with constant stress drop. Later, we calculated six scenarios for a postulate earthquake of M 7.4. From the six scenarios the largest peak ground accelerations for each station were, 83 cm/s2 in Faro de Brucerías , 15.4 cm/s2 in Apatzingán, 23 cm/s2 in Pátzcuaro, 3.7 cm/s2 in Morelia and Maravatio con 3.0 cm/s2 . One limitation of this study is that we used relatively small-magnitude earthquakes. This was a consequence of the relatively short operation period of the temporary network, which had to be limited to 3 months. To improve these simulations it is necessary to have more information about rupture processes of the recorded earthquakes. And likewise, information of future earthquakes in the

  2. Improvements of Visual Based Shoreline Accuracy From Satellite Imagery and Simultaneous Differential GPS Surveys

    NASA Astrophysics Data System (ADS)

    MacKenzie, R. A.; Jaeger, J. M.; Adams, P. N.; Plant, N. G.; Schaub, R.; Kline, S. W.; Lovering, J. L.

    2011-12-01

    Various tools with different sampling density, intervals and uncertainties are now used in decadal -scale coastal change studies. Lidar has rapidly become the preferred tool for coastal studies with its high data density and moderate accuracy (15 - 30 cm), but the typical sampling interval between surveys is ~5 years. GPS measurements have higher accuracy (10 cm) and can be conducted more frequently than Lidar, but data density is lower and surveyed coastline length is minimized. Imagery is used because of its historical significance, but it is the least accurate. High degrees of uncertainty in shoreline position from remotely sensed imagery results from difficulty in choosing an appropriate visual shoreline proxy. This is due, in part to different visual based shoreline (VBS) proxies (e.g. vegetation line, debris line, high water mark, wet/dry line, low water mark) not being visible on all beaches at all times. Geomorphic clues from remotely sensed imagery combined with knowledge of coastal morphology and DBS surveys can minimize uncertainties in using VBS proxies. We established a relationship between VBS and DBS by conducting 32 kinematic differential GPS surveys over a two-year period at Cape Canaveral, FL. Two Geoeye I (0.5 m res.) satellite images were collected 3 days post DBS survey on June 9, 2009 and during a DBS survey on July 28, 2010. June and July beach states were selected for image collection because they have the lowest average monthly significant wave height (HS) and least variability in HS, providing the highest likelihood of stable beach morphology. The 10-km study area exhibits various beach morphologic states and stability histories. During the July survey the visual high water wet/dry line traced with GPS, did not correspond to the wet/dry line on the GeoEye image. Further analysis shows that a wet-dry line related to grain size, composition and ground water seepage face, visible in all images, provides the best VBS proxy. Direct comparison

  3. A Database Selection Expert System Based on Reference Librarian's Database Selection Strategy: A Usability and Empirical Evaluation.

    ERIC Educational Resources Information Center

    Ma, Wei

    2002-01-01

    Describes the development of a prototype Web-based database selection expert system at the University of Illinois at Urbana-Champaign that is based on reference librarians' database selection strategy which allows users to simultaneously search all available databases to identify those most relevant to their search using free-text keywords or…

  4. Relationship of Student Undergraduate Achievement and Personality Characteristics in a Total Web-Based Environment: An Empirical Study

    ERIC Educational Resources Information Center

    Schniederjans, Marc J.; Kim, Eyong B.

    2005-01-01

    Web-based education is a popular format for the delivery of college courses. Research has shown that it may not be the best form of education for all students. Today, many students (and student advisors) face a choice in course delivery format (i.e., Web-based or more traditional classroom courses). This research study examines the relationship…

  5. The National Nursing Assistant Survey: Improving the Evidence Base for Policy Initiatives to Strengthen the Certified Nursing Assistant Workforce

    ERIC Educational Resources Information Center

    Squillace, Marie R.; Remsburg, Robin E.; Harris-Kojetin, Lauren D.; Bercovitz, Anita; Rosenoff, Emily; Han, Beth

    2009-01-01

    Purpose: This study introduces the first National Nursing Assistant Survey (NNAS), a major advance in the data available about certified nursing assistants (CNAs) and a rich resource for evidence-based policy, practice, and applied research initiatives. We highlight potential uses of this new survey using select population estimates as examples of…

  6. Social Media and Evidence-Based Maternity Care: A Cross-Sectional Survey Study.

    PubMed

    Dekker, Rebecca L; King, Sarah; Lester, Kara

    2016-01-01

    The purpose of this study was to describe how people use social media to find and disseminate information about evidence-based maternity care. We used a cross-sectional Internet-based survey design in which 1,661 participants were recruited from childbirth-related blogs. Participants answered questions about how they find, use, and share evidence-based maternity information using social media. Overall, women in this study were highly engaged in using social media to find and share maternity information. Most respondents were very interested in reading evidence-based maternity care articles online. Most intend to use this information that they found, despite the fact that a substantial percentage had no intentions of discussing this information with their childbirth educators or physician. PMID:27445448

  7. Social Media and Evidence-Based Maternity Care: A Cross-Sectional Survey Study.

    PubMed

    Dekker, Rebecca L; King, Sarah; Lester, Kara

    2016-01-01

    The purpose of this study was to describe how people use social media to find and disseminate information about evidence-based maternity care. We used a cross-sectional Internet-based survey design in which 1,661 participants were recruited from childbirth-related blogs. Participants answered questions about how they find, use, and share evidence-based maternity information using social media. Overall, women in this study were highly engaged in using social media to find and share maternity information. Most respondents were very interested in reading evidence-based maternity care articles online. Most intend to use this information that they found, despite the fact that a substantial percentage had no intentions of discussing this information with their childbirth educators or physician.

  8. What Do Stroke Patients Look for in Game-Based Rehabilitation: A Survey Study.

    PubMed

    Hung, Ya-Xuan; Huang, Pei-Chen; Chen, Kuan-Ta; Chu, Woei-Chyn

    2016-03-01

    Stroke is one of the most common causes of physical disability, and early, intensive, and repetitive rehabilitation exercises are crucial to the recovery of stroke survivors. Unfortunately, research shows that only one third of stroke patients actually perform recommended exercises at home, because of the repetitive and mundane nature of conventional rehabilitation exercises. Thus, to motivate stroke survivors to engage in monotonous rehabilitation is a significant issue in the therapy process. Game-based rehabilitation systems have the potential to encourage patients continuing rehabilitation exercises at home. However, these systems are still rarely adopted at patients' places. Discovering and eliminating the obstacles in promoting game-based rehabilitation at home is therefore essential. For this purpose, we conducted a study to collect and analyze the opinions and expectations of stroke patients and clinical therapists. The study is composed of 2 parts: Rehab-preference survey - interviews to both patients and therapists to understand the current practices, challenges, and expectations on game-based rehabilitation systems; and Rehab-compatibility survey - a gaming experiment with therapists to elaborate what commercial games are compatible with rehabilitation. The study is conducted with 30 outpatients with stroke and 19 occupational therapists from 2 rehabilitation centers in Taiwan. Our surveys show that game-based rehabilitation systems can turn the rehabilitation exercises more appealing and provide personalized motivation for various stroke patients. Patients prefer to perform rehabilitation exercises with more diverse and fun games, and need cost-effective rehabilitation systems, which are often built on commodity hardware. Our study also sheds light on incorporating the existing design-for-fun games into rehabilitation system. We envision the results are helpful in developing a platform which enables rehab-compatible (i.e., existing, appropriately

  9. What Do Stroke Patients Look for in Game-Based Rehabilitation: A Survey Study.

    PubMed

    Hung, Ya-Xuan; Huang, Pei-Chen; Chen, Kuan-Ta; Chu, Woei-Chyn

    2016-03-01

    Stroke is one of the most common causes of physical disability, and early, intensive, and repetitive rehabilitation exercises are crucial to the recovery of stroke survivors. Unfortunately, research shows that only one third of stroke patients actually perform recommended exercises at home, because of the repetitive and mundane nature of conventional rehabilitation exercises. Thus, to motivate stroke survivors to engage in monotonous rehabilitation is a significant issue in the therapy process. Game-based rehabilitation systems have the potential to encourage patients continuing rehabilitation exercises at home. However, these systems are still rarely adopted at patients' places. Discovering and eliminating the obstacles in promoting game-based rehabilitation at home is therefore essential. For this purpose, we conducted a study to collect and analyze the opinions and expectations of stroke patients and clinical therapists. The study is composed of 2 parts: Rehab-preference survey - interviews to both patients and therapists to understand the current practices, challenges, and expectations on game-based rehabilitation systems; and Rehab-compatibility survey - a gaming experiment with therapists to elaborate what commercial games are compatible with rehabilitation. The study is conducted with 30 outpatients with stroke and 19 occupational therapists from 2 rehabilitation centers in Taiwan. Our surveys show that game-based rehabilitation systems can turn the rehabilitation exercises more appealing and provide personalized motivation for various stroke patients. Patients prefer to perform rehabilitation exercises with more diverse and fun games, and need cost-effective rehabilitation systems, which are often built on commodity hardware. Our study also sheds light on incorporating the existing design-for-fun games into rehabilitation system. We envision the results are helpful in developing a platform which enables rehab-compatible (i.e., existing, appropriately

  10. Development and validation of a web-based questionnaire for surveying the health and working conditions of high-performance marine craft populations

    PubMed Central

    de Alwis, Manudul Pahansen; Lo Martire, Riccardo; Äng, Björn O; Garme, Karl

    2016-01-01

    Background High-performance marine craft crews are susceptible to various adverse health conditions caused by multiple interactive factors. However, there are limited epidemiological data available for assessment of working conditions at sea. Although questionnaire surveys are widely used for identifying exposures, outcomes and associated risks with high accuracy levels, until now, no validated epidemiological tool exists for surveying occupational health and performance in these populations. Aim To develop and validate a web-based questionnaire for epidemiological assessment of occupational and individual risk exposure pertinent to the musculoskeletal health conditions and performance in high-performance marine craft populations. Method A questionnaire for investigating the association between work-related exposure, performance and health was initially developed by a consensus panel under four subdomains, viz. demography, lifestyle, work exposure and health and systematically validated by expert raters for content relevance and simplicity in three consecutive stages, each iteratively followed by a consensus panel revision. The item content validity index (I-CVI) was determined as the proportion of experts giving a rating of 3 or 4. The scale content validity index (S-CVI/Ave) was computed by averaging the I-CVIs for the assessment of the questionnaire as a tool. Finally, the questionnaire was pilot tested. Results The S-CVI/Ave increased from 0.89 to 0.96 for relevance and from 0.76 to 0.94 for simplicity, resulting in 36 items in the final questionnaire. The pilot test confirmed the feasibility of the questionnaire. Conclusions The present study shows that the web-based questionnaire fulfils previously published validity acceptance criteria and is therefore considered valid and feasible for the empirical surveying of epidemiological aspects among high-performance marine craft crews and similar populations. PMID:27324717

  11. Consultants' perceptions of school counselors' ability to implement an empirically-based intervention for adolescent social anxiety disorder.

    PubMed

    Masia Warner, Carrie; Brice, Chad; Esseling, Petra G; Stewart, Catherine E; Mufson, Laura; Herzig, Kathleen

    2013-11-01

    Social anxiety is highly prevalent but goes untreated. Although school-based CBT programs are efficacious when delivered by specialized psychologists, it is unclear whether school counselors can implement these interventions effectively, which is essential to promote sustainable school programs. We present an initial consultation strategy to support school counselor implementation of group CBT for social anxiety and an evaluation of counselors' treatment fidelity. Counselors were highly adherent to the treatment, but competence varied based on measurement. Counselors and consultants demonstrated good agreement for adherence, but relatively modest correspondence in competence ratings. We discuss future directions for school-based implementation efforts informed by these initial findings.

  12. Consultants’ Perceptions of School Counselors’ Ability to Implement an Empirically-Based Intervention for Adolescent Social Anxiety Disorder

    PubMed Central

    Warner, Carrie Masia; Brice, Chad; Esseling, Petra G.; Stewart, Catherine E.; Mufson, Laura; Herzig, Kathleen

    2013-01-01

    Social anxiety is highly prevalent but goes untreated. Although school-based CBT programs are efficacious when delivered by specialized psychologists, it is unclear whether school counselors can implement these interventions effectively, which is essential to promote sustainable school programs. We present an initial consultation strategy to support school counselor implementation of group CBT for social anxiety and an evaluation of counselors’ treatment fidelity. Counselors were highly adherent to the treatment, but competence varied based on measurement. Counselors and consultants demonstrated good agreement for adherence, but relatively modest correspondence in competence ratings. We discuss future directions for school-based implementation efforts informed by these initial findings. PMID:23716144

  13. An empirical investigation of two competing models of patient satisfaction.

    PubMed

    Mishra, D P; Singh, J; Wood, V

    1991-01-01

    This paper empirically examines two competing models of patient satisfaction. Specifically, a five factor SERVQUAL model proposed by Parasuraman et al. (1988) and a tripartite model posited by Smith, Bloom, and Davis (1986) are examined. The two models are tested via factor analysis based on data collected from a field survey of hospital patients. The results of this study indicate that the five dimensional SERVQUAL model is not supported by data. On the other hand, there is general support for the tripartite model. Implications of our results for health care practitioners and researchers are discussed. Future directions for research are also outlined.

  14. Cleaning sky survey data bases using Hough transform and renewal string approaches

    NASA Astrophysics Data System (ADS)

    Storkey, A. J.; Hambly, N. C.; Williams, C. K. I.; Mann, R. G.

    2004-01-01

    Large astronomical data bases obtained from sky surveys such as the SuperCOSMOS Sky Survey (SSS) invariably suffer from spurious records coming from the artefactual effects of the telescope, satellites and junk objects in orbit around the Earth and physical defects on the photographic plate or CCD. Though relatively small in number, these spurious records present a significant problem in many situations, where they can become a large proportion of the records potentially of interest to a given astronomer. Accurate and robust techniques are needed for locating and flagging such spurious objects, and we are undertaking a programme investigating the use of machine learning techniques in this context. In this paper we focus on the four most common causes of unwanted records in the SSS: satellite or aeroplane tracks, scratches, fibres and other linear phenomena introduced to the plate, circular haloes around bright stars due to internal reflections within the telescope and diffraction spikes near to bright stars. Appropriate techniques are developed for the detection of each of these. The methods are applied to the SSS data to develop a data set of spurious object detections, along with confidence measures, which can allow these unwanted data to be removed from consideration. These methods are general and can be adapted to other astronomical survey data.

  15. Population-based sexual behavior surveys in China: Liuzhou compared with other prefectural cities

    PubMed Central

    Yingying, Huang; Abler, Laurie; Suiming, Pan; Henderson, Gail E.; Xin, Wang; Xingliang, Yao; Parish, William L.

    2013-01-01

    Sexual behaviors in China are rapidly changing; simultaneously, STI/HIV prevalence is increasing in the general population. To investigate these major shifts, we examined sexual behaviors and self-reported sexually transmitted infections (STI) in one prefectural city in southern China, Liuzhou, and compared it to other prefectural cities throughout China. We used adults age 18-39 from two sets of population-based surveys that paralleled each other in both content and method. The first set was the Liuzhou survey conducted in 2008 (n=398). The second set consisted of two national surveys collected in 2006 and 2010 (n=2186). Liuzhou respondents reported more active social and sexual behaviors than their national counterparts, including more socializing, dancing, drinking excessively, sexual activity among never married men and women, purchasing commercial sex among men, one-night stands among men, multiple sexual partnerships and self-reported STI among both men and women. Women in Liuzhou reported greater sexual risk behavior than their national counterparts, although overall they reported less than their male counterparts; they were also more likely to have had an abortion than women in other prefectural cities. Our findings provide a comprehensive overview of the sexual context of Liuzhou among the general population, which may help explain the greater STI/HIV prevalence in Liuzhou. PMID:24174289

  16. Estimation of daily global solar radiation using wavelet regression, ANN, GEP and empirical models: A comparative study of selected temperature-based approaches

    NASA Astrophysics Data System (ADS)

    Sharifi, Sayed Saber; Rezaverdinejad, Vahid; Nourani, Vahid

    2016-11-01

    Although the sunshine-based models generally have a better performance than temperature-based models for estimating solar radiation, the limited availability of sunshine duration records makes the development of temperature-based methods inevitable. This paper presents a comparative study between Artificial Neural Networks (ANNs), Gene Expression Programming (GEP), Wavelet Regression (WR) and 5 selected temperature-based empirical models for estimating the daily global solar radiation. A new combination of inputs including four readily accessible parameters have been employed: daily mean clearness index (KT), temperature range (ΔT), theoretical sunshine duration (N) and extraterrestrial radiation (Ra). Ten statistical indicators in a form of GPI (Global Performance Indicator) is used to ascertain the suitability of the models. The performance of selected models across the range of solar radiation values, was depicted by the quantile-quantile (Q-Q) plots. Comparing these plots makes it evident that ANNs can cover a broader range of solar radiation values. The results shown indicate that the performance of ANN model was clearly superior to the other models. The findings also demonstrated that WR model performed well and presented high accuracy in estimations of daily global solar radiation.

  17. Burnout, psychological morbidity, job satisfaction, and stress: a survey of Canadian hospital based child protection professionals

    PubMed Central

    Bennett, S; Plint, A; Clifford, T

    2005-01-01

    Aims: (1) To measure the prevalence of burnout, psychological morbidity, job satisfaction, job stress, and consideration of alternate work among multidisciplinary hospital based child and youth protection (CYP) professionals; (2) to understand the relations between these variables; and (3) to understand the reasons for leaving among former programme members. Methods: Mailed survey of current and former members of all Canadian academic hospital based CYP programmes. Surveys for current members contained validated measures of burnout, psychological morbidity, job satisfaction/stress, and questions about consideration of alternate work. Surveys for former members examined motivation(s) for leaving. Results: One hundred and twenty six of 165 current members (76.4%) and 13/14 (92.9%) former members responded. Over one third (34.1%) of respondents exhibited burnout while psychological morbidity was present in 13.5%. Job satisfaction was high, with 68.8% finding their job "extremely" or "quite" satisfying, whereas 26.2% found their job "extremely" or "quite" stressful. Psychological morbidity, job satisfaction, and job stress were not associated with any of the demographic variables measured, but burnout was most prevalent among non-physician programme members. Almost two thirds of current members indicated that they had seriously considered a change in work situation. Former members indicated that burnout and high levels of job stress were most responsible for their decision to leave and that increasing the number of programme staff and, consequently, reducing the number of hours worked would have influenced their decision to stay. Conclusions: Current levels of burnout and the large proportion of individuals who have contemplated leaving the service suggest a potential crisis in Canadian hospital based CYP services. PMID:16243862

  18. COOKING APPLIANCE USE IN CALIFORNIA HOMES DATA COLLECTED FROM A WEB-BASED SURVEY

    SciTech Connect

    Klug, Victoria; Lobscheid, Agnes; Singer, Brett

    2011-08-01

    Cooking of food and use of natural gas cooking burners generate pollutants that can have substantial impacts on residential indoor air quality. The extent of these impacts depends on cooking frequency, duration and specific food preparation activities in addition to the extent to which exhaust fans or other ventilation measures (e.g. windows) are used during cooking. With the intent of improving our understanding of indoor air quality impacts of cooking-related pollutants, we created, posted and advertised a web-based survey about cooking activities in residences. The survey included questions similar to those in California's Residential Appliance Saturation Survey (RASS), relating to home, household and cooking appliance characteristics and weekly patterns of meals cooked. Other questions targeted the following information not captured in the RASS: (1) oven vs. cooktop use, the number of cooktop burners used and the duration of burner use when cooking occurs, (2) specific cooking activities, (3) the use of range hood or window to increase ventilation during cooking, and (4) occupancy during cooking. Specific cooking activity questions were asked about the prior 24 hours with the assumption that most people are able to recollect activities over this time period. We examined inter-relationships among cooking activities and patterns and relationships of cooking activities to household demographics. We did not seek to obtain a sample of respondents that is demographically representative of the California population but rather to inexpensively gather information from homes spanning ranges of relevant characteristics including the number of residents and presence or absence of children. This report presents the survey, the responses obtained, and limited analysis of the results.

  19. Agreement on the Level Selection in Laminoplasty among Experienced Surgeons: A Survey-Based Study

    PubMed Central

    Cho, Jae Hwan; Suk, Kyung-Soo; Park, Jong-Beom; Ha, Jung-Ki; Hwang, Chang Ju; Lee, Choon Sung

    2016-01-01

    Study Design Survey based study. Purpose To assess the degree of agreement in level selection of laminoplasty (LP) for the selected cervical myeloradiculopathy cases between experienced spine surgeons. Overview of Literature Although, cervical LP is a widely used surgical technique for multi-level spinal cord compression, until now there is no consensus about how many segments or which segments should be opened to achieve a satisfactory decompression. Methods Thorough clinical and radiographic data (plain X-ray, computed tomography, and magnetic resonance imaging) of 30 patients who had cervical myelopathy were prepared. The data were provided to three independent spine surgeons with over 10 years experience in operation of their own practices. They were questioned about the most preferable surgical method and suitable decompression levels. The second survey was carried out after 6 months with the same cases. If the level difference between respondents was a half level or below, agreement was considered acceptable. The intraobserver and interobserver agreements in level selection were assessed by kappa statistics. Results Three respondents selected LP as an option for 6, 8, and 22 cases in the first survey and 10, 21, and 24 cases in the second survey. The reasons for selection of LP were levels of ossification of the posterior longitudinal ligament (p=0.004), segmental kyphotic deformity (p=0.036) and mean compression score (p=0.041). Intraobserver agreement showed variable results. Interobserver agreement was poor to fair by perfect matching (kappa=0.111–0.304) and fair to moderate by acceptable matching (kappa=0.308–0.625). Conclusions The degree of agreement for level selection of LP was not high even though experienced surgeons would choose the opening segments on the basis of same criteria. These results suggest that more specific guidelines in determination of levels for LP should be required to decrease unnecessary wide decompression according to

  20. PM3 semi-empirical IR spectra simulations for metal complexes of schiff bases of sulfa drugs

    NASA Astrophysics Data System (ADS)

    Topacli, C.; Topacli, A.

    2003-06-01

    The molecular structures and infrared spectra of Co, Ni, Cu and Zn complexes of two schiff base ligands, viz N-( o-vanillinidene)sulfanilamide ( oVSaH) and N-( o-vanillinidene)sulfamerazine ( oVSmrzH) are studied in detail by PM3 method. It has been shown that the proposed structures for the compounds derived from microanalytical, magnetic and various spectral data were consistent with the IR spectra simulated by PM3 method. Coordination effects on ν(CN) and ν(C-O) modes in the schiff base ligands are in close agreement with the observed results.

  1. A Parameter Identification Method for Helicopter Noise Source Identification and Physics-Based Semi-Empirical Modeling

    NASA Technical Reports Server (NTRS)

    Greenwood, Eric, II; Schmitz, Fredric H.

    2010-01-01

    A new physics-based parameter identification method for rotor harmonic noise sources is developed using an acoustic inverse simulation technique. This new method allows for the identification of individual rotor harmonic noise sources and allows them to be characterized in terms of their individual non-dimensional governing parameters. This new method is applied to both wind tunnel measurements and ground noise measurements of two-bladed rotors. The method is shown to match the parametric trends of main rotor Blade-Vortex Interaction (BVI) noise, allowing accurate estimates of BVI noise to be made for operating conditions based on a small number of measurements taken at different operating conditions.

  2. The coverage problem in video-based wireless sensor networks: a survey.

    PubMed

    Costa, Daniel G; Guedes, Luiz Affonso

    2010-01-01

    Wireless sensor networks typically consist of a great number of tiny low-cost electronic devices with limited sensing and computing capabilities which cooperatively communicate to collect some kind of information from an area of interest. When wireless nodes of such networks are equipped with a low-power camera, visual data can be retrieved, facilitating a new set of novel applications. The nature of video-based wireless sensor networks demands new algorithms and solutions, since traditional wireless sensor networks approaches are not feasible or even efficient for that specialized communication scenario. The coverage problem is a crucial issue of wireless sensor networks, requiring specific solutions when video-based sensors are employed. In this paper, it is surveyed the state of the art of this particular issue, regarding strategies, algorithms and general computational solutions. Open research areas are also discussed, envisaging promising investigation considering coverage in video-based wireless sensor networks. PMID:22163651

  3. The Coverage Problem in Video-Based Wireless Sensor Networks: A Survey

    PubMed Central

    Costa, Daniel G.; Guedes, Luiz Affonso

    2010-01-01

    Wireless sensor networks typically consist of a great number of tiny low-cost electronic devices with limited sensing and computing capabilities which cooperatively communicate to collect some kind of information from an area of interest. When wireless nodes of such networks are equipped with a low-power camera, visual data can be retrieved, facilitating a new set of novel applications. The nature of video-based wireless sensor networks demands new algorithms and solutions, since traditional wireless sensor networks approaches are not feasible or even efficient for that specialized communication scenario. The coverage problem is a crucial issue of wireless sensor networks, requiring specific solutions when video-based sensors are employed. In this paper, it is surveyed the state of the art of this particular issue, regarding strategies, algorithms and general computational solutions. Open research areas are also discussed, envisaging promising investigation considering coverage in video-based wireless sensor networks. PMID:22163651

  4. Investigating Measures of Social Context on 2 Population-Based Health Surveys, Hawaii, 2010-2012.

    PubMed

    Pobutsky, Ann M; Baker, Kathleen Kromer; Reyes-Salvail, Florentina

    2015-01-01

    Measures from the Social Context Module of the Centers for Disease Control and Prevention's Behavioral Risk Factor Surveillance System were used on 2 population-based health surveys in Hawaii to explicate the role of the nonmedical and social determinants of health; these measures were also compared with conventional socioeconomic status (SES) variables. Results showed that the self-reported SES vulnerabilities of food and housing insecurity are both linked to demographic factors and physical and mental health status and significant when controlling for the conventional measures of SES. The social context module indicators should be increasingly used so results can inform appropriate interventions for vulnerable populations.

  5. Is cell culture a risky business? Risk analysis based on scientist survey data.

    PubMed

    Shannon, Mark; Capes-Davis, Amanda; Eggington, Elaine; Georghiou, Ronnie; Huschtscha, Lily I; Moy, Elsa; Power, Melinda; Reddel, Roger R; Arthur, Jonathan W

    2016-02-01

    Cell culture is a technique that requires vigilance from the researcher. Common cell culture problems, including contamination with microorganisms or cells from other cultures, can place the reliability and reproducibility of cell culture work at risk. Here we use survey data, contributed by research scientists based in Australia and New Zealand, to assess common cell culture risks and how these risks are managed in practice. Respondents show that sharing of cell lines between laboratories continues to be widespread. Arrangements for mycoplasma and authentication testing are increasingly in place, although scientists are often uncertain how to perform authentication testing. Additional risks are identified for preparation of frozen stocks, storage and shipping. PMID:26365214

  6. Supporting evidence-based medicine: a survey of U.S. medical librarians.

    PubMed

    Li, Ping; Wu, Lin

    2011-01-01

    This study sought to identify medical librarians' roles in supporting evidence-based medicine (EBM) practice; determine whether medical librarians' work settings, work experiences, or job titles made a difference in their EBM responsibilities; and find out medical librarians' perceptions of their roles in EBM practice. An online survey was distributed to U.S. medical librarians. The results showed that medical librarians had positive perceptions of their EBM-related responsibilities, which were diverse and specific. Their work experience, work settings, and job title categories related to some of their EBM responsibilities, as well as the nature of some of the responsibilities. PMID:22040243

  7. Is cell culture a risky business? Risk analysis based on scientist survey data.

    PubMed

    Shannon, Mark; Capes-Davis, Amanda; Eggington, Elaine; Georghiou, Ronnie; Huschtscha, Lily I; Moy, Elsa; Power, Melinda; Reddel, Roger R; Arthur, Jonathan W

    2016-02-01

    Cell culture is a technique that requires vigilance from the researcher. Common cell culture problems, including contamination with microorganisms or cells from other cultures, can place the reliability and reproducibility of cell culture work at risk. Here we use survey data, contributed by research scientists based in Australia and New Zealand, to assess common cell culture risks and how these risks are managed in practice. Respondents show that sharing of cell lines between laboratories continues to be widespread. Arrangements for mycoplasma and authentication testing are increasingly in place, although scientists are often uncertain how to perform authentication testing. Additional risks are identified for preparation of frozen stocks, storage and shipping.

  8. Knowledge based systems: A critical survey of major concepts, issues and techniques. Visuals

    NASA Technical Reports Server (NTRS)

    Dominick, Wayne D. (Editor); Kavi, Srinu

    1984-01-01

    This Working Paper Series entry represents a collection of presentation visuals associated with the companion report entitled, Knowledge Based Systems: A Critical Survey of Major Concepts, Issues, and Techniques, USL/DBMS NASA/RECON Working Paper Series report number DBMS.NASA/RECON-9. The objectives of the report are to: examine various techniques used to build the KBS; to examine at least one KBS in detail, i.e., a case study; to list and identify limitations and problems with the KBS; to suggest future areas of research; and to provide extensive reference materials.

  9. Examining the Potential of Web-Based Multimedia to Support Complex Fine Motor Skill Learning: An Empirical Study

    ERIC Educational Resources Information Center

    Papastergiou, Marina; Pollatou, Elisana; Theofylaktou, Ioannis; Karadimou, Konstantina

    2014-01-01

    Research on the utilization of the Web for complex fine motor skill learning that involves whole body movements is still scarce. The aim of this study was to evaluate the impact of the introduction of a multimedia web-based learning environment, which was targeted at a rhythmic gymnastics routine consisting of eight fine motor skills, into an…

  10. Dynamic Interaction: A Measurement Development and Empirical Evaluation of Knowledge Based Systems and Web 2.0 Decision Support Mashups

    ERIC Educational Resources Information Center

    Beemer, Brandon Alan

    2010-01-01

    The research presented in this dissertation focuses on the organizational and consumer need for knowledge based support in unstructured domains, by developing a measurement scale for dynamic interaction. Addressing this need is approached and evaluated from two different perspectives. The first approach is the development of Knowledge Based…

  11. Analyzing Interactions by an IIS-Map-Based Method in Face-to-Face Collaborative Learning: An Empirical Study

    ERIC Educational Resources Information Center

    Zheng, Lanqin; Yang, Kaicheng; Huang, Ronghuai

    2012-01-01

    This study proposes a new method named the IIS-map-based method for analyzing interactions in face-to-face collaborative learning settings. This analysis method is conducted in three steps: firstly, drawing an initial IIS-map according to collaborative tasks; secondly, coding and segmenting information flows into information items of IIS; thirdly,…

  12. Comparison of Expert-Based and Empirical Evaluation Methodologies in the Case of a CBL Environment: The ''Orestis'' Experience

    ERIC Educational Resources Information Center

    Karoulis, Athanasis; Demetriadis, Stavros; Pombortsis, Andreas

    2006-01-01

    This paper compares several interface evaluation methods applied in the case of a computer based learning (CBL) environment, during a longitudinal study performed in three European countries, Greece, Germany, and Holland, and within the framework of an EC funded Leonardo da Vinci program. The paper firstly considers the particularities of the CBL…

  13. A Cross-Cultural Usability Study on the Internationalization of User Interfaces Based on an Empirical Five Factor Model

    ERIC Educational Resources Information Center

    Chakraborty, Joyram

    2009-01-01

    With the internationalization of e-commerce, it is no longer viable to design one user interface for all environments. Web-based applications and services can be accessed from all over the globe. To account for this globalization process, software developers need to understand that simply accounting for language translation of their websites for…

  14. Quantifying Age-based Forest Carbon Dynamics to Estimate Effects of Wildfire on Carbon Balance: a Multi-scale Empirical Approach

    NASA Astrophysics Data System (ADS)

    Raymond, C. L.; McKenzie, D.

    2009-12-01

    Disturbances affect biomass accumulation and net primary productivity (NPP) across large landscapes by altering the age-class distribution of the landscape. For fire disturbances specifically, a theoretical age-class distribution can be analytically derived from the mean fire return interval using a negative exponential model. However, to determine the consequences of these ecosystem-specific fire return intervals for biomass accumulation and NPP, it is necessary to quantify age-based carbon dynamics at a similar scale. We used chronosequences of Forest Inventory and Analysis (FIA) data to fit empirical models of live biomass carbon accumulation and NPP as a function of stand age. Models were fit at both coarse (ecosections) and fine (potential vegetation types) scales for the forested region of Washington, USA. At the ecosection scale, the Western Cascades and the Coastal ecosection had the highest levels of live biomass C (26.8 and 22.0 kg C/m2 respectively). The fitted maximum live biomass C was lower in the Eastern Cascades (12.2 kg C/m2) and lowest in the Okanogan Highlands (7.56 kg C/m2). However, the order of the ecosections differed for the rate at which these maximums were reached. The Coast and Okanogan Highlands reached maximum live biomass more rapidly than the Eastern and Western Cascades. For the fitted NPP models, maximum NPP was highest in the Coastal ecosection (0.699 kg C/m^2/yr), lowest in the Eastern Cascades (0.196 kg C/m^2/yr) and Okanogan Highlands (0.195 kg C/m^2/yr), and intermediate in the Western Cascades (0.397 kg C/m^2/yr). Surprisingly, all ecosections reached maximum NPP at a similar stand age (approximately 80 years). We then developed similar models at the scale of potential vegetation types within ecosections. These age-based patterns of carbon dynamics, in combination with landscape age-class distributions, provide an empirical approach for estimating the impact of wildfire on biomass accumulation and NPP at the ecosystem scale

  15. Searching transients in large-scale surveys. A method based on the Abbe value

    NASA Astrophysics Data System (ADS)

    Mowlavi, N.

    2014-08-01

    Aims: A new method is presented to identify transient candidates in large-scale surveys based on the variability pattern in their light curves. Methods: The method is based on the Abbe value, Ab, that estimates the smoothness of a light curve, and on a newly introduced value called the excess Abbe and denoted excessAb, that estimates the regularity of the light curve variability pattern over the duration of the observations. Results: Based on simulated light curves, transients are shown to occupy a specific region in the {diagram} diagram, distinct from sources presenting pulsating-like features in their light curves or having featureless light curves. The method is tested on real light curves taken from EROS-2 and OGLE-II surveys in a 0.50° × 0.17° field of the sky in the Large Magellanic Cloud centered at RA(J2000) = 5h25m56.5s and Dec(J2000) = -69d29m43.3s. The method identifies 43 EROS-2 transient candidates out of a total of 1300 variable stars, and 19 more OGLE-II candidates, 10 of which do not have any EROS-2 variable star matches and which would need further confirmation to assess their reliability. The efficiency of the method is further tested by comparing the list of transient candidates with known Be stars in the literature. It is shown that all Be stars known in the studied field of view with detectable bursts or outbursts are successfully extracted by the method. In addition, four new transient candidates displaying bursts and/or outbursts are found in the field, of which at least two are good new Be candidates. Conclusions: The new method proves to be a potentially powerful tool to extract transient candidates from large-scale multi-epoch surveys. The better the photometric measurement uncertainties are, the cleaner the list of detected transient candidates is. In addition, the diagram diagram is shown to be a good diagnostic tool to check the data quality of multi-epoch photometric surveys. A trend of instrumental and/or data reduction origin

  16. Exploring knowledge, attitudes, and practices related to alcohol in Mongolia: a national population-based survey

    PubMed Central

    2013-01-01

    Background The leading cause of mortality in Mongolia is Non-Communicable Disease. Alcohol is recognised by the World Health Organization as one of the four major disease drivers and so, in order to better understand and triangulate recent national burden-of-disease surveys and to inform policy responses to alcohol consumption in Mongolia, a national Knowledge, Attitudes and Practices survey was conducted. Focusing on Non-Communicable Diseases and their risk factors, this publication explores the alcohol-related findings of this national survey. Methods A door-to-door, household-based questionnaire was conducted on 3450 people from across Mongolia. Participants were recruited using a multi-stage random cluster sampling technique, and eligibility was granted to permanent residents of households who were aged between 15 and 64 years. A nationally representative sample size was calculated, based on methodologies aligned with the WHO STEPwise approach to Surveillance. Results Approximately 50% of males and 30% of females were found to be current drinkers of alcohol. Moreover, nine in ten respondents agreed that heavy episodic drinking of alcohol is common among Mongolians, and the harms of daily alcohol consumption were generally perceived to be high. Indeed, 90% of respondents regarded daily alcohol consumption as either ‘harmful’ or ‘very harmful’. Interestingly, morning drinking, suggestive of problematic drinking, was highest in rural men and was associated with lower-levels of education and unemployment. Conclusion This research suggests that Mongolia faces an epidemiological challenge in addressing the burden of alcohol use and related problems. Males, rural populations and those aged 25-34 years exhibited the highest levels of risky drinking practices, while urban populations exhibit higher levels of general alcohol consumption. These findings suggest a focus and context for public health measures addressing alcohol-related harm in Mongolia. PMID

  17. Estimating HIV Prevalence in Zimbabwe Using Population-Based Survey Data

    PubMed Central

    Chinomona, Amos; Mwambi, Henry Godwell

    2015-01-01

    Estimates of HIV prevalence computed using data obtained from sampling a subgroup of the national population may lack the representativeness of all the relevant domains of the population. These estimates are often computed on the assumption that HIV prevalence is uniform across all domains of the population. Use of appropriate statistical methods together with population-based survey data can enhance better estimation of national and subgroup level HIV prevalence and can provide improved explanations of the variation in HIV prevalence across different domains of the population. In this study we computed design-consistent estimates of HIV prevalence, and their respective 95% confidence intervals at both the national and subgroup levels. In addition, we provided a multivariable survey logistic regression model from a generalized linear modelling perspective for explaining the variation in HIV prevalence using demographic, socio-economic, socio-cultural and behavioural factors. Essentially, this study borrows from the proximate determinants conceptual framework which provides guiding principles upon which socio-economic and socio-cultural variables affect HIV prevalence through biological behavioural factors. We utilize the 2010–11 Zimbabwe Demographic and Health Survey (2010–11 ZDHS) data (which are population based) to estimate HIV prevalence in different categories of the population and for constructing the logistic regression model. It was established that HIV prevalence varies greatly with age, gender, marital status, place of residence, literacy level, belief on whether condom use can reduce the risk of contracting HIV and level of recent sexual activity whereas there was no marked variation in HIV prevalence with social status (measured using a wealth index), method of contraceptive and an individual’s level of education. PMID:26624280

  18. Random sample community-based health surveys: does the effort to reach participants matter?

    PubMed Central

    Messiah, Antoine; Castro, Grettel; Rodríguez de la Vega, Pura; Acuna, Juan M

    2014-01-01

    Objectives Conducting health surveys with community-based random samples are essential to capture an otherwise unreachable population, but these surveys can be biased if the effort to reach participants is insufficient. This study determines the desirable amount of effort to minimise such bias. Design A household-based health survey with random sampling and face-to-face interviews. Up to 11 visits, organised by canvassing rounds, were made to obtain an interview. Setting Single-family homes in an underserved and understudied population in North Miami-Dade County, Florida, USA. Participants Of a probabilistic sample of 2200 household addresses, 30 corresponded to empty lots, 74 were abandoned houses, 625 households declined to participate and 265 could not be reached and interviewed within 11 attempts. Analyses were performed on the 1206 remaining households. Primary outcome Each household was asked if any of their members had been told by a doctor that they had high blood pressure, heart disease including heart attack, cancer, diabetes, anxiety/ depression, obesity or asthma. Responses to these questions were analysed by the number of visit attempts needed to obtain the interview. Results Return per visit fell below 10% after four attempts, below 5% after six attempts and below 2% after eight attempts. As the effort increased, household size decreased, while household income and the percentage of interviewees active and employed increased; proportion of the seven health conditions decreased, four of which did so significantly: heart disease 20.4–9.2%, high blood pressure 63.5–58.1%, anxiety/depression 24.4–9.2% and obesity 21.8–12.6%. Beyond the fifth attempt, however, cumulative percentages varied by less than 1% and precision varied by less than 0.1%. Conclusions In spite of the early and steep drop, sustaining at least five attempts to reach participants is necessary to reduce selection bias. PMID:25510887

  19. A national, cross-sectional survey of children's hospital-based safety resource centres

    PubMed Central

    Kendi, Sadiqa; Zonfrillo, Mark R; Seaver Hill, Karen; Arbogast, Kristy B; Gittelman, Michael A

    2014-01-01

    Objective To describe the location, staffing, clientele, safety product disbursement patterns, education provided and sustainability of safety resource centres (SRCs) in US children's hospitals. Methods A cross-sectional survey was distributed to children's hospital-based SRC directors. Survey categories included: funding sources, customer base, items sold, items given free of charge, education provided and directors’ needs. Results 32/38 (84.2%) SRC sites (affiliated with 30 hospitals) completed the survey. SRCs were in many hospital locations including lobby (28.1%), family resource centres (12.5%), gift shop/retail space (18.8%), mobile units (18.8%) and patient clinics (12.5%). 19% of respondents reported that their SRC was financially self-sustainable. Sales to patients predominated (mean of 44%); however, hospital employees made up a mean of 20% (range 0–60%) of sales. 78.1% of SRCs had products for children with special healthcare needs. Documentation kept at SRC sites included items purchased (96.9%), items given free of charge (65.6%) and customer demographics (50%). 56.3% of SRCs provided formal injury prevention education classes. The SRCs’ directors’ most important needs were finances (46.9%), staffing (50%) and space (46.9%). All of the directors were ‘somewhat interested’ or ‘very interested’ in each of the following: creation of a common SRC listserv, national SRC data bank and multisite SRC research platform. Conclusions SRCs are located in many US children's hospitals, and can be characterised as heterogeneous in location, products sold, data kept and ability to be financially sustained. Further research is needed to determine best practices for SRCs to maximise their impact on injury prevention. PMID:24667383

  20. Reliability of Nationwide Prevalence Estimates of Dementia: A Critical Appraisal Based on Brazilian Surveys

    PubMed Central

    2015-01-01

    Background The nationwide dementia prevalence is usually calculated by applying the results of local surveys to countries’ populations. To evaluate the reliability of such estimations in developing countries, we chose Brazil as an example. We carried out a systematic review of dementia surveys, ascertained their risk of bias, and present the best estimate of occurrence of dementia in Brazil. Methods and Findings We carried out an electronic search of PubMed, Latin-American databases, and a Brazilian thesis database for surveys focusing on dementia prevalence in Brazil. The systematic review was registered at PROSPERO (CRD42014008815). Among the 35 studies found, 15 analyzed population-based random samples. However, most of them utilized inadequate criteria for diagnostics. Six studies without these limitations were further analyzed to assess the risk of selection, attrition, outcome and population bias as well as several statistical issues. All the studies presented moderate or high risk of bias in at least two domains due to the following features: high non-response, inaccurate cut-offs, and doubtful accuracy of the examiners. Two studies had limited external validity due to high rates of illiteracy or low income. The three studies with adequate generalizability and the lowest risk of bias presented a prevalence of dementia between 7.1% and 8.3% among subjects aged 65 years and older. However, after adjustment for accuracy of screening, the best available evidence points towards a figure between 15.2% and 16.3%. Conclusions The risk of bias may strongly limit the generalizability of dementia prevalence estimates in developing countries. Extrapolations that have already been made for Brazil and Latin America were based on a prevalence that should have been adjusted for screening accuracy or not used at all due to severe bias. Similar evaluations regarding other developing countries are needed in order to verify the scope of these limitations. PMID:26131563