Science.gov

Sample records for benchmark evaluation project

  1. Criticality safety benchmark evaluation project: Recovering the past

    SciTech Connect

    Trumble, E.F.

    1997-06-01

    A very brief summary of the Criticality Safety Benchmark Evaluation Project of the Westinghouse Savannah River Company is provided in this paper. The purpose of the project is to provide a source of evaluated criticality safety experiments in an easily usable format. Another project goal is to search for any experiments that may have been lost or contain discrepancies, and to determine if they can be used. Results of evaluated experiments are being published as US DOE handbooks.

  2. INTEGRAL BENCHMARKS AVAILABLE THROUGH THE INTERNATIONAL REACTOR PHYSICS EXPERIMENT EVALUATION PROJECT AND THE INTERNATIONAL CRITICALITY SAFETY BENCHMARK EVALUATION PROJECT

    SciTech Connect

    J. Blair Briggs; Lori Scott; Enrico Sartori; Yolanda Rugama

    2008-09-01

    Interest in high-quality integral benchmark data is increasing as efforts to quantify and reduce calculational uncertainties accelerate to meet the demands of next generation reactor and advanced fuel cycle concepts. The International Reactor Physics Experiment Evaluation Project (IRPhEP) and the International Criticality Safety Benchmark Evaluation Project (ICSBEP) continue to expand their efforts and broaden their scope to identify, evaluate, and provide integral benchmark data for method and data validation. Benchmark model specifications provided by these two projects are used heavily by the international reactor physics, nuclear data, and criticality safety communities. Thus far, 14 countries have contributed to the IRPhEP, and 20 have contributed to the ICSBEP. The status of the IRPhEP and ICSBEP is discussed in this paper, and the future of the two projects is outlined and discussed. Selected benchmarks that have been added to the IRPhEP and ICSBEP handbooks since PHYSOR’06 are highlighted, and the future of the two projects is discussed.

  3. Benchmark Data Through The International Reactor Physics Experiment Evaluation Project (IRPHEP)

    SciTech Connect

    J. Blair Briggs; Dr. Enrico Sartori

    2005-09-01

    The International Reactor Physics Experiments Evaluation Project (IRPhEP) was initiated by the Organization for Economic Cooperation and Development (OECD) Nuclear Energy Agency’s (NEA) Nuclear Science Committee (NSC) in June of 2002. The IRPhEP focus is on the derivation of internationally peer reviewed benchmark models for several types of integral measurements, in addition to the critical configuration. While the benchmarks produced by the IRPhEP are of primary interest to the Reactor Physics Community, many of the benchmarks can be of significant value to the Criticality Safety and Nuclear Data Communities. Benchmarks that support the Next Generation Nuclear Plant (NGNP), for example, also support fuel manufacture, handling, transportation, and storage activities and could challenge current analytical methods. The IRPhEP is patterned after the International Criticality Safety Benchmark Evaluation Project (ICSBEP) and is closely coordinated with the ICSBEP. This paper highlights the benchmarks that are currently being prepared by the IRPhEP that are also of interest to the Criticality Safety Community. The different types of measurements and associated benchmarks that can be expected in the first publication and beyond are described. The protocol for inclusion of IRPhEP benchmarks as ICSBEP benchmarks and for inclusion of ICSBEP benchmarks as IRPhEP benchmarks is detailed. The format for IRPhEP benchmark evaluations is described as an extension of the ICSBEP format. Benchmarks produced by the IRPhEP add new dimension to criticality safety benchmarking efforts and expand the collection of available integral benchmarks for nuclear data testing. The first publication of the "International Handbook of Evaluated Reactor Physics Benchmark Experiments" is scheduled for January of 2006.

  4. The Activities of the International Criticality Safety Benchmark Evaluation Project (ICSBEP)

    SciTech Connect

    Briggs, Joseph Blair

    2001-10-01

    The International Criticality Safety Benchmark Evaluation Project (ICSBEP) was initiated in 1992 by the United States Department of Energy. The ICSBEP became an official activity of the Organization for Economic Cooperation and Development (OECD) – Nuclear Energy Agency (NEA) in 1995. Representatives from the United States, United Kingdom, France, Japan, the Russian Federation, Hungary, Republic of Korea, Slovenia, Yugoslavia, Kazakhstan, Spain, and Israel are now participating. The purpose of the ICSBEP is to identify, evaluate, verify, and formally document a comprehensive and internationally peer-reviewed set of criticality safety benchmark data. The work of the ICSBEP is published as an OECD handbook entitled “International Handbook of Evaluated Criticality Safety Benchmark Experiments”. The 2001 Edition of the Handbook contains benchmark specifications for 2642 critical or subcritical configurations that are intended for use in validation efforts and for testing basic nuclear data.

  5. Growth and Expansion of the International Criticality Safety Benchmark Evaluation Project and the Newly Organized International Reactor Physics Experiment Evaluation Project

    SciTech Connect

    J. Blair Briggs; Lori Scott; Yolanda Rugama; Enrico Satori

    2007-05-01

    Since ICNC 2003, the International Criticality Safety Benchmark Evaluation Project (ICSBEP) has continued to expand its efforts and broaden its scope. Criticality-alarm / shielding type benchmarks and fundamental physics measurements that are relevant to criticality safety applications are not only included in the scope of the project, but benchmark data are also included in the latest version of the handbook. A considerable number of improvements have been made to the searchable database, DICE and the criticality-alarm / shielding benchmarks and fundamental physics measurements have been included in the database. There were 12 countries participating on the ICSBEP in 2003. That number has increased to 18 with recent contributions of data and/or resources from Brazil, Czech Republic, Poland, India, Canada, and China. South Africa, Germany, Argentina, and Australia have been invited to participate. Since ICNC 2003, the contents of the “International Handbook of Evaluated Criticality Safety Benchmark Experiments” have increased from 350 evaluations (28,000 pages) containing benchmark specifications for 3070 critical or subcritical configurations to 442 evaluations (over 38,000 pages) containing benchmark specifications for 3957 critical or subcritical configurations, 23 criticality-alarm-placement / shielding configurations with multiple dose points for each, and 20 configurations that have been categorized as fundamental physics measurements that are relevant to criticality safety applications in the 2006 Edition of the ICSBEP Handbook. Approximately 30 new evaluations and 250 additional configurations are expected to be added to the 2007 Edition of the Handbook. Since ICNC 2003, a reactor physics counterpart to the ICSBEP, The International Reactor Physics Experiment Evaluation Project (IRPhEP) was initiated. Beginning in 1999, the IRPhEP was conducted as a pilot activity by the by the Organization of Economic Cooperation and Development (OECD) Nuclear Energy

  6. Benchmarking and validation activities within JEFF project

    NASA Astrophysics Data System (ADS)

    Cabellos, O.; Alvarez-Velarde, F.; Angelone, M.; Diez, C. J.; Dyrda, J.; Fiorito, L.; Fischer, U.; Fleming, M.; Haeck, W.; Hill, I.; Ichou, R.; Kim, D. H.; Klix, A.; Kodeli, I.; Leconte, P.; Michel-Sendis, F.; Nunnenmann, E.; Pecchia, M.; Peneliau, Y.; Plompen, A.; Rochman, D.; Romojaro, P.; Stankovskiy, A.; Sublet, J. Ch.; Tamagno, P.; Marck, S. van der

    2017-09-01

    The challenge for any nuclear data evaluation project is to periodically release a revised, fully consistent and complete library, with all needed data and covariances, and ensure that it is robust and reliable for a variety of applications. Within an evaluation effort, benchmarking activities play an important role in validating proposed libraries. The Joint Evaluated Fission and Fusion (JEFF) Project aims to provide such a nuclear data library, and thus, requires a coherent and efficient benchmarking process. The aim of this paper is to present the activities carried out by the new JEFF Benchmarking and Validation Working Group, and to describe the role of the NEA Data Bank in this context. The paper will also review the status of preliminary benchmarking for the next JEFF-3.3 candidate cross-section files.

  7. National healthcare capital project benchmarking--an owner's perspective.

    PubMed

    Kahn, Noah

    2009-01-01

    Few sectors of the economy have been left unscathed in these economic times. Healthcare construction has been less affected than residential and nonresidential construction sectors, but driven by re-evaluation of healthcare system capital plans, projects are now being put on hold or canceled. The industry is searching for ways to improve the value proposition for project delivery and process controls. In other industries, benchmarking component costs has led to significant, sustainable reductions in costs and cost variations. Kaiser Permanente and the Construction Industry Institute (CII), a research component of the University of Texas at Austin, an industry leader in benchmarking, have joined with several other organizations to work on a national benchmarking and metrics program to gauge the performance of healthcare facility projects. This initiative will capture cost, schedule, delivery method, change, functional, operational, and best practice metrics. This program is the only one of its kind. The CII Web-based interactive reporting system enables a company to view its information and mine industry data. Benchmarking is a tool for continuous improvement that is capable not only of grading outcomes; it can inform all aspects of the healthcare design and construction process and ultimately help moderate the increasing cost of delivering healthcare.

  8. A European benchmarking system to evaluate in-hospital mortality rates in acute coronary syndrome: the EURHOBOP project.

    PubMed

    Dégano, Irene R; Subirana, Isaac; Torre, Marina; Grau, María; Vila, Joan; Fusco, Danilo; Kirchberger, Inge; Ferrières, Jean; Malmivaara, Antti; Azevedo, Ana; Meisinger, Christa; Bongard, Vanina; Farmakis, Dimitros; Davoli, Marina; Häkkinen, Unto; Araújo, Carla; Lekakis, John; Elosua, Roberto; Marrugat, Jaume

    2015-03-01

    Hospital performance models in acute myocardial infarction (AMI) are useful to assess patient management. While models are available for individual countries, mainly US, cross-European performance models are lacking. Thus, we aimed to develop a system to benchmark European hospitals in AMI and percutaneous coronary intervention (PCI), based on predicted in-hospital mortality. We used the EURopean HOspital Benchmarking by Outcomes in ACS Processes (EURHOBOP) cohort to develop the models, which included 11,631 AMI patients and 8276 acute coronary syndrome (ACS) patients who underwent PCI. Models were validated with a cohort of 55,955 European ACS patients. Multilevel logistic regression was used to predict in-hospital mortality in European hospitals for AMI and PCI. Administrative and clinical models were constructed with patient- and hospital-level covariates, as well as hospital- and country-based random effects. Internal cross-validation and external validation showed good discrimination at the patient level and good calibration at the hospital level, based on the C-index (0.736-0.819) and the concordance correlation coefficient (55.4%-80.3%). Mortality ratios (MRs) showed excellent concordance between administrative and clinical models (97.5% for AMI and 91.6% for PCI). Exclusion of transfers and hospital stays ≤1day did not affect in-hospital mortality prediction in sensitivity analyses, as shown by MR concordance (80.9%-85.4%). Models were used to develop a benchmarking system to compare in-hospital mortality rates of European hospitals with similar characteristics. The developed system, based on the EURHOBOP models, is a simple and reliable tool to compare in-hospital mortality rates between European hospitals in AMI and PCI. Copyright © 2015 Elsevier Ireland Ltd. All rights reserved.

  9. [Results of the evaluation of German benchmarking networks funded by the Ministry of Health].

    PubMed

    de Cruppé, Werner; Blumenstock, Gunnar; Fischer, Imma; Selbmann, Hans-Konrad; Geraedts, Max

    2011-01-01

    Nine out of ten demonstration projects on clinical benchmarking funded by the German Ministry of Health were evaluated. Project reports and interviews were uniformly analysed using a list of criteria and a scheme to categorize the realized benchmarking approach. At the end of the funding period four benchmarking networks had implemented all benchmarking steps, and six were continued after funding had expired. The improvement of outcome quality cannot yet be assessed. Factors promoting the introduction of benchmarking networks with regard to organisational and process aspects of benchmarking implementation were derived. Copyright © 2011. Published by Elsevier GmbH.

  10. Providing Nuclear Criticality Safety Analysis Education through Benchmark Experiment Evaluation

    SciTech Connect

    John D. Bess; J. Blair Briggs; David W. Nigg

    2009-11-01

    One of the challenges that today's new workforce of nuclear criticality safety engineers face is the opportunity to provide assessment of nuclear systems and establish safety guidelines without having received significant experience or hands-on training prior to graduation. Participation in the International Criticality Safety Benchmark Evaluation Project (ICSBEP) and/or the International Reactor Physics Experiment Evaluation Project (IRPhEP) provides students and young professionals the opportunity to gain experience and enhance critical engineering skills.

  11. Benchmark Evaluation of Plutonium Nitrate Solution Arrays

    SciTech Connect

    M. A. Marshall; J. D. Bess

    2011-09-01

    In October and November of 1981 thirteen approach-to-critical experiments were performed on a remote split table machine (RSTM) in the Critical Mass Laboratory of Pacific Northwest Laboratory (PNL) in Richland, Washington, using planar arrays of polyethylene bottles filled with plutonium (Pu) nitrate solution. Arrays of up to sixteen bottles were used to measure the critical number of bottles and critical array spacing with a tight fitting Plexiglas{reg_sign} reflector on all sides of the arrays except the top. Some experiments used Plexiglas shells fitted around each bottles to determine the effect of moderation on criticality. Each bottle contained approximately 2.4 L of Pu(NO3)4 solution with a Pu content of 105 g Pu/L and a free acid molarity H+ of 5.1. The plutonium was of low 240Pu (2.9 wt.%) content. These experiments were performed to fill a gap in experimental data regarding criticality limits for storing and handling arrays of Pu solution in reprocessing facilities. Of the thirteen approach-to-critical experiments eleven resulted in extrapolations to critical configurations. Four of the approaches were extrapolated to the critical number of bottles; these were not evaluated further due to the large uncertainty associated with the modeling of a fraction of a bottle. The remaining seven approaches were extrapolated to critical array spacing of 3-4 and 4-4 arrays; these seven critical configurations were evaluation for inclusion as acceptable benchmark experiments in the International Criticality Safety Benchmark Evaluation Project (ICSBEP) Handbook. Detailed and simple models of these configurations were created and the associated bias of these simplifications was determined to range from 0.00116 and 0.00162 {+-} 0.00006 ?keff. Monte Carlo analysis of all models was completed using MCNP5 with ENDF/BVII.0 neutron cross section libraries. A thorough uncertainty analysis of all critical, geometric, and material parameters was performed using parameter

  12. Performance Evaluation and Benchmarking of Intelligent Systems

    SciTech Connect

    Madhavan, Raj; Messina, Elena; Tunstel, Edward

    2009-09-01

    To design and develop capable, dependable, and affordable intelligent systems, their performance must be measurable. Scientific methodologies for standardization and benchmarking are crucial for quantitatively evaluating the performance of emerging robotic and intelligent systems technologies. There is currently no accepted standard for quantitatively measuring the performance of these systems against user-defined requirements; and furthermore, there is no consensus on what objective evaluation procedures need to be followed to understand the performance of these systems. The lack of reproducible and repeatable test methods has precluded researchers working towards a common goal from exchanging and communicating results, inter-comparing system performance, and leveraging previous work that could otherwise avoid duplication and expedite technology transfer. Currently, this lack of cohesion in the community hinders progress in many domains, such as manufacturing, service, healthcare, and security. By providing the research community with access to standardized tools, reference data sets, and open source libraries of solutions, researchers and consumers will be able to evaluate the cost and benefits associated with intelligent systems and associated technologies. In this vein, the edited book volume addresses performance evaluation and metrics for intelligent systems, in general, while emphasizing the need and solutions for standardized methods. To the knowledge of the editors, there is not a single book on the market that is solely dedicated to the subject of performance evaluation and benchmarking of intelligent systems. Even books that address this topic do so only marginally or are out of date. The research work presented in this volume fills this void by drawing from the experiences and insights of experts gained both through theoretical development and practical implementation of intelligent systems in a variety of diverse application domains. The book presents

  13. Ground truth and benchmarks for performance evaluation

    NASA Astrophysics Data System (ADS)

    Takeuchi, Ayako; Shneier, Michael; Hong, Tsai Hong; Chang, Tommy; Scrapper, Christopher; Cheok, Geraldine S.

    2003-09-01

    Progress in algorithm development and transfer of results to practical applications such as military robotics requires the setup of standard tasks, of standard qualitative and quantitative measurements for performance evaluation and validation. Although the evaluation and validation of algorithms have been discussed for over a decade, the research community still faces a lack of well-defined and standardized methodology. The range of fundamental problems include a lack of quantifiable measures of performance, a lack of data from state-of-the-art sensors in calibrated real-world environments, and a lack of facilities for conducting realistic experiments. In this research, we propose three methods for creating ground truth databases and benchmarks using multiple sensors. The databases and benchmarks will provide researchers with high quality data from suites of sensors operating in complex environments representing real problems of great relevance to the development of autonomous driving systems. At NIST, we have prototyped a High Mobility Multi-purpose Wheeled Vehicle (HMMWV) system with a suite of sensors including a Riegl ladar, GDRS ladar, stereo CCD, several color cameras, Global Position System (GPS), Inertial Navigation System (INS), pan/tilt encoders, and odometry . All sensors are calibrated with respect to each other in space and time. This allows a database of features and terrain elevation to be built. Ground truth for each sensor can then be extracted from the database. The main goal of this research is to provide ground truth databases for researchers and engineers to evaluate algorithms for effectiveness, efficiency, reliability, and robustness, thus advancing the development of algorithms.

  14. Monte Carlo Eigenvalue Calculations with ENDF/B-VI.8, JEFF-3.0, and JENDL-3.3 Cross Sections for a Selection of International Criticality Safety Benchmark Evaluation Project Handbook Benchmarks

    SciTech Connect

    Kahler, A.C

    2003-10-15

    Continuous-energy Monte Carlo eigenvalue calculations have been performed for a selection of HEU-MET-FAST, IEU-MET-FAST, HEU-SOL-THERM, LEU-COMP-THERM, and LEU-SOL-THERM benchmarks using ENDF/B (primarily VI.8), JEFF-3.0, and JENDL-3.3 cross sections. These benchmarks allow for testing the cross-section data for both common reactor nuclides such as {sup 1}H, {sup 16}O, and {sup 235,238}U and structural and shielding elements such as Al, Ti, Fe, Ni, and Pb. The latest cross-section libraries yield near-unity eigenvalues for unreflected or water-reflected HEU-SOL-THERM and LEU-SOL-THERM systems. Near-unity eigenvalues are also obtained for bare HEU-MET-FAST and IEU-MET-FAST systems, but small deviations from unity are observed in both FAST and THERM benchmarks as a function of nonhydrogenous reflector material and thickness. The long-standing problem of lower eigenvalues in water-reflected low-enriched-uranium fuel lattice systems remains, regardless of cross-section library.

  15. Performance Evaluation of Supercomputers using HPCC and IMB Benchmarks

    NASA Technical Reports Server (NTRS)

    Saini, Subhash; Ciotti, Robert; Gunney, Brian T. N.; Spelce, Thomas E.; Koniges, Alice; Dossa, Don; Adamidis, Panagiotis; Rabenseifner, Rolf; Tiyyagura, Sunil R.; Mueller, Matthias; Fatoohi, Rod

    2006-01-01

    The HPC Challenge (HPCC) benchmark suite and the Intel MPI Benchmark (IMB) are used to compare and evaluate the combined performance of processor, memory subsystem and interconnect fabric of five leading supercomputers - SGI Altix BX2, Cray XI, Cray Opteron Cluster, Dell Xeon cluster, and NEC SX-8. These five systems use five different networks (SGI NUMALINK4, Cray network, Myrinet, InfiniBand, and NEC IXS). The complete set of HPCC benchmarks are run on each of these systems. Additionally, we present Intel MPI Benchmarks (IMB) results to study the performance of 11 MPI communication functions on these systems.

  16. Benchmark Evaluation of HTR-PROTEUS Pebble Bed Experimental Program

    DOE PAGES

    Bess, John D.; Montierth, Leland; Köberl, Oliver; ...

    2014-10-09

    Benchmark models were developed to evaluate 11 critical core configurations of the HTR-PROTEUS pebble bed experimental program. Various additional reactor physics measurements were performed as part of this program; currently only a total of 37 absorber rod worth measurements have been evaluated as acceptable benchmark experiments for Cores 4, 9, and 10. Dominant uncertainties in the experimental keff for all core configurations come from uncertainties in the ²³⁵U enrichment of the fuel, impurities in the moderator pebbles, and the density and impurity content of the radial reflector. Calculations of keff with MCNP5 and ENDF/B-VII.0 neutron nuclear data are greatermore » than the benchmark values but within 1% and also within the 3σ uncertainty, except for Core 4, which is the only randomly packed pebble configuration. Repeated calculations of keff with MCNP6.1 and ENDF/B-VII.1 are lower than the benchmark values and within 1% (~3σ) except for Cores 5 and 9, which calculate lower than the benchmark eigenvalues within 4σ. The primary difference between the two nuclear data libraries is the adjustment of the absorption cross section of graphite. Simulations of the absorber rod worth measurements are within 3σ of the benchmark experiment values. The complete benchmark evaluation details are available in the 2014 edition of the International Handbook of Evaluated Reactor Physics Benchmark Experiments.« less

  17. Benchmark Evaluation of HTR-PROTEUS Pebble Bed Experimental Program

    SciTech Connect

    Bess, John D.; Montierth, Leland; Köberl, Oliver; Snoj, Luka

    2014-10-09

    Benchmark models were developed to evaluate 11 critical core configurations of the HTR-PROTEUS pebble bed experimental program. Various additional reactor physics measurements were performed as part of this program; currently only a total of 37 absorber rod worth measurements have been evaluated as acceptable benchmark experiments for Cores 4, 9, and 10. Dominant uncertainties in the experimental keff for all core configurations come from uncertainties in the ²³⁵U enrichment of the fuel, impurities in the moderator pebbles, and the density and impurity content of the radial reflector. Calculations of keff with MCNP5 and ENDF/B-VII.0 neutron nuclear data are greater than the benchmark values but within 1% and also within the 3σ uncertainty, except for Core 4, which is the only randomly packed pebble configuration. Repeated calculations of keff with MCNP6.1 and ENDF/B-VII.1 are lower than the benchmark values and within 1% (~3σ) except for Cores 5 and 9, which calculate lower than the benchmark eigenvalues within 4σ. The primary difference between the two nuclear data libraries is the adjustment of the absorption cross section of graphite. Simulations of the absorber rod worth measurements are within 3σ of the benchmark experiment values. The complete benchmark evaluation details are available in the 2014 edition of the International Handbook of Evaluated Reactor Physics Benchmark Experiments.

  18. Demonstration of a Benchmarking Technique to Compare Graduate Education Level of Air Force Project Managers and Selected Benchmarking Partners

    DTIC Science & Technology

    1994-09-01

    the 1,450th file. This 46 technique sampled files on a nonprobability representation basis; some files did not have a nonzero chance of being reviewed...04 ELECTE: 4 t- OF A DEMONSTRATION OF A BENCHMARKING * ’~ TECHNIQUE TO COMPARE GRADUATE EDUCATION LEVELS OF AIR FORCE PROJECT MANAGERS AND SELECTED...8217,- -AIR FORCE INSTITUTE OF TECHNOLOGY .Wright-Patterson Air Force Base, Ohio AFIT/GSM/LAS/94S-3 A DEMONSTRATION OF A BENCHMARKING TECHNIQUE TO COMPARE

  19. Global Gridded Crop Model Evaluation: Benchmarking, Skills, Deficiencies and Implications.

    NASA Technical Reports Server (NTRS)

    Muller, Christoph; Elliott, Joshua; Chryssanthacopoulos, James; Arneth, Almut; Balkovic, Juraj; Ciais, Philippe; Deryng, Delphine; Folberth, Christian; Glotter, Michael; Hoek, Steven; hide

    2017-01-01

    Crop models are increasingly used to simulate crop yields at the global scale, but so far there is no general framework on how to assess model performance. Here we evaluate the simulation results of 14 global gridded crop modeling groups that have contributed historic crop yield simulations for maize, wheat, rice and soybean to the Global Gridded Crop Model Intercomparison (GGCMI) of the Agricultural Model Intercomparison and Improvement Project (AgMIP). Simulation results are compared to reference data at global, national and grid cell scales and we evaluate model performance with respect to time series correlation, spatial correlation and mean bias. We find that global gridded crop models (GGCMs) show mixed skill in reproducing time series correlations or spatial patterns at the different spatial scales. Generally, maize, wheat and soybean simulations of many GGCMs are capable of reproducing larger parts of observed temporal variability (time series correlation coefficients (r) of up to 0.888 for maize, 0.673 for wheat and 0.643 for soybean at the global scale) but rice yield variability cannot be well reproduced by most models. Yield variability can be well reproduced for most major producing countries by many GGCMs and for all countries by at least some. A comparison with gridded yield data and a statistical analysis of the effects of weather variability on yield variability shows that the ensemble of GGCMs can explain more of the yield variability than an ensemble of regression models for maize and soybean, but not for wheat and rice. We identify future research needs in global gridded crop modeling and for all individual crop modeling groups. In the absence of a purely observation-based benchmark for model evaluation, we propose that the best performing crop model per crop and region establishes the benchmark for all others, and modelers are encouraged to investigate how crop model performance can be increased. We make our evaluation system accessible to all

  20. Global gridded crop model evaluation: benchmarking, skills, deficiencies and implications

    NASA Astrophysics Data System (ADS)

    Müller, Christoph; Elliott, Joshua; Chryssanthacopoulos, James; Arneth, Almut; Balkovic, Juraj; Ciais, Philippe; Deryng, Delphine; Folberth, Christian; Glotter, Michael; Hoek, Steven; Iizumi, Toshichika; Izaurralde, Roberto C.; Jones, Curtis; Khabarov, Nikolay; Lawrence, Peter; Liu, Wenfeng; Olin, Stefan; Pugh, Thomas A. M.; Ray, Deepak K.; Reddy, Ashwan; Rosenzweig, Cynthia; Ruane, Alex C.; Sakurai, Gen; Schmid, Erwin; Skalsky, Rastislav; Song, Carol X.; Wang, Xuhui; de Wit, Allard; Yang, Hong

    2017-04-01

    Crop models are increasingly used to simulate crop yields at the global scale, but so far there is no general framework on how to assess model performance. Here we evaluate the simulation results of 14 global gridded crop modeling groups that have contributed historic crop yield simulations for maize, wheat, rice and soybean to the Global Gridded Crop Model Intercomparison (GGCMI) of the Agricultural Model Intercomparison and Improvement Project (AgMIP). Simulation results are compared to reference data at global, national and grid cell scales and we evaluate model performance with respect to time series correlation, spatial correlation and mean bias. We find that global gridded crop models (GGCMs) show mixed skill in reproducing time series correlations or spatial patterns at the different spatial scales. Generally, maize, wheat and soybean simulations of many GGCMs are capable of reproducing larger parts of observed temporal variability (time series correlation coefficients (r) of up to 0.888 for maize, 0.673 for wheat and 0.643 for soybean at the global scale) but rice yield variability cannot be well reproduced by most models. Yield variability can be well reproduced for most major producing countries by many GGCMs and for all countries by at least some. A comparison with gridded yield data and a statistical analysis of the effects of weather variability on yield variability shows that the ensemble of GGCMs can explain more of the yield variability than an ensemble of regression models for maize and soybean, but not for wheat and rice. We identify future research needs in global gridded crop modeling and for all individual crop modeling groups. In the absence of a purely observation-based benchmark for model evaluation, we propose that the best performing crop model per crop and region establishes the benchmark for all others, and modelers are encouraged to investigate how crop model performance can be increased. We make our evaluation system accessible to all

  1. Benchmark Evaluation of the NRAD Reactor LEU Core Startup Measurements

    SciTech Connect

    J. D. Bess; T. L. Maddock; M. A. Marshall

    2011-09-01

    The Neutron Radiography (NRAD) reactor is a 250-kW TRIGA-(Training, Research, Isotope Production, General Atomics)-conversion-type reactor at the Idaho National Laboratory; it is primarily used for neutron radiography analysis of irradiated and unirradiated fuels and materials. The NRAD reactor was converted from HEU to LEU fuel with 60 fuel elements and brought critical on March 31, 2010. This configuration of the NRAD reactor has been evaluated as an acceptable benchmark experiment and is available in the 2011 editions of the International Handbook of Evaluated Criticality Safety Benchmark Experiments (ICSBEP Handbook) and the International Handbook of Evaluated Reactor Physics Benchmark Experiments (IRPhEP Handbook). Significant effort went into precisely characterizing all aspects of the reactor core dimensions and material properties; detailed analyses of reactor parameters minimized experimental uncertainties. The largest contributors to the total benchmark uncertainty were the 234U, 236U, Er, and Hf content in the fuel; the manganese content in the stainless steel cladding; and the unknown level of water saturation in the graphite reflector blocks. A simplified benchmark model of the NRAD reactor was prepared with a keff of 1.0012 {+-} 0.0029 (1s). Monte Carlo calculations with MCNP5 and KENO-VI and various neutron cross section libraries were performed and compared with the benchmark eigenvalue for the 60-fuel-element core configuration; all calculated eigenvalues are between 0.3 and 0.8% greater than the benchmark value. Benchmark evaluations of the NRAD reactor are beneficial in understanding biases and uncertainties affecting criticality safety analyses of storage, handling, or transportation applications with LEU-Er-Zr-H fuel.

  2. Quality in E-Learning--A Conceptual Framework Based on Experiences from Three International Benchmarking Projects

    ERIC Educational Resources Information Center

    Ossiannilsson, E.; Landgren, L.

    2012-01-01

    Between 2008 and 2010, Lund University took part in three international benchmarking projects, "E-xcellence+," the "eLearning Benchmarking Exercise 2009," and the "First Dual-Mode Distance Learning Benchmarking Club." A comparison of these models revealed a rather high level of correspondence. From this finding and…

  3. Quality in E-Learning--A Conceptual Framework Based on Experiences from Three International Benchmarking Projects

    ERIC Educational Resources Information Center

    Ossiannilsson, E.; Landgren, L.

    2012-01-01

    Between 2008 and 2010, Lund University took part in three international benchmarking projects, "E-xcellence+," the "eLearning Benchmarking Exercise 2009," and the "First Dual-Mode Distance Learning Benchmarking Club." A comparison of these models revealed a rather high level of correspondence. From this finding and…

  4. NASA Indexing Benchmarks: Evaluating Text Search Engines

    NASA Technical Reports Server (NTRS)

    Esler, Sandra L.; Nelson, Michael L.

    1997-01-01

    The current proliferation of on-line information resources underscores the requirement for the ability to index collections of information and search and retrieve them in a convenient manner. This study develops criteria for analytically comparing the index and search engines and presents results for a number of freely available search engines. A product of this research is a toolkit capable of automatically indexing, searching, and extracting performance statistics from each of the focused search engines. This toolkit is highly configurable and has the ability to run these benchmark tests against other engines as well. Results demonstrate that the tested search engines can be grouped into two levels. Level one engines are efficient on small to medium sized data collections, but show weaknesses when used for collections 100MB or larger. Level two search engines are recommended for data collections up to and beyond 100MB.

  5. Benchmarking Evaluation Results for Prototype Extravehicular Activity Gloves

    NASA Technical Reports Server (NTRS)

    Aitchison, Lindsay; McFarland, Shane

    2012-01-01

    The Space Suit Assembly (SSA) Development Team at NASA Johnson Space Center has invested heavily in the advancement of rear-entry planetary exploration suit design but largely deferred development of extravehicular activity (EVA) glove designs, and accepted the risk of using the current flight gloves, Phase VI, for unique mission scenarios outside the Space Shuttle and International Space Station (ISS) Program realm of experience. However, as design reference missions mature, the risks of using heritage hardware have highlighted the need for developing robust new glove technologies. To address the technology gap, the NASA Game-Changing Technology group provided start-up funding for the High Performance EVA Glove (HPEG) Project in the spring of 2012. The overarching goal of the HPEG Project is to develop a robust glove design that increases human performance during EVA and creates pathway for future implementation of emergent technologies, with specific aims of increasing pressurized mobility to 60% of barehanded capability, increasing the durability by 100%, and decreasing the potential of gloves to cause injury during use. The HPEG Project focused initial efforts on identifying potential new technologies and benchmarking the performance of current state of the art gloves to identify trends in design and fit leading to establish standards and metrics against which emerging technologies can be assessed at both the component and assembly levels. The first of the benchmarking tests evaluated the quantitative mobility performance and subjective fit of four prototype gloves developed by Flagsuit LLC, Final Frontier Designs, LLC Dover, and David Clark Company as compared to the Phase VI. All of the companies were asked to design and fabricate gloves to the same set of NASA provided hand measurements (which corresponded to a single size of Phase Vi glove) and focus their efforts on improving mobility in the metacarpal phalangeal and carpometacarpal joints. Four test

  6. An evaluation of a benchmarking initiative in extended treatment mental health services.

    PubMed

    Shepherd, Nicole; Meehan, Thomas J; Davidson, Fiona; Stedman, Terry

    2010-08-01

    To evaluate the effectiveness of a benchmarking initiative in facilitating quality improvement and reform in extended care mental health services. Individual interviews and focus group discussions were conducted with 84 staff in 22 extended care mental health services that had previously participated in a State-wide benchmarking exercise in Queensland, Australia. Staff reported positive outcomes from participation in the benchmarking exercise. Information derived from benchmarking provided a different perspective on the strengths and weaknesses of individual services and an opportunity to learn from peer services. Staff in 86% of the services identified issues that needed to be addressed and 64% of services had implemented one or more service improvement projects in response to shortcomings identified through the benchmarking exercise. The collection and reporting of performance data through a process of benchmarking was successful in facilitating service improvement in most of the participating facilities. Engaging services in all stages of the process was considered useful in converting benchmarking data into knowledge that was able to be applied at the local service level.

  7. Benchmarking of municipal waste water treatment plants (an Austrian project).

    PubMed

    Lindtner, S; Kroiss, H; Nowak, O

    2004-01-01

    An Austrian research project focused on the development of process indicators for treatment plants with different process and operation modes. The whole treatment scheme was subdivided into four processes, i.e. mechanical pretreatment (Process 1), mechanical-biological waste water treatment (Process 2), sludge thickening and stabilisation (Process 3) and further sludge treatment and disposal (Process 4). In order to get comparable process indicators it was necessary to subdivide the sample of 76 individual treatment plants all over Austria into five groups according to their mean organic load (COD) in the influent. The specific total yearly costs, the yearly operating costs and the yearly capital costs of the four processes have been related to the yearly average of the measured organic load expressed in COD (110 g COD/pe/d). The specific investment costs for the whole treatment plant and for Process 2 have been related to a calculated standard design capacity of the mechanical-biological part of the treatment plant expressed in COD. The capital costs of processes 1, 3 and 4 have been related to the design capacity of the treatment plant. For each group (related to the size of the plant) a benchmark band has been defined for the total yearly costs, the total yearly operational costs and the total yearly capital costs. For the operational costs of the Processes 1 to 4 one benchmark ([see symbol in text] per pe/year) has been defined for each group. In addition a theoretical cost reduction potential has been calculated. The cost efficiency in regard to water protection and some special sub-processes such as aeration and sludge dewatering has been analysed.

  8. Data Testing CIELO Evaluations with ICSBEP Benchmarks

    SciTech Connect

    Kahler, Albert Comstock

    2016-03-09

    We review criticality data testing performed at Los Alamos with a combination of ENDF/B-VII.1 + potential CIELO nuclear data evaluations. The goal of CIELO is to develop updated, best available evaluated nuclear data files for 1H, 16O, 56Fe, 235,238U and 239Pu. because the major international evaluated nuclear data libraries don’t agree on the internal cross section details of these most important nuclides.

  9. Benchmarking for the Effective Use of Student Evaluation Data

    ERIC Educational Resources Information Center

    Smithson, John; Birks, Melanie; Harrison, Glenn; Nair, Chenicheri Sid; Hitchins, Marnie

    2015-01-01

    Purpose: The purpose of this paper is to examine current approaches to interpretation of student evaluation data and present an innovative approach to developing benchmark targets for the effective and efficient use of these data. Design/Methodology/Approach: This article discusses traditional approaches to gathering and using student feedback…

  10. Benchmarking: A Tool for Web Site Evaluation and Improvement.

    ERIC Educational Resources Information Center

    Misic, Mark M.; Johnson, Kelsey L.

    1999-01-01

    This paper presents a case study on how benchmarking was used to determine how one organization's Web site compared to Web sites of related schools and professional organizations. Highlights include application of metrics, the Web site evaluation form, functional/navigational issues, content and style, and top site generalizations. (Author/LRW)

  11. A comprehensive benchmarking system for evaluating global vegetation models

    NASA Astrophysics Data System (ADS)

    Kelley, D. I.; Prentice, I. C.; Harrison, S. P.; Wang, H.; Simard, M.; Fisher, J. B.; Willis, K. O.

    2013-05-01

    We present a benchmark system for global vegetation models. This system provides a quantitative evaluation of multiple simulated vegetation properties, including primary production; seasonal net ecosystem production; vegetation cover; composition and height; fire regime; and runoff. The benchmarks are derived from remotely sensed gridded datasets and site-based observations. The datasets allow comparisons of annual average conditions and seasonal and inter-annual variability, and they allow the impact of spatial and temporal biases in means and variability to be assessed separately. Specifically designed metrics quantify model performance for each process, and are compared to scores based on the temporal or spatial mean value of the observations and a "random" model produced by bootstrap resampling of the observations. The benchmark system is applied to three models: a simple light-use efficiency and water-balance model (the Simple Diagnostic Biosphere Model: SDBM), the Lund-Potsdam-Jena (LPJ) and Land Processes and eXchanges (LPX) dynamic global vegetation models (DGVMs). In general, the SDBM performs better than either of the DGVMs. It reproduces independent measurements of net primary production (NPP) but underestimates the amplitude of the observed CO2 seasonal cycle. The two DGVMs show little difference for most benchmarks (including the inter-annual variability in the growth rate and seasonal cycle of atmospheric CO2), but LPX represents burnt fraction demonstrably more accurately. Benchmarking also identified several weaknesses common to both DGVMs. The benchmarking system provides a quantitative approach for evaluating how adequately processes are represented in a model, identifying errors and biases, tracking improvements in performance through model development, and discriminating among models. Adoption of such a system would do much to improve confidence in terrestrial model predictions of climate change impacts and feedbacks.

  12. A comprehensive benchmarking system for evaluating global vegetation models

    NASA Astrophysics Data System (ADS)

    Kelley, D. I.; Prentice, I. Colin; Harrison, S. P.; Wang, H.; Simard, M.; Fisher, J. B.; Willis, K. O.

    2012-11-01

    We present a benchmark system for global vegetation models. This system provides a quantitative evaluation of multiple simulated vegetation properties, including primary production; seasonal net ecosystem production; vegetation cover, composition and height; fire regime; and runoff. The benchmarks are derived from remotely sensed gridded datasets and site-based observations. The datasets allow comparisons of annual average conditions and seasonal and inter-annual variability, and they allow the impact of spatial and temporal biases in means and variability to be assessed separately. Specifically designed metrics quantify model performance for each process, and are compared to scores based on the temporal or spatial mean value of the observations and a "random" model produced by bootstrap resampling of the observations. The benchmark system is applied to three models: a simple light-use efficiency and water-balance model (the Simple Diagnostic Biosphere Model: SDBM), and the Lund-Potsdam-Jena (LPJ) and Land Processes and eXchanges (LPX) dynamic global vegetation models (DGVMs). SDBM reproduces observed CO2 seasonal cycles, but its simulation of independent measurements of net primary production (NPP) is too high. The two DGVMs show little difference for most benchmarks (including the inter-annual variability in the growth rate and seasonal cycle of atmospheric CO2), but LPX represents burnt fraction demonstrably more accurately. Benchmarking also identified several weaknesses common to both DGVMs. The benchmarking system provides a quantitative approach for evaluating how adequately processes are represented in a model, identifying errors and biases, tracking improvements in performance through model development, and discriminating among models. Adoption of such a system would do much to improve confidence in terrestrial model predictions of climate change impacts and feedbacks.

  13. COVE 2A Benchmarking calculations using NORIA; Yucca Mountain Site Characterization Project

    SciTech Connect

    Carrigan, C.R.; Bixler, N.E.; Hopkins, P.L.; Eaton, R.R.

    1991-10-01

    Six steady-state and six transient benchmarking calculations have been performed, using the finite element code NORIA, to simulate one-dimensional infiltration into Yucca Mountain. These calculations were made to support the code verification (COVE 2A) activity for the Yucca Mountain Site Characterization Project. COVE 2A evaluates the usefulness of numerical codes for analyzing the hydrology of the potential Yucca Mountain site. Numerical solutions for all cases were found to be stable. As expected, the difficulties and computer-time requirements associated with obtaining solutions increased with infiltration rate. 10 refs., 128 figs., 5 tabs.

  14. A benchmark for fault tolerant flight control evaluation

    NASA Astrophysics Data System (ADS)

    Smaili, H.; Breeman, J.; Lombaerts, T.; Stroosma, O.

    2013-12-01

    A large transport aircraft simulation benchmark (REconfigurable COntrol for Vehicle Emergency Return - RECOVER) has been developed within the GARTEUR (Group for Aeronautical Research and Technology in Europe) Flight Mechanics Action Group 16 (FM-AG(16)) on Fault Tolerant Control (2004 2008) for the integrated evaluation of fault detection and identification (FDI) and reconfigurable flight control strategies. The benchmark includes a suitable set of assessment criteria and failure cases, based on reconstructed accident scenarios, to assess the potential of new adaptive control strategies to improve aircraft survivability. The application of reconstruction and modeling techniques, based on accident flight data, has resulted in high-fidelity nonlinear aircraft and fault models to evaluate new Fault Tolerant Flight Control (FTFC) concepts and their real-time performance to accommodate in-flight failures.

  15. Performance Evaluation and Benchmarking of Next Intelligent Systems

    SciTech Connect

    del Pobil, Angel; Madhavan, Raj; Bonsignorio, Fabio

    2009-10-01

    Performance Evaluation and Benchmarking of Intelligent Systems presents research dedicated to the subject of performance evaluation and benchmarking of intelligent systems by drawing from the experiences and insights of leading experts gained both through theoretical development and practical implementation of intelligent systems in a variety of diverse application domains. This contributed volume offers a detailed and coherent picture of state-of-the-art, recent developments, and further research areas in intelligent systems. The chapters cover a broad range of applications, such as assistive robotics, planetary surveying, urban search and rescue, and line tracking for automotive assembly. Subsystems or components described in this book include human-robot interaction, multi-robot coordination, communications, perception, and mapping. Chapters are also devoted to simulation support and open source software for cognitive platforms, providing examples of the type of enabling underlying technologies that can help intelligent systems to propagate and increase in capabilities. Performance Evaluation and Benchmarking of Intelligent Systems serves as a professional reference for researchers and practitioners in the field. This book is also applicable to advanced courses for graduate level students and robotics professionals in a wide range of engineering and related disciplines including computer science, automotive, healthcare, manufacturing, and service robotics.

  16. 239Pu Resonance Evaluation for Thermal Benchmark System Calculations

    SciTech Connect

    Leal, Luiz C; Noguere, G; De Saint Jean, C; Kahler, A.

    2013-01-01

    Analyses of thermal plutonium solution critical benchmark systems have indicated a deciency in the 239Pu resonance evaluation. To investigate possible solutions to this issue, the Organisation for Economic Co-operation and Development (OECD) Nuclear Energy Agency (NEA) Working Party for Evaluation Cooperation (WPEC) established Subgroup 34 to focus on the reevaluation of the 239Pu resolved resonance parameters. In addition, the impacts of the prompt neutron multiplication (nubar) and the prompt neutron ssion spectrum (PFNS) have been investigated. The objective of this paper is to present the results of the 239Pu resolved resonance evaluation eort.

  17. 239Pu Resonance Evaluation for Thermal Benchmark System Calculations

    NASA Astrophysics Data System (ADS)

    Leal, L. C.; Noguere, G.; de Saint Jean, C.; Kahler, A. C.

    2014-04-01

    Analyses of thermal plutonium solution critical benchmark systems have indicated a deficiency in the 239Pu resonance evaluation. To investigate possible solutions to this issue, the Organisation for Economic Co-operation and Development (OECD) Nuclear Energy Agency (NEA) Working Party for Evaluation Cooperation (WPEC) established Subgroup 34 to focus on the reevaluation of the 239Pu resolved resonance parameters. In addition, the impacts of the prompt neutron multiplicity (νbar) and the prompt neutron fission spectrum (PFNS) have been investigated. The objective of this paper is to present the results of the 239Pu resolved resonance evaluation effort.

  18. Limitations of Community College Benchmarking and Benchmarks

    ERIC Educational Resources Information Center

    Bers, Trudy H.

    2006-01-01

    This chapter distinguishes between benchmarks and benchmarking, describes a number of data and cultural limitations to benchmarking projects, and suggests that external demands for accountability are the dominant reason for growing interest in benchmarking among community colleges.

  19. Limitations of Community College Benchmarking and Benchmarks

    ERIC Educational Resources Information Center

    Bers, Trudy H.

    2006-01-01

    This chapter distinguishes between benchmarks and benchmarking, describes a number of data and cultural limitations to benchmarking projects, and suggests that external demands for accountability are the dominant reason for growing interest in benchmarking among community colleges.

  20. Evaluation of the HTR-10 Reactor as a Benchmark for Physics Code QA

    SciTech Connect

    William K. Terry; Soon Sam Kim; Leland M. Montierth; Joshua J. Cogliati; Abderrafi M. Ougouag

    2006-09-01

    The HTR-10 is a small (10 MWt) pebble-bed research reactor intended to develop pebble-bed reactor (PBR) technology in China. It will be used to test and develop fuel, verify PBR safety features, demonstrate combined electricity production and co-generation of heat, and provide experience in PBR design, operation, and construction. As the only currently operating PBR in the world, the HTR-10 can provide data of great interest to everyone involved in PBR technology. In particular, if it yields data of sufficient quality, it can be used as a benchmark for assessing the accuracy of computer codes proposed for use in PBR analysis. This paper summarizes the evaluation for the International Reactor Physics Experiment Evaluation Project (IRPhEP) of data obtained in measurements of the HTR-10’s initial criticality experiment for use as benchmarks for reactor physics codes.

  1. Thermal Performance Benchmarking (Presentation)

    SciTech Connect

    Moreno, G.

    2014-11-01

    This project will benchmark the thermal characteristics of automotive power electronics and electric motor thermal management systems. Recent vehicle systems will be benchmarked to establish baseline metrics, evaluate advantages and disadvantages of different thermal management systems, and identify areas of improvement to advance the state-of-the-art.

  2. Improving HEI Productivity and Performance through Project Management: Implications from a Benchmarking Case Study

    ERIC Educational Resources Information Center

    Bryde, David; Leighton, Diana

    2009-01-01

    As higher education institutions (HEIs) look to be more commercial in their outlook they are likely to become more dependent on the successful implementation of projects. This article reports a benchmarking survey of PM maturity in a HEI, with the purpose of assessing its capability to implement projects. Data were collected via questionnaires…

  3. Using Web-Based Peer Benchmarking to Manage the Client-Based Project

    ERIC Educational Resources Information Center

    Raska, David; Keller, Eileen Weisenbach; Shaw, Doris

    2013-01-01

    The complexities of integrating client-based projects into marketing courses provide challenges for the instructor but produce richness of context and active learning for the student. This paper explains the integration of Web-based peer benchmarking as a means of improving student performance on client-based projects within a single semester in…

  4. Using Web-Based Peer Benchmarking to Manage the Client-Based Project

    ERIC Educational Resources Information Center

    Raska, David; Keller, Eileen Weisenbach; Shaw, Doris

    2013-01-01

    The complexities of integrating client-based projects into marketing courses provide challenges for the instructor but produce richness of context and active learning for the student. This paper explains the integration of Web-based peer benchmarking as a means of improving student performance on client-based projects within a single semester in…

  5. Middleware Evaluation and Benchmarking for Use in Mission Operations Centers

    NASA Technical Reports Server (NTRS)

    Antonucci, Rob; Waktola, Waka

    2005-01-01

    Middleware technologies have been promoted as timesaving, cost-cutting alternatives to the point-to-point communication used in traditional mission operations systems. However, missions have been slow to adopt the new technology. The lack of existing middleware-based missions has given rise to uncertainty about middleware's ability to perform in an operational setting. Most mission architects are also unfamiliar with the technology and do not know the benefits and detriments to architectural choices - or even what choices are available. We will present the findings of a study that evaluated several middleware options specifically for use in a mission operations system. We will address some common misconceptions regarding the applicability of middleware-based architectures, and we will identify the design decisions and tradeoffs that must be made when choosing a middleware solution. The Middleware Comparison and Benchmark Study was conducted at NASA Goddard Space Flight Center to comprehensively evaluate candidate middleware products, compare and contrast the performance of middleware solutions with the traditional point- to-point socket approach, and assess data delivery and reliability strategies. The study focused on requirements of the Global Precipitation Measurement (GPM) mission, validating the potential use of middleware in the GPM mission ground system. The study was jointly funded by GPM and the Goddard Mission Services Evolution Center (GMSEC), a virtual organization for providing mission enabling solutions and promoting the use of appropriate new technologies for mission support. The study was broken into two phases. To perform the generic middleware benchmarking and performance analysis, a network was created with data producers and consumers passing data between themselves. The benchmark monitored the delay, throughput, and reliability of the data as the characteristics were changed. Measurements were taken under a variety of topologies, data demands

  6. Model evaluation using a community benchmarking system for land surface models

    NASA Astrophysics Data System (ADS)

    Mu, M.; Hoffman, F. M.; Lawrence, D. M.; Riley, W. J.; Keppel-Aleks, G.; Kluzek, E. B.; Koven, C. D.; Randerson, J. T.

    2014-12-01

    Evaluation of atmosphere, ocean, sea ice, and land surface models is an important step in identifying deficiencies in Earth system models and developing improved estimates of future change. For the land surface and carbon cycle, the design of an open-source system has been an important objective of the International Land Model Benchmarking (ILAMB) project. Here we evaluated CMIP5 and CLM models using a benchmarking system that enables users to specify models, data sets, and scoring systems so that results can be tailored to specific model intercomparison projects. Our scoring system used information from four different aspects of global datasets, including climatological mean spatial patterns, seasonal cycle dynamics, interannual variability, and long-term trends. Variable-to-variable comparisons enable investigation of the mechanistic underpinnings of model behavior, and allow for some control of biases in model drivers. Graphics modules allow users to evaluate model performance at local, regional, and global scales. Use of modular structures makes it relatively easy for users to add new variables, diagnostic metrics, benchmarking datasets, or model simulations. Diagnostic results are automatically organized into HTML files, so users can conveniently share results with colleagues. We used this system to evaluate atmospheric carbon dioxide, burned area, global biomass and soil carbon stocks, net ecosystem exchange, gross primary production, ecosystem respiration, terrestrial water storage, evapotranspiration, and surface radiation from CMIP5 historical and ESM historical simulations. We found that the multi-model mean often performed better than many of the individual models for most variables. We plan to publicly release a stable version of the software during fall of 2014 that has land surface, carbon cycle, hydrology, radiation and energy cycle components.

  7. Project Pride Evaluation Report.

    ERIC Educational Resources Information Center

    Jennewein, Marilyn; And Others

    Project PRIDE (Probe, Research, Inquire, Discover, and Evaluate) is evaluated in this report to provide data to be used as a learning tool for project staff and student participants. Major objectives of the project are to provide an inter-disciplinary, objective approach to the study of the American heritage, and to incorporate methods and…

  8. Iowa's Adult Literacy Program Benchmark Projection Report. Program Year 2007, July 1, 2006-June 30, 2007

    ERIC Educational Resources Information Center

    Division of Community Colleges and Workforce Preparation, Iowa Department of Education, 2007

    2007-01-01

    The purpose of this publication is to present Iowa's adult literacy program approved projected benchmark percentage levels for Program Year 2006 (July 1, 2005-June 30, 2006). The passage of the Workforce Investment Act of 1998 (WIA) [Public Law 105-220] by the 105th Congress has ushered in a new era of collaboration, coordinator, cooperation and…

  9. Learning from Follow Up Surveys of Graduates: The Austin Teacher Program and the Benchmark Project. A Discussion Paper.

    ERIC Educational Resources Information Center

    Baker, Thomas E.

    This paper describes Austin College's (Texas) participation in the Benchmark Project, a collaborative followup study of teacher education graduates and their principals, focusing on the second round of data collection. The Benchmark Project was a collaboration of 11 teacher preparation programs that gathered and analyzed data comparing graduates…

  10. BENCHMARK EVALUATION OF THE START-UP CORE REACTOR PHYSICS MEASUREMENTS OF THE HIGH TEMPERATURE ENGINEERING TEST REACTOR

    SciTech Connect

    John Darrell Bess

    2010-05-01

    The benchmark evaluation of the start-up core reactor physics measurements performed with Japan’s High Temperature Engineering Test Reactor, in support of the Next Generation Nuclear Plant Project and Very High Temperature Reactor Program activities at the Idaho National Laboratory, has been completed. The evaluation was performed using MCNP5 with ENDF/B-VII.0 nuclear data libraries and according to guidelines provided for inclusion in the International Reactor Physics Experiment Evaluation Project Handbook. Results provided include updated evaluation of the initial six critical core configurations (five annular and one fully-loaded). The calculated keff eigenvalues agree within 1s of the benchmark values. Reactor physics measurements that were evaluated include reactivity effects measurements such as excess reactivity during the core loading process and shutdown margins for the fully-loaded core, four isothermal temperature reactivity coefficient measurements for the fully-loaded core, and axial reaction rate measurements in the instrumentation columns of three core configurations. The calculated values agree well with the benchmark experiment measurements. Fully subcritical and warm critical configurations of the fully-loaded core were also assessed. The calculated keff eigenvalues for these two configurations also agree within 1s of the benchmark values. The reactor physics measurement data can be used in the validation and design development of future High Temperature Gas-cooled Reactor systems.

  11. Benchmarking Data Sets for the Evaluation of Virtual Ligand Screening Methods: Review and Perspectives.

    PubMed

    Lagarde, Nathalie; Zagury, Jean-François; Montes, Matthieu

    2015-07-27

    Virtual screening methods are commonly used nowadays in drug discovery processes. However, to ensure their reliability, they have to be carefully evaluated. The evaluation of these methods is often realized in a retrospective way, notably by studying the enrichment of benchmarking data sets. To this purpose, numerous benchmarking data sets were developed over the years, and the resulting improvements led to the availability of high quality benchmarking data sets. However, some points still have to be considered in the selection of the active compounds, decoys, and protein structures to obtain optimal benchmarking data sets.

  12. BENCHMARK EVALUATION OF THE INITIAL ISOTHERMAL PHYSICS MEASUREMENTS AT THE FAST FLUX TEST FACILITY

    SciTech Connect

    John Darrell Bess

    2010-05-01

    The benchmark evaluation of the initial isothermal physics tests performed at the Fast Flux Test Facility, in support of Fuel Cycle Research and Development and Generation-IV activities at the Idaho National Laboratory, has been completed. The evaluation was performed using MCNP5 with ENDF/B-VII.0 nuclear data libraries and according to guidelines provided for inclusion in the International Reactor Physics Experiment Evaluation Project Handbook. Results provided include evaluation of the initial fully-loaded core critical, two neutron spectra measurements near the axial core center, 32 reactivity effects measurements (21 control rod worths, two control rod bank worths, six differential control rod worths, two shutdown margins, and one excess reactivity), isothermal temperature coefficient, and low-energy electron and gamma spectra measurements at the core center. All measurements were performed at 400 ºF. There was good agreement between the calculated and benchmark values for the fully-loaded core critical eigenvalue, reactivity effects measurements, and isothermal temperature coefficient. General agreement between benchmark experiment measurements and calculated spectra for neutrons and low-energy gammas at the core midplane exists, but calculations of the neutron spectra below the core and the low-energy gamma spectra at core midplane did not agree well. Homogenization of core components may have had a significant impact upon computational assessment of these effects. Future work includes development of a fully-heterogeneous model for comprehensive evaluation. The reactor physics measurement data can be used in nuclear data adjustment and validation of computational methods for advanced fuel cycle and nuclear reactor systems using Liquid Metal Fast Reactor technology.

  13. Monitoring Based Commissioning: Benchmarking Analysis of 24 UC/CSU/IOU Projects

    SciTech Connect

    Mills, Evan; Mathew, Paul

    2009-04-01

    Buildings rarely perform as intended, resulting in energy use that is higher than anticipated. Building commissioning has emerged as a strategy for remedying this problem in non-residential buildings. Complementing traditional hardware-based energy savings strategies, commissioning is a 'soft' process of verifying performance and design intent and correcting deficiencies. Through an evaluation of a series of field projects, this report explores the efficacy of an emerging refinement of this practice, known as monitoring-based commissioning (MBCx). MBCx can also be thought of as monitoring-enhanced building operation that incorporates three components: (1) Permanent energy information systems (EIS) and diagnostic tools at the whole-building and sub-system level; (2) Retro-commissioning based on the information from these tools and savings accounting emphasizing measurement as opposed to estimation or assumptions; and (3) On-going commissioning to ensure efficient building operations and measurement-based savings accounting. MBCx is thus a measurement-based paradigm which affords improved risk-management by identifying problems and opportunities that are missed with periodic commissioning. The analysis presented in this report is based on in-depth benchmarking of a portfolio of MBCx energy savings for 24 buildings located throughout the University of California and California State University systems. In the course of the analysis, we developed a quality-control/quality-assurance process for gathering and evaluating raw data from project sites and then selected a number of metrics to use for project benchmarking and evaluation, including appropriate normalizations for weather and climate, accounting for variations in central plant performance, and consideration of differences in building types. We performed a cost-benefit analysis of the resulting dataset, and provided comparisons to projects from a larger commissioning 'Meta-analysis' database. A total of 1120

  14. Ready to Retrofit: The Process of Project Team Selection, Building Benchmarking, and Financing Commercial Building Energy Retrofit Projects

    SciTech Connect

    Sanders, Mark D.; Parrish, Kristen; Mathew, Paul

    2012-05-01

    This guide presents a process for three key activities for the building owner in preparing to retrofit existing commercial buildings: selecting project teams, benchmarking the existing building, and financing the retrofit work. Although there are other essential steps in the retrofit process, the three activities presented in this guide are the critical elements where the building owner has the greatest influence on the outcome of the project.

  15. Putting Data to Work: Interim Recommendations from The Benchmarking Project

    ERIC Educational Resources Information Center

    Miles, Marty; Maguire, Sheila; Woodruff-Bolte, Stacy; Clymer, Carol

    2010-01-01

    As public and private funders have focused on evaluating the effectiveness of workforce development programs, a myriad of data collection systems and reporting processes have taken shape. Navigating these systems takes significant time and energy and often saps frontline providers' capacity to use data internally for program improvement.…

  16. Putting Data to Work: Interim Recommendations from The Benchmarking Project

    ERIC Educational Resources Information Center

    Miles, Marty; Maguire, Sheila; Woodruff-Bolte, Stacy; Clymer, Carol

    2010-01-01

    As public and private funders have focused on evaluating the effectiveness of workforce development programs, a myriad of data collection systems and reporting processes have taken shape. Navigating these systems takes significant time and energy and often saps frontline providers' capacity to use data internally for program improvement.…

  17. TPC-V: A Benchmark for Evaluating the Performance of Database Applications in Virtual Environments

    NASA Astrophysics Data System (ADS)

    Sethuraman, Priya; Reza Taheri, H.

    For two decades, TPC benchmarks have been the gold standards for evaluating the performance of database servers. An area that TPC benchmarks had not addressed until now was virtualization. Virtualization is now a major technology in use in data centers, and is the number one technology on Gartner Group's Top Technologies List. In 2009, the TPC formed a Working Group to develop a benchmark specifically intended for virtual environments that run database applications. We will describe the characteristics of this benchmark, and provide a status update on its development.

  18. State Education Agency Communications Process: Benchmark and Best Practices Project. Benchmark and Best Practices Project. Issue No. 01

    ERIC Educational Resources Information Center

    Zavadsky, Heather

    2014-01-01

    The role of state education agencies (SEAs) has shifted significantly from low-profile, compliance activities like managing federal grants to engaging in more complex and politically charged tasks like setting curriculum standards, developing accountability systems, and creating new teacher evaluation systems. The move from compliance-monitoring…

  19. IAEA coordinated research projects on core physics benchmarks for high temperature gas-cooled reactors

    SciTech Connect

    Methnani, M.

    2006-07-01

    High-temperature Gas-Cooled Reactor (HTGR) designs present special computational challenges related to their core physics characteristics, in particular neutron streaming, double heterogeneities, impurities and the random distribution of coated fuel particles in the graphite matrix. In recent years, two consecutive IAEA Coordinated Research Projects (CRP 1 and CRP 5) have focused on code-to-code and code-to-experiment comparisons of representative benchmarks run by several participating international institutes. While the PROTEUS critical HTR experiments provided the test data reference for CRP-1, the more recent CRP-5 data has been made available by the HTTR, HTR-10 and ASTRA test facilities. Other benchmark cases are being considered for the GT-MHR and PBMR core designs. This paper overviews the scope and some sample results of both coordinated research projects. (authors)

  20. Benchmarking Spike-Based Visual Recognition: A Dataset and Evaluation.

    PubMed

    Liu, Qian; Pineda-García, Garibaldi; Stromatias, Evangelos; Serrano-Gotarredona, Teresa; Furber, Steve B

    2016-01-01

    Today, increasing attention is being paid to research into spike-based neural computation both to gain a better understanding of the brain and to explore biologically-inspired computation. Within this field, the primate visual pathway and its hierarchical organization have been extensively studied. Spiking Neural Networks (SNNs), inspired by the understanding of observed biological structure and function, have been successfully applied to visual recognition and classification tasks. In addition, implementations on neuromorphic hardware have enabled large-scale networks to run in (or even faster than) real time, making spike-based neural vision processing accessible on mobile robots. Neuromorphic sensors such as silicon retinas are able to feed such mobile systems with real-time visual stimuli. A new set of vision benchmarks for spike-based neural processing are now needed to measure progress quantitatively within this rapidly advancing field. We propose that a large dataset of spike-based visual stimuli is needed to provide meaningful comparisons between different systems, and a corresponding evaluation methodology is also required to measure the performance of SNN models and their hardware implementations. In this paper we first propose an initial NE (Neuromorphic Engineering) dataset based on standard computer vision benchmarksand that uses digits from the MNIST database. This dataset is compatible with the state of current research on spike-based image recognition. The corresponding spike trains are produced using a range of techniques: rate-based Poisson spike generation, rank order encoding, and recorded output from a silicon retina with both flashing and oscillating input stimuli. In addition, a complementary evaluation methodology is presented to assess both model-level and hardware-level performance. Finally, we demonstrate the use of the dataset and the evaluation methodology using two SNN models to validate the performance of the models and their hardware

  1. Benchmarking Spike-Based Visual Recognition: A Dataset and Evaluation

    PubMed Central

    Liu, Qian; Pineda-García, Garibaldi; Stromatias, Evangelos; Serrano-Gotarredona, Teresa; Furber, Steve B.

    2016-01-01

    Today, increasing attention is being paid to research into spike-based neural computation both to gain a better understanding of the brain and to explore biologically-inspired computation. Within this field, the primate visual pathway and its hierarchical organization have been extensively studied. Spiking Neural Networks (SNNs), inspired by the understanding of observed biological structure and function, have been successfully applied to visual recognition and classification tasks. In addition, implementations on neuromorphic hardware have enabled large-scale networks to run in (or even faster than) real time, making spike-based neural vision processing accessible on mobile robots. Neuromorphic sensors such as silicon retinas are able to feed such mobile systems with real-time visual stimuli. A new set of vision benchmarks for spike-based neural processing are now needed to measure progress quantitatively within this rapidly advancing field. We propose that a large dataset of spike-based visual stimuli is needed to provide meaningful comparisons between different systems, and a corresponding evaluation methodology is also required to measure the performance of SNN models and their hardware implementations. In this paper we first propose an initial NE (Neuromorphic Engineering) dataset based on standard computer vision benchmarksand that uses digits from the MNIST database. This dataset is compatible with the state of current research on spike-based image recognition. The corresponding spike trains are produced using a range of techniques: rate-based Poisson spike generation, rank order encoding, and recorded output from a silicon retina with both flashing and oscillating input stimuli. In addition, a complementary evaluation methodology is presented to assess both model-level and hardware-level performance. Finally, we demonstrate the use of the dataset and the evaluation methodology using two SNN models to validate the performance of the models and their hardware

  2. Applications of Integral Benchmark Data

    SciTech Connect

    Giuseppe Palmiotti; Teruhiko Kugo; Fitz Trumble; Albert C. Kahler; Dale Lancaster

    2014-10-09

    The International Reactor Physics Experiment Evaluation Project (IRPhEP) and the International Criticality Safety Benchmark Evaluation Project (ICSBEP) provide evaluated integral benchmark data that may be used for validation of reactor physics / nuclear criticality safety analytical methods and data, nuclear data testing, advanced modeling and simulation, and safety analysis licensing activities. The handbooks produced by these programs are used in over 30 countries. Five example applications are presented in this paper: (1) Use of IRPhEP Data in Uncertainty Analyses and Cross Section Adjustment, (2) Uncertainty Evaluation Methods for Reactor Core Design at JAEA Using Reactor Physics Experimental Data, (3) Application of Benchmarking Data to a Broad Range of Criticality Safety Problems, (4) Cross Section Data Testing with ICSBEP Benchmarks, and (5) Use of the International Handbook of Evaluated Reactor Physics Benchmark Experiments to Support the Power Industry.

  3. NRC-BNL Benchmark Program on Evaluation of Methods for Seismic Analysis of Coupled Systems

    SciTech Connect

    Chokshi, N.; DeGrassi, G.; Xu, J.

    1999-03-24

    A NRC-BNL benchmark program for evaluation of state-of-the-art analysis methods and computer programs for seismic analysis of coupled structures with non-classical damping is described. The program includes a series of benchmarking problems designed to investigate various aspects of complexities, applications and limitations associated with methods for analysis of non-classically damped structures. Discussions are provided on the benchmarking process, benchmark structural models, and the evaluation approach, as well as benchmarking ground rules. It is expected that the findings and insights, as well as recommendations from this program will be useful in developing new acceptance criteria and providing guidance for future regulatory activities involving licensing applications of these alternate methods to coupled systems.

  4. NRC-BNL BENCHMARK PROGRAM ON EVALUATION OF METHODS FOR SEISMIC ANALYSIS OF COUPLED SYSTEMS.

    SciTech Connect

    XU,J.

    1999-08-15

    A NRC-BNL benchmark program for evaluation of state-of-the-art analysis methods and computer programs for seismic analysis of coupled structures with non-classical damping is described. The program includes a series of benchmarking problems designed to investigate various aspects of complexities, applications and limitations associated with methods for analysis of non-classically damped structures. Discussions are provided on the benchmarking process, benchmark structural models, and the evaluation approach, as well as benchmarking ground rules. It is expected that the findings and insights, as well as recommendations from this program will be useful in developing new acceptance criteria and providing guidance for future regulatory activities involving licensing applications of these alternate methods to coupled systems.

  5. Development of an ICSBEP Benchmark Evaluation, Nearly 20 Years of Experience

    SciTech Connect

    J. Blair Briggs; John D. Bess

    2011-06-01

    The basic structure of all ICSBEP benchmark evaluations is essentially the same and includes (1) a detailed description of the experiment; (2) an evaluation of the experiment, including an exhaustive effort to quantify the effects of uncertainties on measured quantities; (3) a concise presentation of benchmark-model specifications; (4) sample calculation results; and (5) a summary of experimental references. Computer code input listings and other relevant information are generally preserved in appendixes. Details of an ICSBEP evaluation is presented.

  6. Benchmark for evaluation and validation of reactor simulations (BEAVRS)

    SciTech Connect

    Horelik, N.; Herman, B.; Forget, B.; Smith, K.

    2013-07-01

    Advances in parallel computing have made possible the development of high-fidelity tools for the design and analysis of nuclear reactor cores, and such tools require extensive verification and validation. This paper introduces BEAVRS, a new multi-cycle full-core Pressurized Water Reactor (PWR) depletion benchmark based on two operational cycles of a commercial nuclear power plant that provides a detailed description of fuel assemblies, burnable absorbers, in-core fission detectors, core loading patterns, and numerous in-vessel components. This benchmark enables analysts to develop extremely detailed reactor core models that can be used for testing and validation of coupled neutron transport, thermal-hydraulics, and fuel isotopic depletion. The benchmark also provides measured reactor data for Hot Zero Power (HZP) physics tests, boron letdown curves, and three-dimensional in-core flux maps from fifty-eight instrumented assemblies. Initial comparisons between calculations performed with MIT's OpenMC Monte Carlo neutron transport code and measured cycle 1 HZP test data are presented, and these results display an average deviation of approximately 100 pcm for the various critical configurations and control rod worth measurements. Computed HZP radial fission detector flux maps also agree reasonably well with the available measured data. All results indicate that this benchmark will be extremely useful in validation of coupled-physics codes and uncertainty quantification of in-core physics computational predictions. The detailed BEAVRS specification and its associated data package is hosted online at the MIT Computational Reactor Physics Group web site (http://crpg.mit.edu/), where future revisions and refinements to the benchmark specification will be made publicly available. (authors)

  7. Specifications for the Large Core Code Evaluation Working Group Benchmark Problem Four. [LMFBR

    SciTech Connect

    Cowan, C.L.; Protsik, R.

    1981-09-01

    Benchmark studies have been carried out by the members of the Large Core Code Evaluation Working Group (LCCEWG) as part of a broad effort to systematically evaluate the important steps in the reactor design and analysis process for large fast breeder reactors. The specific objectives of the LCCEWG benchmark studies have been: to quantify the accuracy and efficiency of current neutronics methods for large cores; to identify neutronic design problems unique to large breeder reactors; to identify computer code development requirements; and to provide support for large core critical benchmark experiments.

  8. Key findings of the US Cystic Fibrosis Foundation's clinical practice benchmarking project.

    PubMed

    Boyle, Michael P; Sabadosa, Kathryn A; Quinton, Hebe B; Marshall, Bruce C; Schechter, Michael S

    2014-04-01

    Benchmarking is the process of using outcome data to identify high-performing centres and determine practices associated with their outstanding performance. The US Cystic Fibrosis Foundation (CFF) Patient Registry contains centre-specific outcomes data for all CFF-certified paediatric and adult cystic fibrosis (CF) care programmes in the USA. The CFF benchmarking project analysed these registry data, adjusting for differences in patient case mix known to influence outcomes, and identified the top-performing US paediatric and adult CF care programmes for pulmonary and nutritional outcomes. Separate multidisciplinary paediatric and adult benchmarking teams each visited 10 CF care programmes, five in the top quintile for pulmonary outcomes and five in the top quintile for nutritional outcomes. Key practice patterns and approaches present in both paediatric and adult programmes with outstanding clinical outcomes were identified and could be summarised as systems, attitudes, practices, patient/family empowerment and projects. These included: (1) the presence of strong leadership and a well-functioning care team working with a systematic approach to providing consistent care; (2) high expectations for outcomes among providers and families; (3) early and aggressive management of clinical declines, avoiding reliance on 'rescues'; and (4) patients/families that were engaged, empowered and well informed on disease management and its rationale. In summary, assessment of practice patterns at CF care centres with top-quintile pulmonary and nutritional outcomes provides insight into characteristic practices that may aid in optimising patient outcomes.

  9. Towards a common benchmark for long-term process control and monitoring performance evaluation.

    PubMed

    Rosen, C; Jeppsson, U; Vanrolleghem, P A

    2004-01-01

    The COST/IWA benchmark simulation model has been available for seven years. Its primary purpose has been to create a platform for control strategy benchmarking of biological wastewater treatment processes. The fact that the benchmark has resulted in more than 100 publications, not only in Europe but also worldwide, demonstrates the interest for such a tool in the research community. In this paper, an extension of the benchmark simulation model no. 1 (BSM1) is proposed. It aims at facilitating evaluation of two closely related operational tasks: long-term control strategy performance and process monitoring performance. The motivation for the extension is that these two tasks typically act on longer time scales. The extension proposed here consists of 1) prolonging the evaluation period to one year (including influent files), 2) specifying time varying process parameters and 3) including sensor and actuator failures. The prolonged evaluation period is necessary to obtain a relevant and realistic assessment of the effects of such disturbances. Also, a prolonged evaluation period allows for a number of long-term control actions/handles that cannot be evaluated in a realistic fashion in the one week BSM1 evaluation period. In the paper, models for influent file design, parameter changes and sensor failures, initialization procedure and evaluation criteria are discussed. Important remaining topics, for which consensus is required, are identified. The potential of a long-term benchmark is illustrated with an example of process monitoring algorithm benchmarking.

  10. Evaluating the Information Power Grid using the NAS Grid Benchmarks

    NASA Technical Reports Server (NTRS)

    VanderWijngaartm Rob F.; Frumkin, Michael A.

    2004-01-01

    The NAS Grid Benchmarks (NGB) are a collection of synthetic distributed applications designed to rate the performance and functionality of computational grids. We compare several implementations of the NGB to determine programmability and efficiency of NASA's Information Power Grid (IPG), whose services are mostly based on the Globus Toolkit. We report on the overheads involved in porting existing NGB reference implementations to the IPG. No changes were made to the component tasks of the NGB can still be improved.

  11. An Overview of the International Reactor Physics Experiment Evaluation Project

    SciTech Connect

    Briggs, J. Blair; Gulliford, Jim

    2014-10-09

    Interest in high-quality integral benchmark data is increasing as efforts to quantify and reduce calculational uncertainties associated with advanced modeling and simulation accelerate to meet the demands of next generation reactor and advanced fuel cycle concepts. Two Organization for Economic Cooperation and Development (OECD) Nuclear Energy Agency (NEA) activities, the International Criticality Safety Benchmark Evaluation Project (ICSBEP), initiated in 1992, and the International Reactor Physics Experiment Evaluation Project (IRPhEP), initiated in 2003, have been identifying existing integral experiment data, evaluating those data, and providing integral benchmark specifications for methods and data validation for nearly two decades. Data provided by those two projects will be of use to the international reactor physics, criticality safety, and nuclear data communities for future decades. An overview of the IRPhEP and a brief update of the ICSBEP are provided in this paper.

  12. Evaluation of microfinance projects.

    PubMed

    Johnson, S

    1999-08-01

    This paper criticizes the quick system proposed by Henk Moll for evaluating microfinance projects in the article ¿How to Pre-Evaluate Credit Projects in Ten Minutes¿. The author contended that there is a need to emphasize the objectives of the project. The procedure used by Moll, he contended, is applicable only to projects that have only two key objectives, such as credit operations, and the provision of services. Arguments are presented on the three specific questions proposed by Moll, ranging from the availability of externally audited financial reports, the performance of interest rate on loans vis-a-vis the inflation rate, and the provision of loans according to the individual requirements of the borrowers. Lastly, the author emphasizes that the overall approach is not useful and suggests that careful considerations should be observed in the use or abuse of a simple scoring system or checklist such as the one proposed by Moll.

  13. Preliminary Benchmark Evaluation of Japan’s High Temperature Engineering Test Reactor

    SciTech Connect

    John Darrell Bess

    2009-05-01

    A benchmark model of the initial fully-loaded start-up core critical of Japan’s High Temperature Engineering Test Reactor (HTTR) was developed to provide data in support of ongoing validation efforts of the Very High Temperature Reactor Program using publicly available resources. The HTTR is a 30 MWt test reactor utilizing graphite moderation, helium coolant, and prismatic TRISO fuel. The benchmark was modeled using MCNP5 with various neutron cross-section libraries. An uncertainty evaluation was performed by perturbing the benchmark model and comparing the resultant eigenvalues. The calculated eigenvalues are approximately 2-3% greater than expected with an uncertainty of ±0.70%. The primary sources of uncertainty are the impurities in the core and reflector graphite. The release of additional HTTR data could effectively reduce the benchmark model uncertainties and bias. Sensitivity of the results to the graphite impurity content might imply that further evaluation of the graphite content could significantly improve calculated results. Proper characterization of graphite for future Next Generation Nuclear Power reactor designs will improve computational modeling capabilities. Current benchmarking activities include evaluation of the annular HTTR cores and assessment of the remaining start-up core physics experiments, including reactivity effects, reactivity coefficient, and reaction-rate distribution measurements. Long term benchmarking goals might include analyses of the hot zero-power critical, rise-to-power tests, and other irradiation, safety, and technical evaluations performed with the HTTR.

  14. Concept of using a benchmark part to evaluate rapid prototype processes

    NASA Technical Reports Server (NTRS)

    Cariapa, Vikram

    1994-01-01

    A conceptual benchmark part for guiding manufacturers and users of rapid prototyping technologies is proposed. This is based on a need to have some tool to evaluate the development of this technology and to assist the user in judiciously selecting a process. The benchmark part is designed to have unique product details and features. The extent to which a rapid prototyping process can reproduce these features becomes a measure of the capability of the process. Since rapid prototyping is a dynamic technology, this benchmark part should be used to continuously monitor process capability of existing and developing technologies. Development of this benchmark part is, therefore, based on an understanding of the properties required from prototypes and characteristics of various rapid prototyping processes and measuring equipment that is used for evaluation.

  15. Transient Eddy-Current Nondestructive Evaluation: Benchmark Data for Backface Slots in a Plate

    DTIC Science & Technology

    2011-12-01

    DO and Chimenti DE, (Plenum Press, New York), p. 241-248. [9] Bowler J and Johnson M (1997), "Pulsed eddy-current response to a conducting half...UNCLASSIFIED Transient Eddy-Current Nondestructive Evaluation: Benchmark Data for Backface Slots in a Plate S K Burke and M E...Ibrahim Maritime Platforms Division Defence Science and Technology Organisation DSTO-TN-1047 ABSTRACT The results of a benchmark

  16. Benchmark Evaluation of Uranium Metal Annuli and Cylinders with Beryllium Reflectors

    SciTech Connect

    John D. Bess

    2010-06-01

    An extensive series of delayed critical experiments were performed at the Oak Ridge Critical Experiments Facility using enriched uranium metal during the 1960s and 1970s in support of criticality safety operations at the Y-12 Plant. These experiments were designed to evaluate the storage, casting, and handling limits of the Y-12 Plant and to provide data for the verification of cross sections and calculation methods utilized in nuclear criticality safety applications. Many of these experiments have already been evaluated and included in the International Criticality Safety Benchmark Evaluation Project (ICSBEP) Handbook: unreflected (HEU-MET-FAST-051), graphite-reflected (HEU-MET-FAST-071), and polyethylene-reflected (HEU-MET-FAST-076). Three of the experiments consisted of highly-enriched uranium (HEU, ~93.2% 235U) metal parts reflected by beryllium metal discs. The first evaluated experiment was constructed from a stack of 7-in.-diameter, 4-1/8-in.-high stack of HEU discs top-reflected by a 7-in.-diameter, 5-9/16-in.-high stack of beryllium discs. The other two experiments were formed from stacks of concentric HEU metal annular rings surrounding a 7-in.diameter beryllium core. The nominal outer diameters were 13 and 15 in. with a nominal stack height of 5 and 4 in., respectively. These experiments have been evaluated for inclusion in the ICSBEP Handbook.

  17. TOSPAC calculations in support of the COVE 2A benchmarking activity; Yucca Mountain Site Characterization Project

    SciTech Connect

    Gauthier, J.H.; Zieman, N.B.; Miller, W.B.

    1991-10-01

    The purpose of the the Code Verification (COVE) 2A benchmarking activity is to assess the numerical accuracy of several computer programs for the Yucca Mountain Site Characterization Project of the Department of Energy. This paper presents a brief description of the computer program TOSPAC and a discussion of the calculational effort and results generated by TOSPAC for the COVE 2A problem set. The calculations were performed twice. The initial calculations provided preliminary results for comparison with the results from other COVE 2A participants. TOSPAC was modified in response to the comparison and the final calculations included a correction and several enhancements to improve efficiency. 8 refs.

  18. Towards a benchmark simulation model for plant-wide control strategy performance evaluation of WWTPs.

    PubMed

    Jeppsson, U; Rosen, C; Alex, J; Copp, J; Gernaey, K V; Pons, M N; Vanrolleghem, P A

    2006-01-01

    The COST/IWA benchmark simulation model has been available for seven years. Its primary purpose has been to create a platform for control strategy benchmarking of activated sludge processes. The fact that the benchmark has resulted in more than 100 publications, not only in Europe but also worldwide, demonstrates the interest in such a tool within the research community In this paper, an extension of the benchmark simulation model no 1 (BSM1) is proposed. This extension aims at facilitating control strategy development and performance evaluation at a plant-wide level and, consequently, includes both pre-treatment of wastewater as well as the processes describing sludge treatment. The motivation for the extension is the increasing interest and need to operate and control wastewater treatment systems not only at an individual process level but also on a plant-wide basis. To facilitate the changes, the evaluation period has been extended to one year. A prolonged evaluation period allows for long-term control strategies to be assessed and enables the use of control handles that cannot be evaluated in a realistic fashion in the one-week BSM1 evaluation period. In the paper, the extended plant layout is proposed and the new suggested process models are described briefly. Models for influent file design, the benchmarking procedure and the evaluation criteria are also discussed. And finally, some important remaining topics, for which consensus is required, are identified.

  19. Evaluation of Project Trend.

    ERIC Educational Resources Information Center

    Unco, Inc., Washington, DC.

    This report is a descriptive evaluation of the five pilot sites of Project TREND (Targeting Resources on the Educational Needs of the Disadvantaged). The five Local Education Agency (LEA) pilot sites are the educational systems of: (1) Akron, Ohio; (2) El Paso, Texas; (3) Newark, New Jersey; (4) Portland, Oregon; and, (5) San Jose (Unified),…

  20. Overview of the 2014 Edition of the International Handbook of Evaluated Reactor Physics Benchmark Experiments (IRPhEP Handbook)

    SciTech Connect

    John D. Bess; J. Blair Briggs; Jim Gulliford; Ian Hill

    2014-10-01

    The International Reactor Physics Experiment Evaluation Project (IRPhEP) is a widely recognized world class program. The work of the IRPhEP is documented in the International Handbook of Evaluated Reactor Physics Benchmark Experiments (IRPhEP Handbook). Integral data from the IRPhEP Handbook is used by reactor safety and design, nuclear data, criticality safety, and analytical methods development specialists, worldwide, to perform necessary validations of their calculational techniques. The IRPhEP Handbook is among the most frequently quoted reference in the nuclear industry and is expected to be a valuable resource for future decades.

  1. Two h-Index Benchmarks for Evaluating the Publication Performance of Medical Informatics Researchers

    PubMed Central

    Arbuckle, Luk; Jonker, Elizabeth; Anderson, Kevin

    2012-01-01

    Background The h-index is a commonly used metric for evaluating the publication performance of researchers. However, in a multidisciplinary field such as medical informatics, interpreting the h-index is a challenge because researchers tend to have diverse home disciplines, ranging from clinical areas to computer science, basic science, and the social sciences, each with different publication performance profiles. Objective To construct a reference standard for interpreting the h-index of medical informatics researchers based on the performance of their peers. Methods Using a sample of authors with articles published over the 5-year period 2006–2011 in the 2 top journals in medical informatics (as determined by impact factor), we computed their h-index using the Scopus database. Percentiles were computed to create a 6-level benchmark, similar in scheme to one used by the US National Science Foundation, and a 10-level benchmark. Results The 2 benchmarks can be used to place medical informatics researchers in an ordered category based on the performance of their peers. A validation exercise mapped the benchmark levels to the ranks of medical informatics academic faculty in the United States. The 10-level benchmark tracked academic rank better (with no ties) and is therefore more suitable for practical use. Conclusions Our 10-level benchmark provides an objective basis to evaluate and compare the publication performance of medical informatics researchers with that of their peers using the h-index. PMID:23079075

  2. The ORSphere Benchmark Evaluation and Its Potential Impact on Nuclear Criticality Safety

    SciTech Connect

    John D. Bess; Margaret A. Marshall; J. Blair Briggs

    2013-10-01

    In the early 1970’s, critical experiments using an unreflected metal sphere of highly enriched uranium (HEU) were performed with the focus to provide a “very accurate description…as an ideal benchmark for calculational methods and cross-section data files.” Two near-critical configurations of the Oak Ridge Sphere (ORSphere) were evaluated as acceptable benchmark experiments for inclusion in the International Handbook of Evaluated Criticality Safety Benchmark Experiments (ICSBEP Handbook). The results from those benchmark experiments were then compared with additional unmoderated and unreflected HEU metal benchmark experiment configurations currently found in the ICSBEP Handbook. For basic geometries (spheres, cylinders, and slabs) the eigenvalues calculated using MCNP5 and ENDF/B-VII.0 were within 3 of their respective benchmark values. There appears to be generally a good agreement between calculated and benchmark values for spherical and slab geometry systems. Cylindrical geometry configurations tended to calculate low, including more complex bare HEU metal systems containing cylinders. The ORSphere experiments do not calculate within their 1s uncertainty and there is a possibility that the effect of the measured uncertainties for the GODIVA I benchmark may need reevaluated. There is significant scatter in the calculations for the highly-correlated ORCEF cylinder experiments, which are constructed from close-fitting HEU discs and annuli. Selection of a nuclear data library can have a larger impact on calculated eigenvalue results than the variation found within calculations of a given experimental series, such as the ORCEF cylinders, using a single nuclear data set.

  3. Benchmark Evaluation of the Neutron Radiography (NRAD) Reactor Upgraded LEU-Fueled Core

    SciTech Connect

    John D. Bess

    2001-09-01

    Benchmark models were developed to evaluate the cold-critical start-up measurements performed during the fresh core reload of the Neutron Radiography (NRAD) reactor with Low Enriched Uranium (LEU) fuel. The final upgraded core configuration with 64 fuel elements has been completed. Evaluated benchmark measurement data include criticality, control-rod worth measurements, shutdown margin, and excess reactivity. Dominant uncertainties in keff include the manganese content and impurities contained within the stainless steel cladding of the fuel and the 236U and erbium poison content in the fuel matrix. Calculations with MCNP5 and ENDF/B-VII.0 nuclear data are approximately 1.4% greater than the benchmark model eigenvalue, supporting contemporary research regarding errors in the cross section data necessary to simulate TRIGA-type reactors. Uncertainties in reactivity effects measurements are estimated to be ~10% with calculations in agreement with benchmark experiment values within 2s. The completed benchmark evaluation de-tails are available in the 2014 edition of the International Handbook of Evaluated Reactor Physics Experiments (IRPhEP Handbook). Evaluation of the NRAD LEU cores containing 56, 60, and 62 fuel elements have also been completed, including analysis of their respective reactivity effects measurements; they are also available in the IRPhEP Handbook but will not be included in this summary paper.

  4. A web application for evaluating Phase I methods using a non-parametric optimal benchmark.

    PubMed

    Wages, Nolan A; Varhegyi, Nikole

    2017-10-01

    In evaluating the performance of Phase I dose-finding designs, simulation studies are typically conducted to assess how often a method correctly selects the true maximum tolerated dose under a set of assumed dose-toxicity curves. A necessary component of the evaluation process is to have some concept for how well a design can possibly perform. The notion of an upper bound on the accuracy of maximum tolerated dose selection is often omitted from the simulation study, and the aim of this work is to provide researchers with accessible software to quickly evaluate the operating characteristics of Phase I methods using a benchmark. The non-parametric optimal benchmark is a useful theoretical tool for simulations that can serve as an upper limit for the accuracy of maximum tolerated dose identification based on a binary toxicity endpoint. It offers researchers a sense of the plausibility of a Phase I method's operating characteristics in simulation. We have developed an R shiny web application for simulating the benchmark. The web application has the ability to quickly provide simulation results for the benchmark and requires no programming knowledge. The application is free to access and use on any device with an Internet browser. The application provides the percentage of correct selection of the maximum tolerated dose and an accuracy index, operating characteristics typically used in evaluating the accuracy of dose-finding designs. We hope this software will facilitate the use of the non-parametric optimal benchmark as an evaluation tool in dose-finding simulation.

  5. Benchmark Testing of a New 56Fe Evaluation for Criticality Safety Applications

    SciTech Connect

    Leal, Luiz C; Ivanov, E.

    2015-01-01

    The SAMMY code was used to evaluate resonance parameters of the 56Fe cross section in the resolved resonance energy range of 0–2 MeV using transmission data, capture, elastic, inelastic, and double differential elastic cross sections. The resonance analysis was performed with the code SAMMY that fits R-matrix resonance parameters using the generalized least-squares technique (Bayes’ theory). The evaluation yielded a set of resonance parameters that reproduced the experimental data very well, along with a resonance parameter covariance matrix for data uncertainty calculations. Benchmark tests were conducted to assess the evaluation performance in benchmark calculations.

  6. Reactor Physics and Criticality Benchmark Evaluations for Advanced Nuclear Fuel - Final Technical Report

    SciTech Connect

    William Anderson; James Tulenko; Bradley Rearden; Gary Harms

    2008-09-11

    The nuclear industry interest in advanced fuel and reactor design often drives towards fuel with uranium enrichments greater than 5 wt% 235U. Unfortunately, little data exists, in the form of reactor physics and criticality benchmarks, for uranium enrichments ranging between 5 and 10 wt% 235U. The primary purpose of this project is to provide benchmarks for fuel similar to what may be required for advanced light water reactors (LWRs). These experiments will ultimately provide additional information for application to the criticality-safety bases for commercial fuel facilities handling greater than 5 wt% 235U fuel.

  7. [Benchmarking in health care: conclusions and recommendations].

    PubMed

    Geraedts, Max; Selbmann, Hans-Konrad

    2011-01-01

    The German Health Ministry funded 10 demonstration projects and accompanying research of benchmarking in health care. The accompanying research work aimed to infer generalisable findings and recommendations. We performed a meta-evaluation of the demonstration projects and analysed national and international approaches to benchmarking in health care. It was found that the typical benchmarking sequence is hardly ever realised. Most projects lack a detailed analysis of structures and processes of the best performers as a starting point for the process of learning from and adopting best practice. To tap the full potential of benchmarking in health care, participation in voluntary benchmarking projects should be promoted that have been demonstrated to follow all the typical steps of a benchmarking process. Copyright © 2011. Published by Elsevier GmbH.

  8. Benchmark Evaluation of Dounreay Prototype Fast Reactor Minor Actinide Depletion Measurements

    SciTech Connect

    Hess, J. D.; Gauld, I. C.; Gulliford, J.; Hill, I.; Okajima, S.

    2017-01-01

    Historic measurements of actinide samples in the Dounreay Prototype Fast Reactor (PFR) are of interest for modern nuclear data and simulation validation. Samples of various higher-actinide isotopes were irradiated for 492 effective full-power days and radiochemically assayed at Oak Ridge National Laboratory (ORNL) and Japan Atomic Energy Research Institute (JAERI). Limited data were available regarding the PFR irradiation; a six-group neutron spectra was available with some power history data to support a burnup depletion analysis validation study. Under the guidance of the Organisation for Economic Co-Operation and Development Nuclear Energy Agency (OECD NEA), the International Reactor Physics Experiment Evaluation Project (IRPhEP) and Spent Fuel Isotopic Composition (SFCOMPO) Project are collaborating to recover all measurement data pertaining to these measurements, including collaboration with the United Kingdom to obtain pertinent reactor physics design and operational history data. These activities will produce internationally peer-reviewed benchmark data to support validation of minor actinide cross section data and modern neutronic simulation of fast reactors with accompanying fuel cycle activities such as transportation, recycling, storage, and criticality safety.

  9. Evaluation of the Aleph PIC Code on Benchmark Simulations

    NASA Astrophysics Data System (ADS)

    Boerner, Jeremiah; Pacheco, Jose; Grillet, Anne

    2016-09-01

    Aleph is a massively parallel, 3D unstructured mesh, Particle-in-Cell (PIC) code, developed to model low temperature plasma applications. In order to verify and validate performance, Aleph is benchmarked against a series of canonical problems to demonstrate statistical indistinguishability in the results. Here, a series of four problems is studied: Couette flows over a range of Knudsen number, sheath formation in an undriven plasma, the two-stream instability, and a capacitive discharge. These problems respectively exercise collisional processes, particle motion in electrostatic fields, electrostatic field solves coupled to particle motion, and a fully coupled reacting plasma. Favorable comparison with accepted results establishes confidence in Aleph's capability and accuracy as a general purpose PIC code. Finally, Aleph is used to investigate the sensitivity of a triggered vacuum gap switch to the particle injection conditions associated with arc breakdown at the trigger. Sandia National Laboratories is a multi-program laboratory managed and operated by Sandia Corporation, a wholly owned subsidiary of Lockheed Martin Corporation, for the U.S. Department of Energy's National Nuclear Security Administration under contract DE-AC04-94AL85000.

  10. Benchmarking and Performance Measurement.

    ERIC Educational Resources Information Center

    Town, J. Stephen

    This paper defines benchmarking and its relationship to quality management, describes a project which applied the technique in a library context, and explores the relationship between performance measurement and benchmarking. Numerous benchmarking methods contain similar elements: deciding what to benchmark; identifying partners; gathering…

  11. Benchmarking the Remote-Handled Waste Facility at the West Valley Demonstration Project

    SciTech Connect

    O. P. Mendiratta; D. K. Ploetz

    2000-02-29

    ABSTRACT Facility decontamination activities at the West Valley Demonstration Project (WVDP), the site of a former commercial nuclear spent fuel reprocessing facility near Buffalo, New York, have resulted in the removal of radioactive waste. Due to high dose and/or high contamination levels of this waste, it needs to be handled remotely for processing and repackaging into transport/disposal-ready containers. An initial conceptual design for a Remote-Handled Waste Facility (RHWF), completed in June 1998, was estimated to cost $55 million and take 11 years to process the waste. Benchmarking the RHWF with other facilities around the world, completed in November 1998, identified unique facility design features and innovative waste pro-cessing methods. Incorporation of the benchmarking effort has led to a smaller yet fully functional, $31 million facility. To distinguish it from the June 1998 version, the revised design is called the Rescoped Remote-Handled Waste Facility (RRHWF) in this topical report. The conceptual design for the RRHWF was completed in June 1999. A design-build contract was approved by the Department of Energy in September 1999.

  12. Extension of the IWA/COST simulation benchmark to include expert reasoning for system performance evaluation.

    PubMed

    Comas, J; Rodríguez-Roda, I; Poch, M; Gernaey, K V; Rosen, C; Jeppsson, U

    2006-01-01

    In this paper the development of an extension module to the IWA/COST simulation benchmark to include expert reasoning is presented. This module enables the detection of suitable conditions for the development of settling problems of biological origin (filamentous bulking, foaming and rising sludge) when applying activated sludge control strategies to the simulation benchmark. Firstly, a flow diagram is proposed for each settling problem, and secondly, the outcome of its application is shown. Results of the benchmark for two evaluated control strategies illustrate that, once applied to the simulation outputs, this module provides supplementary criteria for plant performance assessment. Therefore, simulated control strategies can be evaluated in a more realistic framework, and results can be recognised as more realistic and satisfactory from the point of view of operators and real facilities.

  13. How Can the eCampus Be Organized and Run To Address Traditional Concerns, but Maintain an Innovative Approach to Providing Educational Access? Project Eagle Evaluation Question #3. Benchmarking St. Petersburg College: A Report to Leadership.

    ERIC Educational Resources Information Center

    Burkhart, Joyce

    This paper discusses the findings of St. Petersburg College's (SPC) (Florida) evaluation question: "How can the eCampus be organized and run to address traditional faculty concerns, but maintain an innovative approach to providing educational access?" In order to evaluate this question, a list was compiled of faculty issues identified by…

  14. Benchmark Evaluation of Start-Up and Zero-Power Measurements at the High-Temperature Engineering Test Reactor

    DOE PAGES

    Bess, John D.; Fujimoto, Nozomu

    2014-10-09

    Benchmark models were developed to evaluate six cold-critical and two warm-critical, zero-power measurements of the HTTR. Additional measurements of a fully-loaded subcritical configuration, core excess reactivity, shutdown margins, six isothermal temperature coefficients, and axial reaction-rate distributions were also evaluated as acceptable benchmark experiments. Insufficient information is publicly available to develop finely-detailed models of the HTTR as much of the design information is still proprietary. However, the uncertainties in the benchmark models are judged to be of sufficient magnitude to encompass any biases and bias uncertainties incurred through the simplification process used to develop the benchmark models. Dominant uncertainties in themore » experimental keff for all core configurations come from uncertainties in the impurity content of the various graphite blocks that comprise the HTTR. Monte Carlo calculations of keff are between approximately 0.9 % and 2.7 % greater than the benchmark values. Reevaluation of the HTTR models as additional information becomes available could improve the quality of this benchmark and possibly reduce the computational biases. High-quality characterization of graphite impurities would significantly improve the quality of the HTTR benchmark assessment. Simulation of the other reactor physics measurements are in good agreement with the benchmark experiment values. The complete benchmark evaluation details are available in the 2014 edition of the International Handbook of Evaluated Reactor Physics Benchmark Experiments.« less

  15. Benchmark Evaluation of Start-Up and Zero-Power Measurements at the High-Temperature Engineering Test Reactor

    SciTech Connect

    Bess, John D.; Fujimoto, Nozomu

    2014-10-09

    Benchmark models were developed to evaluate six cold-critical and two warm-critical, zero-power measurements of the HTTR. Additional measurements of a fully-loaded subcritical configuration, core excess reactivity, shutdown margins, six isothermal temperature coefficients, and axial reaction-rate distributions were also evaluated as acceptable benchmark experiments. Insufficient information is publicly available to develop finely-detailed models of the HTTR as much of the design information is still proprietary. However, the uncertainties in the benchmark models are judged to be of sufficient magnitude to encompass any biases and bias uncertainties incurred through the simplification process used to develop the benchmark models. Dominant uncertainties in the experimental keff for all core configurations come from uncertainties in the impurity content of the various graphite blocks that comprise the HTTR. Monte Carlo calculations of keff are between approximately 0.9 % and 2.7 % greater than the benchmark values. Reevaluation of the HTTR models as additional information becomes available could improve the quality of this benchmark and possibly reduce the computational biases. High-quality characterization of graphite impurities would significantly improve the quality of the HTTR benchmark assessment. Simulation of the other reactor physics measurements are in good agreement with the benchmark experiment values. The complete benchmark evaluation details are available in the 2014 edition of the International Handbook of Evaluated Reactor Physics Benchmark Experiments.

  16. How Can St. Petersburg College Leverage Technology To Increase Access to Courses and Programs for an Expanded Pool of Learners? Project Eagle Evaluation Question #4. Benchmarking St. Petersburg College: A Report to Leadership.

    ERIC Educational Resources Information Center

    Burkhart, Joyce

    This report discusses St. Petersburg College's (SPC) (Florida) evaluation question, "How can St. Petersburg College leverage technology to increase access to courses and programs for an expanded pool of learners?" The report summarizes both nationwide/worldwide best practices and current SPC efforts related to four strategies: (1) an…

  17. What Are the Appropriate Models for St. Petersburg College and the University Partnership Center To Expand Access to Bachelor's and Master's Degrees? Project Eagle Evaluation Question #5. Benchmarking St. Petersburg College: A Report to Leadership.

    ERIC Educational Resources Information Center

    Burkhart, Joyce

    St. Petersburg College (SPC) (Florida), formerly a two-year community college, now offers four-year degrees. This paper discusses the findings of SPC's evaluation question focusing on what the appropriate models are for St. Petersburg College and the University Partnership Center (UPC) to increase access to bachelor's and master's programs.…

  18. Benchmarks for evaluation and comparison of udder health status using monthly individual somatic cell count.

    PubMed

    Fauteux, Véronique; Roy, Jean-Philippe; Scholl, Daniel T; Bouchard, Émile

    2014-08-01

    The objectives of this study were to propose benchmarks for the interpretation of herd udder health using monthly individual somatic cell counts (SCC) from dairy herds in Quebec, Canada and to evaluate the association of risk factors with intramammary infection (IMI) dynamics relative to these benchmarks. The mean and percentiles of indices related to udder infection status [e.g., proportion of healthy or chronically infected cows, cows cured and new IMI (NIMI) rate] during lactation and over the dry period were calculated using a threshold of ≥ 200 000 cells/mL at test day. Mean NIMI proportion and proportion of cows cured during lactation were 0.11 and 0.27. Benchmarks of 0.70 and 0.03 for healthy and chronically infected cows over the dry period were proposed. Season and herd mean SCC were risk factors influencing IMI dynamics during lactation and over the dry period.

  19. Benchmark Evaluation of the HTR-PROTEUS Absorber Rod Worths (Core 4)

    SciTech Connect

    John D. Bess; Leland M. Montierth

    2014-06-01

    PROTEUS was a zero-power research reactor at the Paul Scherrer Institute (PSI) in Switzerland. The critical assembly was constructed from a large graphite annulus surrounding a central cylindrical cavity. Various experimental programs were investigated in PROTEUS; during the years 1992 through 1996, it was configured as a pebble-bed reactor and designated HTR-PROTEUS. Various critical configurations were assembled with each accompanied by an assortment of reactor physics experiments including differential and integral absorber rod measurements, kinetics, reaction rate distributions, water ingress effects, and small sample reactivity effects [1]. Four benchmark reports were previously prepared and included in the March 2013 edition of the International Handbook of Evaluated Reactor Physics Benchmark Experiments (IRPhEP Handbook) [2] evaluating eleven critical configurations. A summary of that effort was previously provided [3] and an analysis of absorber rod worth measurements for Cores 9 and 10 have been performed prior to this analysis and included in PROTEUS-GCR-EXP-004 [4]. In the current benchmark effort, absorber rod worths measured for Core Configuration 4, which was the only core with a randomly-packed pebble loading, have been evaluated for inclusion as a revision to the HTR-PROTEUS benchmark report PROTEUS-GCR-EXP-002.

  20. Evaluation of mobile phone camera benchmarking using objective camera speed and image quality metrics

    NASA Astrophysics Data System (ADS)

    Peltoketo, Veli-Tapani

    2014-11-01

    When a mobile phone camera is tested and benchmarked, the significance of image quality metrics is widely acknowledged. There are also existing methods to evaluate the camera speed. However, the speed or rapidity metrics of the mobile phone's camera system has not been used with the quality metrics even if the camera speed has become a more and more important camera performance feature. There are several tasks in this work. First, the most important image quality and speed-related metrics of a mobile phone's camera system are collected from the standards and papers and, also, novel speed metrics are identified. Second, combinations of the quality and speed metrics are validated using mobile phones on the market. The measurements are done toward application programming interface of different operating systems. Finally, the results are evaluated and conclusions are made. The paper defines a solution to combine different image quality and speed metrics to a single benchmarking score. A proposal of the combined benchmarking metric is evaluated using measurements of 25 mobile phone cameras on the market. The paper is a continuation of a previous benchmarking work expanded with visual noise measurement and updates of the latest mobile phone versions.

  1. Cloud-Based Evaluation of Anatomical Structure Segmentation and Landmark Detection Algorithms: VISCERAL Anatomy Benchmarks.

    PubMed

    Jimenez-Del-Toro, Oscar; Muller, Henning; Krenn, Markus; Gruenberg, Katharina; Taha, Abdel Aziz; Winterstein, Marianne; Eggel, Ivan; Foncubierta-Rodriguez, Antonio; Goksel, Orcun; Jakab, Andras; Kontokotsios, Georgios; Langs, Georg; Menze, Bjoern H; Salas Fernandez, Tomas; Schaer, Roger; Walleyo, Anna; Weber, Marc-Andre; Dicente Cid, Yashin; Gass, Tobias; Heinrich, Mattias; Jia, Fucang; Kahl, Fredrik; Kechichian, Razmig; Mai, Dominic; Spanier, Assaf B; Vincent, Graham; Wang, Chunliang; Wyeth, Daniel; Hanbury, Allan

    2016-11-01

    Variations in the shape and appearance of anatomical structures in medical images are often relevant radiological signs of disease. Automatic tools can help automate parts of this manual process. A cloud-based evaluation framework is presented in this paper including results of benchmarking current state-of-the-art medical imaging algorithms for anatomical structure segmentation and landmark detection: the VISCERAL Anatomy benchmarks. The algorithms are implemented in virtual machines in the cloud where participants can only access the training data and can be run privately by the benchmark administrators to objectively compare their performance in an unseen common test set. Overall, 120 computed tomography and magnetic resonance patient volumes were manually annotated to create a standard Gold Corpus containing a total of 1295 structures and 1760 landmarks. Ten participants contributed with automatic algorithms for the organ segmentation task, and three for the landmark localization task. Different algorithms obtained the best scores in the four available imaging modalities and for subsets of anatomical structures. The annotation framework, resulting data set, evaluation setup, results and performance analysis from the three VISCERAL Anatomy benchmarks are presented in this article. Both the VISCERAL data set and Silver Corpus generated with the fusion of the participant algorithms on a larger set of non-manually-annotated medical images are available to the research community.

  2. Epitope prediction based on random peptide library screening: benchmark dataset and prediction tools evaluation.

    PubMed

    Sun, Pingping; Chen, Wenhan; Huang, Yanxin; Wang, Hongyan; Ma, Zhiqiang; Lv, Yinghua

    2011-06-16

    Epitope prediction based on random peptide library screening has become a focus as a promising method in immunoinformatics research. Some novel software and web-based servers have been proposed in recent years and have succeeded in given test cases. However, since the number of available mimotopes with the relevant structure of template-target complex is limited, a systematic evaluation of these methods is still absent. In this study, a new benchmark dataset was defined. Using this benchmark dataset and a representative dataset, five examples of the most popular epitope prediction software products which are based on random peptide library screening have been evaluated. Using the benchmark dataset, in no method did performance exceed a 0.42 precision and 0.37 sensitivity, and the MCC scores suggest that the epitope prediction results of these software programs are greater than random prediction about 0.09-0.13; while using the representative dataset, most of the values of these performance measures are slightly improved, but the overall performance is still not satisfactory. Many test cases in the benchmark dataset cannot be applied to these pieces of software due to software limitations. Moreover chances are that these software products are overfitted to the small dataset and will fail in other cases. Therefore finding the correlation between mimotopes and genuine epitope residues is still far from resolved and much larger dataset for mimotope-based epitope prediction is desirable.

  3. A benchmarking tool to evaluate computer tomography perfusion infarct core predictions against a DWI standard.

    PubMed

    Cereda, Carlo W; Christensen, Søren; Campbell, Bruce Cv; Mishra, Nishant K; Mlynash, Michael; Levi, Christopher; Straka, Matus; Wintermark, Max; Bammer, Roland; Albers, Gregory W; Parsons, Mark W; Lansberg, Maarten G

    2016-10-01

    Differences in research methodology have hampered the optimization of Computer Tomography Perfusion (CTP) for identification of the ischemic core. We aim to optimize CTP core identification using a novel benchmarking tool. The benchmarking tool consists of an imaging library and a statistical analysis algorithm to evaluate the performance of CTP. The tool was used to optimize and evaluate an in-house developed CTP-software algorithm. Imaging data of 103 acute stroke patients were included in the benchmarking tool. Median time from stroke onset to CT was 185 min (IQR 180-238), and the median time between completion of CT and start of MRI was 36 min (IQR 25-79). Volumetric accuracy of the CTP-ROIs was optimal at an rCBF threshold of <38%; at this threshold, the mean difference was 0.3 ml (SD 19.8 ml), the mean absolute difference was 14.3 (SD 13.7) ml, and CTP was 67% sensitive and 87% specific for identification of DWI positive tissue voxels. The benchmarking tool can play an important role in optimizing CTP software as it provides investigators with a novel method to directly compare the performance of alternative CTP software packages. © The Author(s) 2015.

  4. Evaluation methods for hospital projects.

    PubMed

    Buelow, Janet R; Zuckweiler, Kathryn M; Rosacker, Kirsten M

    2010-01-01

    The authors report the findings of a survey of hospital managers on the utilization of various project selection and evaluation methodologies. The focus of the analysis was the empirical relationship between a portfolio of project evaluation(1) methods actually utilized for a given project and several measures of perceived project success. The analysis revealed that cost-benefit analysis and top management support were the two project evaluation methods used most often by the hospital managers. The authors' empirical assessment provides evidence that top management support is associated with overall project success.

  5. RESULTS FOR THE INTERMEDIATE-SPECTRUM ZEUS BENCHMARK OBTAINED WITH NEW 63,65Cu CROSS-SECTION EVALUATIONS

    SciTech Connect

    Sobes, Vladimir; Leal, Luiz C

    2014-01-01

    The four HEU, intermediate-spectrum, copper-reflected Zeus experiments have shown discrepant results between measurement and calculation for the last several major releases of the ENDF library. The four benchmarks show a trend in reported C/E values with increasing energy of average lethargy causing fission. Recently, ORNL has made improvements to the evaluations of three key isotopes involved in the benchmark cases in question. Namely, an updated evaluation for 235U and evaluations of 63,65Cu. This paper presents the benchmarking results of the four intermediate-spectrum Zeus cases using the three updated evaluations.

  6. Benchmarking and Its Relevance to the Library and Information Sector. Interim Findings of "Best Practice Benchmarking in the Library and Information Sector," a British Library Research and Development Department Project.

    ERIC Educational Resources Information Center

    Kinnell, Margaret; Garrod, Penny

    This British Library Research and Development Department study assesses current activities and attitudes toward quality management in library and information services (LIS) in the academic sector as well as the commercial/industrial sector. Definitions and types of benchmarking are described, and the relevance of benchmarking to LIS is evaluated.…

  7. Evaluation of HEU-Beryllium Benchmark Experiments to Improve Computational Analysis of Space Reactors

    SciTech Connect

    John D. Bess; Keith C. Bledsoe; Bradley T. Rearden

    2011-02-01

    An assessment was previously performed to evaluate modeling capabilities and quantify preliminary biases and uncertainties associated with the modeling methods and data utilized in designing a nuclear reactor such as a beryllium-reflected, highly-enriched-uranium (HEU)-O2 fission surface power (FSP) system for space nuclear power. The conclusion of the previous study was that current capabilities could preclude the necessity of a cold critical test of the FSP; however, additional testing would reduce uncertainties in the beryllium and uranium cross-section data and the overall uncertainty in the computational models. A series of critical experiments using HEU metal were performed in the 1960s and 1970s in support of criticality safety operations at the Y-12 Plant. Of the hundreds of experiments, three were identified as fast-fission configurations reflected by beryllium metal. These experiments have been evaluated as benchmarks for inclusion in the International Handbook of Evaluated Criticality Safety Benchmark Experiments (IHECSBE). Further evaluation of the benchmark experiments was performed using the sensitivity and uncertainty analysis capabilities of SCALE 6. The data adjustment methods of SCALE 6 have been employed in the validation of an example FSP design model to reduce the uncertainty due to the beryllium cross section data.

  8. Evaluation of HEU-Beryllium Benchmark Experiments to Improve Computational Analysis of Space Reactors

    SciTech Connect

    Bess, John; Bledsoe, Keith C; Rearden, Bradley T

    2011-01-01

    An assessment was previously performed to evaluate modeling capabilities and quantify preliminary biases and uncertainties associated with the modeling methods and data utilized in designing a nuclear reactor such as a beryllium-reflected, highly-enriched-uranium (HEU)-O2 fission surface power (FSP) system for space nuclear power. The conclusion of the previous study was that current capabilities could preclude the necessity of a cold critical test of the FSP; however, additional testing would reduce uncertainties in the beryllium and uranium cross-section data and the overall uncertainty in the computational models. A series of critical experiments using HEU metal were performed in the 1960s and 1970s in support of criticality safety operations at the Y-12 Plant. Of the hundreds of experiments, three were identified as fast-fission configurations reflected by beryllium metal. These experiments have been evaluated as benchmarks for inclusion in the International Handbook of Evaluated Criticality Safety Benchmark Experiments (IHECSBE). Further evaluation of the benchmark experiments was performed using the sensitivity and uncertainty analysis capabilities of SCALE 6. The data adjustment methods of SCALE 6 have been employed in the validation of an example FSP design model to reduce the uncertainty due to the beryllium cross section data.

  9. Development of Conceptual Benchmark Models to Evaluate Complex Hydrologic Model Calibration in Managed Basins Using Python

    NASA Astrophysics Data System (ADS)

    Hughes, J. D.; White, J.

    2013-12-01

    For many numerical hydrologic models it is a challenge to quantitatively demonstrate that complex models are preferable to simpler models. Typically, a decision is made to develop and calibrate a complex model at the beginning of a study. The value of selecting a complex model over simpler models is commonly inferred from use of a model with fewer simplifications of the governing equations because it can be time consuming to develop another numerical code with data processing and parameter estimation functionality. High-level programming languages like Python can greatly reduce the effort required to develop and calibrate simple models that can be used to quantitatively demonstrate the increased value of a complex model. We have developed and calibrated a spatially-distributed surface-water/groundwater flow model for managed basins in southeast Florida, USA, to (1) evaluate the effect of municipal groundwater pumpage on surface-water/groundwater exchange, (2) investigate how the study area will respond to sea-level rise, and (3) explore combinations of these forcing functions. To demonstrate the increased value of this complex model, we developed a two-parameter conceptual-benchmark-discharge model for each basin in the study area. The conceptual-benchmark-discharge model includes seasonal scaling and lag parameters and is driven by basin rainfall. The conceptual-benchmark-discharge models were developed in the Python programming language and used weekly rainfall data. Calibration was implemented with the Broyden-Fletcher-Goldfarb-Shanno method available in the Scientific Python (SciPy) library. Normalized benchmark efficiencies calculated using output from the complex model and the corresponding conceptual-benchmark-discharge model indicate that the complex model has more explanatory power than the simple model driven only by rainfall.

  10. Windows NT Workstation Performance Evaluation Based on Pro/E 2000i BENCHMARK

    SciTech Connect

    DAVIS,SEAN M.

    2000-08-02

    A performance evaluation of several computers was necessary, so an evaluation program, or benchmark, was run on each computer to determine maximum possible performance. The program was used to test the Computer Aided Drafting (CAD) ability of each computer by monitoring the speed with which several functions were executed. The main objective of the benchmarking program was to record assembly loading times and image regeneration times and then compile a composite score that could be compared with the same tests on other computers. The three computers that were tested were the Compaq AP550, the SGI 230, and the Hewlett-PackardP750C. The Compaq and SGI computers each had a Pentium III 733mhz processor, while the Hewlett-Packard had a Pentium III 750mhz processor. The size and speed of Random Access Memory (RAM) in each computer varied, as did the type of graphics card. Each computer that was tested was using Windows NT 4.0 and Pro/ENGINEER{trademark} 2000i CAD benchmark software provided by Standard Performance Evaluation Corporation (SPEC). The benchmarking program came with its own assembly, automatically loaded and ran tests on the assembly, then compiled the time each test took to complete. Due to the automation of the tests, any sort of user error affecting test scores was virtually eliminated. After all the tests were completed, scores were then compiled and compared. The Silicon Graphics 230 was by far the overall winner with a composite score of 8.57. The Compaq AP550 was next with a score of 5.19, while the Hewlett-Packard P750C performed dismally, achieving a score of 3.34. Several factors, including motherboard chipset, graphics card, and the size and speed of RAM, were involved in the differing scores of the three machines. Surprisingly the Hewlett-Packard, which had the fastest processor, came back with the lowest score. The above factors most likely contributed to the poor performance of the Hewlett-Packard. Based on the results of the benchmark test

  11. Evaluating Productivity Predictions Under Elevated CO2 Conditions: Multi-Model Benchmarking Across FACE Experiments

    NASA Astrophysics Data System (ADS)

    Cowdery, E.; Dietze, M.

    2016-12-01

    As atmospheric levels of carbon dioxide levels continue to increase, it is critical that terrestrial ecosystem models can accurately predict ecological responses to the changing environment. Current predictions of net primary productivity (NPP) in response to elevated atmospheric CO2 concentration are highly variable and contain a considerable amount of uncertainty.The Predictive Ecosystem Analyzer (PEcAn) is an informatics toolbox that wraps around an ecosystem model and can be used to help identify which factors drive uncertainty. We tested a suite of models (LPJ-GUESS, MAESPA, GDAY, CLM5, DALEC, ED2), which represent a range from low to high structural complexity, across a range of Free-Air CO2 Enrichment (FACE) experiments: the Kennedy Space Center Open Top Chamber Experiment, the Rhinelander FACE experiment, the Duke Forest FACE experiment and the Oak Ridge Experiment on CO2 Enrichment. These tests were implemented in a novel benchmarking workflow that is automated, repeatable, and generalized to incorporate different sites and ecological models. Observational data from the FACE experiments represent a first test of this flexible, extensible approach aimed at providing repeatable tests of model process representation.To identify and evaluate the assumptions causing inter-model differences we used PEcAn to perform model sensitivity and uncertainty analysis, not only to assess the components of NPP, but also to examine system processes such nutrient uptake and and water use. Combining the observed patterns of uncertainty between multiple models with results of the recent FACE-model data synthesis project (FACE-MDS) can help identify which processes need further study and additional data constraints. These findings can be used to inform future experimental design and in turn can provide informative starting point for data assimilation.

  12. Benchmark Evaluation of the Medium-Power Reactor Experiment Program Critical Configurations

    SciTech Connect

    Margaret A. Marshall; John D. Bess

    2013-02-01

    A series of small, compact critical assembly (SCCA) experiments were performed in 1962-1965 at the Oak Ridge National Laboratory Critical Experiments Facility (ORCEF) for the Medium-Power Reactor Experiment (MPRE) program. The MPRE was a stainless-steel clad, highly enriched uranium (HEU)-O2 fuelled, BeO reflected reactor design to provide electrical power to space vehicles. Cooling and heat transfer were to be achieved by boiling potassium in the reactor core and passing vapor directly through a turbine. Graphite- and beryllium-reflected assemblies were constructed at ORCEF to verify the critical mass, power distribution, and other reactor physics measurements needed to validate reactor calculations and reactor physics methods. The experimental series was broken into three parts, with the third portion of the experiments representing the beryllium-reflected measurements. The latter experiments are of interest for validating current reactor design efforts for a fission surface power reactor. The entire series has been evaluated as acceptable benchmark experiments and submitted for publication in the International Handbook of Evaluated Criticality Safety Benchmark Experiments and in the International Handbook of Evaluated Reactor Physics Benchmark Experiments.

  13. External Evaluation of Project ACT.

    ERIC Educational Resources Information Center

    Kreitlow, Burton W.; Kreitlow, Doris J.

    Third party evaluation of Project ACT (Adult Competency Training), a United States Office of Education staff development project in Region 8, has examined the achievement of the project's goal of developing a self-generating and self-supporting adult staff development system, thus helping adult educators to become better prepared to serve their…

  14. Gifted Science Project: Evaluation Report.

    ERIC Educational Resources Information Center

    Ott, Susan L.; Emanuel, Elizabeth, Ed.

    The document contains the evaluation report on the Gifted Science Project in Montgomery County, Maryland, a program to identify resources for students in grades 3-8 who are motivated in science. The Project's primary product is a Project Resource File (PRF) listing people, places, and published materials that can be used by individual students. An…

  15. Outpatient echocardiography in the evaluation of innocent murmurs in children: utilisation benchmarking.

    PubMed

    Frias, Patricio A; Oster, Matthew; Daley, Patricia A; Boris, Jeffrey R

    2016-03-01

    We sought to benchmark the utilisation of echocardiography in the outpatient evaluation of heart murmurs by evaluating two large paediatric cardiology centres. Although criteria exist for appropriate use of echocardiography, there are no benchmarking data demonstrating its utilisation. We performed a retrospective cohort study of outpatients aged between 0 and 18 years at the Sibley Heart Center Cardiology and the Children's Hospital of Philadelphia Division of Cardiology, given a sole diagnosis of "innocent murmur" from 1 July, 2007 to 31 October, 2010. Using internal claims data, we compared the utilisation of echocardiography according to centre, patient age, and physician years of service. Of 23,114 eligible patients (Sibley Heart Center Cardiology: 12,815, Children's Hospital of Philadelphia Division of Cardiology: 10,299), 43.1% (Sibley Heart Center Cardiology: 45.2%, Children's Hospital of Philadelphia Division of Cardiology: 40.4%; p1-5 years had the lowest utilisation (32.7%). In two large paediatric cardiology practices, the overall utilisation of echocardiography by physicians with a sole diagnosis of innocent murmur was similar. There was significant and similar variability in utilisation by provider at both centres. Although these data serve as initial benchmarking, the variability in utilisation highlights the importance of appropriate use criteria.

  16. Automated Generation of Message-Passing Programs: An Evaluation of CAPTools using NAS Benchmarks

    NASA Technical Reports Server (NTRS)

    Hribar, Michelle R.; Jin, Hao-Qiang; Yan, Jerry C.; Bailey, David (Technical Monitor)

    1998-01-01

    Scientists at NASA Ames Research Center have been developing computational aeroscience applications on highly parallel architectures over the past ten years. During the same time period, a steady transition of hardware and system software also occurred, forcing us to expand great efforts into migrating and receding our applications. As applications and machine architectures continue to become increasingly complex, the cost and time required for this process will become prohibitive. Various attempts to exploit software tools to assist and automate the parallelization process have not produced favorable results. In this paper, we evaluate an interactive parallelization tool, CAPTools, for parallelizing serial versions of the NAB Parallel Benchmarks. Finally, we compare the performance of the resulting CAPTools generated code to the hand-coded benchmarks on the Origin 2000 and IBM SP2. Based on these results, a discussion on the feasibility of automated parallelization of aerospace applications is presented along with suggestions for future work.

  17. Team Projects and Peer Evaluations

    ERIC Educational Resources Information Center

    Doyle, John Kevin; Meeker, Ralph D.

    2008-01-01

    The authors assign semester- or quarter-long team-based projects in several Computer Science and Finance courses. This paper reports on our experience in designing, managing, and evaluating such projects. In particular, we discuss the effects of team size and of various peer evaluation schemes on team performance and student learning. We report…

  18. Team Projects and Peer Evaluations

    ERIC Educational Resources Information Center

    Doyle, John Kevin; Meeker, Ralph D.

    2008-01-01

    The authors assign semester- or quarter-long team-based projects in several Computer Science and Finance courses. This paper reports on our experience in designing, managing, and evaluating such projects. In particular, we discuss the effects of team size and of various peer evaluation schemes on team performance and student learning. We report…

  19. Project Change Evaluation Research Brief.

    ERIC Educational Resources Information Center

    Leiderman, Sally A.; Dupree, David M.

    Project Change is a community-driven anti-racism initiative operating in four communities: Albuquerque, New Mexico; El Paso, Texas; Knoxville, Tennessee; and Valdosta, Georgia. The formative evaluation of Project Change began in 1994 when all of the sites were still in planning or early action phases. Findings from the summative evaluation will be…

  20. Project financial evaluation

    SciTech Connect

    None, None

    2009-01-18

    The project financial section of the Renewable Energy Technology Characterizations describes structures and models to support the technical and economic status of emerging renewable energy options for electricity supply.

  1. Implications of the Trauma Quality Improvement Project inclusion of nonsurvivable injuries in performance benchmarking.

    PubMed

    Heaney, Jiselle Bock; Schroll, Rebecca; Turney, Jennifer; Stuke, Lance; Marr, Alan B; Greiffenstein, Patrick; Robledo, Rosemarie; Theriot, Amanda; Duchesne, Juan; Hunt, John

    2017-10-01

    The Trauma Quality Improvement Project (TQIP) uses an injury prediction model for performance benchmarking. We hypothesize that at a Level I high-volume penetrating trauma center, performance outcomes will be biased due to inclusion of patients with nonsurvivable injuries. Retrospective chart review was conducted for all patients included in the institutional TQIP analysis from 2013 to 2014 with length of stay (LOS) less than 1 day to determine survivability of the injuries. Observed (O)/expected (E) mortality ratios were calculated before and after exclusion of these patients. Completeness of data reported to TQIP was examined. Eight hundred twenty-six patients were reported to TQIP including 119 deaths. Nonsurvivable injuries accounted 90.9% of the deaths in patients with an LOS of 1 day or less. The O/E mortality ratio for all patients was 1.061, and the O/E ratio after excluding all patients with LOS less than 1 day found to have nonsurvivable injuries was 0.895. Data for key variables were missing in 63.3% of patients who died in the emergency department, 50% of those taken to the operating room and 0% of those admitted to the intensive care unit. Charts for patients who died with LOS less than 1 day were significantly more likely than those who lived to be missing crucial. This study shows TQIP inclusion of patients with nonsurvivable injuries biases outcomes at an urban trauma center. Missing data results in imputation of values, increasing inaccuracy. Further investigation is needed to determine if these findings exist at other institutions, and whether the current TQIP model needs revision to accurately identify and exclude patients with nonsurvivable injuries. Prognostic and epidemiological, level III.

  2. MPI performance evaluation and characterization using a compact application benchmark code

    SciTech Connect

    Worley, P.H.

    1996-06-01

    In this paper the parallel benchmark code PSTSWM is used to evaluate the performance of the vendor-supplied implementations of the MPI message-passing standard on the Intel Paragon, IBM SP2, and Cray Research T3D. This study is meant to complement the performance evaluation of individual MPI commands by providing information on the practical significance of MPI performance on the execution of a communication-intensive application code. In particular, three performance questions are addressed: how important is the communication protocol in determining performance when using MPI, how does MPI performance compare with that of the native communication library, and how efficient are the collective communication routines.

  3. Performance evaluation of tile-based Fisher Ratio analysis using a benchmark yeast metabolome dataset.

    PubMed

    Watson, Nathanial E; Parsons, Brendon A; Synovec, Robert E

    2016-08-12

    Performance of tile-based Fisher Ratio (F-ratio) data analysis, recently developed for discovery-based studies using comprehensive two-dimensional gas chromatography coupled with time-of-flight mass spectrometry (GC×GC-TOFMS), is evaluated with a metabolomics dataset that had been previously analyzed in great detail, but while taking a brute force approach. The previously analyzed data (referred to herein as the benchmark dataset) were intracellular extracts from Saccharomyces cerevisiae (yeast), either metabolizing glucose (repressed) or ethanol (derepressed), which define the two classes in the discovery-based analysis to find metabolites that are statistically different in concentration between the two classes. Beneficially, this previously analyzed dataset provides a concrete means to validate the tile-based F-ratio software. Herein, we demonstrate and validate the significant benefits of applying tile-based F-ratio analysis. The yeast metabolomics data are analyzed more rapidly in about one week versus one year for the prior studies with this dataset. Furthermore, a null distribution analysis is implemented to statistically determine an adequate F-ratio threshold, whereby the variables with F-ratio values below the threshold can be ignored as not class distinguishing, which provides the analyst with confidence when analyzing the hit table. Forty-six of the fifty-four benchmarked changing metabolites were discovered by the new methodology while consistently excluding all but one of the benchmarked nineteen false positive metabolites previously identified.

  4. Project OUTREACH Evaluation.

    ERIC Educational Resources Information Center

    Hollis, Patricia A.; Newton, Josephine K.

    Described is a 4-week summer workshop, Project OUTREACH, designed to train Head Start personnel in the knowledge and skills necessary to identify handicapped or potentially handicapped children and to develop specific teaching strategies for the preschool handicapped child. It is explained that a unique aspect of the workshop was the coordination…

  5. The DLESE Evaluation Toolkit Project

    NASA Astrophysics Data System (ADS)

    Buhr, S. M.; Barker, L. J.; Marlino, M.

    2002-12-01

    The Evaluation Toolkit and Community project is a new Digital Library for Earth System Education (DLESE) collection designed to raise awareness of project evaluation within the geoscience education community, and to enable principal investigators, teachers, and evaluators to implement project evaluation more readily. This new resource is grounded in the needs of geoscience educators, and will provide a virtual home for a geoscience education evaluation community. The goals of the project are to 1) provide a robust collection of evaluation resources useful for Earth systems educators, 2) establish a forum and community for evaluation dialogue within DLESE, and 3) disseminate the resources through the DLESE infrastructure and through professional society workshops and proceedings. Collaboration and expertise in education, geoscience and evaluation are necessary if we are to conduct the best possible geoscience education. The Toolkit allows users to engage in evaluation at whichever level best suits their needs, get more evaluation professional development if desired, and access the expertise of other segments of the community. To date, a test web site has been built and populated, initial community feedback from the DLESE and broader community is being garnered, and we have begun to heighten awareness of geoscience education evaluation within our community. The web site contains features that allow users to access professional development about evaluation, search and find evaluation resources, submit resources, find or offer evaluation services, sign up for upcoming workshops, take the user survey, and submit calendar items. The evaluation resource matrix currently contains resources that have met our initial review. The resources are currently organized by type; they will become searchable on multiple dimensions of project type, audience, objectives and evaluation resource type as efforts to develop a collection-specific search engine mature. The peer review

  6. Project Proposals Evaluation

    NASA Astrophysics Data System (ADS)

    Encheva, Sylvia; Tumin, Sharil

    2009-08-01

    Collaboration among various firms has been traditionally used trough single project joint ventures for bonding purposes. Eventhough the performed work is usually beneficial to some extend to all participants, the type of collaboration option to be adapted is strongly influenced by overall purposes and goals that can be achieved. In order to facilitate a choice of collaboration option best suited to a firm's need a computer based model is proposed.

  7. Surfactant EOR project evaluated

    SciTech Connect

    Holm, L.W.

    1984-07-16

    The Union Oil Co.'s Uniflood process has successfully mobilized and produced tertiary oil from a micellar-polymer pilot project on the Hegberg lease in the El Dorado field, Kansas. This half-completed EOR flood has recovered over 11% of the waterflood residual oil and is currently producing at an oil cut of 10%. Oil recovery has been limited by (1) the presence of gypsum in portions of the reservoir which adversly affects injected chemicals, (2) poor quality reservoir rock in one quadrant of the pilot, and (3) a substantial fluid drift (30 ft/year) which causes a portion of the injected chemicals to flow out of the pilot pattern. The El Dorado demonstration project is a joint experiment covered by a cost-sharing contract between the U.S. Department of Energy and Cities Service Company. It was proposed as a micellar-polymer process in a highly saline (10 wt % salts) reservoir that had been waterflooded to residual oil. Despite the extended project life, and indications that total recovery efficiency will be less than originally predicted, oil response in the Hegberg pattern is encouraging for application of the micellar-polymer process in high brine reservoirs.

  8. Evaluation of DFT-D3 dispersion corrections for various structural benchmark sets

    NASA Astrophysics Data System (ADS)

    Schröder, Heiner; Hühnert, Jens; Schwabe, Tobias

    2017-01-01

    We present an evaluation of our newly developed density functional theory (DFT)-D3 dispersion correction D3(CSO) in comparison to its predecessor D3(BJ) for geometry optimizations. Therefore, various benchmark sets covering bond lengths, rotational constants, and center of mass distances of supramolecular complexes have been chosen. Overall both corrections give accurate structures and show no systematic differences. Additionally, we present an optimized algorithm for the computation of the DFT-D3 gradient, which reduces the formal scaling of the gradient calculation from O (N3) to O (N2) .

  9. The NAS parallel benchmarks

    NASA Technical Reports Server (NTRS)

    Bailey, David (Editor); Barton, John (Editor); Lasinski, Thomas (Editor); Simon, Horst (Editor)

    1993-01-01

    A new set of benchmarks was developed for the performance evaluation of highly parallel supercomputers. These benchmarks consist of a set of kernels, the 'Parallel Kernels,' and a simulated application benchmark. Together they mimic the computation and data movement characteristics of large scale computational fluid dynamics (CFD) applications. The principal distinguishing feature of these benchmarks is their 'pencil and paper' specification - all details of these benchmarks are specified only algorithmically. In this way many of the difficulties associated with conventional benchmarking approaches on highly parallel systems are avoided.

  10. GEAR UP Aspirations Project Evaluation

    ERIC Educational Resources Information Center

    Trimble, Brad A.

    2013-01-01

    The purpose of this study was to conduct a formative evaluation of the first two years of the Gaining Early Awareness and Readiness for Undergraduate Programs (GEAR UP) Aspirations Project (Aspirations) using a Context, Input, Process, and Product (CIPP) model so as to gain an in-depth understanding of the project during the middle school…

  11. GEAR UP Aspirations Project Evaluation

    ERIC Educational Resources Information Center

    Trimble, Brad A.

    2013-01-01

    The purpose of this study was to conduct a formative evaluation of the first two years of the Gaining Early Awareness and Readiness for Undergraduate Programs (GEAR UP) Aspirations Project (Aspirations) using a Context, Input, Process, and Product (CIPP) model so as to gain an in-depth understanding of the project during the middle school…

  12. Incorporating specificity into optimization: evaluation of SPA using CSAR 2014 and CASF 2013 benchmarks

    NASA Astrophysics Data System (ADS)

    Yan, Zhiqiang; Wang, Jin

    2016-03-01

    Scoring functions of protein-ligand interactions are widely used in computationally docking software and structure-based drug discovery. Accurate prediction of the binding energy between the protein and the ligand is the main task of the scoring function. The accuracy of a scoring function is normally evaluated by testing it on the benchmarks of protein-ligand complexes. In this work, we report the evaluation analysis of an improved version of scoring function SPecificity and Affinity (SPA). By testing on two independent benchmarks Community Structure-Activity Resource (CSAR) 2014 and Comparative Assessment of Scoring Functions (CASF) 2013, the assessment shows that SPA is relatively more accurate than other compared scoring functions in predicting the interactions between the protein and the ligand. We conclude that the inclusion of the specificity in the optimization can effectively suppress the competitive state on the funnel-like binding energy landscape, and make SPA more accurate in identifying the "native" conformation and scoring the binding decoys. The evaluation of SPA highlights the importance of binding specificity in improving the accuracy of the scoring functions.

  13. Incorporating specificity into optimization: evaluation of SPA using CSAR 2014 and CASF 2013 benchmarks.

    PubMed

    Yan, Zhiqiang; Wang, Jin

    2016-03-01

    Scoring functions of protein-ligand interactions are widely used in computationally docking software and structure-based drug discovery. Accurate prediction of the binding energy between the protein and the ligand is the main task of the scoring function. The accuracy of a scoring function is normally evaluated by testing it on the benchmarks of protein-ligand complexes. In this work, we report the evaluation analysis of an improved version of scoring function SPecificity and Affinity (SPA). By testing on two independent benchmarks Community Structure-Activity Resource (CSAR) 2014 and Comparative Assessment of Scoring Functions (CASF) 2013, the assessment shows that SPA is relatively more accurate than other compared scoring functions in predicting the interactions between the protein and the ligand. We conclude that the inclusion of the specificity in the optimization can effectively suppress the competitive state on the funnel-like binding energy landscape, and make SPA more accurate in identifying the "native" conformation and scoring the binding decoys. The evaluation of SPA highlights the importance of binding specificity in improving the accuracy of the scoring functions.

  14. A Quantitative Methodology for Determining the Critical Benchmarks for Project 2061 Strand Maps

    ERIC Educational Resources Information Center

    Kuhn, G.

    2008-01-01

    The American Association for the Advancement of Science (AAAS) was tasked with identifying the key science concepts for science literacy in K-12 students in America (AAAS, 1990, 1993). The AAAS Atlas of Science Literacy (2001) has organized roughly half of these science concepts or benchmarks into fifty flow charts. Each flow chart or strand map…

  15. A Quantitative Methodology for Determining the Critical Benchmarks for Project 2061 Strand Maps

    ERIC Educational Resources Information Center

    Kuhn, G.

    2008-01-01

    The American Association for the Advancement of Science (AAAS) was tasked with identifying the key science concepts for science literacy in K-12 students in America (AAAS, 1990, 1993). The AAAS Atlas of Science Literacy (2001) has organized roughly half of these science concepts or benchmarks into fifty flow charts. Each flow chart or strand map…

  16. Grass Roots Project Evaluation.

    ERIC Educational Resources Information Center

    Wick, John W.

    Some aspects of a grass roots evaluation training program are presented. The program consists of two elements: (1) a series of 11 slide/tape individualized self-paced units, and (2) a six-week summer program. Three points of view on this program are: (1) University graduate programs in quantitative areas are usually consumed by specialists; (2)…

  17. Conceptual Soundness, Metric Development, Benchmarking, and Targeting for PATH Subprogram Evaluation

    SciTech Connect

    Mosey. G.; Doris, E.; Coggeshall, C.; Antes, M.; Ruch, J.; Mortensen, J.

    2009-01-01

    The objective of this study is to evaluate the conceptual soundness of the U.S. Department of Housing and Urban Development (HUD) Partnership for Advancing Technology in Housing (PATH) program's revised goals and establish and apply a framework to identify and recommend metrics that are the most useful for measuring PATH's progress. This report provides an evaluative review of PATH's revised goals, outlines a structured method for identifying and selecting metrics, proposes metrics and benchmarks for a sampling of individual PATH programs, and discusses other metrics that potentially could be developed that may add value to the evaluation process. The framework and individual program metrics can be used for ongoing management improvement efforts and to inform broader program-level metrics for government reporting requirements.

  18. NASA PC software evaluation project

    NASA Technical Reports Server (NTRS)

    Dominick, Wayne D. (Editor); Kuan, Julie C.

    1986-01-01

    The USL NASA PC software evaluation project is intended to provide a structured framework for facilitating the development of quality NASA PC software products. The project will assist NASA PC development staff to understand the characteristics and functions of NASA PC software products. Based on the results of the project teams' evaluations and recommendations, users can judge the reliability, usability, acceptability, maintainability and customizability of all the PC software products. The objective here is to provide initial, high-level specifications and guidelines for NASA PC software evaluation. The primary tasks to be addressed in this project are as follows: to gain a strong understanding of what software evaluation entails and how to organize a structured software evaluation process; to define a structured methodology for conducting the software evaluation process; to develop a set of PC software evaluation criteria and evaluation rating scales; and to conduct PC software evaluations in accordance with the identified methodology. Communication Packages, Network System Software, Graphics Support Software, Environment Management Software, General Utilities. This report represents one of the 72 attachment reports to the University of Southwestern Louisiana's Final Report on NASA Grant NGT-19-010-900. Accordingly, appropriate care should be taken in using this report out of context of the full Final Report.

  19. [Chest pain evaluation project].

    PubMed

    Filippo, Ottani; Nicola, Binetti; Casagranda, Ivo; Cassin, Matteo; Cavazza, Mario; Grifoni, Stefano; Lenzi, Tiziano; Lorenzoni, Roberto; Sbrojavacca, Rodolfo; Tanzi, Pietro; Vergara, Giuseppe

    2009-01-01

    The evaluation of acute chest pain remains challenging, despite many insights and innovations over the past two decades. The percentage of patients presenting at the emergency department with acute chest pain who are subsequently admitted to the hospital appears to be increasing. Patients with acute coronary syndromes who are inadvertently discharged from the emergency department have an adverse short-term prognosis. However, the admission of a patient with chest pain who is at low risk for acute coronary syndrome can lead to unnecessary tests and procedures, with their burden of costs and complications. Therefore, with increasing economic pressures on health care, physicians and administrators are interested in improving the efficiency of care for patients with acute chest pain. Since the emergency department organization (i.e. the availability of an intensive observational area) and integration of care and treatment between emergency physicians and cardiologists greatly differ over the national territory, the purpose of the present position paper is two-fold: first, to review the evidence-based efficacy and utility of various diagnostic tools, and, second, to delineate the basic critical pathways (describing key steps for care and treatment) that need to be implemented in order to standardize and expedite the evaluation of chest pain patients, making their diagnosis and treatment as uniform as possible across the country.

  20. GROWTH OF THE INTERNATIONAL CRITICALITY SAFETY AND REACTOR PHYSICS EXPERIMENT EVALUATION PROJECTS

    SciTech Connect

    J. Blair Briggs; John D. Bess; Jim Gulliford

    2011-09-01

    Since the International Conference on Nuclear Criticality Safety (ICNC) 2007, the International Criticality Safety Benchmark Evaluation Project (ICSBEP) and the International Reactor Physics Experiment Evaluation Project (IRPhEP) have continued to expand their efforts and broaden their scope. Eighteen countries participated on the ICSBEP in 2007. Now, there are 20, with recent contributions from Sweden and Argentina. The IRPhEP has also expanded from eight contributing countries in 2007 to 16 in 2011. Since ICNC 2007, the contents of the 'International Handbook of Evaluated Criticality Safety Benchmark Experiments1' have increased from 442 evaluations (38000 pages), containing benchmark specifications for 3955 critical or subcritical configurations to 516 evaluations (nearly 55000 pages), containing benchmark specifications for 4405 critical or subcritical configurations in the 2010 Edition of the ICSBEP Handbook. The contents of the Handbook have also increased from 21 to 24 criticality-alarm-placement/shielding configurations with multiple dose points for each, and from 20 to 200 configurations categorized as fundamental physics measurements relevant to criticality safety applications. Approximately 25 new evaluations and 150 additional configurations are expected to be added to the 2011 edition of the Handbook. Since ICNC 2007, the contents of the 'International Handbook of Evaluated Reactor Physics Benchmark Experiments2' have increased from 16 different experimental series that were performed at 12 different reactor facilities to 53 experimental series that were performed at 30 different reactor facilities in the 2011 edition of the Handbook. Considerable effort has also been made to improve the functionality of the searchable database, DICE (Database for the International Criticality Benchmark Evaluation Project) and verify the accuracy of the data contained therein. DICE will be discussed in separate papers at ICNC 2011. The status of the ICSBEP and the IRPh

  1. Benchmarking specialty hospitals, a scoping review on theory and practice.

    PubMed

    Wind, A; van Harten, W H

    2017-04-04

    Although benchmarking may improve hospital processes, research on this subject is limited. The aim of this study was to provide an overview of publications on benchmarking in specialty hospitals and a description of study characteristics. We searched PubMed and EMBASE for articles published in English in the last 10 years. Eligible articles described a project stating benchmarking as its objective and involving a specialty hospital or specific patient category; or those dealing with the methodology or evaluation of benchmarking. Of 1,817 articles identified in total, 24 were included in the study. Articles were categorized into: pathway benchmarking, institutional benchmarking, articles on benchmark methodology or -evaluation and benchmarking using a patient registry. There was a large degree of variability:(1) study designs were mostly descriptive and retrospective; (2) not all studies generated and showed data in sufficient detail; and (3) there was variety in whether a benchmarking model was just described or if quality improvement as a consequence of the benchmark was reported upon. Most of the studies that described a benchmark model described the use of benchmarking partners from the same industry category, sometimes from all over the world. Benchmarking seems to be more developed in eye hospitals, emergency departments and oncology specialty hospitals. Some studies showed promising improvement effects. However, the majority of the articles lacked a structured design, and did not report on benchmark outcomes. In order to evaluate the effectiveness of benchmarking to improve quality in specialty hospitals, robust and structured designs are needed including a follow up to check whether the benchmark study has led to improvements.

  2. Benchmark Calculations for Reflector Effect in Fast Cores by Using the Latest Evaluated Nuclear Data Libraries

    NASA Astrophysics Data System (ADS)

    Fukushima, M.; Ishikawa, M.; Numata, K.; Jin, T.; Kugo, T.

    2014-04-01

    Benchmark calculations for reflector effects in fast cores were performed to validate the reliability of scattering data of structural materials in the major evaluated nuclear data libraries, JENDL-4.0, ENDF/B-VII.1 and JEFF-3.1.2. The criticalities of two FCA and two ZPR cores were analyzed by using a continuous energy Monte Carlo calculation code. The ratios of calculation to experimental values were compared between these cores and the sensitivity analyses were performed. From the results, the replacement reactivity from blanket to SS and Na reflector is better evaluated by JENDL-4.0 than by ENDF/B-VII.1 mainly due to the μbar values of Na and 52Cr.

  3. Use of Benchmark Methodology in Environmental Impact Assessment

    NASA Astrophysics Data System (ADS)

    Pubule, Jelena; Blumberga, Dagnija

    2010-01-01

    Every industrial activity and procedure influences the environment and climate change. This impact has to be assessed and therefore the procedure of Environmental Impact Assessment (EIA) including the application of a benchmark methodology has been developed. The developed benchmark methodology can be used in the initial assessment as a screening method. The article surveys the developed benchmark methodology for impact assessment of the projects providing extraction of dolomite in the deposits of mineral resources. The benchmark methodology developed makes it possible to assess the impacts caused by the provided activity by objective considerations, to conduct comparison of different projects and evaluate whether the provided activity corresponds to the principles of sustainable development.

  4. Model benchmarking and reference signals for angled-beam shear wave ultrasonic nondestructive evaluation (NDE) inspections

    NASA Astrophysics Data System (ADS)

    Aldrin, John C.; Hopkins, Deborah; Datuin, Marvin; Warchol, Mark; Warchol, Lyudmila; Forsyth, David S.; Buynak, Charlie; Lindgren, Eric A.

    2017-02-01

    For model benchmark studies, the accuracy of the model is typically evaluated based on the change in response relative to a selected reference signal. The use of a side drilled hole (SDH) in a plate was investigated as a reference signal for angled beam shear wave inspection for aircraft structure inspections of fastener sites. Systematic studies were performed with varying SDH depth and size, and varying the ultrasonic probe frequency, focal depth, and probe height. Increased error was observed with the simulation of angled shear wave beams in the near-field. Even more significant, asymmetry in real probes and the inherent sensitivity of signals in the near-field to subtle test conditions were found to provide a greater challenge with achieving model agreement. To achieve quality model benchmark results for this problem, it is critical to carefully align the probe with the part geometry, to verify symmetry in probe response, and ideally avoid using reference signals from the near-field response. Suggested reference signals for angled beam shear wave inspections include using the `through hole' corner specular reflection signal and the full skip' signal off of the far wall from the side drilled hole.

  5. The impact of incomplete knowledge on evaluation: an experimental benchmark for protein function prediction

    PubMed Central

    Huttenhower, Curtis; Hibbs, Matthew A.; Myers, Chad L.; Caudy, Amy A.; Hess, David C.; Troyanskaya, Olga G.

    2009-01-01

    Motivation: Rapidly expanding repositories of highly informative genomic data have generated increasing interest in methods for protein function prediction and inference of biological networks. The successful application of supervised machine learning to these tasks requires a gold standard for protein function: a trusted set of correct examples, which can be used to assess performance through cross-validation or other statistical approaches. Since gene annotation is incomplete for even the best studied model organisms, the biological reliability of such evaluations may be called into question. Results: We address this concern by constructing and analyzing an experimentally based gold standard through comprehensive validation of protein function predictions for mitochondrion biogenesis in Saccharomyces cerevisiae. Specifically, we determine that (i) current machine learning approaches are able to generalize and predict novel biology from an incomplete gold standard and (ii) incomplete functional annotations adversely affect the evaluation of machine learning performance. While computational approaches performed better than predicted in the face of incomplete data, relative comparison of competing approaches—even those employing the same training data—is problematic with a sparse gold standard. Incomplete knowledge causes individual methods' performances to be differentially underestimated, resulting in misleading performance evaluations. We provide a benchmark gold standard for yeast mitochondria to complement current databases and an analysis of our experimental results in the hopes of mitigating these effects in future comparative evaluations. Availability: The mitochondrial benchmark gold standard, as well as experimental results and additional data, is available at http://function.princeton.edu/mitochondria Contact: ogt@cs.princeton.edu Supplementary information: Supplementary data are available at Bioinformatics online. PMID:19561015

  6. A real-world evidence-based approach to laboratory reorganization using e-Valuate benchmarking data.

    PubMed

    Bonini, Fabio; Barletta, Giovanni; Plebani, Mario

    2017-03-01

    Pressure to cut health-care costs has involved clinical laboratories underpinning the need to reduce cost per test through programs designed to consolidate activities and increase volumes. Currently, however, there is little evidence of the effectiveness of these measures. The aim of the present study was to verify whether a rational, evidence-based decision-making process might be achieved based on an activity-based cost analysis performed by collecting the data of all variables affecting cost per test. An activity-based costing analysis was performed using a program that provides collected data on performance indicators, benchmark between different laboratories based on performance indicators, and information on reorganization initiatives. The data provided were used in two different settings to (1) verify the results of the internal re-organization of specific protein assay and (2) simulate some scenarios for the reorganization of autoimmune testing in the network of clinical laboratories in a large territory. The data produced by the e-Valuate project enabled the quantification of variation in costs, the utilization of human and technological resources and efficiency, both as final result of a reorganization project (proteins) and as a simulation of a possible future organization (autoimmune tests).

  7. SU-E-J-30: Benchmark Image-Based TCP Calculation for Evaluation of PTV Margins for Lung SBRT Patients

    SciTech Connect

    Li, M; Chetty, I; Zhong, H

    2014-06-01

    Purpose: Tumor control probability (TCP) calculated with accumulated radiation doses may help design appropriate treatment margins. Image registration errors, however, may compromise the calculated TCP. The purpose of this study is to develop benchmark CT images to quantify registration-induced errors in the accumulated doses and their corresponding TCP. Methods: 4DCT images were registered from end-inhale (EI) to end-exhale (EE) using a “demons” algorithm. The demons DVFs were corrected by an FEM model to get realistic deformation fields. The FEM DVFs were used to warp the EI images to create the FEM-simulated images. The two images combined with the FEM DVF formed a benchmark model. Maximum intensity projection (MIP) images, created from the EI and simulated images, were used to develop IMRT plans. Two plans with 3 and 5 mm margins were developed for each patient. With these plans, radiation doses were recalculated on the simulated images and warped back to the EI images using the FEM DVFs to get the accumulated doses. The Elastix software was used to register the FEM-simulated images to the EI images. TCPs calculated with the Elastix-accumulated doses were compared with those generated by the FEM to get the TCP error of the Elastix registrations. Results: For six lung patients, the mean Elastix registration error ranged from 0.93 to 1.98 mm. Their relative dose errors in PTV were between 0.28% and 6.8% for 3mm margin plans, and between 0.29% and 6.3% for 5mm-margin plans. As the PTV margin reduced from 5 to 3 mm, the mean TCP error of the Elastix-reconstructed doses increased from 2.0% to 2.9%, and the mean NTCP errors decreased from 1.2% to 1.1%. Conclusion: Patient-specific benchmark images can be used to evaluate the impact of registration errors on the computed TCPs, and may help select appropriate PTV margins for lung SBRT patients.

  8. Evaluation of CRISTO II Storage Arrays Benchmark with TRIPOLI-4.2 Criticality Calculations

    NASA Astrophysics Data System (ADS)

    Lee, Y. K.

    The new lattice feature of TRIPOLI-4.2 geometry package was applied to model the CRISTO II storage arrays of PWR fuels with various kinds of neutron absorber plates. The new `Kcoll' collision estimator of TRIPOLI-4.2 code was utilized to evaluate the infinite multiplication factors, Kinf. Comparing with the published ICSBEP benchmark results of CRISTO II experiments and of three different continuousenergy Monte Carlo codes - TRIPOLI-4.1 (JEF2.2), MCNP4B2 (ENDF/B-V) and MCNP4XS (ENDF/B-VI.r4), the present study using cost-effective modeling, JEF2.2 and ENDF/B-VI.r4 libraries obtained satisfactory results.

  9. Benchmark Data for Evaluation of Aeroacoustic Propagation Codes With Grazing Flow

    NASA Technical Reports Server (NTRS)

    Jones, Michael G.; Watson, Willie R.; Parrott, Tony L.

    2005-01-01

    Increased understanding of the effects of acoustic treatment on the propagation of sound through commercial aircraft engine nacelles is a requirement for more efficient liner design. To this end, one of NASA s goals is to further the development of duct propagation and impedance reduction codes. A number of these codes have been developed over the last three decades. These codes are typically divided into two categories: (1) codes that use the measured complex acoustic pressure field to reduce the acoustic impedance of treatment that is positioned along the wall of the duct, and (2) codes that use the acoustic impedance of the treatment as input and compute the sound field throughout the duct. Clearly, the value of these codes is dependent upon the quality of the data used for their validation. Over the past two decades, data acquired in the NASA Langley Research Center Grazing Incidence Tube have been used by a number of researchers for comparison with their propagation codes. Many of these comparisons have been based upon Grazing Incidence Tube tests that were conducted to study specific liner technology components, and were incomplete for general propagation code validation. Thus, the objective of the current investigation is to provide a quality data set that can be used as a benchmark for evaluation of duct propagation and impedance reduction codes. In order to achieve this objective, two parallel efforts have been undertaken. The first of these is the development of an enhanced impedance eduction code that uses data acquired in the Grazing Incidence Tube. This enhancement is intended to place the benchmark data on as firm a foundation as possible. The second key effort is the acquisition of a comprehensive set of data selected to allow propagation code evaluations over a range of test conditions.

  10. Managing for Results in America's Great City Schools 2014: Results from Fiscal Year 2012-13. A Report of the Performance Measurement and Benchmarking Project

    ERIC Educational Resources Information Center

    Council of the Great City Schools, 2014

    2014-01-01

    In 2002 the "Council of the Great City Schools" and its members set out to develop performance measures that could be used to improve business operations in urban public school districts. The Council launched the "Performance Measurement and Benchmarking Project" to achieve these objectives. The purposes of the project was to:…

  11. Block Transfer Agreement Evaluation Project

    ERIC Educational Resources Information Center

    Bastedo, Helena

    2010-01-01

    The objective of this project is to evaluate for the British Columbia Council on Admissions and Transfer (BCCAT) the effectiveness of block transfer agreements (BTAs) in the BC Transfer System and recommend steps to be taken to improve their effectiveness. Findings of this study revealed that institutions want to expand block credit transfer;…

  12. A benchmarking program to reduce red blood cell outdating: implementation, evaluation, and a conceptual framework.

    PubMed

    Barty, Rebecca L; Gagliardi, Kathleen; Owens, Wendy; Lauzon, Deborah; Scheuermann, Sheena; Liu, Yang; Wang, Grace; Pai, Menaka; Heddle, Nancy M

    2015-07-01

    Benchmarking is a quality improvement tool that compares an organization's performance to that of its peers for selected indicators, to improve practice. Processes to develop evidence-based benchmarks for red blood cell (RBC) outdating in Ontario hospitals, based on RBC hospital disposition data from Canadian Blood Services, have been previously reported. These benchmarks were implemented in 160 hospitals provincewide with a multifaceted approach, which included hospital education, inventory management tools and resources, summaries of best practice recommendations, recognition of high-performing sites, and audit tools on the Transfusion Ontario website (http://transfusionontario.org). In this study we describe the implementation process and the impact of the benchmarking program on RBC outdating. A conceptual framework for continuous quality improvement of a benchmarking program was also developed. The RBC outdating rate for all hospitals trended downward continuously from April 2006 to February 2012, irrespective of hospitals' transfusion rates or their distance from the blood supplier. The highest annual outdating rate was 2.82%, at the beginning of the observation period. Each year brought further reductions, with a nadir outdating rate of 1.02% achieved in 2011. The key elements of the successful benchmarking strategy included dynamic targets, a comprehensive and evidence-based implementation strategy, ongoing information sharing, and a robust data system to track information. The Ontario benchmarking program for RBC outdating resulted in continuous and sustained quality improvement. Our conceptual iterative framework for benchmarking provides a guide for institutions implementing a benchmarking program. © 2015 AABB.

  13. ISLES 2015 - A public evaluation benchmark for ischemic stroke lesion segmentation from multispectral MRI.

    PubMed

    Maier, Oskar; Menze, Bjoern H; von der Gablentz, Janina; Häni, Levin; Heinrich, Mattias P; Liebrand, Matthias; Winzeck, Stefan; Basit, Abdul; Bentley, Paul; Chen, Liang; Christiaens, Daan; Dutil, Francis; Egger, Karl; Feng, Chaolu; Glocker, Ben; Götz, Michael; Haeck, Tom; Halme, Hanna-Leena; Havaei, Mohammad; Iftekharuddin, Khan M; Jodoin, Pierre-Marc; Kamnitsas, Konstantinos; Kellner, Elias; Korvenoja, Antti; Larochelle, Hugo; Ledig, Christian; Lee, Jia-Hong; Maes, Frederik; Mahmood, Qaiser; Maier-Hein, Klaus H; McKinley, Richard; Muschelli, John; Pal, Chris; Pei, Linmin; Rangarajan, Janaki Raman; Reza, Syed M S; Robben, David; Rueckert, Daniel; Salli, Eero; Suetens, Paul; Wang, Ching-Wei; Wilms, Matthias; Kirschke, Jan S; Krämer, Ulrike M; Münte, Thomas F; Schramm, Peter; Wiest, Roland; Handels, Heinz; Reyes, Mauricio

    2017-01-01

    Ischemic stroke is the most common cerebrovascular disease, and its diagnosis, treatment, and study relies on non-invasive imaging. Algorithms for stroke lesion segmentation from magnetic resonance imaging (MRI) volumes are intensely researched, but the reported results are largely incomparable due to different datasets and evaluation schemes. We approached this urgent problem of comparability with the Ischemic Stroke Lesion Segmentation (ISLES) challenge organized in conjunction with the MICCAI 2015 conference. In this paper we propose a common evaluation framework, describe the publicly available datasets, and present the results of the two sub-challenges: Sub-Acute Stroke Lesion Segmentation (SISS) and Stroke Perfusion Estimation (SPES). A total of 16 research groups participated with a wide range of state-of-the-art automatic segmentation algorithms. A thorough analysis of the obtained data enables a critical evaluation of the current state-of-the-art, recommendations for further developments, and the identification of remaining challenges. The segmentation of acute perfusion lesions addressed in SPES was found to be feasible. However, algorithms applied to sub-acute lesion segmentation in SISS still lack accuracy. Overall, no algorithmic characteristic of any method was found to perform superior to the others. Instead, the characteristics of stroke lesion appearances, their evolution, and the observed challenges should be studied in detail. The annotated ISLES image datasets continue to be publicly available through an online evaluation system to serve as an ongoing benchmarking resource (www.isles-challenge.org).

  14. Evaluating Soil Health Using Remotely Sensed Evapotranspiration on the Benchmark Barnes Soils of North Dakota

    NASA Astrophysics Data System (ADS)

    Bohn, Meyer; Hopkins, David; Steele, Dean; Tuscherer, Sheldon

    2017-04-01

    The benchmark Barnes soil series is an extensive upland Hapludoll of the northern Great Plains that is both economically and ecologically vital to the region. Effects of tillage erosion coupled with wind and water erosion have degraded Barnes soil quality, but with unknown extent, distribution, or severity. Evidence of soil degradation documented for a half century warrants that the assumption of productivity be tested. Soil resilience is linked to several dynamic soil properties and National Cooperative Soil Survey initiatives are now focused on identifying those properties for benchmark soils. Quantification of soil degradation is dependent on a reliable method for broad-scale evaluation. The soil survey community is currently developing rapid and widespread soil property assessment technologies. Improvements in satellite based remote-sensing and image analysis software have stimulated the application of broad-scale resource assessment. Furthermore, these technologies have fostered refinement of land-based surface energy balance algorithms, i.e. Mapping Evapotranspiration at High Resolution with Internalized Calibration (METRIC) algorithm for evapotranspiration (ET) mapping. The hypothesis of this study is that ET mapping technology can differentiate soil function on extensive landscapes and identify degraded areas. A recent soil change study in eastern North Dakota resampled legacy Barnes pedons sampled prior to 1960 and found significant decreases in organic carbon. An ancillary study showed that evapotranspiration (ET) estimates from METRIC decreased with Barnes erosion class severity. An ET raster map has been developed for three eastern North Dakota counties using METRIC and Landsat 5 imagery. ET pixel candidates on major Barnes soil map units were stratified into tertiles and classified as ranked ET subdivisions. A sampling population of randomly selected points stratified by ET class and county proportion was established. Morphologic and chemical data will

  15. Canadian Language Benchmarks-TOEFL Research Project: A Comparison Study of the Canadian Language Benchmarks Assessment and the Test of English as a Foreign Language.

    ERIC Educational Resources Information Center

    Watt, David L. E.; Lake, Deidre M.

    This study is an examination of the test results of 90 academically oriented adult participants on the Test of English as a Foreign Language (TOEFL) and the Canadian Language Benchmarks Assessment (CBLA), to determine the comparability of performance on the two tests and the possibility of using two tests in the academic admissions process for…

  16. Integral Benchmark Data for Nuclear Data Testing Through the ICSBEP & IRPhEP

    SciTech Connect

    J. Blair Briggs; John D. Bess; Jim Gulliford; Ian Hill

    2013-10-01

    The status of the International Criticality Safety Benchmark Evaluation Project (ICSBEP) and International Reactor Physics Experiment Evaluation Project (IRPhEP) was last discussed directly with the nuclear data community at ND2007. Since ND2007, integral benchmark data that are available for nuclear data testing have increased significantly. The status of the ICSBEP and the IRPhEP is discussed and selected benchmark configurations that have been added to the ICSBEP and IRPhEP Handbooks since ND2007 are highlighted.

  17. Integral Benchmark Data for Nuclear Data Testing Through the ICSBEP & IRPhEP

    NASA Astrophysics Data System (ADS)

    Briggs, J. B.; Bess, J. D.; Gulliford, J.

    2014-04-01

    The status of the International Criticality Safety Benchmark Evaluation Project (ICSBEP) and International Reactor Physics Experiment Evaluation Project (IRPhEP) was last discussed directly with the nuclear data community at ND2007. Since ND2007, integral benchmark data that are available for nuclear data testing have increased significantly. The status of the ICSBEP and the IRPhEP is discussed and selected benchmark configurations that have been added to the ICSBEP and IRPhEP Handbooks since ND2007 are highlighted.

  18. Yucca Mountain Project thermal and mechanical codes first benchmark exercise: Part 3, Jointed rock mass analysis; Yucca Mountain Site Characterization Project

    SciTech Connect

    Costin, L.S.; Bauer, S.J.

    1991-10-01

    Thermal and mechanical models for intact and jointed rock mass behavior are being developed, verified, and validated at Sandia National Laboratories for the Yucca Mountain Site Characterization Project. Benchmarking is an essential part of this effort and is one of the tools used to demonstrate verification of engineering software used to solve thermomechanical problems. This report presents the results of the third (and final) phase of the first thermomechanical benchmark exercise. In the first phase of this exercise, nonlinear heat conduction code were used to solve the thermal portion of the benchmark problem. The results from the thermal analysis were then used as input to the second and third phases of the exercise, which consisted of solving the structural portion of the benchmark problem. In the second phase of the exercise, a linear elastic rock mass model was used. In the third phase of the exercise, two different nonlinear jointed rock mass models were used to solve the thermostructural problem. Both models, the Sandia compliant joint model and the RE/SPEC joint empirical model, explicitly incorporate the effect of the joints on the response of the continuum. Three different structural codes, JAC, SANCHO, and SPECTROM-31, were used with the above models in the third phase of the study. Each model was implemented in two different codes so that direct comparisons of results from each model could be made. The results submitted by the participants showed that the finite element solutions using each model were in reasonable agreement. Some consistent differences between the solutions using the two different models were noted but are not considered important to verification of the codes. 9 refs., 18 figs., 8 tabs.

  19. Analysis of a benchmark suite to evaluate mixed numeric and symbolic processing

    NASA Technical Reports Server (NTRS)

    Ragharan, Bharathi; Galant, David

    1992-01-01

    The suite of programs that formed the benchmark for a proposed advanced computer is described and analyzed. The features of the processor and its operating system that are tested by the benchmark are discussed. The computer codes and the supporting data for the analysis are given as appendices.

  20. BENCHMARK DOSES FOR CHEMICAL MIXTURES: EVALUATION OF A MIXTURE OF 18 PHAHS.

    EPA Science Inventory

    Benchmark doses (BMDs), defined as doses of a substance that are expected to result in a pre-specified level of "benchmark" response (BMR), have been used for quantifying the risk associated with exposure to environmental hazards. The lower confidence limit of the BMD is used as...

  1. BENCHMARK DOSES FOR CHEMICAL MIXTURES: EVALUATION OF A MIXTURE OF 18 PHAHS.

    EPA Science Inventory

    Benchmark doses (BMDs), defined as doses of a substance that are expected to result in a pre-specified level of "benchmark" response (BMR), have been used for quantifying the risk associated with exposure to environmental hazards. The lower confidence limit of the BMD is used as...

  2. Evaluation for 4S core nuclear design method through integration of benchmark data

    SciTech Connect

    Nagata, A.; Tsuboi, Y.; Moriki, Y.; Kawashima, M.

    2012-07-01

    The 4S is a sodium-cooled small fast reactor which is reflector-controlled for operation through core lifetime about 30 years. The nuclear design method has been selected to treat neutron leakage with high accuracy. It consists of a continuous-energy Monte Carlo code, discrete ordinate transport codes and JENDL-3.3. These two types of neutronic analysis codes are used for the design in a complementary manner. The accuracy of the codes has been evaluated by analysis of benchmark critical experiments and the experimental reactor data. The measured data used for the evaluation is critical experimental data of the FCA XXIII as a physics mockup assembly of the 4S core, FCA XVI, FCA XIX, ZPR, and data of experimental reactor JOYO MK-1. Evaluated characteristics are criticality, reflector reactivity worth, power distribution, absorber reactivity worth, and sodium void worth. A multi-component bias method was applied, especially to improve the accuracy of sodium void reactivity worth. As the result, it has been confirmed that the 4S core nuclear design method provides good accuracy, and typical bias factors and their uncertainties are determined. (authors)

  3. Using false discovery rates to benchmark SNP-callers in next-generation sequencing projects.

    PubMed

    Farrer, Rhys A; Henk, Daniel A; MacLean, Dan; Studholme, David J; Fisher, Matthew C

    2013-01-01

    Sequence alignments form the basis for many comparative and population genomic studies. Alignment tools provide a range of accuracies dependent on the divergence between the sequences and the alignment methods. Despite widespread use, there is no standard method for assessing the accuracy of a dataset and alignment strategy after resequencing. We present a framework and tool for determining the overall accuracies of an input read dataset, alignment and SNP-calling method providing an isolate in that dataset has a corresponding, or closely related reference sequence available. In addition to this tool for comparing False Discovery Rates (FDR), we include a method for determining homozygous and heterozygous positions from an alignment using binomial probabilities for an expected error rate. We benchmark this method against other SNP callers using our FDR method with three fungal genomes, finding that it was able achieve a high level of accuracy. These tools are available at http://cfdr.sourceforge.net/.

  4. Using False Discovery Rates to Benchmark SNP-callers in next-generation sequencing projects

    PubMed Central

    Farrer, Rhys A.; Henk, Daniel A.; MacLean, Dan; Studholme, David J.; Fisher, Matthew C.

    2013-01-01

    Sequence alignments form the basis for many comparative and population genomic studies. Alignment tools provide a range of accuracies dependent on the divergence between the sequences and the alignment methods. Despite widespread use, there is no standard method for assessing the accuracy of a dataset and alignment strategy after resequencing. We present a framework and tool for determining the overall accuracies of an input read dataset, alignment and SNP-calling method providing an isolate in that dataset has a corresponding, or closely related reference sequence available. In addition to this tool for comparing False Discovery Rates (FDR), we include a method for determining homozygous and heterozygous positions from an alignment using binomial probabilities for an expected error rate. We benchmark this method against other SNP callers using our FDR method with three fungal genomes, finding that it was able achieve a high level of accuracy. These tools are available at http://cfdr.sourceforge.net/. PMID:23518929

  5. 5 CFR 470.317 - Project evaluation.

    Code of Federal Regulations, 2012 CFR

    2012-01-01

    ... 5 Administrative Personnel 1 2012-01-01 2012-01-01 false Project evaluation. 470.317 Section 470... MANAGEMENT RESEARCH PROGRAMS AND DEMONSTRATIONS PROJECTS Regulatory Requirements Pertaining to Demonstration Projects § 470.317 Project evaluation. (a) Compliance evaluation. OPM will review the operation of...

  6. 5 CFR 470.317 - Project evaluation.

    Code of Federal Regulations, 2014 CFR

    2014-01-01

    ... 5 Administrative Personnel 1 2014-01-01 2014-01-01 false Project evaluation. 470.317 Section 470... MANAGEMENT RESEARCH PROGRAMS AND DEMONSTRATIONS PROJECTS Regulatory Requirements Pertaining to Demonstration Projects § 470.317 Project evaluation. (a) Compliance evaluation. OPM will review the operation of...

  7. An evaluation, comparison, and accurate benchmarking of several publicly available MS/MS search algorithms: Sensitivity and Specificity analysis.

    SciTech Connect

    Kapp, Eugene; Schutz, Frederick; Connolly, Lisa M.; Chakel, John A.; Meza, Jose E.; Miller, Christine A.; Fenyo, David; Eng, Jimmy K.; Adkins, Joshua N.; Omenn, Gilbert; Simpson, Richard

    2005-08-01

    MS/MS and associated database search algorithms are essential proteomic tools for identifying peptides. Due to their widespread use, it is now time to perform a systematic analysis of the various algorithms currently in use. Using blood specimens used in the HUPO Plasma Proteome Project, we have evaluated five search algorithms with respect to their sensitivity and specificity, and have also accurately benchmarked them based on specified false-positive (FP) rates. Spectrum Mill and SEQUEST performed well in terms of sensitivity, but were inferior to MASCOT, X-Tandem, and Sonar in terms of specificity. Overall, MASCOT, a probabilistic search algorithm, correctly identified most peptides based on a specified FP rate. The rescoring algorithm, Peptide Prophet, enhanced the overall performance of the SEQUEST algorithm, as well as provided predictable FP error rates. Ideally, score thresholds should be calculated for each peptide spectrum or minimally, derived from a reversed-sequence search as demonstrated in this study based on a validated data set. The availability of open-source search algorithms, such as X-Tandem, makes it feasible to further improve the validation process (manual or automatic) on the basis of ''consensus scoring'', i.e., the use of multiple (at least two) search algorithms to reduce the number of FPs. complement.

  8. ICSBEP Benchmarks For Nuclear Data Applications

    NASA Astrophysics Data System (ADS)

    Briggs, J. Blair

    2005-05-01

    The International Criticality Safety Benchmark Evaluation Project (ICSBEP) was initiated in 1992 by the United States Department of Energy. The ICSBEP became an official activity of the Organization for Economic Cooperation and Development (OECD) — Nuclear Energy Agency (NEA) in 1995. Representatives from the United States, United Kingdom, France, Japan, the Russian Federation, Hungary, Republic of Korea, Slovenia, Serbia and Montenegro (formerly Yugoslavia), Kazakhstan, Spain, Israel, Brazil, Poland, and the Czech Republic are now participating. South Africa, India, China, and Germany are considering participation. The purpose of the ICSBEP is to identify, evaluate, verify, and formally document a comprehensive and internationally peer-reviewed set of criticality safety benchmark data. The work of the ICSBEP is published as an OECD handbook entitled "International Handbook of Evaluated Criticality Safety Benchmark Experiments." The 2004 Edition of the Handbook contains benchmark specifications for 3331 critical or subcritical configurations that are intended for use in validation efforts and for testing basic nuclear data. New to the 2004 Edition of the Handbook is a draft criticality alarm / shielding type benchmark that should be finalized in 2005 along with two other similar benchmarks. The Handbook is being used extensively for nuclear data testing and is expected to be a valuable resource for code and data validation and improvement efforts for decades to come. Specific benchmarks that are useful for testing structural materials such as iron, chromium, nickel, and manganese; beryllium; lead; thorium; and 238U are highlighted.

  9. Evaluating the Effect of Labeled Benchmarks on Children's Number Line Estimation Performance and Strategy Use.

    PubMed

    Peeters, Dominique; Sekeris, Elke; Verschaffel, Lieven; Luwel, Koen

    2017-01-01

    Some authors argue that age-related improvements in number line estimation (NLE) performance result from changes in strategy use. More specifically, children's strategy use develops from only using the origin of the number line, to using the origin and the endpoint, to eventually also relying on the midpoint of the number line. Recently, Peeters et al. (unpublished) investigated whether the provision of additional unlabeled benchmarks at 25, 50, and 75% of the number line, positively affects third and fifth graders' NLE performance and benchmark-based strategy use. It was found that only the older children benefitted from the presence of these benchmarks at the quartiles of the number line (i.e., 25 and 75%), as they made more use of these benchmarks, leading to more accurate estimates. A possible explanation for this lack of improvement in third graders might be their inability to correctly link the presented benchmarks with their corresponding numerical values. In the present study, we investigated whether labeling these benchmarks with their corresponding numerical values, would have a positive effect on younger children's NLE performance and quartile-based strategy use as well. Third and sixth graders were assigned to one of three conditions: (a) a control condition with an empty number line bounded by 0 at the origin and 1,000 at the endpoint, (b) an unlabeled condition with three additional external benchmarks without numerical labels at 25, 50, and 75% of the number line, and (c) a labeled condition in which these benchmarks were labeled with 250, 500, and 750, respectively. Results indicated that labeling the benchmarks has a positive effect on third graders' NLE performance and quartile-based strategy use, whereas sixth graders already benefited from the mere provision of unlabeled benchmarks. These findings imply that children's benchmark-based strategy use can be stimulated by adding additional externally provided benchmarks on the number line, but that

  10. Evaluation of anode (electro)catalytic materials for the direct borohydride fuel cell: Methods and benchmarks

    NASA Astrophysics Data System (ADS)

    Olu, Pierre-Yves; Job, Nathalie; Chatenet, Marian

    2016-09-01

    In this paper, different methods are discussed for the evaluation of the potential of a given catalyst, in view of an application as a direct borohydride fuel cell DBFC anode material. Characterizations results in DBFC configuration are notably analyzed at the light of important experimental variables which influence the performances of the DBFC. However, in many practical DBFC-oriented studies, these various experimental variables prevent one to isolate the influence of the anode catalyst on the cell performances. Thus, the electrochemical three-electrode cell is a widely-employed and useful tool to isolate the DBFC anode catalyst and to investigate its electrocatalytic activity towards the borohydride oxidation reaction (BOR) in the absence of other limitations. This article reviews selected results for different types of catalysts in electrochemical cell containing a sodium borohydride alkaline electrolyte. In particular, propositions of common experimental conditions and benchmarks are given for practical evaluation of the electrocatalytic activity towards the BOR in three-electrode cell configuration. The major issue of gaseous hydrogen generation and escape upon DBFC operation is also addressed through a comprehensive review of various results depending on the anode composition. At last, preliminary concerns are raised about the stability of potential anode catalysts upon DBFC operation.

  11. Using a Project Portfolio: Empowerment Evaluation for Model Demonstration Projects.

    ERIC Educational Resources Information Center

    Baggett, David

    For model postsecondary demonstration projects serving individuals with disabilities, a portfolio of project activities may serve as a method for program evaluation, program replication, and program planning. Using a portfolio for collecting, describing, and documenting a project's successes, efforts, and failures enables project staff to take…

  12. A Benchmark Data Set to Evaluate the Illumination Robustness of Image Processing Algorithms for Object Segmentation and Classification.

    PubMed

    Khan, Arif Ul Maula; Mikut, Ralf; Reischl, Markus

    2015-01-01

    Developers of image processing routines rely on benchmark data sets to give qualitative comparisons of new image analysis algorithms and pipelines. Such data sets need to include artifacts in order to occlude and distort the required information to be extracted from an image. Robustness, the quality of an algorithm related to the amount of distortion is often important. However, using available benchmark data sets an evaluation of illumination robustness is difficult or even not possible due to missing ground truth data about object margins and classes and missing information about the distortion. We present a new framework for robustness evaluation. The key aspect is an image benchmark containing 9 object classes and the required ground truth for segmentation and classification. Varying levels of shading and background noise are integrated to distort the data set. To quantify the illumination robustness, we provide measures for image quality, segmentation and classification success and robustness. We set a high value on giving users easy access to the new benchmark, therefore, all routines are provided within a software package, but can as well easily be replaced to emphasize other aspects.

  13. A Benchmark Data Set to Evaluate the Illumination Robustness of Image Processing Algorithms for Object Segmentation and Classification

    PubMed Central

    Khan, Arif ul Maula; Mikut, Ralf; Reischl, Markus

    2015-01-01

    Developers of image processing routines rely on benchmark data sets to give qualitative comparisons of new image analysis algorithms and pipelines. Such data sets need to include artifacts in order to occlude and distort the required information to be extracted from an image. Robustness, the quality of an algorithm related to the amount of distortion is often important. However, using available benchmark data sets an evaluation of illumination robustness is difficult or even not possible due to missing ground truth data about object margins and classes and missing information about the distortion. We present a new framework for robustness evaluation. The key aspect is an image benchmark containing 9 object classes and the required ground truth for segmentation and classification. Varying levels of shading and background noise are integrated to distort the data set. To quantify the illumination robustness, we provide measures for image quality, segmentation and classification success and robustness. We set a high value on giving users easy access to the new benchmark, therefore, all routines are provided within a software package, but can as well easily be replaced to emphasize other aspects. PMID:26191792

  14. Technical Requirements for Benchmark Simulator-Based Terminal Instrument Procedures (TERPS) Evaluation.

    DTIC Science & Technology

    1986-05-01

    ARD-RA169 947 TECHNICAL -REQUIREMENTS FOR BENCHMARK SIMULATOR-BASED 1.𔃼 TERMINAL INSTRUMENT ..(U) ANALYTICAL MECHANICS ASSOCIATES INC MOUNTAIN VIEW...14 10 U1es DOTI FAA/ PM-86/14 NASA CR-177407 Technical Requirements for Benchmark Program Engineering Simulator-Based Terminal Instrument and...16 Abstract ---In order to take full advantage of the helicopter’s unique flight characteristics, enhanced terminal instrument procedures (TERPS) need

  15. Towards improved quality benchmarking and shelf life evaluation of black tiger shrimp (Penaeus monodon).

    PubMed

    Le, Nhat Tam; Doan, Nhu Khue; Nguyen Ba, Thanh; Tran, Thi Van Thi

    2017-11-15

    An improved quality benchmarking and shelf life evaluation of freshly harvested black tiger shrimp (Penaeus monodon) was pursued by combining sensory and chemical methods. This involved developing a quality index method (QIM) to further assess both freshness and shelf life of the studied shrimp samples. The quality index included the use of trimethylamine (TMA-N), total volatile basis nitrogen (TVB-N), histamine, and hypoxanthine, which were performed at scheduled times during the ten days of ice storage (0°C). Shelf life of the studied shrimp was most likely to be 8days, and there were positive linear correlations between quality indices (QI) and storage period. The quality of shrimp decreased over storage time. In fact, significant changes of chemical and sensory characteristics of the shrimp samples would become more obvious from day 5 onwards. Besides, quality classification of black tiger shrimp involved four main levels, namely: excellent, good, moderately acceptable, and just acceptable. Copyright © 2017 Elsevier Ltd. All rights reserved.

  16. Comparative assessment of scoring functions on an updated benchmark: 2. Evaluation methods and general results.

    PubMed

    Li, Yan; Han, Li; Liu, Zhihai; Wang, Renxiao

    2014-06-23

    Our comparative assessment of scoring functions (CASF) benchmark is created to provide an objective evaluation of current scoring functions. The key idea of CASF is to compare the general performance of scoring functions on a diverse set of protein-ligand complexes. In order to avoid testing scoring functions in the context of molecular docking, the scoring process is separated from the docking (or sampling) process by using ensembles of ligand binding poses that are generated in prior. Here, we describe the technical methods and evaluation results of the latest CASF-2013 study. The PDBbind core set (version 2013) was employed as the primary test set in this study, which consists of 195 protein-ligand complexes with high-quality three-dimensional structures and reliable binding constants. A panel of 20 scoring functions, most of which are implemented in main-stream commercial software, were evaluated in terms of "scoring power" (binding affinity prediction), "ranking power" (relative ranking prediction), "docking power" (binding pose prediction), and "screening power" (discrimination of true binders from random molecules). Our results reveal that the performance of these scoring functions is generally more promising in the docking/screening power tests than in the scoring/ranking power tests. Top-ranked scoring functions in the scoring power test, such as X-Score(HM), ChemScore@SYBYL, ChemPLP@GOLD, and PLP@DS, are also top-ranked in the ranking power test. Top-ranked scoring functions in the docking power test, such as ChemPLP@GOLD, Chemscore@GOLD, GlidScore-SP, LigScore@DS, and PLP@DS, are also top-ranked in the screening power test. Our results obtained on the entire test set and its subsets suggest that the real challenge in protein-ligand binding affinity prediction lies in polar interactions and associated desolvation effect. Nonadditive features observed among high-affinity protein-ligand complexes also need attention.

  17. Ada compiler evaluation on the Space Station Freedom Software Support Environment project

    NASA Technical Reports Server (NTRS)

    Badal, D. L.

    1989-01-01

    This paper describes the work in progress to select the Ada compilers for the Space Station Freedom Program (SSFP) Software Support Environment (SSE) project. The purpose of the SSE Ada compiler evaluation team is to establish the criteria, test suites, and benchmarks to be used for evaluating Ada compilers for the mainframes, workstations, and the realtime target for flight- and ground-based computers. The combined efforts and cooperation of the customer, subcontractors, vendors, academia and SIGAda groups made it possible to acquire the necessary background information, benchmarks, test suites, and criteria used.

  18. Vermont Rural and Farm Family Rehabilitation Project. A Benchmark Report. Research Report MP73.

    ERIC Educational Resources Information Center

    Tompkins, E. H.; And Others

    The report presents information about client families and their farms during their contact with the Vermont Rural and Farm Family Rehabilitation (RFFR) project from March 1, 1969 to June 30, 1971. Data are from 450 family case histories which include 2,089 members. Most were from northern Vermont. Families averaged 4.64 persons each, about 1 more…

  19. Oregon's Technical, Human, and Organizational Networking Infrastructure for Science and Mathematics: A Planning Project. Benchmark Reports.

    ERIC Educational Resources Information Center

    Lamb, William G., Ed.

    This compilation of reports is part of a planning project that aims to establish a coalition of organizations and key people who can work together to bring computerized telecommunications (CT) to Oregon as a teaching tool for science and mathematics teachers and students, and to give that coalition practical ideas for proposals to make CT a…

  20. Evaluation and optimization of virtual screening workflows with DEKOIS 2.0--a public library of challenging docking benchmark sets.

    PubMed

    Bauer, Matthias R; Ibrahim, Tamer M; Vogel, Simon M; Boeckler, Frank M

    2013-06-24

    The application of molecular benchmarking sets helps to assess the actual performance of virtual screening (VS) workflows. To improve the efficiency of structure-based VS approaches, the selection and optimization of various parameters can be guided by benchmarking. With the DEKOIS 2.0 library, we aim to further extend and complement the collection of publicly available decoy sets. Based on BindingDB bioactivity data, we provide 81 new and structurally diverse benchmark sets for a wide variety of different target classes. To ensure a meaningful selection of ligands, we address several issues that can be found in bioactivity data. We have improved our previously introduced DEKOIS methodology with enhanced physicochemical matching, now including the consideration of molecular charges, as well as a more sophisticated elimination of latent actives in the decoy set (LADS). We evaluate the docking performance of Glide, GOLD, and AutoDock Vina with our data sets and highlight existing challenges for VS tools. All DEKOIS 2.0 benchmark sets will be made accessible at http://www.dekois.com.

  1. Benchmarking the Selection and Projection Operations, and Ordering Capabilities of Relational Database Machines.

    DTIC Science & Technology

    1983-09-01

    814M 6?h’ekI 6aThis thesis describes th( performance- measurement experiments designed for a number of back- end, relational database machine...SELECTION MEASUREMENTS . . . . . . . . . . . . 30 1. The Percentage of Selection . . . . . . . 30 2. Effects of Clustare1 and Non-Clustered Indicies... MEASUREMENTS . . ......... 47 1. P.rcgntage of Projections on Non-Key A ttributes . . . . . . o . . o o . . . . o 47 2. Comparison of the Equivalent Qur4es Cn

  2. BioBenchmark Toyama 2012: an evaluation of the performance of triple stores on biological data.

    PubMed

    Wu, Hongyan; Fujiwara, Toyofumi; Yamamoto, Yasunori; Bolleman, Jerven; Yamaguchi, Atsuko

    2014-01-01

    Biological databases vary enormously in size and data complexity, from small databases that contain a few million Resource Description Framework (RDF) triples to large databases that contain billions of triples. In this paper, we evaluate whether RDF native stores can be used to meet the needs of a biological database provider. Prior evaluations have used synthetic data with a limited database size. For example, the largest BSBM benchmark uses 1 billion synthetic e-commerce knowledge RDF triples on a single node. However, real world biological data differs from the simple synthetic data much. It is difficult to determine whether the synthetic e-commerce data is efficient enough to represent biological databases. Therefore, for this evaluation, we used five real data sets from biological databases. We evaluated five triple stores, 4store, Bigdata, Mulgara, Virtuoso, and OWLIM-SE, with five biological data sets, Cell Cycle Ontology, Allie, PDBj, UniProt, and DDBJ, ranging in size from approximately 10 million to 8 billion triples. For each database, we loaded all the data into our single node and prepared the database for use in a classical data warehouse scenario. Then, we ran a series of SPARQL queries against each endpoint and recorded the execution time and the accuracy of the query response. Our paper shows that with appropriate configuration Virtuoso and OWLIM-SE can satisfy the basic requirements to load and query biological data less than 8 billion or so on a single node, for the simultaneous access of 64 clients. OWLIM-SE performs best for databases with approximately 11 million triples; For data sets that contain 94 million and 590 million triples, OWLIM-SE and Virtuoso perform best. They do not show overwhelming advantage over each other; For data over 4 billion Virtuoso works best. 4store performs well on small data sets with limited features when the number of triples is less than 100 million, and our test shows its scalability is poor; Bigdata

  3. BioBenchmark Toyama 2012: an evaluation of the performance of triple stores on biological data

    PubMed Central

    2014-01-01

    Background Biological databases vary enormously in size and data complexity, from small databases that contain a few million Resource Description Framework (RDF) triples to large databases that contain billions of triples. In this paper, we evaluate whether RDF native stores can be used to meet the needs of a biological database provider. Prior evaluations have used synthetic data with a limited database size. For example, the largest BSBM benchmark uses 1 billion synthetic e-commerce knowledge RDF triples on a single node. However, real world biological data differs from the simple synthetic data much. It is difficult to determine whether the synthetic e-commerce data is efficient enough to represent biological databases. Therefore, for this evaluation, we used five real data sets from biological databases. Results We evaluated five triple stores, 4store, Bigdata, Mulgara, Virtuoso, and OWLIM-SE, with five biological data sets, Cell Cycle Ontology, Allie, PDBj, UniProt, and DDBJ, ranging in size from approximately 10 million to 8 billion triples. For each database, we loaded all the data into our single node and prepared the database for use in a classical data warehouse scenario. Then, we ran a series of SPARQL queries against each endpoint and recorded the execution time and the accuracy of the query response. Conclusions Our paper shows that with appropriate configuration Virtuoso and OWLIM-SE can satisfy the basic requirements to load and query biological data less than 8 billion or so on a single node, for the simultaneous access of 64 clients. OWLIM-SE performs best for databases with approximately 11 million triples; For data sets that contain 94 million and 590 million triples, OWLIM-SE and Virtuoso perform best. They do not show overwhelming advantage over each other; For data over 4 billion Virtuoso works best. 4store performs well on small data sets with limited features when the number of triples is less than 100 million, and our test shows its

  4. Criticality experiments and benchmarks for cross section evaluation: the neptunium case

    NASA Astrophysics Data System (ADS)

    Leong, L. S.; Tassan-Got, L.; Audouin, L.; Paradela, C.; Wilson, J. N.; Tarrio, D.; Berthier, B.; Duran, I.; Le Naour, C.; Stéphan, C.

    2013-03-01

    The 237Np neutron-induced fission cross section has been recently measured in a large energy range (from eV to GeV) at the n_TOF facility at CERN. When compared to previous measurement the n_TOF fission cross section appears to be higher by 5-7% beyond the fission threshold. To check the relevance of n_TOF data, we apply a criticality experiment performed at Los Alamos with a 6 kg sphere of 237Np, surrounded by enriched uranium 235U so as to approach criticality with fast neutrons. The multiplication factor ke f f of the calculation is in better agreement with the experiment (the deviation of 750 pcm is reduced to 250 pcm) when we replace the ENDF/B-VII.0 evaluation of the 237Np fission cross section by the n_TOF data. We also explore the hypothesis of deficiencies of the inelastic cross section in 235U which has been invoked by some authors to explain the deviation of 750 pcm. With compare to inelastic large distortion calculation, it is incompatible with existing measurements. Also we show that the v of 237Np can hardly be incriminated because of the high accuracy of the existing data. Fission rate ratios or averaged fission cross sections measured in several fast neutron fields seem to give contradictory results on the validation of the 237Np cross section but at least one of the benchmark experiments, where the active deposits have been well calibrated for the number of atoms, favors the n_TOF data set. These outcomes support the hypothesis of a higher fission cross section of 237Np.

  5. Benchmarking studies for the DESCARTES and CIDER codes. Hanford Environmental Dose Reconstruction Project

    SciTech Connect

    Eslinger, P.W.; Ouderkirk, S.J.; Nichols, W.E.

    1993-01-01

    The Hanford Envirorunental Dose Reconstruction (HEDR) project is developing several computer codes to model the airborne release, transport, and envirormental accumulation of radionuclides resulting from Hanford operations from 1944 through 1972. In order to calculate the dose of radiation a person may have received in any given location, the geographic area addressed by the HEDR Project will be divided into a grid. The grid size suggested by the draft requirements contains 2091 units called nodes. Two of the codes being developed are DESCARTES and CIDER. The DESCARTES code will be used to estimate the concentration of radionuclides in environmental pathways from the output of the air transport code RATCHET. The CIDER code will use information provided by DESCARTES to estimate the dose received by an individual. The requirements that Battelle (BNW) set for these two codes were released to the HEDR Technical Steering Panel (TSP) in a draft document on November 10, 1992. This document reports on the preliminary work performed by the code development team to determine if the requirements could be met.

  6. Benchmarking Clinical Speech Recognition and Information Extraction: New Data, Methods, and Evaluations

    PubMed Central

    Zhou, Liyuan; Hanlen, Leif; Ferraro, Gabriela

    2015-01-01

    Background Over a tenth of preventable adverse events in health care are caused by failures in information flow. These failures are tangible in clinical handover; regardless of good verbal handover, from two-thirds to all of this information is lost after 3-5 shifts if notes are taken by hand, or not at all. Speech recognition and information extraction provide a way to fill out a handover form for clinical proofing and sign-off. Objective The objective of the study was to provide a recorded spoken handover, annotated verbatim transcriptions, and evaluations to support research in spoken and written natural language processing for filling out a clinical handover form. This dataset is based on synthetic patient profiles, thereby avoiding ethical and legal restrictions, while maintaining efficacy for research in speech-to-text conversion and information extraction, based on realistic clinical scenarios. We also introduce a Web app to demonstrate the system design and workflow. Methods We experiment with Dragon Medical 11.0 for speech recognition and CRF++ for information extraction. To compute features for information extraction, we also apply CoreNLP, MetaMap, and Ontoserver. Our evaluation uses cross-validation techniques to measure processing correctness. Results The data provided were a simulation of nursing handover, as recorded using a mobile device, built from simulated patient records and handover scripts, spoken by an Australian registered nurse. Speech recognition recognized 5276 of 7277 words in our 100 test documents correctly. We considered 50 mutually exclusive categories in information extraction and achieved the F1 (ie, the harmonic mean of Precision and Recall) of 0.86 in the category for irrelevant text and the macro-averaged F1 of 0.70 over the remaining 35 nonempty categories of the form in our 101 test documents. Conclusions The significance of this study hinges on opening our data, together with the related performance benchmarks and some

  7. Helical screw expander evaluation project

    NASA Technical Reports Server (NTRS)

    Mckay, R.

    1982-01-01

    A one MW helical rotary screw expander power system for electric power generation from geothermal brine was evaluated. The technology explored in the testing is simple, potentially very efficient, and ideally suited to wellhead installations in moderate to high enthalpy, liquid dominated field. A functional one MW geothermal electric power plant that featured a helical screw expander was produced and then tested with a demonstrated average performance of approximately 45% machine efficiency over a wide range of test conditions in noncondensing, operation on two-phase geothermal fluids. The Project also produced a computer equipped data system, an instrumentation and control van, and a 1000 kW variable load bank, all integrated into a test array designed for operation at a variety of remote test sites. Data are presented for the Utah testing and for the noncondensing phases of the testing in Mexico. Test time logged was 437 hours during the Utah tests and 1101 hours during the Mexico tests.

  8. Helical screw expander evaluation project

    NASA Astrophysics Data System (ADS)

    McKay, R.

    1982-03-01

    A one MW helical rotary screw expander power system for electric power generation from geothermal brine was evaluated. The technology explored in the testing is simple, potentially very efficient, and ideally suited to wellhead installations in moderate to high enthalpy, liquid dominated field. A functional one MW geothermal electric power plant that featured a helical screw expander was produced and then tested with a demonstrated average performance of approximately 45% machine efficiency over a wide range of test conditions in noncondensing, operation on two-phase geothermal fluids. The Project also produced a computer equipped data system, an instrumentation and control van, and a 1000 kW variable load bank, all integrated into a test array designed for operation at a variety of remote test sites. Data are presented for the Utah testing and for the noncondensing phases of the testing in Mexico. Test time logged was 437 hours during the Utah tests and 1101 hours during the Mexico tests.

  9. Creating Robust Evaluation of ATE Projects

    ERIC Educational Resources Information Center

    Eddy, Pamela L.

    2017-01-01

    Funded grant projects all involve some form of evaluation, and Advanced Technological Education (ATE) grants are no exception. Program evaluation serves as a critical component not only for evaluating if a project has met its intended and desired outcomes, but the evaluation process is also a central feature of the grant application itself.…

  10. ICSBEP Criticality Benchmark Eigenvalues with ENDF/B-VII.1 Cross Sections

    SciTech Connect

    Kahler, Albert C. III; MacFarlane, Robert

    2012-06-28

    We review MCNP eigenvalue calculations from a suite of International Criticality Safety Benchmark Evaluation Project (ICSBEP) Handbook evaluations with the recently distributed ENDF/B-VII.1 cross section library.

  11. Benchmark simulation Model no 2 in Matlab-simulink: towards plant-wide WWTP control strategy evaluation.

    PubMed

    Vreck, D; Gernaey, K V; Rosen, C; Jeppsson, U

    2006-01-01

    In this paper, implementation of the Benchmark Simulation Model No 2 (BSM2) within Matlab-Simulink is presented. The BSM2 is developed for plant-wide WWTP control strategy evaluation on a long-term basis. It consists of a pre-treatment process, an activated sludge process and sludge treatment processes. Extended evaluation criteria are proposed for plant-wide control strategy assessment. Default open-loop and closed-loop strategies are also proposed to be used as references with which to compare other control strategies. Simulations indicate that the BM2 is an appropriate tool for plant-wide control strategy evaluation.

  12. The PIE Institute Project: Final Evaluation Report

    ERIC Educational Resources Information Center

    St. John, Mark; Carroll, Becky; Helms, Jen; Smith, Anita

    2008-01-01

    The Playful Invention and Exploration (PIE) Institute project was funded in 2005 by the National Science Foundation (NSF). For the past three years, Inverness Research has served as the external evaluator for the PIE project. The authors' evaluation efforts have included extensive observation and documentation of PIE project activities; ongoing…

  13. Evaluation of Project Symbiosis: An Interdisciplinary Science Education Project.

    ERIC Educational Resources Information Center

    Altschuld, James W.

    1993-01-01

    The goal of this report is to provide a summary of the evaluation of Project Symbiosis which focused on enhancing the teaching of science principles in high school agriculture courses. The project initially involved 15 teams of science and agriculture teachers and was characterized by an extensive evaluation component consisting of six formal…

  14. Evaluation of the ACEC Benchmark Suite for Real-Time Applications

    DTIC Science & Technology

    1990-07-23

    abstract. Enter either UL (unlimited) or SAP be published in.... When a report is revised. (same as report). An entry in this block is include a...results to the Real-time benchmarks (listed in [4]) shows that the two sets of results are in sync for the common features measured by those

  15. Evaluation of the potential of benchmarking to facilitate the measurement of chemical persistence in lakes.

    PubMed

    Zou, Hongyan; MacLeod, Matthew; McLachlan, Michael S

    2014-01-01

    The persistence of chemicals in the environment is rarely measured in the field due to a paucity of suitable methods. Here we explore the potential of chemical benchmarking to facilitate the measurement of persistence in lake systems using a multimedia chemical fate model. The model results show that persistence in a lake can be assessed by quantifying the ratio of test chemical and benchmark chemical at as few as two locations: the point of emission and the outlet of the lake. Appropriate selection of benchmark chemicals also allows pseudo-first-order rate constants for physical removal processes such as volatilization and sediment burial to be quantified. We use the model to explore how the maximum persistence that can be measured in a particular lake depends on the partitioning properties of the test chemical of interest and the characteristics of the lake. Our model experiments demonstrate that combining benchmarking techniques with good experimental design and sensitive environmental analytical chemistry may open new opportunities for quantifying chemical persistence, particularly for relatively slowly degradable chemicals for which current methods do not perform well.

  16. The PATH project in eight European countries: an evaluation.

    PubMed

    Veillard, Jeremy Henri Maurice; Schiøtz, Michaela Louise; Guisset, Ann-Lise; Brown, Adalsteinn Davidson; Klazinga, Niek S

    2013-01-01

    This paper's aim is to evaluate the perceived impact and the enabling factors and barriers experienced by hospital staff participating in an international hospital performance measurement project focused on internal quality improvement. Semi-structured interviews involving international hospital performance measurement project coordinators, including 140 hospitals from eight European countries (Belgium, Estonia, France, Germany, Hungary, Poland, Slovakia and Slovenia). Inductively analyzing the interview transcripts was carried out using the grounded theory approach. Even when public reporting is absent, the project was perceived as having stimulated performance measurement and quality improvement initiatives in participating hospitals. Attention should be paid to leadership/ownership, context, content (project intrinsic features) and processes supporting elements. Generalizing the findings is limited by the study's small sample size. Possible implications for the WHO European Regional Office and for participating hospitals would be to assess hospital preparedness to participate in the PATH project, depending on context, process and structural elements; and enhance performance and practice benchmarking through suggested approaches. This research gathered rich and unique material related to an international performance measurement project. It derived actionable findings.

  17. Neutron Cross Section Processing Methods for Improved Integral Benchmarking of Unresolved Resonance Region Evaluations

    NASA Astrophysics Data System (ADS)

    Walsh, Jonathan A.; Forget, Benoit; Smith, Kord S.; Brown, Forrest B.

    2016-03-01

    In this work we describe the development and application of computational methods for processing neutron cross section data in the unresolved resonance region (URR). These methods are integrated with a continuous-energy Monte Carlo neutron transport code, thereby enabling their use in high-fidelity analyses. Enhanced understanding of the effects of URR evaluation representations on calculated results is then obtained through utilization of the methods in Monte Carlo integral benchmark simulations of fast spectrum critical assemblies. First, we present a so-called on-the-fly (OTF) method for calculating and Doppler broadening URR cross sections. This method proceeds directly from ENDF-6 average unresolved resonance parameters and, thus, eliminates any need for a probability table generation pre-processing step in which tables are constructed at several energies for all desired temperatures. Significant memory reduction may be realized with the OTF method relative to a probability table treatment if many temperatures are needed. Next, we examine the effects of using a multi-level resonance formalism for resonance reconstruction in the URR. A comparison of results obtained by using the same stochastically-generated realization of resonance parameters in both the single-level Breit-Wigner (SLBW) and multi-level Breit-Wigner (MLBW) formalisms allows for the quantification of level-level interference effects on integrated tallies such as keff and energy group reaction rates. Though, as is well-known, cross section values at any given incident energy may differ significantly between single-level and multi-level formulations, the observed effects on integral results are minimal in this investigation. Finally, we demonstrate the calculation of true expected values, and the statistical spread of those values, through independent Monte Carlo simulations, each using an independent realization of URR cross section structure throughout. It is observed that both probability table

  18. Linking user and staff perspectives in the evaluation of innovative transition projects for youth with disabilities.

    PubMed

    McAnaney, Donal F; Wynne, Richard F

    2016-06-01

    A key challenge in formative evaluation is to gather appropriate evidence to inform the continuous improvement of initiatives. In the absence of outcome data, the programme evaluator often must rely on the perceptions of beneficiaries and staff in generating insight into what is making a difference. The article describes the approach adopted in an evaluation of 15 innovative projects supporting school-leavers with disabilities in making the transition to education, work and life in community settings. Two complementary processes provided an insight into what project staff and leadership viewed as the key project activities and features that facilitated successful transition as well as the areas of quality of life (QOL) that participants perceived as having been impacted positively by the projects. A comparison was made between participants' perceptions of QOL impact with the views of participants in services normally offered by the wider system. This revealed that project participants were significantly more positive in their views than participants in traditional services. In addition, the processes and activities of the more highly rated projects were benchmarked against less highly rated projects and also with usually available services. Even in the context of a range of intervening variables such as level and complexity of participant needs and variations in the stage of development of individual projects, the benchmarking process indicated a number of project characteristics that were highly valued by participants. © The Author(s) 2016.

  19. Project Performance Evaluation Using Deep Belief Networks

    NASA Astrophysics Data System (ADS)

    Nguvulu, Alick; Yamato, Shoso; Honma, Toshihisa

    A Project Assessment Indicator (PAI) Model has recently been applied to evaluate monthly project performance based on 15 project elements derived from the project management (PM) knowledge areas. While the PAI Model comprehensively evaluates project performance, it lacks objectivity and universality. It lacks objectivity because experts assign model weights intuitively based on their PM skills and experience. It lacks universality because the allocation of ceiling scores to project elements is done ad hoc based on the empirical rule without taking into account the interactions between the project elements. This study overcomes these limitations by applying a DBN approach where the model automatically assigns weights and allocates ceiling scores to the project elements based on the DBN weights which capture the interaction between the project elements. We train our DBN on 5 IT projects of 12 months duration and test it on 8 IT projects with less than 12 months duration. We completely eliminate the manual assigning of weights and compute ceiling scores of project elements based on DBN weights. Our trained DBN evaluates monthly project performance of the 8 test projects based on the 15 project elements to within a monthly relative error margin of between ±1.03 and ±3.30%.

  20. Evaluation of a High-Accuracy MacCormack-Type Scheme Using Benchmark Problems

    NASA Technical Reports Server (NTRS)

    Hixon, R.

    1997-01-01

    Due to their inherent dissipation and stability, the MacCormack scheme and its variants have been widely used in the computation of unsteady flow and acoustic problems. However, these schemes require many points per wavelength in order to propagate waves with a reasonable amount of accuracy. In this work, the linear wave propagation characteristics of MacCormack-type schemes are shown by solving several of the CAA Benchmark Problems.

  1. Establishing Benchmarks for DOE Commercial Building R&D and Program Evaluation: Preprint

    SciTech Connect

    Deru, M.; Griffith, B.; Torcellini, P.

    2006-06-01

    The U.S. Department of Energy (DOE) Building Technologies Program and the DOE research laboratories conduct a great deal of research on building technologies. However, differences in models and simulation tools used by various research groups make it difficult to compare results among studies. The authors have developed a set of 22 hypothetical benchmark buildings and weighting factors for nine locations across the country, for a total of 198 buildings.

  2. Evaluating the Joint Theater Trauma Registry as a Data Source to Benchmark Casualty Care

    DTIC Science & Technology

    2012-05-01

    in casualties with polytrauma and a moderate blunt TBI. Secondary insults after TBI, especially hypother- mia and hypoxemia, increased the odds of 24...combat casualty care.3 Benchmark ana - lyses can be used to document the effectiveness of the combat care provided but may also reveal gaps in care...increased mortality when hypother- mia accompanies polytrauma in the civilian sector,36–38 our data indicate that combat injured individuals with hypother

  3. Benchmark IMRT evaluation of a Co-60 MRI-guided radiation therapy system.

    PubMed

    Wooten, H Omar; Rodriguez, Vivian; Green, Olga; Kashani, Rojano; Santanam, Lakshmi; Tanderup, Kari; Mutic, Sasa; Li, H Harold

    2015-03-01

    A device for MRI-guided radiation therapy (MR-IGRT) that uses cobalt-60 sources to deliver intensity modulated radiation therapy is now commercially available. We investigated the performance of the treatment planning and delivery system against the benchmark recommended by the American Association of Physicists in Medicine (AAPM) Task Group 119 for IMRT commissioning and demonstrated that the device plans and delivers IMRT treatments within recommended confidence limits and with similar accuracy as linac IMRT.

  4. Evaluation of the Pool Critical Assembly Benchmark with Explicitly-Modeled Geometry using MCNP6

    DOE PAGES

    Kulesza, Joel A.; Martz, Roger Lee

    2017-03-01

    Despite being one of the most widely used benchmarks for qualifying light water reactor (LWR) radiation transport methods and data, no benchmark calculation of the Oak Ridge National Laboratory (ORNL) Pool Critical Assembly (PCA) pressure vessel wall benchmark facility (PVWBF) using MCNP6 with explicitly modeled core geometry exists. As such, this paper provides results for such an analysis. First, a criticality calculation is used to construct the fixed source term. Next, ADVANTG-generated variance reduction parameters are used within the final MCNP6 fixed source calculations. These calculations provide unadjusted dosimetry results using three sets of dosimetry reaction cross sections of varyingmore » ages (those packaged with MCNP6, from the IRDF-2002 multi-group library, and from the ACE-formatted IRDFF v1.05 library). These results are then compared to two different sets of measured reaction rates. The comparison agrees in an overall sense within 2% and on a specific reaction- and dosimetry location-basis within 5%. Except for the neptunium dosimetry, the individual foil raw calculation-to-experiment comparisons usually agree within 10% but is typically greater than unity. Finally, in the course of developing these calculations, geometry that has previously not been completely specified is provided herein for the convenience of future analysts.« less

  5. A modified ATP benchmark for evaluating the cleaning of some hospital environmental surfaces.

    PubMed

    Lewis, T; Griffith, C; Gallo, M; Weinbren, M

    2008-06-01

    Hospital cleaning continues to attract patient, media and political attention. In the UK it is still primarily assessed via visual inspection, which can be misleading. Calls have therefore been made for a more objective approach to assessing surface cleanliness. To improve the management of hospital cleaning the use of adenosine triphosphate (ATP) in combination with microbiological analysis has been proposed, with a general ATP benchmark value of 500 relative light units (RLU) for one combination of test and equipment. In this study, the same test combination was used to assess cleaning effectiveness in a 1300-bed teaching hospital after routine and modified cleaning protocols. Based upon the ATP results a revised stricter pass/fail benchmark of 250 RLU is proposed for the range of surfaces used in this study. This was routinely achieved using modified best practice cleaning procedures which also gave reduced surface counts with, for example, aerobic colony counts reduced from >100 to <2.5 cfu/cm(2), and counts of Staphylococcus aureus reduced from up to 2.5 to <1 cfu/cm(2) (95% of the time). Benchmarking is linked to incremental quality improvements and both the original suggestion of 500 RLU and the revised figure of 250 RLU can be used by hospitals as part of this process. They can also be used in the assessment of novel cleaning methods, such as steam cleaning and microfibre cloths, which have potential use in the National Health Service.

  6. Comprehensive Evaluation Project. Final Report.

    ERIC Educational Resources Information Center

    1969

    This project sought to develop a set of tests for the assessment of the basic literacy and occupational cognizance of pupils in those public elementary and secondary schools, including vocational schools, receiving services through Federally supported educational programs and projects. The assessment is to produce generalizable average scores for…

  7. Implementation of patient blood management remains extremely variable in Europe and Canada: the NATA benchmark project: An observational study.

    PubMed

    Van der Linden, Philippe; Hardy, Jean-François

    2016-12-01

    Preoperative anaemia is associated with increased postoperative morbidity and mortality. Patient blood management (PBM) is advocated to improve patient outcomes. NATA, the 'Network for the advancement of patient blood management, haemostasis and thrombosis', initiated a benchmark project with the aim of providing the basis for educational strategies to implement optimal PBM in participating centres. Prospective, observational study with online data collection in 11 secondary and tertiary care institutions interested in developing PBM. Ten European centres (Austria, Spain, England, Denmark, Belgium, Netherlands, Romania, Greece, France, and Germany) and one Canadian centre participated between January 2010 and June 2011. A total of 2470 patients undergoing total hip (THR) or knee replacement, or coronary artery bypass grafting (CABG), were registered in the study. Data from 2431 records were included in the final analysis. Primary outcome measures were the incidence and volume of red blood cells (RBC) transfused. Logistic regression analysis identified variables independently associated with RBC transfusions. The incidence of transfusion was significantly different between centres for THR (range 7 to 95%), total knee replacement (range 3 to 100%) and CABG (range 20 to 95%). The volume of RBC transfused was significantly different between centres for THR and CABG. The incidence of preoperative anaemia ranged between 3 and 40% and its treatment between 0 and 40%, the latter not being related to the former. Patient characteristics, evolution of haemoglobin concentrations and blood losses were also different between centres. Variables independently associated with RBC transfusion were preoperative haemoglobin concentration, lost volume of RBC and female sex. Implementation of PBM remains extremely variable across centres. The relative importance of factors explaining RBC transfusion differs across institutions, some being patient related whereas others are related to

  8. Thermal Performance Benchmarking: Annual Report

    SciTech Connect

    Moreno, Gilbert

    2016-04-08

    The goal for this project is to thoroughly characterize the performance of state-of-the-art (SOA) automotive power electronics and electric motor thermal management systems. Information obtained from these studies will be used to: Evaluate advantages and disadvantages of different thermal management strategies; establish baseline metrics for the thermal management systems; identify methods of improvement to advance the SOA; increase the publicly available information related to automotive traction-drive thermal management systems; help guide future electric drive technologies (EDT) research and development (R&D) efforts. The performance results combined with component efficiency and heat generation information obtained by Oak Ridge National Laboratory (ORNL) may then be used to determine the operating temperatures for the EDT components under drive-cycle conditions. In FY15, the 2012 Nissan LEAF power electronics and electric motor thermal management systems were benchmarked. Testing of the 2014 Honda Accord Hybrid power electronics thermal management system started in FY15; however, due to time constraints it was not possible to include results for this system in this report. The focus of this project is to benchmark the thermal aspects of the systems. ORNL's benchmarking of electric and hybrid electric vehicle technology reports provide detailed descriptions of the electrical and packaging aspects of these automotive systems.

  9. Research Administration Benchmarking Project

    SciTech Connect

    Arnest, Paul

    2003-04-20

    The National Science Foundation (NSF) is conducting a collection of baseline data that will serve as an index for describing the way in which grant applications/proposals are developed and tracked at grant recipient organizations. The type of information to be collected includes (1) estimate of burden, both time and cost, to the applicant organization (2) descriptiion of the overall development and tracking systems used by the applicant, and (3) data that can be compared after the implementation of either a streamlined paper or electronic system.

  10. Benchmarking expert system tools

    NASA Technical Reports Server (NTRS)

    Riley, Gary

    1988-01-01

    As part of its evaluation of new technologies, the Artificial Intelligence Section of the Mission Planning and Analysis Div. at NASA-Johnson has made timing tests of several expert system building tools. Among the production systems tested were Automated Reasoning Tool, several versions of OPS5, and CLIPS (C Language Integrated Production System), an expert system builder developed by the AI section. Also included in the test were a Zetalisp version of the benchmark along with four versions of the benchmark written in Knowledge Engineering Environment, an object oriented, frame based expert system tool. The benchmarks used for testing are studied.

  11. Benchmarking expert system tools

    NASA Technical Reports Server (NTRS)

    Riley, Gary

    1988-01-01

    As part of its evaluation of new technologies, the Artificial Intelligence Section of the Mission Planning and Analysis Div. at NASA-Johnson has made timing tests of several expert system building tools. Among the production systems tested were Automated Reasoning Tool, several versions of OPS5, and CLIPS (C Language Integrated Production System), an expert system builder developed by the AI section. Also included in the test were a Zetalisp version of the benchmark along with four versions of the benchmark written in Knowledge Engineering Environment, an object oriented, frame based expert system tool. The benchmarks used for testing are studied.

  12. Hospital benchmarking: are U.S. eye hospitals ready?

    PubMed

    de Korne, Dirk F; van Wijngaarden, Jeroen D H; Sol, Kees J C A; Betz, Robert; Thomas, Richard C; Schein, Oliver D; Klazinga, Niek S

    2012-01-01

    Benchmarking is increasingly considered a useful management instrument to improve quality in health care, but little is known about its applicability in hospital settings. The aims of this study were to assess the applicability of a benchmarking project in U.S. eye hospitals and compare the results with an international initiative. We evaluated multiple cases by applying an evaluation frame abstracted from the literature to five U.S. eye hospitals that used a set of 10 indicators for efficiency benchmarking. Qualitative analysis entailed 46 semistructured face-to-face interviews with stakeholders, document analyses, and questionnaires. The case studies only partially met the conditions of the evaluation frame. Although learning and quality improvement were stated as overall purposes, the benchmarking initiative was at first focused on efficiency only. No ophthalmic outcomes were included, and clinicians were skeptical about their reporting relevance and disclosure. However, in contrast with earlier findings in international eye hospitals, all U.S. hospitals worked with internal indicators that were integrated in their performance management systems and supported benchmarking. Benchmarking can support performance management in individual hospitals. Having a certain number of comparable institutes provide similar services in a noncompetitive milieu seems to lay fertile ground for benchmarking. International benchmarking is useful only when these conditions are not met nationally. Although the literature focuses on static conditions for effective benchmarking, our case studies show that it is a highly iterative and learning process. The journey of benchmarking seems to be more important than the destination. Improving patient value (health outcomes per unit of cost) requires, however, an integrative perspective where clinicians and administrators closely cooperate on both quality and efficiency issues. If these worlds do not share such a relationship, the added

  13. A study on operation efficiency evaluation based on firm's financial index and benchmark selection: take China Unicom as an example

    NASA Astrophysics Data System (ADS)

    Wu, Zu-guang; Tian, Zhan-jun; Liu, Hui; Huang, Rui; Zhu, Guo-hua

    2009-07-01

    Being the only listed telecom operators of A share market, China Unicom has always been attracted many institutional investors under the concept of 3G recent years,which itself is a great technical progress expectation.Do the institutional investors or the concept of technical progress have signficant effect on the improving of firm's operating efficiency?Though reviewing the documentary about operating efficiency we find that schoolars study this problem useing the regress analyzing based on traditional production function and data envelopment analysis(DEA) and financial index anayzing and marginal function and capital labor ratio coefficient etc. All the methods mainly based on macrodata. This paper we use the micro-data of company to evaluate the operating efficiency.Using factor analyzing based on financial index and comparing the factor score of three years from 2005 to 2007, we find that China Unicom's operating efficiency is under the averge level of benchmark corporates and has't improved under the concept of 3G from 2005 to 2007.In other words,institutional investor or the conception of technical progress expectation have faint effect on the changes of China Unicom's operating efficiency. Selecting benchmark corporates as post to evaluate the operating efficiency is a characteristic of this method ,which is basicallly sipmly and direct.This method is suit for the operation efficiency evaluation of agriculture listed companies because agriculture listed also face technical progress and marketing concept such as tax-free etc.

  14. Benchmarking foreign electronics technologies

    SciTech Connect

    Bostian, C.W.; Hodges, D.A.; Leachman, R.C.; Sheridan, T.B.; Tsang, W.T.; White, R.M.

    1994-12-01

    This report has been drafted in response to a request from the Japanese Technology Evaluation Center`s (JTEC) Panel on Benchmarking Select Technologies. Since April 1991, the Competitive Semiconductor Manufacturing (CSM) Program at the University of California at Berkeley has been engaged in a detailed study of quality, productivity, and competitiveness in semiconductor manufacturing worldwide. The program is a joint activity of the College of Engineering, the Haas School of Business, and the Berkeley Roundtable on the International Economy, under sponsorship of the Alfred P. Sloan Foundation, and with the cooperation of semiconductor producers from Asia, Europe and the United States. Professors David A. Hodges and Robert C. Leachman are the project`s Co-Directors. The present report for JTEC is primarily based on data and analysis drawn from that continuing program. The CSM program is being conducted by faculty, graduate students and research staff from UC Berkeley`s Schools of Engineering and Business, and Department of Economics. Many of the participating firms are represented on the program`s Industry Advisory Board. The Board played an important role in defining the research agenda. A pilot study was conducted in 1991 with the cooperation of three semiconductor plants. The research plan and survey documents were thereby refined. The main phase of the CSM benchmarking study began in mid-1992 and will continue at least through 1997. reports are presented on the manufacture of integrated circuits; data storage; wireless technology; human-machine interfaces; and optoelectronics. Selected papers are indexed separately for inclusion in the Energy Science and Technology Database.

  15. MicroRNA array normalization: an evaluation using a randomized dataset as the benchmark.

    PubMed

    Qin, Li-Xuan; Zhou, Qin

    2014-01-01

    MicroRNA arrays possess a number of unique data features that challenge the assumption key to many normalization methods. We assessed the performance of existing normalization methods using two microRNA array datasets derived from the same set of tumor samples: one dataset was generated using a blocked randomization design when assigning arrays to samples and hence was free of confounding array effects; the second dataset was generated without blocking or randomization and exhibited array effects. The randomized dataset was assessed for differential expression between two tumor groups and treated as the benchmark. The non-randomized dataset was assessed for differential expression after normalization and compared against the benchmark. Normalization improved the true positive rate significantly in the non-randomized data but still possessed a false discovery rate as high as 50%. Adding a batch adjustment step before normalization further reduced the number of false positive markers while maintaining a similar number of true positive markers, which resulted in a false discovery rate of 32% to 48%, depending on the specific normalization method. We concluded the paper with some insights on possible causes of false discoveries to shed light on how to improve normalization for microRNA arrays.

  16. Establishing the Geomagnetic Disturbance Benchmark Event for Evaluation of the Space Weather Hazard on Power Grids

    NASA Astrophysics Data System (ADS)

    Pulkkinen, A. A.; Bernabeu, E.; Eichner, J.

    2014-12-01

    The awareness about potential major impact geomagnetically induced currents (GIC) can have on the North American high-voltage power transmission system has prompted Federal Energy Regulatory Commission (FERC) to launch a geomagnetic disturbances (GMD) standards drafting process. The goals of the GMD standards are to quantify and mitigate the GMD hazard on the North American grid. North American Electric Reliability Corporation's (NERC) is coordinating the standards drafting process that is now entering Phase II involving quantification of the impact GIC can have on individual parts of the North American grid. As a part of the Phase II GMD standards drafting process, substantial effort has been made for generating benchmark GMD scenarios. These scenarios that quantify extreme geoelectric field magnitudes and temporal waveforms of the field fluctuations are the foundation for subsequent engineering and impacts analyses. The engineering analyses will include the transmission system voltage stability and transformer heating assessments. The work on the GMD scenarios has been a major collaboration between a number of international entities involved in GMD research and transmission system operations. We will discuss in this paper the key elements of the benchmark GMD generation process and show the latest results from our work on the topic.

  17. Interlog protein network: an evolutionary benchmark of protein interaction networks for the evaluation of clustering algorithms.

    PubMed

    Jafari, Mohieddin; Mirzaie, Mehdi; Sadeghi, Mehdi

    2015-10-05

    In the field of network science, exploring principal and crucial modules or communities is critical in the deduction of relationships and organization of complex networks. This approach expands an arena, and thus allows further study of biological functions in the field of network biology. As the clustering algorithms that are currently employed in finding modules have innate uncertainties, external and internal validations are necessary. Sequence and network structure alignment, has been used to define the Interlog Protein Network (IPN). This network is an evolutionarily conserved network with communal nodes and less false-positive links. In the current study, the IPN is employed as an evolution-based benchmark in the validation of the module finding methods. The clustering results of five algorithms; Markov Clustering (MCL), Restricted Neighborhood Search Clustering (RNSC), Cartographic Representation (CR), Laplacian Dynamics (LD) and Genetic Algorithm; to find communities in Protein-Protein Interaction networks (GAPPI) are assessed by IPN in four distinct Protein-Protein Interaction Networks (PPINs). The MCL shows a more accurate algorithm based on this evolutionary benchmarking approach. Also, the biological relevance of proteins in the IPN modules generated by MCL is compatible with biological standard databases such as Gene Ontology, KEGG and Reactome. In this study, the IPN shows its potential for validation of clustering algorithms due to its biological logic and straightforward implementation.

  18. Benchmark evaluation of the RELAP code to calculate boiling in narrow channels

    SciTech Connect

    Kunze, J.F.; Loyalka, S.K. ); McKibben, J.C.; Hultsch, R.; Oladiran, O.

    1990-06-01

    The RELAP code has been tested with benchmark experiments (such as the loss-of-fluid test experiments at the Idaho National Engineering Laboratory) at high pressures and temperatures characteristic of those encountered in loss-of-coolant accidents (LOCAs) in commercial light water power reactors. Application of RELAP to the LOCA analysis of a low pressure (< 7 atm) and low temperature (< 100{degree}C), plate-type research reactor, such as the University of Missouri Research Reactor (MURR), the high-flux breeder reactor, high-flux isotope reactor, and Advanced Test Reactor, requires resolution of questions involving overextrapolation to very low pressures and low temperatures, and calculations of the pulsed boiling/reflood conditions in the narrow rectangular cross-section channels (typically 2 mm thick) of the plate fuel elements. The practical concern of this problem is that plate fuel temperatures predicted by RELAP5 (MOD2, version 3) during the pulsed boiling period can reach high enough temperatures to cause plate (clad) weakening, though not melting. Since an experimental benchmark of RELAP under such LOCA conditions is not available and since such conditions present substantial challenges to the code, it is important to verify the code predictions. The comparison of the pulsed boiling experiments with the RELAP calculations involves both visual observations of void fraction versus time and measurements of temperatures near the fuel plate surface.

  19. Techniques for Project Evaluation. A Selected Bibliography.

    ERIC Educational Resources Information Center

    Boston, Guy D.

    This annotated bibliography of documents discussing program evaluation methodologies was compiled in order to help federal, state, and local law enforcement agencies improve their evaluation activities. The three major categories include: (1) techniques and methodology for evaluation of criminal justice projects, (2) evaluation methods and…

  20. Evaluating Learning: An Action Project.

    ERIC Educational Resources Information Center

    Eiss, Albert F.

    This booklet was developed to help teachers evaluate student courses. It is not intended to provide the basis for a total evaluation, but as a "quick and easy" method to describe the basis of awarding student grades; identify ways in which to evaluate the effectiveness of a course of study; analyze the cognitive levels of test items; list the real…

  1. Benchmarking the evaluated proton differential cross sections suitable for the EBS analysis of natSi and 16O

    NASA Astrophysics Data System (ADS)

    Kokkoris, M.; Dede, S.; Kantre, K.; Lagoyannis, A.; Ntemou, E.; Paneta, V.; Preketes-Sigalas, K.; Provatas, G.; Vlastou, R.; Bogdanović-Radović, I.; Siketić, Z.; Obajdin, N.

    2017-08-01

    The evaluated proton differential cross sections suitable for the Elastic Backscattering Spectroscopy (EBS) analysis of natSi and 16O, as obtained from SigmaCalc 2.0, have been benchmarked over a wide energy and angular range at two different accelerator laboratories, namely at N.C.S.R. 'Demokritos', Athens, Greece and at Ruđer Bošković Institute (RBI), Zagreb, Croatia, using a variety of high-purity thick targets of known stoichiometry. The results are presented in graphical and tabular forms, while the observed discrepancies, as well as, the limits in accuracy of the benchmarking procedure, along with target related effects, are thoroughly discussed and analysed. In the case of oxygen the agreement between simulated and experimental spectra was generally good, while for silicon serious discrepancies were observed above Ep,lab = 2.5 MeV, suggesting that a further tuning of the appropriate nuclear model parameters in the evaluated differential cross-section datasets is required.

  2. An Evaluation of Project PLAN.

    ERIC Educational Resources Information Center

    Patterson, Eldon

    Project Plan, a computer managed individualized learning system developed by the Westinghouse Learning Corporation, was introduced into the St. Louis Public Schools under a Title III grant of the Elementary and Secondary Education Act. The program, offering individualized education in reading, language arts, mathematics, science, and social…

  3. Project HEED. Final Evaluation Report.

    ERIC Educational Resources Information Center

    Hughes, Orval D.

    During 1972-73, Project HEED (Heed Ethnic Educational Depolarization) involved 1,350 Indian students in 60 classrooms at Sells, Topowa, San Carlos, Rice, Many Farms, Hotevilla, Peach Springs, and Sacaton. Primary objectives were: (1) improvement in reading skills, (2) development of cultural awareness, and (3) providing for the Special Education…

  4. INTEGRAL BENCHMARK DATA FOR NUCLEAR DATA TESTING THROUGH THE ICSBEP AND THE NEWLY ORGANIZED IRPHEP

    SciTech Connect

    J. Blair Briggs; Lori Scott; Yolanda Rugama; Enrico Satori

    2007-04-01

    The status of the International Criticality Safety Benchmark Evaluation Project (ICSBEP) was last reported in a nuclear data conference at the International Conference on Nuclear Data for Science and Technology, ND-2004, in Santa Fe, New Mexico. Since that time the number and type of integral benchmarks have increased significantly. Included in the ICSBEP Handbook are criticality-alarm / shielding and fundamental physic benchmarks in addition to the traditional critical / subcritical benchmark data. Since ND 2004, a reactor physics counterpart to the ICSBEP, the International Reactor Physics Experiment Evaluation Project (IRPhEP) was initiated. The IRPhEP is patterned after the ICSBEP, but focuses on other integral measurements, such as buckling, spectral characteristics, reactivity effects, reactivity coefficients, kinetics measurements, reaction-rate and power distributions, nuclide compositions, and other miscellaneous-type measurements in addition to the critical configuration. The status of these two projects is discussed and selected benchmarks highlighted in this paper.

  5. Automatic benchmarking of homogenization packages applied to synthetic monthly series within the frame of the MULTITEST project

    NASA Astrophysics Data System (ADS)

    Guijarro, José A.; López, José A.; Aguilar, Enric; Domonkos, Peter; Venema, Victor; Sigró, Javier; Brunet, Manola

    2017-04-01

    After the successful inter-comparison of homogenization methods carried out in the COST Action ES0601 (HOME), many methods kept improving their algorithms, suggesting the need of performing new inter-comparison exercises. However, manual applications of the methodologies to a large number of testing networks cannot be afforded without involving the work of many researchers over an extended time. The alternative is to make the comparisons as automatic as possible, as in the MULTITEST project, which, funded by the Spanish Ministry of Economy and Competitiveness, tests homogenization methods by applying them to a large number of synthetic networks of monthly temperature and precipitation. One hundred networks of 10 series were sampled from different master networks containing 100 series of 720 values (60 years times 12 months). Three master temperature networks were built with different degree of cross-correlations between the series in order to simulate conditions of different station densities or climatic heterogeneity. Also three master synthetic networks were developed for precipitation, this time mimicking the characteristics of three different climates: Atlantic temperate, Mediterranean and monsoonal. Inhomogeneities were introduced in every network sampled from the master networks, and all publicly available homogenization methods that we could run in an automatic way were applied to them: ACMANT 3.0, Climatol 3.0, MASH 3.03, RHTestV4, USHCN v52d and HOMER 2.6. Most of them were tested with different settings, and their comparative results can be inspected in box-plot graphics of Root Mean Squared Errors and trend biases computed between the homogenized data and their original homogeneous series. In a first stage, inhomogeneities were applied to the synthetic homogeneous series with five different settings with increasing difficulty and realism: i) big shifts in half of the series; ii) the same with a strong seasonality; iii) short term platforms and local

  6. Evaluated benchmark experiments at critical assemblies simulating features of an HTHR at the ASTRA facility

    SciTech Connect

    Garin, V. P.; Glushkov, A. E.; Glushkov, E. S.; Gomin, E. A.; Gurevich, M. I.; Zimin, A. A.; Kompaniets, G. V.; Kukharkin, N. E.; Lobyntsev, V. A.; Nosov, V. I.; Polyakov, D. N.; Ponomarev-Stepnoi, N. N.; Smirnov, O. N.; Tel'kovskaya, O. V.; Chunyaev, E. I.

    2010-12-15

    The design of the ASTRA facility and critical assemblies that simulate physics features of modular high-temperature reactors (HTHR-Ms) with a graphite moderator and reflectors loaded with fuel particles having multilayer ceramic coatings is described in detail. Geometrical dimensions of the main elements and regions of the critical assemblies, composition of the materials used, and experimental results for various configurations of the critical assemblies are presented. A detailed computational benchmark model allowing for the structural and compositional features of the critical assembly configurations in question is developed on the basis of all the above data. The results are to be used for verification of the neutronics codes used for calculations of high-temperature helium-cooled reactors.

  7. Performance evaluation of firefly algorithm with variation in sorting for non-linear benchmark problems

    NASA Astrophysics Data System (ADS)

    Umbarkar, A. J.; Balande, U. T.; Seth, P. D.

    2017-06-01

    The field of nature inspired computing and optimization techniques have evolved to solve difficult optimization problems in diverse fields of engineering, science and technology. The firefly attraction process is mimicked in the algorithm for solving optimization problems. In Firefly Algorithm (FA) sorting of fireflies is done by using sorting algorithm. The original FA is proposed with bubble sort for ranking the fireflies. In this paper, the quick sort replaces bubble sort to decrease the time complexity of FA. The dataset used is unconstrained benchmark functions from CEC 2005 [22]. The comparison of FA using bubble sort and FA using quick sort is performed with respect to best, worst, mean, standard deviation, number of comparisons and execution time. The experimental result shows that FA using quick sort requires less number of comparisons but requires more execution time. The increased number of fireflies helps to converge into optimal solution whereas by varying dimension for algorithm performed better at a lower dimension than higher dimension.

  8. Evaluation of Pancreatic Cancer Clinical Trials and Benchmarks for Clinically Meaningful Future Trials: A Systematic Review.

    PubMed

    Rahib, Lola; Fleshman, Julie M; Matrisian, Lynn M; Berlin, Jordan D

    2016-09-01

    Progress in the treatment of pancreatic adenocarcinoma has been minimal; it remains the only major cancer type with a 5-year survival rate of less than 10%. To explore why a large proportion of advanced pancreatic cancer clinical trials executed over the past 25 years have had negative results and to identify benchmarks that could have predicted success. Phase 3 studies of patients with advanced pancreatic cancer were identified by searching clinicaltrials.gov and the scientific literature. Thirty-two phase 3 studies in 13 675 chemotherapy-naive patients resulted in 3 agents or combinations being considered clinically meaningful. Nineteen agents or combinations (70%) were tested in phase 2 trials preceding the phase 3 trial. In cases with paired phase 2 and 3 results, meeting the primary end point of the phase 2 trial predicted the outcome of the phase 3 trial 76% of the time but proceeded despite phase 2 negative results in 10 cases. We applied criteria for a clinically meaningful result identified by the American Society of Clinical Oncology (ASCO) Cancer Research Committee to these historical cases. Overall, progression-free and 1-year survival of experimental arms was compared with time period-controlled median values of control arms to normalize for the observed increase in response to gemcitabine over time. Applying the benchmark of a 50% improvement in overall survival as the primary end point to phase 2 data, or secondary end points of a 90% increase in 1-year survival or an 80% to 100% increase in progression-free survival, showed the greatest ability to predict a clinically meaningful phase 3 trial. Had these criteria been applied to these trials over the past 25 years, more than 11 571 patients enrolled in phase 3 trials that did not meet the primary end point could theoretically have been diverted to earlier-stage trials in an attempt to more rapidly advance the field.

  9. Benchmarking reference services: an introduction.

    PubMed

    Marshall, J G; Buchanan, H S

    1995-01-01

    Benchmarking is based on the common sense idea that someone else, either inside or outside of libraries, has found a better way of doing certain things and that your own library's performance can be improved by finding out how others do things and adopting the best practices you find. Benchmarking is one of the tools used for achieving continuous improvement in Total Quality Management (TQM) programs. Although benchmarking can be done on an informal basis, TQM puts considerable emphasis on formal data collection and performance measurement. Used to its full potential, benchmarking can provide a common measuring stick to evaluate process performance. This article introduces the general concept of benchmarking, linking it whenever possible to reference services in health sciences libraries. Data collection instruments that have potential application in benchmarking studies are discussed and the need to develop common measurement tools to facilitate benchmarking is emphasized.

  10. Benchmark Evaluation of Fuel Effect and Material Worth Measurements for a Beryllium-Reflected Space Reactor Mockup

    SciTech Connect

    Marshall, Margaret A.; Bess, John D.

    2015-02-01

    The critical configuration of the small, compact critical assembly (SCCA) experiments performed at the Oak Ridge Critical Experiments Facility (ORCEF) in 1962-1965 have been evaluated as acceptable benchmark experiments for inclusion in the International Handbook of Evaluated Criticality Safety Benchmark Experiments. The initial intent of these experiments was to support the design of the Medium Power Reactor Experiment (MPRE) program, whose purpose was to study “power plants for the production of electrical power in space vehicles.” The third configuration in this series of experiments was a beryllium-reflected assembly of stainless-steel-clad, highly enriched uranium (HEU)-O2 fuel mockup of a potassium-cooled space power reactor. Reactivity measurements cadmium ratio spectral measurements and fission rate measurements were measured through the core and top reflector. Fuel effect worth measurements and neutron moderating and absorbing material worths were also measured in the assembly fuel region. The cadmium ratios, fission rate, and worth measurements were evaluated for inclusion in the International Handbook of Evaluated Criticality Safety Benchmark Experiments. The fuel tube effect and neutron moderating and absorbing material worth measurements are the focus of this paper. Additionally, a measurement of the worth of potassium filling the core region was performed but has not yet been evaluated Pellets of 93.15 wt.% enriched uranium dioxide (UO2) were stacked in 30.48 cm tall stainless steel fuel tubes (0.3 cm tall end caps). Each fuel tube had 26 pellets with a total mass of 295.8 g UO2 per tube. 253 tubes were arranged in 1.506-cm triangular lattice. An additional 7-tube cluster critical configuration was also measured but not used for any physics measurements. The core was surrounded on all side by a beryllium reflector. The fuel effect worths were measured by removing fuel tubes at various radius. An accident scenario

  11. Toxicological Benchmarks for Wildlife

    SciTech Connect

    Sample, B.E. Opresko, D.M. Suter, G.W.

    1993-01-01

    Ecological risks of environmental contaminants are evaluated by using a two-tiered process. In the first tier, a screening assessment is performed where concentrations of contaminants in the environment are compared to no observed adverse effects level (NOAEL)-based toxicological benchmarks. These benchmarks represent concentrations of chemicals (i.e., concentrations presumed to be nonhazardous to the biota) in environmental media (water, sediment, soil, food, etc.). While exceedance of these benchmarks does not indicate any particular level or type of risk, concentrations below the benchmarks should not result in significant effects. In practice, when contaminant concentrations in food or water resources are less than these toxicological benchmarks, the contaminants may be excluded from further consideration. However, if the concentration of a contaminant exceeds a benchmark, that contaminant should be retained as a contaminant of potential concern (COPC) and investigated further. The second tier in ecological risk assessment, the baseline ecological risk assessment, may use toxicological benchmarks as part of a weight-of-evidence approach (Suter 1993). Under this approach, based toxicological benchmarks are one of several lines of evidence used to support or refute the presence of ecological effects. Other sources of evidence include media toxicity tests, surveys of biota (abundance and diversity), measures of contaminant body burdens, and biomarkers. This report presents NOAEL- and lowest observed adverse effects level (LOAEL)-based toxicological benchmarks for assessment of effects of 85 chemicals on 9 representative mammalian wildlife species (short-tailed shrew, little brown bat, meadow vole, white-footed mouse, cottontail rabbit, mink, red fox, and whitetail deer) or 11 avian wildlife species (American robin, rough-winged swallow, American woodcock, wild turkey, belted kingfisher, great blue heron, barred owl, barn owl, Cooper's hawk, and red-tailed hawk

  12. Evaluating success levels of mega-projects

    NASA Technical Reports Server (NTRS)

    Kumaraswamy, Mohan M.

    1994-01-01

    Today's mega-projects transcend the traditional trajectories traced within national and technological limitations. Powers unleashed by internationalization of initiatives, in for example space exploration and environmental protection, are arguably only temporarily suppressed by narrower national, economic, and professional disagreements as to how best they should be harnessed. While the world gets its act together there is time to develop the technologies of such supra-mega-project management that will synergize truly diverse resources and smoothly mesh their interfaces. Such mega-projects and their management need to be realistically evaluated, when implementing such improvements. This paper examines current approaches to evaluating mega-projects and questions the validity of extrapolations to the supra-mega-projects of the future. Alternatives to improve such evaluations are proposed and described.

  13. Evaluating a Project on Roma Education

    ERIC Educational Resources Information Center

    Georgiadis, Fokion; Nikolajevic, Dragana; van Driel, Barry

    2011-01-01

    This research note is based on the evaluation of the Comenius project Teacher-IN-SErvice-Training-for-Roma-inclusion ("INSETRom"). The project represented an international effort that was undertaken to bridge the gap between Roma and non-Roma communities and to improve the educational attainment of Roma children in the mainstream…

  14. Evaluation Report: The Home Visit Project.

    ERIC Educational Resources Information Center

    Cowan, Geni; Bobby, Kim; St. Roseman, Paul; Echandia, Adriana

    This paper summarizes the "Year 3 Evaluation" of the Home Visit Project currently operating in the Sacramento City Unified School District. The project's purpose is to strengthen the relationship between home and school to enhance and enrich student-teacher interactions and relationships, to improve the frequency and quality of…

  15. Evaluation of the Law Focus Curriculum Project.

    ERIC Educational Resources Information Center

    Watson, Patricia J.; Workman, Eva Mae

    1974-01-01

    This evaluation of the Law Focused Curriculum Project of the Oklahoma Public Schools analyzes the human and nonhuman resources utilized in the project, and the nature and extent of activities. The first part of the document examines the program and its objectives. School-age citizens are to become acquainted with the law, the functions and…

  16. Evaluating a Project on Roma Education

    ERIC Educational Resources Information Center

    Georgiadis, Fokion; Nikolajevic, Dragana; van Driel, Barry

    2011-01-01

    This research note is based on the evaluation of the Comenius project Teacher-IN-SErvice-Training-for-Roma-inclusion ("INSETRom"). The project represented an international effort that was undertaken to bridge the gap between Roma and non-Roma communities and to improve the educational attainment of Roma children in the mainstream…

  17. 5 CFR 470.317 - Project evaluation.

    Code of Federal Regulations, 2013 CFR

    2013-01-01

    ... 5 Administrative Personnel 1 2013-01-01 2013-01-01 false Project evaluation. 470.317 Section 470.317 Administrative Personnel OFFICE OF PERSONNEL MANAGEMENT CIVIL SERVICE REGULATIONS PERSONNEL MANAGEMENT RESEARCH PROGRAMS AND DEMONSTRATIONS PROJECTS Regulatory Requirements Pertaining to...

  18. 5 CFR 470.317 - Project evaluation.

    Code of Federal Regulations, 2011 CFR

    2011-01-01

    ... 5 Administrative Personnel 1 2011-01-01 2011-01-01 false Project evaluation. 470.317 Section 470.317 Administrative Personnel OFFICE OF PERSONNEL MANAGEMENT CIVIL SERVICE REGULATIONS PERSONNEL MANAGEMENT RESEARCH PROGRAMS AND DEMONSTRATIONS PROJECTS Regulatory Requirements Pertaining to...

  19. 5 CFR 470.317 - Project evaluation.

    Code of Federal Regulations, 2010 CFR

    2010-01-01

    ... 5 Administrative Personnel 1 2010-01-01 2010-01-01 false Project evaluation. 470.317 Section 470.317 Administrative Personnel OFFICE OF PERSONNEL MANAGEMENT CIVIL SERVICE REGULATIONS PERSONNEL MANAGEMENT RESEARCH PROGRAMS AND DEMONSTRATIONS PROJECTS Regulatory Requirements Pertaining to...

  20. Evaluation of potential factors affecting deriving conductivity benchmark by utilizing weighting methods in Hun-Tai River Basin, Northeastern China.

    PubMed

    Jia, Xiaobo; Zhao, Qian; Guo, Fen; Ma, Shuqin; Zhang, Yuan; Zang, Xiaomiao

    2017-03-01

    Specific conductivity is an increasingly important stressor for freshwater ecosystems. Interacting with other environmental factors, it may lead to habitat degradation and biodiversity loss. However, it is still poorly understood how the effect of specific conductivity on freshwater organisms is confounded by other environmental factors. In this study, a weight-of-evidence method was applied to evaluate the potential environmental factors that may confound the effect of specific conductivity on macroinvertebrate structure communities and identify the confounders affecting deriving conductivity benchmark in Hun-Tai River Basin, China. A total of seven potential environmental factors were assessed by six types of evidence (i.e., correlation of cause and confounder, correlation of effect and confounder, the contingency of high level cause and confounder, the removal of confounder, levels of confounder known to cause effects, and multivariate statistics for confounding). Results showed that effects of dissolved oxygen (DO), fecal coliform, habitat score, total phosphorus (TP), pH, and temperature on the relationship between sensitive genera loss and specific conductivity were minimal and manageable. NH3-N was identified as a confounder affecting deriving conductivity benchmark for macroinvertebrate. The potential confounding by high NH3-N was minimized by removing sites with NH3-N > 2.0 mg/L from the data set. Our study tailored the weighting method previously developed by USEPA to use field data to develop causal relationships for basin-scale applications and may provide useful information for pollution remediation and natural resource management.

  1. Monitoring and Evaluating Nonpoint Source Watershed Projects

    EPA Pesticide Factsheets

    This guide is written primarily for those who develop and implement monitoring plans for watershed management projects. it can also be used evaluate the technical merits of monitoring proposals they might sponsor. It is an update to the 1997 Guide.

  2. A benchmarking project on the quality of previous guidelines about the management of malignant pleural effusion from the European Society of Thoracic Surgeons (ESTS) Pleural Diseases Working Group.

    PubMed

    Bertolaccini, Luca; Bedetti, Benedetta; Brunelli, Alessandro; Marinova, Katerina; Raveglia, Federico; Rocco, Gaetano; Shargall, Yaron; Solli, Piergiorgio; Varela, Gonzalo; Papagiannopoulos, Kostas; Kuzdzal, Jaroslaw; Massard, Gilbert; Ruffini, Enrico; Falcoz, Pierre-Emmanuel; Martinez-Barenys, Carlos; Opitz, Isabelle; Batirel, Hasan F; Toker, Alper; Scarci, Marco

    2017-08-01

    In the European Society of Thoracic Surgeons (ESTS) survey about management of malignant pleural effusions (MPE), 56% of respondents are not informed of any relevant clinical guidelines and 52%, who are aware of the existence of guidelines, declared that they are in need of updating or revision. The ESTS Pleural Diseases Working Group developed a benchmarking project on quality of previous guidelines on the management of MPE. The Appraisal of Guidelines for Research and Evaluation (AGREE) II instrument was used to assess each guideline. Each item was scored on a 7-point scale. Scores for each domain were calculated. Economic data for the nations which have issued the guidelines were collected from the Organisation for Economic Cooperation and Development health statistics database. Six guidelines fitted the inclusion criteria and were assessed. Five out of 6 guidelines were produced by a multinational collaboration. Observers would recommend only 2 guidelines with minimal modification. Two areas that received the best score were clarity of presentation and scope and purpose (objectives and health questions target population). The applicability of guideline domain had the lowest score. Multivariate analysis demonstrated that clarity of presentation, international guidelines and publication through medical journal were related to improved scores. A strong correlation was observed between the measures of economic status. The quality of guidelines assessed by the AGREE II criteria was found to be extremely variable. Guidelines achieving higher AGREE II scores were more likely to come from the European Union with the direct involvement of scientific societies in their development. It was also recognized that some fundamental unanswered questions remain about the management of MPE.

  3. The integrated hydrologic model intercomparison project, IH-MIP2: A second set of benchmark results to diagnose integrated hydrology and feedbacks

    NASA Astrophysics Data System (ADS)

    Kollet, Stefan; Sulis, Mauro; Maxwell, Reed M.; Paniconi, Claudio; Putti, Mario; Bertoldi, Giacomo; Coon, Ethan T.; Cordano, Emanuele; Endrizzi, Stefano; Kikinzon, Evgeny; Mouche, Emmanuel; Mügler, Claude; Park, Young-Jin; Refsgaard, Jens C.; Stisen, Simon; Sudicky, Edward

    2017-01-01

    Emphasizing the physical intricacies of integrated hydrology and feedbacks in simulating connected, variably saturated groundwater-surface water systems, the Integrated Hydrologic Model Intercomparison Project initiated a second phase (IH-MIP2), increasing the complexity of the benchmarks of the first phase. The models that took part in the intercomparison were ATS, Cast3M, CATHY, GEOtop, HydroGeoSphere, MIKE-SHE, and ParFlow. IH-MIP2 benchmarks included a tilted v-catchment with 3-D subsurface; a superslab case expanding the slab case of the first phase with an additional horizontal subsurface heterogeneity; and the Borden field rainfall-runoff experiment. The analyses encompassed time series of saturated, unsaturated, and ponded storages, as well as discharge. Vertical cross sections and profiles were also inspected in the superslab and Borden benchmarks. An analysis of agreement was performed including systematic and unsystematic deviations between the different models. Results show generally good agreement between the different models, which lends confidence in the fundamental physical and numerical implementation of the governing equations in the different models. Differences can be attributed to the varying level of detail in the mathematical and numerical representation or in the parameterization of physical processes, in particular with regard to ponded storage and friction slope in the calculation of overland flow. These differences may become important for specific applications such as detailed inundation modeling or when strong inhomogeneities are present in the simulation domain.

  4. Current Reactor Physics Benchmark Activities at the Idaho National Laboratory

    SciTech Connect

    John D. Bess; Margaret A. Marshall; Mackenzie L. Gorham; Joseph Christensen; James C. Turnbull; Kim Clark

    2011-11-01

    The International Reactor Physics Experiment Evaluation Project (IRPhEP) [1] and the International Criticality Safety Benchmark Evaluation Project (ICSBEP) [2] were established to preserve integral reactor physics and criticality experiment data for present and future research. These valuable assets provide the basis for recording, developing, and validating our integral nuclear data, and experimental and computational methods. These projects are managed through the Idaho National Laboratory (INL) and the Organisation for Economic Co-operation and Development Nuclear Energy Agency (OECD-NEA). Staff and students at the Department of Energy - Idaho (DOE-ID) and INL are engaged in the development of benchmarks to support ongoing research activities. These benchmarks include reactors or assemblies that support Next Generation Nuclear Plant (NGNP) research, space nuclear Fission Surface Power System (FSPS) design validation, and currently operational facilities in Southeastern Idaho.

  5. In response to an open invitation for comments on AAAS project 2061's Benchmark books on science. Part 1: documentation of serious errors in cell biology.

    PubMed

    Ling, Gilbert

    2006-01-01

    Project 2061 was founded by the American Association for the Advancement of Science (AAAS) to improve secondary school science education. An in-depth study of ten 9 to 12th grade biology textbooks led to the verdict that none conveyed "Big Ideas" that would give coherence and meaning to the profusion of lavishly illustrated isolated details. However, neither the Project report itself nor the Benchmark books put out earlier by the Project carries what deserves the designation of "Big Ideas." Worse, in the two earliest-published Benchmark books, the basic unit of all life forms--the living cell--is described as a soup enclosed by a cell membrane, that determines what can enter or leave the cell. This is astonishing since extensive experimental evidence has unequivocally disproved this idea 60 years ago. A "new" version of the membrane theory brought in to replace the discredited (sieve) version is the pump model--currently taught as established truth in all high-school and college biology textbooks--was also unequivocally disproved 40 years ago. This comment is written partly in response to Bechmark's gracious open invitation for ideas to improve the books and through them, to improve US secondary school science education.

  6. Residual activity evaluation: a benchmark between ANITA, FISPACT, FLUKA and PHITS codes

    NASA Astrophysics Data System (ADS)

    Firpo, Gabriele; Viberti, Carlo Maria; Ferrari, Anna; Frisoni, Manuela

    2017-09-01

    The activity of residual nuclides dictates the radiation fields in periodic inspections/repairs (maintenance periods) and dismantling operations (decommissioning phase) of accelerator facilities (i.e., medical, industrial, research) and nuclear reactors. Therefore, the correct prediction of the material activation allows for a more accurate planning of the activities, in line with the ALARA (As Low As Reasonably Achievable) principles. The scope of the present work is to show the results of a comparison between residual total specific activity versus a set of cooling time instants (from zero up to 10 years after irradiation) as obtained by two analytical (FISPACT and ANITA) and two Monte Carlo (FLUKA and PHITS) codes, making use of their default nuclear data libraries. A set of 40 irradiating scenarios is considered, i.e. neutron and proton particles of different energies, ranging from zero to many hundreds MeV, impinging on pure elements or materials of standard composition typically used in industrial applications (namely, AISI SS316 and Portland concrete). In some cases, experimental results were also available for a more thorough benchmark.

  7. AID: A Benchmark Data Set for Performance Evaluation of Aerial Scene Classification

    NASA Astrophysics Data System (ADS)

    Xia, Gui-Song; Hu, Jingwen; Hu, Fan; Shi, Baoguang; Bai, Xiang; Zhong, Yanfei; Zhang, Liangpei; Lu, Xiaoqiang

    2017-07-01

    Aerial scene classification, which aims to automatically label an aerial image with a specific semantic category, is a fundamental problem for understanding high-resolution remote sensing imagery. In recent years, it has become an active task in remote sensing area and numerous algorithms have been proposed for this task, including many machine learning and data-driven approaches. However, the existing datasets for aerial scene classification like UC-Merced dataset and WHU-RS19 are with relatively small sizes, and the results on them are already saturated. This largely limits the development of scene classification algorithms. This paper describes the Aerial Image Dataset (AID): a large-scale dataset for aerial scene classification. The goal of AID is to advance the state-of-the-arts in scene classification of remote sensing images. For creating AID, we collect and annotate more than ten thousands aerial scene images. In addition, a comprehensive review of the existing aerial scene classification techniques as well as recent widely-used deep learning methods is given. Finally, we provide a performance analysis of typical aerial scene classification and deep learning approaches on AID, which can be served as the baseline results on this benchmark.

  8. Teaching Medical Students at a Distance: Using Distance Learning Benchmarks to Plan and Evaluate a Web-Enhanced Medical Student Curriculum

    ERIC Educational Resources Information Center

    Olney, Cynthia A.; Chumley, Heidi; Parra, Juan M.

    2004-01-01

    A team designing a Web-enhanced third-year medical education didactic curriculum based their course planning and evaluation activities on the Institute for Higher Education Policy's (2000) 24 benchmarks for online distance learning. The authors present the team's blueprint for planning and evaluating the Web-enhanced curriculum, which incorporates…

  9. The DLESE Evaluation Core Services Project

    NASA Astrophysics Data System (ADS)

    Buhr, S. M.; Barker, L. J.; Reeves, T. C.

    2003-12-01

    The DLESE Evaluation Core Service project will conduct evaluation of DLESE and provide evaluation consultation, resources and services to the DLESE community. Through this work we anticipate that we will learn more about the impact and use of digital libraries, and will promote an evaluation mindset within the geoscience education community. Activities of the DLESE Evaluation Service team include 1) evaluation planning for and of DLESE, 2) conducting formative evaluation of DLESE (user needs, data access, collections, outreach), 3) conducting classroom evaluation of DLESE use on teaching practices and learning outcomes, and 4) collection, synthesis, and reporting of evaluation findings garnered from all core teams and major projects. Many opportunities for community involvement exist. A strand group convened during the 2004 DLESE Annual Meeting took DLESE Evaluation as their topic, provided recommendations and will continue their activities through the year. The related Evaluation Toolkit collection is now discoverable through DLESE, and upcoming activities of all the core teams will provide evaluation opportunities to be described. Other community opportunities include consulting with Evaluation Service for education grant proposals, attending an evaluation workshop,and applying for an Evaluation Minigrant (up to \\$5K per award) Progress to date will be discussed, the Evaluation Core Services team members will be introduced, and plans and opportunities will be described in more detail.

  10. International decay data evaluation project

    SciTech Connect

    Helmer, R.G.

    1996-10-01

    Basic concepts of, and information from, radionuclide decay are used in many applications. The author limits this discussion to the data needed for applied {gamma}-ray spectrometry; this includes applications such as nuclide identification and quantitative assay. Many of these applications require a knowledge of half-lives and radiation energies and emission probabilities. For over 50 years, people have compiled and evaluated measured data with the goal of obtaining the best values of these quantities. This has resulted in numerous sets of recommended values, many of which still have scientific, historical, or national reasons for existing. These sets show varying degrees of agreement and disagreement in the quoted values and varying time lags in incorporating new and improved experimental results. A new informational international group has been formed to carry out evaluations for radionuclides of importance in applications; it is expected that the results will become an authoritative and widely accepted set of decay data.

  11. EVALUATION OF U10MO FUEL PLATE IRRADIATION BEHAVIOR VIA NUMERICAL AND EXPERIMENTAL BENCHMARKING

    SciTech Connect

    Samuel J. Miller; Hakan Ozaltun

    2012-11-01

    This article analyzes dimensional changes due to irradiation of monolithic plate-type nuclear fuel and compares results with finite element analysis of the plates during fabrication and irradiation. Monolithic fuel plates tested in the Advanced Test Reactor (ATR) at Idaho National Lab (INL) are being used to benchmark proposed fuel performance for several high power research reactors. Post-irradiation metallographic images of plates sectioned at the midpoint were analyzed to determine dimensional changes of the fuel and the cladding response. A constitutive model of the fabrication process and irradiation behavior of the tested plates was developed using the general purpose commercial finite element analysis package, Abaqus. Using calculated burn-up profiles of irradiated plates to model the power distribution and including irradiation behaviors such as swelling and irradiation enhanced creep, model simulations allow analysis of plate parameters that are either impossible or infeasible in an experimental setting. The development and progression of fabrication induced stress concentrations at the plate edges was of primary interest, as these locations have a unique stress profile during irradiation. Additionally, comparison between 2D and 3D models was performed to optimize analysis methodology. In particular, the ability of 2D and 3D models account for out of plane stresses which result in 3-dimensional creep behavior that is a product of these components. Results show that assumptions made in 2D models for the out-of-plane stresses and strains cannot capture the 3-dimensional physics accurately and thus 2D approximations are not computationally accurate. Stress-strain fields are dependent on plate geometry and irradiation conditions, thus, if stress based criteria is used to predict plate behavior (as opposed to material impurities, fine micro-structural defects, or sharp power gradients), unique 3D finite element formulation for each plate is required.

  12. An Evaluation of Benchmarks for Esthetic Orientation of the Occlusal Plane.

    PubMed

    Priest, George; Wilson, Michael G

    2017-04-01

    This patient study was designed to measure the validity of both the horizontal and Camper's planes, which are used as benchmarks to reestablish the sagittal orientation of the occlusal plane angles in dental rehabilitation. Profile digital photographs were made of the first 100 consenting dentate patients as they closed on an occlusal plane analyzer while maintaining natural head posture. Using a digital screen protractor, three angles were measured: the occlusal plane angle relative to the horizontal plane, and the angle between the occlusal plane and Camper's plane from both the superior and inferior borders of the tragus of the ear. The angle between the occlusal plane and the horizontal reference plane for the 100 patients ranged from -8.72° to +18.08° (mean +3.25°); the angle between the occlusal plane and Camper's plane, from the superior border of the tragus to the ala of the nose demonstrated a range from -8.49° to +15.16° (mean +3.03°); and the angle between Camper's plane, from the inferior border of the tragus to the ala of the nose and the occlusal plane demonstrated a range from -15.57° to +9.15° (mean -4.09°). Occlusal plane angles measured in this patient population with natural dentitions demonstrated a relatively small mean deviation from both the horizontal plane and Camper's plane when using the superior border of the tragus as the distal reference point, but the range was quite broad and could result in unacceptable occlusal plane angles in many patients undergoing dental rehabilitation. Both Camper's plane and the horizontal reference plane may be acceptable initial reference planes for oral rehabilitation, but additional anatomic and esthetic parameters are required for verification of an esthetically pleasing occlusal plane angle. © 2016 by the American College of Prosthodontists.

  13. Benchmarking techniques for evaluation of compression transform performance in ATR applications

    NASA Astrophysics Data System (ADS)

    Schmalz, Mark S.

    2004-10-01

    Image compression is increasingly employed in applications such as medical imaging, for reducing data storage requirement, and Internet video transmission, to effectively increase channel bandwidth. Similarly, military applications such as automated target recognition (ATR) often employ compression to achieve storage and communication efficiencies, particularly to enhance the effective bandwidth of communication channels whose throughput suffers, for example, from overhead due to error correction/detection or encryption. In the majority of cases, lossy compression is employed due the resultant low bit rates (high compression ratio). However, lossy compression produces artifacts in decompressed imagery that can confound ATR processes applied to such imagery, thereby reducing the probability of detection (Pd) and possibly increasing the rate or number of false alarms (Rfa or Nfa). In this paper, the authors' previous research in performance measurement of compression transforms is extended to include (a) benchmarking algorithms and software tools, (b) a suite of error exemplars that are designed to elicit compression transform behavior in an operationally relevant context, and (c) a posteriori analysis of performance data. The following transforms are applied to a suite of 64 error exemplars: Visual Pattern Image Coding (VPIC [1]), Vector Quantization with a fast codebook search algorithm (VQ [2,3]), JPEG and a preliminary implementation of JPEG 2000 [4,5], and EBLAST [6-8]. Compression ratios range from 2:1 to 200:1, and various noise levels and types are added to the error exemplars to produce a database of 7,680 synthetic test images. Several global and local (e.g., featural) distortion measures are applied to the decompressed test imagery to provide a basis for rate-distortion and rate-performance analysis as a function of noise and compression transform type.

  14. Toward a benchmarking data set able to evaluate ligand- and structure-based virtual screening using public HTS data.

    PubMed

    Lindh, Martin; Svensson, Fredrik; Schaal, Wesley; Zhang, Jin; Sköld, Christian; Brandt, Peter; Karlén, Anders

    2015-02-23

    Virtual screening has the potential to accelerate and reduce costs of probe development and drug discovery. To develop and benchmark virtual screening methods, validation data sets are commonly used. Over the years, such data sets have been constructed to overcome the problems of analogue bias and artificial enrichment. With the rapid growth of public domain databases containing high-throughput screening data, such as the PubChem BioAssay database, there is an increased possibility to use such data for validation. In this study, we identify PubChem data sets suitable for validation of both structure- and ligand-based virtual screening methods. To achieve this, high-throughput screening data for which a crystal structure of the bioassay target was available in the PDB were identified. Thereafter, the data sets were inspected to identify structures and data suitable for use in validation studies. In this work, we present seven data sets (MMP13, DUSP3, PTPN22, EPHX2, CTDSP1, MAPK10, and CDK5) compiled using this method. In the seven data sets, the number of active compounds varies between 19 and 369 and the number of inactive compounds between 59 405 and 337 634. This gives a higher ratio of the number of inactive to active compounds than what is found in most benchmark data sets. We have also evaluated the screening performance using docking and 3D shape similarity with default settings. To characterize the data sets, we used physicochemical similarity and 2D fingerprint searches. We envision that these data sets can be a useful complement to current data sets used for method evaluation.

  15. Project SAVE: Evaluation of Pilot Test Results

    ERIC Educational Resources Information Center

    Bell, Mary Lou; Bliss, Kappie

    The long-term goal of Project SAVE (Stop Alcohol Violations Early) is to reduce underage drinking. When a major revision of the program was initiated, the pilot program was evaluated for statistically measurable changes against short-term goals. The results of that evaluation are presented here. Four elements were included in the evaluation…

  16. Risk variables in evaluation of transport projects

    NASA Astrophysics Data System (ADS)

    Vařbuchta, Petr; Kovářová, Hana; Hromádka, Vít; Vítková, Eva

    2017-09-01

    Depending on the constantly increasing demands on assessment of investment projects, especially assessment of large-scale projects in transport and important European projects with wide impacts, there is constantly increasing focus on risk management, whether to find mitigations, creating corrective measures or their implementation in assessment, especially in the context of Cost-Benefit analysis. To project assessment is often used implementation of certain risk variables, which can generate negative impacts of project outputs in framework of assess. Especially in case of transportation infrastructure projects is taken much emphasis on the influence of risk variables. However, currently in case of assessment of transportation projects is in Czech Republic used a few risk variables, which occur in the most projects. This leads to certain limitation in framework of impact assessment of risk variables. This papers aims to specify a new risk variables and process of applying them to already executed project assessment. Based on changes generated by new risk variables will be evaluated differences between original and adapted assessment.

  17. Stakeholder approach for evaluating organizational change projects.

    PubMed

    Peltokorpi, Antti; Alho, Antti; Kujala, Jaakko; Aitamurto, Johanna; Parvinen, Petri

    2008-01-01

    This paper aims to create a model for evaluating organizational change initiatives from a stakeholder resistance viewpoint. The paper presents a model to evaluate change projects and their expected benefits. Factors affecting the challenge to implement change were defined based on stakeholder theory literature. The authors test the model's practical validity for screening change initiatives to improve operating room productivity. Change initiatives can be evaluated using six factors: the effect of the planned intervention on stakeholders' actions and position; stakeholders' capability to influence the project's implementation; motivation to participate; capability to change; change complexity; and management capability. The presented model's generalizability should be explored by filtering presented factors through a larger number of historical cases operating in different healthcare contexts. The link between stakeholders, the change challenge and the outcomes of change projects needs to be empirically tested. The proposed model can be used to prioritize change projects, manage stakeholder resistance and establish a better organizational and professional competence for managing healthcare organization change projects. New insights into existing stakeholder-related understanding of change project successes are provided.

  18. Training Evaluation Based on Cases of Taiwanese Benchmarked High-Tech Companies

    ERIC Educational Resources Information Center

    Lien, Bella Ya Hui; Hung, Richard Yu Yuan; McLean, Gary N.

    2007-01-01

    Although the influence of workplace practices and employees' experiences with training effectiveness has received considerable attention, less is known of the influence of workplace practices on training evaluation methods. The purposes of this study were to: (1) explore and understand the training evaluation methods used by seven Taiwanese…

  19. Strategic evaluation central to LNG project formation

    SciTech Connect

    Nissen, D.; DiNapoli, R.N.; Yost, C.C.

    1995-07-03

    An efficient-scale, grassroots LNG facility of about 6 million metric tons/year capacity requires a prestart-up outlay of $5 billion or more for the supply facilities--production, feedgas pipeline, liquefaction, and shipping. The demand side of the LNG chain requires a similar outlay, counting the import-regasification terminal and a combination of 5 gigawatts or more of electric power generation or the equivalent in city gas and industrial gas-using facilities. There exist no well-developed commodity markets for free-on-board (fob) or delivered LNG. A new LNG supply project is dedicated to its buyers. Indeed, the buyers` revenue commitment is the project`s only bankable asset. For the buyer to make this commitment, the supply venture`s capability and commitment must be credible: to complete the project and to deliver the LNG reliably over the 20+ years required to recover capital committed on both sides. This requirement has technical, economic, and business dimensions. In this article the authors describe a LNG project evaluation system and show its application to typical tasks: project cost of service and participant shares; LNG project competition; alternative project structures; and market competition for LNG-supplied electric power generation.

  20. HPC Analytics Support. Requirements for Uncertainty Quantification Benchmarks

    SciTech Connect

    Paulson, Patrick R.; Purohit, Sumit; Rodriguez, Luke R.

    2015-05-01

    This report outlines techniques for extending benchmark generation products so they support uncertainty quantification by benchmarked systems. We describe how uncertainty quantification requirements can be presented to candidate analytical tools supporting SPARQL. We describe benchmark data sets for evaluating uncertainty quantification, as well as an approach for using our benchmark generator to produce data sets for generating benchmark data sets.

  1. Self-benchmarking Guide for Data Centers: Metrics, Benchmarks, Actions

    SciTech Connect

    Mathew, Paul; Ganguly, Srirupa; Greenberg, Steve; Sartor, Dale

    2009-07-13

    This guide describes energy efficiency metrics and benchmarks that can be used to track the performance of and identify potential opportunities to reduce energy use in data centers. This guide is primarily intended for personnel who have responsibility for managing energy use in existing data centers - including facilities managers, energy managers, and their engineering consultants. Additionally, data center designers may also use the metrics and benchmarks described in this guide for goal-setting in new construction or major renovation. This guide provides the following information: (1) A step-by-step outline of the benchmarking process. (2) A set of performance metrics for the whole building as well as individual systems. For each metric, the guide provides a definition, performance benchmarks, and potential actions that can be inferred from evaluating this metric. (3) A list and descriptions of the data required for computing the metrics. This guide is complemented by spreadsheet templates for data collection and for computing the benchmarking metrics. This guide builds on prior data center benchmarking studies supported by the California Energy Commission. Much of the benchmarking data are drawn from the LBNL data center benchmarking database that was developed from these studies. Additional benchmark data were obtained from engineering experts including facility designers and energy managers. This guide also builds on recent research supported by the U.S. Department of Energy's Save Energy Now program.

  2. HANFORD DST THERMAL & SEISMIC PROJECT ANSYS BENCHMARK ANALYSIS OF SEISMIC INDUCED FLUID STRUCTURE INTERACTION IN A HANFORD DOUBLE SHELL PRIMARY TANK

    SciTech Connect

    MACKEY, T.C.

    2006-03-14

    M&D Professional Services, Inc. (M&D) is under subcontract to Pacific Northwest National Laboratories (PNNL) to perform seismic analysis of the Hanford Site Double-Shell Tanks (DSTs) in support of a project entitled ''Double-Shell Tank (DSV Integrity Project-DST Thermal and Seismic Analyses)''. The overall scope of the project is to complete an up-to-date comprehensive analysis of record of the DST System at Hanford in support of Tri-Party Agreement Milestone M-48-14. The work described herein was performed in support of the seismic analysis of the DSTs. The thermal and operating loads analysis of the DSTs is documented in Rinker et al. (2004). The overall seismic analysis of the DSTs is being performed with the general-purpose finite element code ANSYS. The overall model used for the seismic analysis of the DSTs includes the DST structure, the contained waste, and the surrounding soil. The seismic analysis of the DSTs must address the fluid-structure interaction behavior and sloshing response of the primary tank and contained liquid. ANSYS has demonstrated capabilities for structural analysis, but the capabilities and limitations of ANSYS to perform fluid-structure interaction are less well understood. The purpose of this study is to demonstrate the capabilities and investigate the limitations of ANSYS for performing a fluid-structure interaction analysis of the primary tank and contained waste. To this end, the ANSYS solutions are benchmarked against theoretical solutions appearing in BNL 1995, when such theoretical solutions exist. When theoretical solutions were not available, comparisons were made to theoretical solutions of similar problems and to the results from Dytran simulations. The capabilities and limitations of the finite element code Dytran for performing a fluid-structure interaction analysis of the primary tank and contained waste were explored in a parallel investigation (Abatt 2006). In conjunction with the results of the global ANSYS analysis

  3. Medico-economic evaluation of healthcare products. Methodology for defining a significant impact on French health insurance costs and selection of benchmarks for interpreting results.

    PubMed

    Dervaux, Benoît; Baseilhac, Eric; Fagon, Jean-Yves; Biot, Claire; Blachier, Corinne; Braun, Eric; Debroucker, Frédérique; Detournay, Bruno; Ferretti, Carine; Granger, Muriel; Jouan-Flahault, Chrystel; Lussier, Marie-Dominique; Meyer, Arlette; Muller, Sophie; Pigeon, Martine; De Sahb, Rima; Sannié, Thomas; Sapède, Claudine; Vray, Muriel

    2014-01-01

    Decree No. 2012-1116 of 2 October 2012 on medico-economic assignments of the French National Authority for Health (Haute autorité de santé, HAS) significantly alters the conditions for accessing the health products market in France. This paper presents a theoretical framework for interpreting the results of the economic evaluation of health technologies and summarises the facts available in France for developing benchmarks that will be used to interpret incremental cost-effectiveness ratios. This literature review shows that it is difficult to determine a threshold value but it is also difficult to interpret then incremental cost effectiveness ratio (ICER) results without a threshold value. In this context, round table participants favour a pragmatic approach based on "benchmarks" as opposed to a threshold value, based on an interpretative and normative perspective, i.e. benchmarks that can change over time based on feedback.

  4. Evaluation of various LandFlux evapotranspiration algorithms using the LandFlux-EVAL synthesis benchmark products and observational data

    NASA Astrophysics Data System (ADS)

    Michel, Dominik; Hirschi, Martin; Jimenez, Carlos; McCabe, Mathew; Miralles, Diego; Wood, Eric; Seneviratne, Sonia

    2014-05-01

    Research on climate variations and the development of predictive capabilities largely rely on globally available reference data series of the different components of the energy and water cycles. Several efforts aimed at producing large-scale and long-term reference data sets of these components, e.g. based on in situ observations and remote sensing, in order to allow for diagnostic analyses of the drivers of temporal variations in the climate system. Evapotranspiration (ET) is an essential component of the energy and water cycle, which can not be monitored directly on a global scale by remote sensing techniques. In recent years, several global multi-year ET data sets have been derived from remote sensing-based estimates, observation-driven land surface model simulations or atmospheric reanalyses. The LandFlux-EVAL initiative presented an ensemble-evaluation of these data sets over the time periods 1989-1995 and 1989-2005 (Mueller et al. 2013). Currently, a multi-decadal global reference heat flux data set for ET at the land surface is being developed within the LandFlux initiative of the Global Energy and Water Cycle Experiment (GEWEX). This LandFlux v0 ET data set comprises four ET algorithms forced with a common radiation and surface meteorology. In order to estimate the agreement of this LandFlux v0 ET data with existing data sets, it is compared to the recently available LandFlux-EVAL synthesis benchmark product. Additional evaluation of the LandFlux v0 ET data set is based on a comparison to in situ observations of a weighing lysimeter from the hydrological research site Rietholzbach in Switzerland. These analyses serve as a test bed for similar evaluation procedures that are envisaged for ESA's WACMOS-ET initiative (http://wacmoset.estellus.eu). Reference: Mueller, B., Hirschi, M., Jimenez, C., Ciais, P., Dirmeyer, P. A., Dolman, A. J., Fisher, J. B., Jung, M., Ludwig, F., Maignan, F., Miralles, D. G., McCabe, M. F., Reichstein, M., Sheffield, J., Wang, K

  5. 23 CFR 505.11 - Project evaluation and rating.

    Code of Federal Regulations, 2011 CFR

    2011-04-01

    ... preliminary rating and evaluation at any point in the project development after the project's concept plan is... 23 Highways 1 2011-04-01 2011-04-01 false Project evaluation and rating. 505.11 Section 505.11... MANAGEMENT PROJECTS OF NATIONAL AND REGIONAL SIGNIFICANCE EVALUATION AND RATING § 505.11 Project...

  6. A comparative evaluation of risk-adjustment models for benchmarking amputation-free survival after lower extremity bypass.

    PubMed

    Simons, Jessica P; Goodney, Philip P; Flahive, Julie; Hoel, Andrew W; Hallett, John W; Kraiss, Larry W; Schanzer, Andres

    2016-04-01

    Providing patients and payers with publicly reported risk-adjusted quality metrics for the purpose of benchmarking physicians and institutions has become a national priority. Several prediction models have been developed to estimate outcomes after lower extremity revascularization for critical limb ischemia, but the optimal model to use in contemporary practice has not been defined. We sought to identify the highest-performing risk-adjustment model for amputation-free survival (AFS) at 1 year after lower extremity bypass (LEB). We used the national Society for Vascular Surgery Vascular Quality Initiative (VQI) database (2003-2012) to assess the performance of three previously validated risk-adjustment models for AFS. The Bypass versus Angioplasty in Severe Ischaemia of the Leg (BASIL), Finland National Vascular (FINNVASC) registry, and the modified Project of Ex-vivo vein graft Engineering via Transfection III (PREVENT III [mPIII]) risk scores were applied to the VQI cohort. A novel model for 1-year AFS was also derived using the VQI data set and externally validated using the PIII data set. The relative discrimination (Harrell c-index) and calibration (Hosmer-May goodness-of-fit test) of each model were compared. Among 7754 patients in the VQI who underwent LEB for critical limb ischemia, the AFS was 74% at 1 year. Each of the previously published models for AFS demonstrated similar discriminative performance: c-indices for BASIL, FINNVASC, mPIII were 0.66, 0.60, and 0.64, respectively. The novel VQI-derived model had improved discriminative ability with a c-index of 0.71 and appropriate generalizability on external validation with a c-index of 0.68. The model was well calibrated in both the VQI and PIII data sets (goodness of fit P = not significant). Currently available prediction models for AFS after LEB perform modestly when applied to national contemporary VQI data. Moreover, the performance of each model was inferior to that of the novel VQI-derived model

  7. Benchmarking Quality in Online Teaching and Learning: A Rubric for Course Construction and Evaluation

    ERIC Educational Resources Information Center

    Ternus, Mona P.; Palmer, Kay L.; Faulk, Debbie R.

    2007-01-01

    Online courses have many components and dimensions. Both the form (structure) and the content (expression) are situated in an overall environment. The sum of these elements results in student outcomes and learning. In order to facilitate construction and evaluate the quality of an online course, a four-part rubric was designed to reflect:…

  8. Implementing Cognitive Behavioral Therapy for Chronic Fatigue Syndrome in a Mental Health Center: A Benchmarking Evaluation

    ERIC Educational Resources Information Center

    Scheeres, Korine; Wensing, Michel; Knoop, Hans; Bleijenberg, Gijs

    2008-01-01

    Objective: This study evaluated the success of implementing cognitive behavioral therapy (CBT) for chronic fatigue syndrome (CFS) in a representative clinical practice setting and compared the patient outcomes with those of previously published randomized controlled trials (RCTs) of CBT for CFS. Method: The implementation interventions were the…

  9. Implementing Cognitive Behavioral Therapy for Chronic Fatigue Syndrome in a Mental Health Center: A Benchmarking Evaluation

    ERIC Educational Resources Information Center

    Scheeres, Korine; Wensing, Michel; Knoop, Hans; Bleijenberg, Gijs

    2008-01-01

    Objective: This study evaluated the success of implementing cognitive behavioral therapy (CBT) for chronic fatigue syndrome (CFS) in a representative clinical practice setting and compared the patient outcomes with those of previously published randomized controlled trials (RCTs) of CBT for CFS. Method: The implementation interventions were the…

  10. Evaluating Bias of Sequential Mixed-Mode Designs against Benchmark Surveys

    ERIC Educational Resources Information Center

    Klausch, Thomas; Schouten, Barry; Hox, Joop J.

    2017-01-01

    This study evaluated three types of bias--total, measurement, and selection bias (SB)--in three sequential mixed-mode designs of the Dutch Crime Victimization Survey: telephone, mail, and web, where nonrespondents were followed up face-to-face (F2F). In the absence of true scores, all biases were estimated as mode effects against two different…

  11. Project BACSTOP Evaluation Report 1974-1975.

    ERIC Educational Resources Information Center

    Nelson, Neil; Martin, William

    Designed to observe changes in biracial student behavior brought about by Project BACSTOP (a series of structured experiences in a variety of wilderness settings meant to bring students of different races together in stressful adventure activities geared to promote interaction, communication, and cooperation), this evaluation studied five…

  12. The GLAD Project Evaluation Summary: 1994 Report.

    ERIC Educational Resources Information Center

    Boloz, Sigmund A.

    This report evaluates the Ganado Learning Arts Development Project (GLAD), a program implemented at the Ganado Primary School in Arizona. The school serves K-2 students from the Navajo reservation and emphasizes integration of traditional Navajo teachings and culture with Western knowledge. The report covers the following data: (1) student average…

  13. Federal Workplace Literacy Project. Internal Evaluation Report.

    ERIC Educational Resources Information Center

    Matuszak, David J.

    This report describes the following components of the Nestle Workplace Literacy Project: six job task analyses, curricula for six workplace basic skills training programs, delivery of courses using these curricula, and evaluation of the process. These six job categories were targeted for training: forklift loader/checker, BB's processing systems…

  14. Project ALERT. Workplace Education. External Evaluators Reports.

    ERIC Educational Resources Information Center

    Philippi, Jorie W.; Mikulecky, Larry; Lloyd, Paul

    This document contains four evaluations of Project ALERT (Adult Literacy Enhanced & Redefined through Training), a workplace literacy partnership of Wayne State University, the Detroit Public Schools, and several city organizations, unions, and manufacturers in the automobile industry that was formed to meet employees' job-specific basic skills…

  15. An Evaluation of the Connected Mathematics Project.

    ERIC Educational Resources Information Center

    Cain, Judith S.

    2002-01-01

    Evaluated the Connected Mathematics Project (CMP), a middle school reform mathematics curriculum used in Louisiana's Lafayette parish. Analysis of Iowa Test of Basic Skills and Louisiana Education Assessment Program mathematics data indicated that CMP schools significantly outperformed non-CMP schools. Surveys of teachers and students showed that…

  16. Workforce development and effective evaluation of projects.

    PubMed

    Dickerson, Claire; Green, Tess; Blass, Eddie

    The success of a project or programme is typically determined in relation to outputs. However, there is a commitment among UK public services to spending public funds efficiently and on activities that provide the greatest benefit to society. Skills for Health recognised the need for a tool to manage the complex process of evaluating project benefits. An integrated evaluation framework was developed to help practitioners identify, describe, measure and evaluate the benefits of workforce development projects. Practitioners tested the framework on projects within three NHS trusts and provided valuable feedback to support its development. The prospective approach taken to identify benefits and collect baseline data to support evaluation was positively received and the clarity and completeness of the framework, as well as the relevance of the questions, were commended. Users reported that the framework was difficult to complete; an online version could be developed, which might help to improve usability. Effective implementation of this approach will depend on the quality and usability of the framework, the willingness of organisations to implement it, and the presence or establishment of an effective change management culture.

  17. 'Score to Door Time', a benchmarking tool for rapid response systems: a pilot multi-centre service evaluation

    PubMed Central

    2011-01-01

    Introduction Rapid Response Systems were created to minimise delays in recognition and treatment of deteriorating patients on general wards. Physiological 'track and trigger' systems are used to alert a team with critical care skills to stabilise patients and expedite admission to intensive care units. No benchmarking tool exists to facilitate comparison for quality assurance. This study was designed to create and test a tool to analyse the efficiency of intensive care admission processes. Methods We conducted a pilot multicentre service evaluation of patients admitted to 17 intensive care units from the United Kingdom, Ireland, Denmark, United States of America and Australia. Physiological abnormalities were recorded via a standardised track and trigger score (VitalPAC™ Early Warning Score). The period between the time of initial physiological abnormality (Score) and admission to intensive care (Door) was recorded as 'Score to Door Time'. Participants subsequently suggested causes for admission delays. Results Score to Door Time for 177 admissions was a median of 4:10 hours (interquartile range (IQR) 1:49 to 9:10). Time from physiological trigger to activation of a Rapid Response System was a median 0:47 hours (IQR 0:00 to 2:15). Time from call-out to intensive care admission was a median of 2:45 hours (IQR 1:19 to 6:32). A total of 127 (71%) admissions were deemed to have been delayed. Stepwise linear regression analysis yielded three significant predictors of longer Score to Door Time: being treated in a British centre, higher Acute Physiology and Chronic Health Evaluation (APACHE) II score and increasing age. Binary regression analysis demonstrated a significant association (P < 0.045) of APACHE II scores >20 with Score to Door Times greater than the median 4:10 hours. Conclusions Score to Door Time seemed to be largely independent of illness severity and, when combined with qualitative feedback from centres, suggests that admission delays could be due to

  18. 'Score to Door Time', a benchmarking tool for rapid response systems: a pilot multi-centre service evaluation.

    PubMed

    Oglesby, Kieran J; Durham, Lesley; Welch, John; Subbe, Christian P

    2011-07-27

    Rapid Response Systems were created to minimise delays in recognition and treatment of deteriorating patients on general wards. Physiological 'track and trigger' systems are used to alert a team with critical care skills to stabilise patients and expedite admission to intensive care units. No benchmarking tool exists to facilitate comparison for quality assurance. This study was designed to create and test a tool to analyse the efficiency of intensive care admission processes. We conducted a pilot multicentre service evaluation of patients admitted to 17 intensive care units from the United Kingdom, Ireland, Denmark, United States of America and Australia. Physiological abnormalities were recorded via a standardised track and trigger score (VitalPAC™ Early Warning Score). The period between the time of initial physiological abnormality (Score) and admission to intensive care (Door) was recorded as 'Score to Door Time'. Participants subsequently suggested causes for admission delays. Score to Door Time for 177 admissions was a median of 4:10 hours (interquartile range (IQR) 1:49 to 9:10). Time from physiological trigger to activation of a Rapid Response System was a median 0:47 hours (IQR 0:00 to 2:15). Time from call-out to intensive care admission was a median of 2:45 hours (IQR 1:19 to 6:32). A total of 127 (71%) admissions were deemed to have been delayed. Stepwise linear regression analysis yielded three significant predictors of longer Score to Door Time: being treated in a British centre, higher Acute Physiology and Chronic Health Evaluation (APACHE) II score and increasing age. Binary regression analysis demonstrated a significant association (P < 0.045) of APACHE II scores >20 with Score to Door Times greater than the median 4:10 hours. Score to Door Time seemed to be largely independent of illness severity and, when combined with qualitative feedback from centres, suggests that admission delays could be due to organisational issues, rather than patient

  19. 23 CFR 505.11 - Project evaluation and rating.

    Code of Federal Regulations, 2010 CFR

    2010-04-01

    ... 23 Highways 1 2010-04-01 2010-04-01 false Project evaluation and rating. 505.11 Section 505.11... MANAGEMENT PROJECTS OF NATIONAL AND REGIONAL SIGNIFICANCE EVALUATION AND RATING § 505.11 Project evaluation and rating. (a) The Secretary shall evaluate and rate each proposed project as “highly recommended...

  20. Reactor Physics and Criticality Benchmark Evaluations for Advanced Nuclear Fuel, Progress Report for Work through August 31, 2002, First Annual/4th Quarterly Report

    SciTech Connect

    Anderson, William J.; Ake, Timothy N.; Punatar, Mahendra; Pitts, Michelle L.; Harms, Gary A.; Rearden, Bradley T.; Parks, Cecil V.; Tulenko, James S.; Dugan, Edward; Smith, Robert M.

    2002-09-23

    OAK B204 The objective of this Nuclear Energy Research Initiative (NERI) project is to design, perform, and analyze critical benchmark experiments for validating reactor physics methods and models for fuel enrichments greater than 5-wt% 235U. These experiments will also provide additional information for application to the criticality-safety bases for commercial fuel facilities handling greater than 5-wt% 235U fuel. These experiments are designed as reactor physics benchmarks, to include measurements of critical boron concentration, burnable absorber worth, relative pin powers, and relative average powers.The first year focused primarily on designing the experiments using available fuel, preparing the necessary plans, procedures and authorization basis for performing the experiments, and preparing for the transportation, receipt and storage of the Pathfinder fuel currently stored at Pennsylvania State University.Framatome ANP, Inc. leads the project with the collaboration of Oak Ridge National Laboratory (ORNL), Sandia National Laboratories (SNL), and the University of Florida (UF). The project is organized into 5 tasks:Task 1: Framatome ANP, Inc., ORNL, and SNL will design the specific experiments, establish the safety authorization, and obtain approvals to perform these experiments at the SNL facility. ORNL will apply their sensitivity/uncertainty methodology to verify the need for particular experiments and the parameters that these experiments need to explore.Task 2: Framatome ANP, Inc., ORNL, and UF will analyze the proposed experiments using a variety of reactor-physics methods employed in the nuclear industry. These analyses will support the operation of the experiments by predicting the expected experimental values for the criticality and physics parameters.Task 3: This task encompasses the experiments to be performed. The Pathfinder fuel will be transported from Penn State to SNL for use in the experiments. The experiments will be performed and the

  1. Kenya's Radio Language Arts Project: evaluation results.

    PubMed

    Oxford, R L

    1985-01-01

    The Kenya Radio Language Arts Project (RLAP), which has just been completed, documents the effectiveness of interactive radio-based educational instruction. Analyses in the areas of listening, reading, speaking, and writing show that children in radio classrooms consistently scored better than children in nonradio classrooms in every test. An evaluation of the project was conducted with the assistance of the Center for Applied Linguistics (CAL). Evaluation results came from a variety of sources, including language tests, observations, interviews, demographic and administrative records, and an attitude survey. A large proportion of the project's students were considerably transient. Only 22% of the total student population of 3908 were "normal progression" students -- that is, they advanced regularly through their education during the life of the project. Students who moved from the area, failed a standard (grade), dropped out, or were otherwise untrackable, comprised the remaining 78% of the total. 7 districts were included in the project. Tests were developed for listening and reading in Standards 1, 2, and 3 and in speaking and writing in Standards 2 and 3. The achievement tests were based on the official Kenya curriculum for those standards, so as to measure achievement against the curriculum. Nearly all the differences were highly significant statistically, with a probability of less than 1 in 1000 that the findings could have occurred by chance. Standard 1 radio students scored nearly 8 points higher than did their counterparts in the control group. Standard 2 and 3 radio students outperformed the control students by 4 points. The radio group consistently outperformed the control group in reading, writing, and speaking. Unstructured interviews and observations were conducted by the RLAP field staff. Overwhelmingly positive attitudes about the project prevailed among project teachers and headmasters. The data demonstrate that RLAP works. In fact, it works so

  2. Evaluation of the influence of the definition of an isolated hip fracture as an exclusion criterion for trauma system benchmarking: a multicenter cohort study.

    PubMed

    Tiao, J; Moore, L; Porgo, T V; Belcaid, A

    2016-06-01

    To assess whether the definition of an IHF used as an exclusion criterion influences the results of trauma center benchmarking. We conducted a multicenter retrospective cohort study with data from an integrated Canadian trauma system. The study population included all patients admitted between 1999 and 2010 to any of the 57 adult trauma centers. Seven definitions of IHF based on diagnostic codes, age, mechanism of injury, and secondary injuries, identified in a systematic review, were used. Trauma centers were benchmarked using risk-adjusted mortality estimates generated using the Trauma Risk Adjustment Model. The agreement between benchmarking results generated under different IHF definitions was evaluated with correlation coefficients on adjusted mortality estimates. Correlation coefficients >0.95 were considered to convey acceptable agreement. The study population consisted of 172,872 patients before exclusion of IHF and between 128,094 and 139,588 patients after exclusion. Correlation coefficients between risk-adjusted mortality estimates generated in populations including and excluding IHF varied between 0.86 and 0.90. Correlation coefficients of estimates generated under different definitions of IHF varied between 0.97 and 0.99, even when analyses were restricted to patients aged ≥65 years. Although the exclusion of patients with IHF has an influence on the results of trauma center benchmarking based on mortality, the definition of IHF in terms of diagnostic codes, age, mechanism of injury and secondary injury has no significant impact on benchmarking results. Results suggest that there is no need to obtain formal consensus on the definition of IHF for benchmarking activities.

  3. A simple benchmark for evaluating quality of care of patients following acute myocardial infarction

    PubMed Central

    Dorsch, M; Lawrance, R; Sapsford, R; Oldham, J; Greenwood, D; Jackson, B; Morrell, C; Ball, S; Robinson, M; Hall, A

    2001-01-01

    OBJECTIVE—To develop a simple risk model as a basis for evaluating care of patients admitted with acute myocardial infarction.
METHODS—From coronary care registers, biochemistry records and hospital management systems, 2153 consecutive patients with confirmed acute myocardial infarction were identified. With 30 day all cause mortality as the end point, a multivariable logistic regression model of risk was constructed and validated in independent patient cohorts. The areas under receiver operating characteristic curves were calculated as an assessment of sensitivity and specificity. The model was reapplied to a number of commonly studied subgroups for further assessment of robustness.
RESULTS—A three variable model was developed based on age, heart rate, and systolic blood pressure on admission. This produced an individual probability of death by 30 days (P30) where P30 = 1/(1 + exp(−L30)) and L30 = −5.624 + (0.085 × age) + (0.014 × heart rate) − (0.022 × systolic blood pressure). The areas under the receiver operating characteristic curves for the reference and test cohorts were 0.79 (95% CI 0.76 to 0.82) and 0.76 (95% CI 0.72 to 0.79), respectively. To aid application of the model to routine clinical audit, a normogram relating observed mortality and sample size to the likelihood of a significant deviation from the expected 30 day mortality rate was constructed.
CONCLUSIONS—This risk model is simple, reproducible, and permits quality of care of acute myocardial infarction patients to be reliably evaluated both within and between centres.


Keywords: acute myocardial infarction; risk model PMID:11454829

  4. Optimizing and benchmarking de novo transcriptome sequencing: from library preparation to assembly evaluation.

    PubMed

    Hara, Yuichiro; Tatsumi, Kaori; Yoshida, Michio; Kajikawa, Eriko; Kiyonari, Hiroshi; Kuraku, Shigehiro

    2015-11-18

    RNA-seq enables gene expression profiling in selected spatiotemporal windows and yields massive sequence information with relatively low cost and time investment, even for non-model species. However, there remains a large room for optimizing its workflow, in order to take full advantage of continuously developing sequencing capacity. Transcriptome sequencing for three embryonic stages of Madagascar ground gecko (Paroedura picta) was performed with the Illumina platform. The output reads were assembled de novo for reconstructing transcript sequences. In order to evaluate the completeness of transcriptome assemblies, we prepared a reference gene set consisting of vertebrate one-to-one orthologs. To take advantage of increased read length of >150 nt, we demonstrated shortened RNA fragmentation time, which resulted in a dramatic shift of insert size distribution. To evaluate products of multiple de novo assembly runs incorporating reads with different RNA sources, read lengths, and insert sizes, we introduce a new reference gene set, core vertebrate genes (CVG), consisting of 233 genes that are shared as one-to-one orthologs by all vertebrate genomes examined (29 species)., The completeness assessment performed by the computational pipelines CEGMA and BUSCO referring to CVG, demonstrated higher accuracy and resolution than with the gene set previously established for this purpose. As a result of the assessment with CVG, we have derived the most comprehensive transcript sequence set of the Madagascar ground gecko by means of assembling individual libraries followed by clustering the assembled sequences based on their overall similarities. Our results provide several insights into optimizing de novo RNA-seq workflow, including the coordination between library insert size and read length, which manifested in improved connectivity of assemblies. The approach and assembly assessment with CVG demonstrated here would be applicable to transcriptome analysis of other species as

  5. Self-benchmarking Guide for Laboratory Buildings: Metrics, Benchmarks, Actions

    SciTech Connect

    Mathew, Paul; Greenberg, Steve; Sartor, Dale

    2009-07-13

    This guide describes energy efficiency metrics and benchmarks that can be used to track the performance of and identify potential opportunities to reduce energy use in laboratory buildings. This guide is primarily intended for personnel who have responsibility for managing energy use in existing laboratory facilities - including facilities managers, energy managers, and their engineering consultants. Additionally, laboratory planners and designers may also use the metrics and benchmarks described in this guide for goal-setting in new construction or major renovation. This guide provides the following information: (1) A step-by-step outline of the benchmarking process. (2) A set of performance metrics for the whole building as well as individual systems. For each metric, the guide provides a definition, performance benchmarks, and potential actions that can be inferred from evaluating this metric. (3) A list and descriptions of the data required for computing the metrics. This guide is complemented by spreadsheet templates for data collection and for computing the benchmarking metrics. This guide builds on prior research supported by the national Laboratories for the 21st Century (Labs21) program, supported by the U.S. Department of Energy and the U.S. Environmental Protection Agency. Much of the benchmarking data are drawn from the Labs21 benchmarking database and technical guides. Additional benchmark data were obtained from engineering experts including laboratory designers and energy managers.

  6. Self-benchmarking Guide for Cleanrooms: Metrics, Benchmarks, Actions

    SciTech Connect

    Mathew, Paul; Sartor, Dale; Tschudi, William

    2009-07-13

    This guide describes energy efficiency metrics and benchmarks that can be used to track the performance of and identify potential opportunities to reduce energy use in laboratory buildings. This guide is primarily intended for personnel who have responsibility for managing energy use in existing laboratory facilities - including facilities managers, energy managers, and their engineering consultants. Additionally, laboratory planners and designers may also use the metrics and benchmarks described in this guide for goal-setting in new construction or major renovation. This guide provides the following information: (1) A step-by-step outline of the benchmarking process. (2) A set of performance metrics for the whole building as well as individual systems. For each metric, the guide provides a definition, performance benchmarks, and potential actions that can be inferred from evaluating this metric. (3) A list and descriptions of the data required for computing the metrics. This guide is complemented by spreadsheet templates for data collection and for computing the benchmarking metrics. This guide builds on prior research supported by the national Laboratories for the 21st Century (Labs21) program, supported by the U.S. Department of Energy and the U.S. Environmental Protection Agency. Much of the benchmarking data are drawn from the Labs21 benchmarking database and technical guides. Additional benchmark data were obtained from engineering experts including laboratory designers and energy managers.

  7. Valuing Metal-Organic Frameworks for Postcombustion Carbon Capture: A Benchmark Study for Evaluating Physical Adsorbents.

    PubMed

    Adil, Karim; Bhatt, Prashant M; Belmabkhout, Youssef; Abtab, Sk Md Towsif; Jiang, Hao; Assen, Ayalew H; Mallick, Arijit; Cadiau, Amandine; Aqil, Jamal; Eddaoudi, Mohamed

    2017-08-22

    The development of practical solutions for the energy-efficient capture of carbon dioxide is of prime importance and continues to attract intensive research interest. Conceivably, the implementation of adsorption-based processes using different cycling modes, e.g., pressure-swing adsorption or temperature-swing adsorption, offers great prospects to address this challenge. Practically, the successful deployment of practical adsorption-based technologies depends on the development of made-to-order adsorbents expressing mutually two compulsory requisites: i) high selectivity/affinity for CO2 and ii) excellent chemical stability in the presence of impurities. This study presents a new comprehensive experimental protocol apposite for assessing the prospects of a given physical adsorbent for carbon capture under flue gas stream conditions. The protocol permits: i) the baseline performance of commercial adsorbents such as zeolite 13X, activated carbon versus liquid amine scrubbing to be ascertained, and ii) a standardized evaluation of the best reported metal-organic framework (MOF) materials for carbon dioxide capture from flue gas to be undertaken. This extensive study corroborates the exceptional CO2 capture performance of the recently isolated second-generation fluorinated MOF material, NbOFFIVE-1-Ni, concomitant with an impressive chemical stability and a low energy for regeneration. Essentially, the NbOFFIVE-1-Ni adsorbent presents the best compromise by satisfying all the required metrics for efficient CO2 scrubbing. © 2017 WILEY-VCH Verlag GmbH & Co. KGaA, Weinheim.

  8. Testing (Validating?) Cross Sections with ICSBEP Benchmarks

    SciTech Connect

    Kahler, Albert C. III

    2012-06-28

    We discuss how to use critical benchmarks from the International Handbook of Evaluated Criticality Safety Benchmark Experiments to determine the applicability of specific cross sections to the end-user's problem of interest. Particular attention is paid to making sure the selected suite of benchmarks includes the user's range of applicability (ROA).

  9. Benchmarking pathology services: implementing a longitudinal study.

    PubMed

    Gordon, M; Holmes, S; McGrath, K; Neil, A

    1999-05-01

    This paper details the benchmarking process and its application to the activities of pathology laboratories participating in a benchmark pilot study [the Royal College of Pathologists of Australasian (RCPA) Benchmarking Project]. The discussion highlights the primary issues confronted in collecting, processing, analysing and comparing benchmark data. The paper outlines the benefits of engaging in a benchmarking exercise and provides a framework which can be applied across a range of public health settings. This information is then applied to a review of the development of the RCPA Benchmarking Project. Consideration is also given to the nature of the preliminary results of the project and the implications of these results to the on-going conduct of the study.

  10. Multi-class computational evolution: development, benchmark evaluation and application to RNA-Seq biomarker discovery.

    PubMed

    Crabtree, Nathaniel M; Moore, Jason H; Bowyer, John F; George, Nysia I

    2017-01-01

    A computational evolution system (CES) is a knowledge discovery engine that can identify subtle, synergistic relationships in large datasets. Pareto optimization allows CESs to balance accuracy with model complexity when evolving classifiers. Using Pareto optimization, a CES is able to identify a very small number of features while maintaining high classification accuracy. A CES can be designed for various types of data, and the user can exploit expert knowledge about the classification problem in order to improve discrimination between classes. These characteristics give CES an advantage over other classification and feature selection algorithms, particularly when the goal is to identify a small number of highly relevant, non-redundant biomarkers. Previously, CESs have been developed only for binary class datasets. In this study, we developed a multi-class CES. The multi-class CES was compared to three common feature selection and classification algorithms: support vector machine (SVM), random k-nearest neighbor (RKNN), and random forest (RF). The algorithms were evaluated on three distinct multi-class RNA sequencing datasets. The comparison criteria were run-time, classification accuracy, number of selected features, and stability of selected feature set (as measured by the Tanimoto distance). The performance of each algorithm was data-dependent. CES performed best on the dataset with the smallest sample size, indicating that CES has a unique advantage since the accuracy of most classification methods suffer when sample size is small. The multi-class extension of CES increases the appeal of its application to complex, multi-class datasets in order to identify important biomarkers and features.

  11. Classification of malnutrition in cystic fibrosis: implications for evaluating and benchmarking clinical practice performance2

    PubMed Central

    HuiChuan, J Lai; Suzanne, M Shoff

    2008-01-01

    Background In 2005, the Cystic Fibrosis Foundation (CFF) revised the nutrition classification guidelines to eliminate the use of percentage of ideal body weight (%IBW) to define “nutritional failure”; the CFF also recommended that children with cystic fibrosis maintain a body mass index percentile (BMIp) ≥ 50th. Objective We assessed the effect of the 2005 CFF nutrition classification guidelines on evaluating the performance of nutritional care practices. Design Data from 14 702 children reported to the 2002 CFF Patient Registry were analyzed to compare malnutrition rates in 113 cystic fibrosis centers in the United States. Nutritional failure was defined according to the 2002 CFF criteria—ie, height < 5th percentile, %IBW < 90%, or BMIp < 10th. “Below BMI goal” was defined according to the 2005 CFF criterion, ie BMIp < 50th. Results Eliminating %IBW resulted in a 6% reduction (from 33% to 27%) in the nutritional failure rate in the United States. The use of BMIp < 50th led to the classification of 57% of children as below the BMI goal. Misclassification of nutritional failure according to %IBW ranged from 1% to 16% among 113 centers and was greater in the centers with a larger proportion of tall patients. After the elimination of %IBW, one-third of centers changed to a different tertile ranking for nutritional failure rates (kappa = 0.50, moderate-to-poor agreement). More than half the centers changed to a different tertile ranking, from nutritional failure to below BMI goal (kappa = 0.22, poor agreement). Conclusion Eliminating misclassification by %IBW and implementing the new BMI goal led to profound and unequal changes in malnutrition rates across cystic fibrosis centers. PMID:18614737

  12. Small Commercial Program DOE Project: Impact evaluation

    SciTech Connect

    Bathgate, R.; Faust, S. )

    1992-08-12

    In 1991, Washington Electric Cooperative (WEC) implemented a Department of Energy grant to conduct a small commercial energy conservation project. The small commercial Mom, and Pop'' grocery stores within WEC's service territory were selected as the target market for the project. Energy Solid Waste Consultant's (E SWC) Impact Evaluation is documented here. The evaluation was based on data gathered from a variety of sources, including load profile metering, kWh submeters, elapsed time indicators, and billing histories. Five stores were selected to receive measures under this program: Waits River General Store, Joe's Pond Store, Hastings Store, Walden General Store, and Adamant Cooperative. Specific measures installed in each store and description of each are included.

  13. Implementation of Benchmarking Transportation Logistics Practices and Future Benchmarking Organizations

    SciTech Connect

    Thrower, A.W.; Patric, J.; Keister, M.

    2008-07-01

    The purpose of the Office of Civilian Radioactive Waste Management's (OCRWM) Logistics Benchmarking Project is to identify established government and industry practices for the safe transportation of hazardous materials which can serve as a yardstick for design and operation of OCRWM's national transportation system for shipping spent nuclear fuel and high-level radioactive waste to the proposed repository at Yucca Mountain, Nevada. The project will present logistics and transportation practices and develop implementation recommendations for adaptation by the national transportation system. This paper will describe the process used to perform the initial benchmarking study, highlight interim findings, and explain how these findings are being implemented. It will also provide an overview of the next phase of benchmarking studies. The benchmarking effort will remain a high-priority activity throughout the planning and operational phases of the transportation system. The initial phase of the project focused on government transportation programs to identify those practices which are most clearly applicable to OCRWM. These Federal programs have decades of safe transportation experience, strive for excellence in operations, and implement effective stakeholder involvement, all of which parallel OCRWM's transportation mission and vision. The initial benchmarking project focused on four business processes that are critical to OCRWM's mission success, and can be incorporated into OCRWM planning and preparation in the near term. The processes examined were: transportation business model, contract management/out-sourcing, stakeholder relations, and contingency planning. More recently, OCRWM examined logistics operations of AREVA NC's Business Unit Logistics in France. The next phase of benchmarking will focus on integrated domestic and international commercial radioactive logistic operations. The prospective companies represent large scale shippers and have vast experience in

  14. The NIEHS Predictive-Toxicology Evaluation Project.

    PubMed Central

    Bristol, D W; Wachsman, J T; Greenwell, A

    1996-01-01

    The Predictive-Toxicology Evaluation (PTE) project conducts collaborative experiments that subject the performance of predictive-toxicology (PT) methods to rigorous, objective evaluation in a uniquely informative manner. Sponsored by the National Institute of Environmental Health Sciences, it takes advantage of the ongoing testing conducted by the U.S. National Toxicology Program (NTP) to estimate the true error of models that have been applied to make prospective predictions on previously untested, noncongeneric-chemical substances. The PTE project first identifies a group of standardized NTP chemical bioassays either scheduled to be conducted or are ongoing, but not yet complete. The project then announces and advertises the evaluation experiment, disseminates information about the chemical bioassays, and encourages researchers from a wide variety of disciplines to publish their predictions in peer-reviewed journals, using whatever approaches and methods they feel are best. A collection of such papers is published in this Environmental Health Perspectives Supplement, providing readers the opportunity to compare and contrast PT approaches and models, within the context of their prospective application to an actual-use situation. This introduction to this collection of papers on predictive toxicology summarizes the predictions made and the final results obtained for the 44 chemical carcinogenesis bioassays of the first PTE experiment (PTE-1) and presents information that identifies the 30 chemical carcinogenesis bioassays of PTE-2, along with a table of prediction sets that have been published to date. It also provides background about the origin and goals of the PTE project, outlines the special challenge associated with estimating the true error of models that aspire to predict open-system behavior, and summarizes what has been learned to date. PMID:8933048

  15. NASA Countermeasures Evaluation and Validation Project

    NASA Technical Reports Server (NTRS)

    Lundquist, Charlie M.; Paloski, William H. (Technical Monitor)

    2000-01-01

    To support its ISS and exploration class mission objectives, NASA has developed a Countermeasure Evaluation and Validation Project (CEVP). The goal of this project is to evaluate and validate the optimal complement of countermeasures required to maintain astronaut health, safety, and functional ability during and after short- and long-duration space flight missions. The CEVP is the final element of the process in which ideas and concepts emerging from basic research evolve into operational countermeasures. The CEVP is accomplishing these objectives by conducting operational/clinical research to evaluate and validate countermeasures to mitigate these maladaptive responses. Evaluation is accomplished by testing in space flight analog facilities, and validation is accomplished by space flight testing. Both will utilize a standardized complement of integrated physiological and psychological tests, termed the Integrated Testing Regimen (ITR) to examine candidate countermeasure efficacy and intersystem effects. The CEVP emphasis is currently placed on validating the initial complement of ISS countermeasures targeting bone, muscle, and aerobic fitness; followed by countermeasures for neurological, psychological, immunological, nutrition and metabolism, and radiation risks associated with space flight. This presentation will review the processes, plans, and procedures that will enable CEVP to play a vital role in transitioning promising research results into operational countermeasures necessary to maintain crew health and performance during long duration space flight.

  16. NASA Countermeasures Evaluation and Validation Project

    NASA Technical Reports Server (NTRS)

    Lundquist, Charlie M.; Paloski, William H. (Technical Monitor)

    2000-01-01

    To support its ISS and exploration class mission objectives, NASA has developed a Countermeasure Evaluation and Validation Project (CEVP). The goal of this project is to evaluate and validate the optimal complement of countermeasures required to maintain astronaut health, safety, and functional ability during and after short- and long-duration space flight missions. The CEVP is the final element of the process in which ideas and concepts emerging from basic research evolve into operational countermeasures. The CEVP is accomplishing these objectives by conducting operational/clinical research to evaluate and validate countermeasures to mitigate these maladaptive responses. Evaluation is accomplished by testing in space flight analog facilities, and validation is accomplished by space flight testing. Both will utilize a standardized complement of integrated physiological and psychological tests, termed the Integrated Testing Regimen (ITR) to examine candidate countermeasure efficacy and intersystem effects. The CEVP emphasis is currently placed on validating the initial complement of ISS countermeasures targeting bone, muscle, and aerobic fitness; followed by countermeasures for neurological, psychological, immunological, nutrition and metabolism, and radiation risks associated with space flight. This presentation will review the processes, plans, and procedures that will enable CEVP to play a vital role in transitioning promising research results into operational countermeasures necessary to maintain crew health and performance during long duration space flight.

  17. Managing for Results in America's Great City Schools. A Report of the Performance Measurement and Benchmarking Project

    ERIC Educational Resources Information Center

    Council of the Great City Schools, 2012

    2012-01-01

    "Managing for Results in America's Great City Schools, 2012" is presented by the Council of the Great City Schools to its members and the public. The purpose of the project was and is to develop performance measures that can improve the business operations of urban public school districts nationwide. This year's report includes data from 61 of the…

  18. GRID-based three-dimensional pharmacophores II: PharmBench, a benchmark data set for evaluating pharmacophore elucidation methods.

    PubMed

    Cross, Simon; Ortuso, Francesco; Baroni, Massimo; Costa, Giosuè; Distinto, Simona; Moraca, Federica; Alcaro, Stefano; Cruciani, Gabriele

    2012-10-22

    To date, published pharmacophore elucidation approaches typically use a handful of data sets for validation: here, we have assembled a data set for 81 targets, containing 960 ligands aligned using their cocrystallized protein targets, to provide the experimental "gold standard". The two-dimensional structures are also assembled to remove conformational bias; an ideal method would be able to take these structures as input, find the common features, and reproduce the bioactive conformations and their alignments to correspond with the X-ray-determined gold standard alignments. Here we present this data set and describe three objective measures to evaluate performance: the ability to identify the bioactive conformation, the ability to identify and correctly align this conformation for 50% of the molecules in each data set, and the pharmacophoric field similarity. We have applied this validation methodology to our pharmacophore elucidation method FLAPpharm, that is published in the first paper of this series and discuss the limitations of the data set and objective success criteria. Starting from two-dimensional structures and producing unbiased models, FLAPpharm was able to identify the bioactive conformations for 67% of the ligands and also to produce successful models according to the second metric for 67% of the Pharmbench data sets. Inspection of the unsuccessful models highlighted the limitation of this root mean square (rms)-derived metric, since many were found to be pharmacophorically reasonable, increasing the overall success rate to 83%. The PharmBench data set is available at http://www.moldiscovery.com/PharmBench , along with a web service to enable users to score model alignments coming from external methods in the same way that we have presented here and, therefore, establishes a pharmacophore elucidation benchmark data set available to be used by the community.

  19. Wildlife habitat evaluation demonstration project. [Michigan

    NASA Technical Reports Server (NTRS)

    Burgoyne, G. E., Jr.; Visser, L. G.

    1981-01-01

    To support the deer range improvement project in Michigan, the capability of LANDSAT data in assessing deer habitat in terms of areas and mixes of species and age classes of vegetation is being examined to determine whether such data could substitute for traditional cover type information sources. A second goal of the demonstration project is to determine whether LANDSAT data can be used to supplement and improve the information normally used for making deer habitat management decisions, either by providing vegetative cover for private land or by providing information about the interspersion and juxtaposition of valuable vegetative cover types. The procedure to be used for evaluating in LANDSAT data of the Lake County test site is described.

  20. Wildlife habitat evaluation demonstration project. [Michigan

    NASA Technical Reports Server (NTRS)

    Burgoyne, G. E., Jr.; Visser, L. G.

    1981-01-01

    To support the deer range improvement project in Michigan, the capability of LANDSAT data in assessing deer habitat in terms of areas and mixes of species and age classes of vegetation is being examined to determine whether such data could substitute for traditional cover type information sources. A second goal of the demonstration project is to determine whether LANDSAT data can be used to supplement and improve the information normally used for making deer habitat management decisions, either by providing vegetative cover for private land or by providing information about the interspersion and juxtaposition of valuable vegetative cover types. The procedure to be used for evaluating in LANDSAT data of the Lake County test site is described.

  1. Benchmarking in Student Affairs.

    ERIC Educational Resources Information Center

    Mosier, Robert E.; Schwarzmueller, Gary J.

    2002-01-01

    Discusses the use of benchmarking in student affairs, focusing on issues related to student housing. Provides examples of how benchmarking has influenced administrative practice at many institutions. (EV)

  2. Color back projection for fruit maturity evaluation

    NASA Astrophysics Data System (ADS)

    Zhang, Dong; Lee, Dah-Jye; Desai, Alok

    2013-12-01

    In general, fruits and vegetables such as tomatoes and dates are harvested before they fully ripen. After harvesting, they continue to ripen and their color changes. Color is a good indicator of fruit maturity. For example, tomatoes change color from dark green to light green and then pink, light red, and dark red. Assessing tomato maturity helps maximize its shelf life. Color is used to determine the length of time the tomatoes can be transported. Medjool dates change color from green to yellow, and the orange, light red and dark red. Assessing date maturity helps determine the length of drying process to help ripen the dates. Color evaluation is an important step in the processing and inventory control of fruits and vegetables that directly affects profitability. This paper presents an efficient color back projection and image processing technique that is designed specifically for real-time maturity evaluation of fruits. This color processing method requires very simple training procedure to obtain the frequencies of colors that appear in each maturity stage. This color statistics is used to back project colors to predefined color indexes. Fruit maturity is then evaluated by analyzing the reprojected color indexes. This method has been implemented and used for commercial production.

  3. Toxicological Benchmarks for Screening Potential Contaminants of Concern for Effects on Sediment-Associated Biota

    SciTech Connect

    Hull, R.N.

    1993-01-01

    A hazardous waste site may contain hundreds of chemicals; therefore, it is important to screen contaminants of potential concern for the ecological risk assessment. Often this screening is done as part of a screening assessment, the purpose of which is to evaluate the available data, identify data gaps, and screen contaminants of potential concern. Screening may be accomplished by using a set of toxicological benchmarks. These benchmarks are helpful in determining whether contaminants warrant further assessment or are at a level that requires no further attention. If a chemical concentration or the reported detection limit exceeds a proposed lower benchmark, further analysis is needed to determine the hazards posed by that chemical. If, however, the chemical concentration falls below the lower benchmark value, the chemical may be eliminated from further study. The use of multiple benchmarks is recommended for screening chemicals of concern in sediments. Integrative benchmarks developed for the National Oceanic and Atmospheric Administration and the Florida Department of Environmental Protection are included for inorganic and organic chemicals. Equilibrium partitioning benchmarks are included for screening nonionic organic chemicals. Freshwater sediment effect concentrations developed as part of the U.S. Environmental Protection Agency's (EPA's) Assessment and Remediation of Contaminated Sediment Project are included for inorganic and organic chemicals (EPA 1996). Field survey benchmarks developed for the Ontario Ministry of the Environment are included for inorganic and organic chemicals. In addition, EPA-proposed sediment quality criteria are included along with screening values from EPA Region IV and Ecotox Threshold values from the EPA Office of Solid Waste and Emergency Response. Pore water analysis is recommended for ionic organic compounds; comparisons are then made against water quality benchmarks. This report is an update of three prior reports (Jones et al

  4. ELAPSE - NASA AMES LISP AND ADA BENCHMARK SUITE: EFFICIENCY OF LISP AND ADA PROCESSING - A SYSTEM EVALUATION

    NASA Technical Reports Server (NTRS)

    Davis, G. J.

    1994-01-01

    One area of research of the Information Sciences Division at NASA Ames Research Center is devoted to the analysis and enhancement of processors and advanced computer architectures, specifically in support of automation and robotic systems. To compare systems' abilities to efficiently process Lisp and Ada, scientists at Ames Research Center have developed a suite of non-parallel benchmarks called ELAPSE. The benchmark suite was designed to test a single computer's efficiency as well as alternate machine comparisons on Lisp, and/or Ada languages. ELAPSE tests the efficiency with which a machine can execute the various routines in each environment. The sample routines are based on numeric and symbolic manipulations and include two-dimensional fast Fourier transformations, Cholesky decomposition and substitution, Gaussian elimination, high-level data processing, and symbol-list references. Also included is a routine based on a Bayesian classification program sorting data into optimized groups. The ELAPSE benchmarks are available for any computer with a validated Ada compiler and/or Common Lisp system. Of the 18 routines that comprise ELAPSE, provided within this package are 14 developed or translated at Ames. The others are readily available through literature. The benchmark that requires the most memory is CHOLESKY.ADA. Under VAX/VMS, CHOLESKY.ADA requires 760K of main memory. ELAPSE is available on either two 5.25 inch 360K MS-DOS format diskettes (standard distribution) or a 9-track 1600 BPI ASCII CARD IMAGE format magnetic tape. The contents of the diskettes are compressed using the PKWARE archiving tools. The utility to unarchive the files, PKUNZIP.EXE, is included. The ELAPSE benchmarks were written in 1990. VAX and VMS are trademarks of Digital Equipment Corporation. MS-DOS is a registered trademark of Microsoft Corporation.

  5. ELAPSE - NASA AMES LISP AND ADA BENCHMARK SUITE: EFFICIENCY OF LISP AND ADA PROCESSING - A SYSTEM EVALUATION

    NASA Technical Reports Server (NTRS)

    Davis, G. J.

    1994-01-01

    One area of research of the Information Sciences Division at NASA Ames Research Center is devoted to the analysis and enhancement of processors and advanced computer architectures, specifically in support of automation and robotic systems. To compare systems' abilities to efficiently process Lisp and Ada, scientists at Ames Research Center have developed a suite of non-parallel benchmarks called ELAPSE. The benchmark suite was designed to test a single computer's efficiency as well as alternate machine comparisons on Lisp, and/or Ada languages. ELAPSE tests the efficiency with which a machine can execute the various routines in each environment. The sample routines are based on numeric and symbolic manipulations and include two-dimensional fast Fourier transformations, Cholesky decomposition and substitution, Gaussian elimination, high-level data processing, and symbol-list references. Also included is a routine based on a Bayesian classification program sorting data into optimized groups. The ELAPSE benchmarks are available for any computer with a validated Ada compiler and/or Common Lisp system. Of the 18 routines that comprise ELAPSE, provided within this package are 14 developed or translated at Ames. The others are readily available through literature. The benchmark that requires the most memory is CHOLESKY.ADA. Under VAX/VMS, CHOLESKY.ADA requires 760K of main memory. ELAPSE is available on either two 5.25 inch 360K MS-DOS format diskettes (standard distribution) or a 9-track 1600 BPI ASCII CARD IMAGE format magnetic tape. The contents of the diskettes are compressed using the PKWARE archiving tools. The utility to unarchive the files, PKUNZIP.EXE, is included. The ELAPSE benchmarks were written in 1990. VAX and VMS are trademarks of Digital Equipment Corporation. MS-DOS is a registered trademark of Microsoft Corporation.

  6. Nuclear Data Performance Testing Using Sensitive, but Less Frequently Used ICSBEP Benchmarks

    SciTech Connect

    J. Blair Briggs; John D. Bess

    2011-08-01

    The International Criticality Safety Benchmark Evaluation Project (ICSBEP) has published the International Handbook of Evaluated Criticality Safety Benchmark Experiments annually since 1995. The Handbook now spans over 51,000 pages with benchmark specifications for 4,283 critical, near critical, or subcritical configurations; 24 criticality alarm placement/shielding configurations with multiple dose points for each; and 200 configurations that have been categorized as fundamental physics measurements relevant to criticality safety applications. Benchmark data in the ICSBEP Handbook were originally intended for validation of criticality safety methods and data; however, the benchmark specifications are now used extensively for nuclear data testing. There are several, less frequently used benchmarks within the Handbook that are very sensitive to thorium and certain key structural and moderating materials. Calculated results for many of those benchmarks using modern nuclear data libraries suggest there is still room for improvement. These and other highly sensitive, but rarely quoted benchmarks are highlighted and data testing results provided using the Monte Carlo N-Particle Version 5 (MCNP5) code and continuous energy ENDF/B-V, VI.8, and VII.0, JEFF-3.1, and JENDL-3.3 nuclear data libraries.

  7. Evaluation of Title I ESEA Projects: 1975-76.

    ERIC Educational Resources Information Center

    Philadelphia School District, PA. Office of Research and Evaluation.

    Evaluation services to be provided during 1975-76 to projects funded under the Elementary and Secondary Education Act Title I are listed in this annual booklet. For each project, the following information is provided: goals to be assessed, evaluation techniques (design), and evaluation milestones. Regular term and summer term projects reported on…

  8. Evaluation in Adult Literacy Research. Project ALERT. Phase II.

    ERIC Educational Resources Information Center

    Ntiri, Daphne Williams, Ed.

    This document contains an evaluation handbook for adult literacy programs and feedback from/regarding the evaluation instruments developed during the project titled Adult Literacy and Evaluation Research Team (also known as Project ALERT), a two-phase project initiated by the Detroit Literacy Coalition (DLC) for the purpose of developing and…

  9. Processor Emulator with Benchmark Applications

    SciTech Connect

    Lloyd, G. Scott; Pearce, Roger; Gokhale, Maya

    2015-11-13

    A processor emulator and a suite of benchmark applications have been developed to assist in characterizing the performance of data-centric workloads on current and future computer architectures. Some of the applications have been collected from other open source projects. For more details on the emulator and an example of its usage, see reference [1].

  10. Benchmarking on Tsunami Currents with ComMIT

    NASA Astrophysics Data System (ADS)

    Sharghi vand, N.; Kanoglu, U.

    2015-12-01

    There were no standards for the validation and verification of tsunami numerical models before 2004 Indian Ocean tsunami. Even, number of numerical models has been used for inundation mapping effort, evaluation of critical structures, etc. without validation and verification. After 2004, NOAA Center for Tsunami Research (NCTR) established standards for the validation and verification of tsunami numerical models (Synolakis et al. 2008 Pure Appl. Geophys. 165, 2197-2228), which will be used evaluation of critical structures such as nuclear power plants against tsunami attack. NCTR presented analytical, experimental and field benchmark problems aimed to estimate maximum runup and accepted widely by the community. Recently, benchmark problems were suggested by the US National Tsunami Hazard Mitigation Program Mapping & Modeling Benchmarking Workshop: Tsunami Currents on February 9-10, 2015 at Portland, Oregon, USA (http://nws.weather.gov/nthmp/index.html). These benchmark problems concentrated toward validation and verification of tsunami numerical models on tsunami currents. Three of the benchmark problems were: current measurement of the Japan 2011 tsunami in Hilo Harbor, Hawaii, USA and in Tauranga Harbor, New Zealand, and single long-period wave propagating onto a small-scale experimental model of the town of Seaside, Oregon, USA. These benchmark problems were implemented in the Community Modeling Interface for Tsunamis (ComMIT) (Titov et al. 2011 Pure Appl. Geophys. 168, 2121-2131), which is a user-friendly interface to the validated and verified Method of Splitting Tsunami (MOST) (Titov and Synolakis 1995 J. Waterw. Port Coastal Ocean Eng. 121, 308-316) model and is developed by NCTR. The modeling results are compared with the required benchmark data, providing good agreements and results are discussed. Acknowledgment: The research leading to these results has received funding from the European Union's Seventh Framework Programme (FP7/2007-2013) under grant

  11. Airway Science Curriculum Demonstration Project: Summary of Initial Evaluation Findings

    DTIC Science & Technology

    1988-10-01

    DEMONSTRATION PROJECT: Or C988 SUMMARY OF INITIAL EVALUATION FINDINGS 8. Performn 9 Organ zaton Report No. 7. Author’ s$ Debora L. Clough 9...Airway Science project objectives for which data were available. Two limitations associated with the project evaluation at this time were described... EVALUATION FINDINGS INTRODUCTION The Airway Science Curriculum Demonstration Project was designed to investigate the effectiveness of an alternative approach

  12. Project Tomorrow. Special Project Performance Evaluation. Final Report.

    ERIC Educational Resources Information Center

    Northern Nevada Community Coll., Elko.

    In August 1984, Northern Nevada Community College (NNCC) began a project to develop a practical plan for a program in personal literacy to be incorporated into its Adult Basic Education Program. The first months of the project were spent in structuring the test model, developing intake procedures, research, interagency networking, and coordinating…

  13. [Benchmarking in gastrointestinal endoscopy].

    PubMed

    Rosien, U; Leffmann, C

    2007-12-01

    Quality and costs of endoscopic procedures in gastroenterology have recently come into focus of professionals and patients alike. The project presented here combines benchmarking of indicators of quality and simultaneous assessment of costs in departments of endoscopy. In all 9400 examinations from 14 departments were analysed. First and most obvious results are presented: the machine-readable data sheet was filled out correctly in 95 % (compulsory data fields); endoscopy costs differed by factor two for diagnostic gastroscopy and by factor five for therapeutic ERCP; in some departments a polypectomy was performed in up to every second diagnostic colonoscopy while in others polypectomy occurred only in 10 % in some departments patient monitoring like pulse oxymetry was used only in half of the procedures. Interpretation of the data requires further validation by resumption of the project to obtain more data from possibly more departments.

  14. Cleanroom energy benchmarking results

    SciTech Connect

    Tschudi, William; Xu, Tengfang

    2001-09-01

    A utility market transformation project studied energy use and identified energy efficiency opportunities in cleanroom HVAC design and operation for fourteen cleanrooms. This paper presents the results of this work and relevant observations. Cleanroom owners and operators know that cleanrooms are energy intensive but have little information to compare their cleanroom's performance over time, or to others. Direct comparison of energy performance by traditional means, such as watts/ft{sup 2}, is not a good indicator with the wide range of industrial processes and cleanliness levels occurring in cleanrooms. In this project, metrics allow direct comparison of the efficiency of HVAC systems and components. Energy and flow measurements were taken to determine actual HVAC system energy efficiency. The results confirm a wide variation in operating efficiency and they identify other non-energy operating problems. Improvement opportunities were identified at each of the benchmarked facilities. Analysis of the best performing systems and components is summarized, as are areas for additional investigation.

  15. Global and local scale flood discharge simulations in the Rhine River basin for flood risk reduction benchmarking in the Flagship Project

    NASA Astrophysics Data System (ADS)

    Gädeke, Anne; Gusyev, Maksym; Magome, Jun; Sugiura, Ai; Cullmann, Johannes; Takeuchi, Kuniyoshi

    2015-04-01

    The global flood risk assessment is prerequisite to set global measurable targets of post-Hyogo Framework for Action (HFA) that mobilize international cooperation and national coordination towards disaster risk reduction (DRR) and requires the establishment of a uniform flood risk assessment methodology on various scales. To address these issues, the International Flood Initiative (IFI) has initiated a Flagship Project, which was launched in year 2013, to support flood risk reduction benchmarking at global, national and local levels. In the Flagship Project road map, it is planned to identify the original risk (1), to identify the reduced risk (2), and to facilitate the risk reduction actions (3). In order to achieve this goal at global, regional and local scales, international research collaboration is absolutely necessary involving domestic and international institutes, academia and research networks such as UNESCO International Centres. The joint collaboration by ICHARM and BfG was the first attempt that produced the first step (1a) results on the flood discharge estimates with inundation maps under way. As a result of this collaboration, we demonstrate the outcomes of the first step of the IFI Flagship Project to identify flood hazard in the Rhine river basin on the global and local scale. In our assessment, we utilized a distributed hydrological Block-wise TOP (BTOP) model on 20-km and 0.5-km scales with local precipitation and temperature input data between 1980 and 2004. We utilized existing 20-km BTOP model, which is applied globally, and constructed the local scale 0.5-km BTOP model for the Rhine River basin. For the BTOP model results, both calibrated 20-km and 0.5-km BTOP models had similar statistical performance and represented observed flood river discharges, epecially for 1993 and 1995 floods. From 20-km and 0.5-km BTOP simulation, the flood discharges of the selected return period were estimated using flood frequency analysis and were comparable to

  16. Evaluation and comparison of benchmark QSAR models to predict a relevant REACH endpoint: The bioconcentration factor (BCF)

    SciTech Connect

    Gissi, Andrea; Lombardo, Anna; Roncaglioni, Alessandra; Gadaleta, Domenico; Mangiatordi, Giuseppe Felice; Nicolotti, Orazio; Benfenati, Emilio

    2015-02-15

    }=0.85) and sensitivity (average>0.70) for new compounds in the AD but not present in the training set. However, no single optimal model exists and, thus, it would be wise a case-by-case assessment. Yet, integrating the wealth of information from multiple models remains the winner approach. - Highlights: • REACH encourages the use of in silico methods in the assessment of chemicals safety. • The performances of nine BCF models were evaluated on a benchmark database of 851 chemicals. • We compared the models on the basis of both regression and classification performance. • Statistics on chemicals out of the training set and/or within the applicability domain were compiled. • The results show that QSAR models are useful as weight-of-evidence in support to other methods.

  17. Towards Systematic Benchmarking of Climate Model Performance

    NASA Astrophysics Data System (ADS)

    Gleckler, P. J.

    2014-12-01

    The process by which climate models are evaluated has evolved substantially over the past decade, with the Coupled Model Intercomparison Project (CMIP) serving as a centralizing activity for coordinating model experimentation and enabling research. Scientists with a broad spectrum of expertise have contributed to the CMIP model evaluation process, resulting in many hundreds of publications that have served as a key resource for the IPCC process. For several reasons, efforts are now underway to further systematize some aspects of the model evaluation process. First, some model evaluation can now be considered routine and should not require "re-inventing the wheel" or a journal publication simply to update results with newer models. Second, the benefit of CMIP research to model development has not been optimal because the publication of results generally takes several years and is usually not reproducible for benchmarking newer model versions. And third, there are now hundreds of model versions and many thousands of simulations, but there is no community-based mechanism for routinely monitoring model performance changes. An important change in the design of CMIP6 can help address these limitations. CMIP6 will include a small set standardized experiments as an ongoing exercise (CMIP "DECK": ongoing Diagnostic, Evaluation and Characterization of Klima), so that modeling groups can submit them at any time and not be overly constrained by deadlines. In this presentation, efforts to establish routine benchmarking of existing and future CMIP simulations will be described. To date, some benchmarking tools have been made available to all CMIP modeling groups to enable them to readily compare with CMIP5 simulations during the model development process. A natural extension of this effort is to make results from all CMIP simulations widely available, including the results from newer models as soon as the simulations become available for research. Making the results from routine

  18. Framework for the Evaluation of an IT Project Portfolio

    ERIC Educational Resources Information Center

    Tai, W. T.

    2010-01-01

    The basis for evaluating projects in an organizational IT project portfolio includes complexity factors, arguments/criteria, and procedures, with various implications. The purpose of this research was to develop a conceptual framework for IT project proposal evaluation. The research involved using a heuristic roadmap and the mind-mapping method to…

  19. Framework for the Evaluation of an IT Project Portfolio

    ERIC Educational Resources Information Center

    Tai, W. T.

    2010-01-01

    The basis for evaluating projects in an organizational IT project portfolio includes complexity factors, arguments/criteria, and procedures, with various implications. The purpose of this research was to develop a conceptual framework for IT project proposal evaluation. The research involved using a heuristic roadmap and the mind-mapping method to…

  20. Modelling in Evaluating a Working Life Project in Higher Education

    ERIC Educational Resources Information Center

    Sarja, Anneli; Janhonen, Sirpa; Havukainen, Pirjo; Vesterinen, Anne

    2012-01-01

    This article describes an evaluation method based on collaboration between the higher education, a care home and university, in a R&D project. The aim of the project was to elaborate modelling as a tool of developmental evaluation for innovation and competence in project cooperation. The approach was based on activity theory. Modelling enabled a…

  1. Modelling in Evaluating a Working Life Project in Higher Education

    ERIC Educational Resources Information Center

    Sarja, Anneli; Janhonen, Sirpa; Havukainen, Pirjo; Vesterinen, Anne

    2012-01-01

    This article describes an evaluation method based on collaboration between the higher education, a care home and university, in a R&D project. The aim of the project was to elaborate modelling as a tool of developmental evaluation for innovation and competence in project cooperation. The approach was based on activity theory. Modelling enabled a…

  2. Critical evaluation of climate syntheses to benchmark CMIP6/PMIP4 127 ka Last Interglacial simulations in the high-latitude regions

    NASA Astrophysics Data System (ADS)

    Capron, E.; Govin, A.; Feng, R.; Otto-Bliesner, B. L.; Wolff, E. W.

    2017-07-01

    The Last Interglacial (LIG, ∼129-116 thousand years ago, ka) represents an excellent case study to investigate the response of sensitive components of the Earth System and mechanisms of high-latitude amplification to a climate warmer than present-day. The Paleoclimate Model Intercomparison Project (Phase 4, hereafter referred as PMIP4) and the Coupled Model Intercomparison Project (Phase 6, hereafter referred as CMIP6) are coordinating the design of (1) a LIG Tier 1 equilibrium simulation to simulate the climate response at 127 ka, a time interval associated with a strong orbital forcing and greenhouse gas concentrations close to preindustrial levels and (2) associated Tier 2 sensitivity experiments to examine the role of the ocean, vegetation and dust feedbacks in modulating the response to this orbital forcing. Evaluating the capability of the CMIP6/PMIP4 models to reproduce the 127 ka polar and sub-polar climate will require appropriate data-based benchmarks which are currently missing. Based on a recent data synthesis that offers the first spatio-temporal representation of high-latitude (i.e. poleward of 40°N and 40°S) surface temperature evolution during the LIG, we produce a new 126-128 ka time slab, hereafter named 127 ka time slice. This 127 ka time slice represents surface temperature anomalies relative to preindustrial and is associated with quantitative estimates of the uncertainties related to relative dating and surface temperature reconstruction methods. It illustrates warmer-than-preindustrial conditions in the high-latitude regions of both hemispheres. In particular, summer sea surface temperatures (SST) in the North Atlantic region were on average 1.1 °C (with a standard error of the mean of 0.7 °C) warmer relative to preindustrial and 1.8 °C (with a standard error of the mean of 0.8 °C) in the Southern Ocean. In Antarctica, average 127 ka annual surface air temperature was 2.2 °C (with a standard error of the mean of 1.4 °C) warmer

  3. RESEARCH DESIGN FOR EVALUATING PROJECT MISSION.

    ERIC Educational Resources Information Center

    FURNO, ORLANDO F.; AND OTHERS

    THIS REPORT OUTLINES DESIGNS FOR 8 POSSIBLE RESEARCH STUDIES WHICH COULD BE UNDERTAKEN WITH REGARD TO PROJECT MISSION, A PROGRAM TO PREPARE TEACHERS FOR ASSIGNMENT TO INNER CITY SCHOOLS. THEY ARE (1) A STUDY OF ATTRITION RATES OF STUDENT-INTERN-TEACHER ENROLLEES IN TRAINING IN PROJECT MISSION, (2) TEACHER CHARACTERISTICS OF PROJECT MISSION INTERNS…

  4. BN-600 full MOX core benchmark analysis.

    SciTech Connect

    Kim, Y. I.; Hill, R. N.; Grimm, K.; Rimpault, G.; Newton, T.; Li, Z. H.; Rineiski, A.; Mohanakrishan, P.; Ishikawa, M.; Lee, K. B.; Danilytchev, A.; Stogov, V.; Nuclear Engineering Division; International Atomic Energy Agency; CEA SERCO Assurance; China Inst. of Atomic Energy; Forschnungszentrum Karlsruhe; Indira Gandhi Centre for Atomic Research; Japan Nuclear Cycle Development Inst.; Korea Atomic Energy Research Inst.; Inst. of Physics and Power Engineering

    2004-01-01

    As a follow-up of the BN-600 hybrid core benchmark, a full MOX core benchmark was performed within the framework of the IAEA co-ordinated research project. Discrepancies between the values of main reactivity coefficients obtained by the participants for the BN-600 full MOX core benchmark appear to be larger than those in the previous hybrid core benchmarks on traditional core configurations. This arises due to uncertainties in the proper modelling of the axial sodium plenum above the core. It was recognized that the sodium density coefficient strongly depends on the core model configuration of interest (hybrid core vs. fully MOX fuelled core with sodium plenum above the core) in conjunction with the calculation method (diffusion vs. transport theory). The effects of the discrepancies revealed between the participants results on the ULOF and UTOP transient behaviours of the BN-600 full MOX core were investigated in simplified transient analyses. Generally the diffusion approximation predicts more benign consequences for the ULOF accident but more hazardous ones for the UTOP accident when compared with the transport theory results. The heterogeneity effect does not have any significant effect on the simulation of the transient. The comparison of the transient analyses results concluded that the fuel Doppler coefficient and the sodium density coefficient are the two most important coefficients in understanding the ULOF transient behaviour. In particular, the uncertainty in evaluating the sodium density coefficient distribution has the largest impact on the description of reactor dynamics. This is because the maximum sodium temperature rise takes place at the top of the core and in the sodium plenum.

  5. Computational Chemistry Comparison and Benchmark Database

    National Institute of Standards and Technology Data Gateway

    SRD 101 NIST Computational Chemistry Comparison and Benchmark Database (Web, free access)   The NIST Computational Chemistry Comparison and Benchmark Database is a collection of experimental and ab initio thermochemical properties for a selected set of molecules. The goals are to provide a benchmark set of molecules for the evaluation of ab initio computational methods and allow the comparison between different ab initio computational methods for the prediction of thermochemical properties.

  6. All inclusive benchmarking.

    PubMed

    Ellis, Judith

    2006-07-01

    The aim of this article is to review published descriptions of benchmarking activity and synthesize benchmarking principles to encourage the acceptance and use of Essence of Care as a new benchmarking approach to continuous quality improvement, and to promote its acceptance as an integral and effective part of benchmarking activity in health services. The Essence of Care, was launched by the Department of Health in England in 2001 to provide a benchmarking tool kit to support continuous improvement in the quality of fundamental aspects of health care, for example, privacy and dignity, nutrition and hygiene. The tool kit is now being effectively used by some frontline staff. However, use is inconsistent, with the value of the tool kit, or the support clinical practice benchmarking requires to be effective, not always recognized or provided by National Health Service managers, who are absorbed with the use of quantitative benchmarking approaches and measurability of comparative performance data. This review of published benchmarking literature, was obtained through an ever-narrowing search strategy commencing from benchmarking within quality improvement literature through to benchmarking activity in health services and including access to not only published examples of benchmarking approaches and models used but the actual consideration of web-based benchmarking data. This supported identification of how benchmarking approaches have developed and been used, remaining true to the basic benchmarking principles of continuous improvement through comparison and sharing (Camp 1989). Descriptions of models and exemplars of quantitative and specifically performance benchmarking activity in industry abound (Camp 1998), with far fewer examples of more qualitative and process benchmarking approaches in use in the public services and then applied to the health service (Bullivant 1998). The literature is also in the main descriptive in its support of the effectiveness of

  7. Design Alternatives for Evaluating the Impact of Conservation Projects

    ERIC Educational Resources Information Center

    Margoluis, Richard; Stem, Caroline; Salafsky, Nick; Brown, Marcia

    2009-01-01

    Historically, examples of project evaluation in conservation were rare. In recent years, however, conservation professionals have begun to recognize the importance of evaluation both for accountability and for improving project interventions. Even with this growing interest in evaluation, the conservation community has paid little attention to…

  8. Alternate Methods for Assuring Credibility of Research and Evaluation Findings in Project Evaluation.

    ERIC Educational Resources Information Center

    Denton, William T.; Murray, Wayne R.

    This paper describes six existing evaluator-auditor working formats and the conditions which foster credibility of evaluation findings. Evaluators were classified as: (1) member of project developmental team, accountable to project director; (2) independent internal evaluator, accountable to system in general but not to project directors, and (3)…

  9. Evolving Our Evaluation of Lighting Environments Project

    NASA Technical Reports Server (NTRS)

    Terrier, Douglas; Clayton, Ronald; Clark, Toni Anne

    2016-01-01

    Imagine you are an astronaut on their 100th day of your three year exploration mission. During your daily routine to the small hygiene compartment of the spacecraft, you realize that no matter what you do, your body blocks the light from the lamp. You can clearly see your hands or your toes but not both! What were those design engineers thinking! It would have been nice if they could have made the walls glow instead! The reason the designers were not more innovative is that their interpretation of the system lighting requirements didn't allow them to be so! Currently, our interior spacecraft lighting standards and requirements are written around the concept of a quantity of light illuminating a spacecraft surface. The natural interpretation for the engineer is that a lamp that throws light to the surface is required. Because of certification costs, only one lamp is designed and small rooms can wind up with lamps that may be inappropriate for the room architecture. The advances in solid state light emitting technologies and optics for lighting and visual communication necessitates the evaluation of how NASA envisions spacecraft lighting architectures and how NASA uses industry standards for the design and evaluation of lighting system. Current NASA lighting standards and requirements for existing architectures focus on the separate ability of a lighting system to throw light against a surface or the ability of a display system to provide the appropriate visual contrast. Realization that these systems can be integrated is not realized. The result is that the systems are developed independent from one another and potential efficiencies that could be realized from borrowing from the concept of one technology and applying it for the purpose of the other does not occur. This project investigated the possibility of incorporating large luminous surface lamps as an alternative or supplement to overhead lighting. We identified existing industry standards for architectural

  10. Benchmarking concentrating photovoltaic systems

    NASA Astrophysics Data System (ADS)

    Duerr, Fabian; Muthirayan, Buvaneshwari; Meuret, Youri; Thienpont, Hugo

    2010-08-01

    Integral to photovoltaics is the need to provide improved economic viability. To achieve this goal, photovoltaic technology has to be able to harness more light at less cost. A large variety of concentrating photovoltaic concepts has provided cause for pursuit. To obtain a detailed profitability analysis, a flexible evaluation is crucial for benchmarking the cost-performance of this variety of concentrating photovoltaic concepts. To save time and capital, a way to estimate the cost-performance of a complete solar energy system is to use computer aided modeling. In this work a benchmark tool is introduced based on a modular programming concept. The overall implementation is done in MATLAB whereas Advanced Systems Analysis Program (ASAP) is used for ray tracing calculations. This allows for a flexible and extendable structuring of all important modules, namely an advanced source modeling including time and local dependence, and an advanced optical system analysis of various optical designs to obtain an evaluation of the figure of merit. An important figure of merit: the energy yield for a given photovoltaic system at a geographical position over a specific period, can be calculated.

  11. A BENCHMARK PROGRAM FOR EVALUATION OF METHODS FOR COMPUTING SEISMIC RESPONSE OF COUPLED BUILDING-PIPING/EQUIPMENT WITH NON-CLASSICAL DAMPING.

    SciTech Connect

    Xu, J.; Degrassi, G.; Chokshi, N.

    2001-03-22

    Under the auspices of the US Nuclear Regulatory Commission (NRC), Brookhaven National Laboratory (BNL) developed a comprehensive program to evaluate state-of-the-art methods and computer programs for seismic analysis of typical coupled nuclear power plant (NPP) systems with nonclassical damping. In this program, four benchmark models of coupled building-piping/equipment systems with different damping characteristics were analyzed for a suite of earthquakes by program participants applying their uniquely developed methods and computer programs. This paper presents the results of their analyses, and their comparison to the benchmark solutions generated by BNL using time domain direct integration methods. The participant's analysis results established using complex modal time history methods showed good comparison with the BNL solutions, while the analyses produced with either complex-mode response spectrum methods or classical normal-mode response spectrum method, in general, produced more conservative results, when averaged over a suite of earthquakes. However, when coupling due to damping is significant, complex-mode response spectrum methods performed better than the classical normal-mode response spectrum method. Furthermore, as part of the program objectives, a parametric assessment is also presented in this paper, aimed at evaluation of the applicability of various analysis methods to problems with different dynamic characteristics unique to coupled NPP systems. It is believed that the findings and insights learned from this program will be useful in developing new acceptance criteria and providing guidance for future regulatory activities involving licensing applications of these alternate methods to coupled systems.

  12. Benchmarking for Higher Education.

    ERIC Educational Resources Information Center

    Jackson, Norman, Ed.; Lund, Helen, Ed.

    The chapters in this collection explore the concept of benchmarking as it is being used and developed in higher education (HE). Case studies and reviews show how universities in the United Kingdom are using benchmarking to aid in self-regulation and self-improvement. The chapters are: (1) "Introduction to Benchmarking" (Norman Jackson…

  13. Benchmarking: The New Tool.

    ERIC Educational Resources Information Center

    Stralser, Steven

    1995-01-01

    This article suggests that benchmarking, the process of comparing one's own operation with the very best, can be used to make improvements in colleges and universities. Six steps are outlined: determining what to benchmark, forming a team, discovering who to benchmark, collecting and analyzing data, using the data to redesign one's own operation,…

  14. Benchmarking for Higher Education.

    ERIC Educational Resources Information Center

    Jackson, Norman, Ed.; Lund, Helen, Ed.

    The chapters in this collection explore the concept of benchmarking as it is being used and developed in higher education (HE). Case studies and reviews show how universities in the United Kingdom are using benchmarking to aid in self-regulation and self-improvement. The chapters are: (1) "Introduction to Benchmarking" (Norman Jackson…

  15. Benchmarking the performance of daily temperature homogenisation algorithms

    NASA Astrophysics Data System (ADS)

    Warren, Rachel; Bailey, Trevor; Jolliffe, Ian; Willett, Kate

    2015-04-01

    This work explores the creation of realistic synthetic data and its use as a benchmark for comparing the performance of different homogenisation algorithms on daily temperature data. Four different regions in the United States have been selected and three different inhomogeneity scenarios explored for each region. These benchmark datasets are beneficial as, unlike in the real world, the underlying truth is known a priori, thus allowing definite statements to be made about the performance of the algorithms run on them. Performance can be assessed in terms of the ability of algorithms to detect changepoints and also their ability to correctly remove inhomogeneities. The focus is on daily data, thus presenting new challenges in comparison to monthly data and pushing the boundaries of previous studies. The aims of this work are to evaluate and compare the performance of various homogenisation algorithms, aiding their improvement and enabling a quantification of the uncertainty remaining in the data even after they have been homogenised. An important outcome is also to evaluate how realistic the created benchmarks are. It is essential that any weaknesses in the benchmarks are taken into account when judging algorithm performance against them. This information in turn will help to improve future versions of the benchmarks. I intend to present a summary of this work including the method of benchmark creation, details of the algorithms run and some preliminary results. This work forms a three year PhD and feeds into the larger project of the International Surface Temperature Initiative which is working on a global scale and with monthly instead of daily data.

  16. Wais-III norms for working-age adults: a benchmark for conducting vocational, career, and employment-related evaluations.

    PubMed

    Fjordbak, Timothy; Fjordbak, Bess Sirmon

    2005-02-01

    The Wechsler Intelligence Scales are routinely used to assess threshold variables which correlate with subsequent job performance. Intellectual testing within educational and clinical settings accommodates natural developmental changes by referencing results to restricted age-band norms. However, accuracy in vocational and career consultation, as well as equity in hiring and promotion requires the application of a single normative benchmark unbiased by chronological age. Such unitary norms for working-age adults (18- to 64-yr.-olds) were derived from the WAIS-III standardization sample in accord with the proportional representation of the seven age-bands subsumed within this age range. Tabular summaries of results are given for the conversion of raw scores to scaled scores for the working-age population which can be used to derive IQ values and Index Scores.

  17. Evaluation of the School Administration Manager Project

    ERIC Educational Resources Information Center

    Turnbull, Brenda J.; Haslam, M. Bruce; Arcaira, Erikson R.; Riley, Derek L.; Sinclair, Beth; Coleman, Stephen

    2009-01-01

    The School Administration Manager (SAM) project, supported by The Wallace Foundation as part of its education initiative, focuses on changing the conditions in schools that prevent principals from devoting more time to instructional leadership. In schools participating in the National SAM Project, principals have made a commitment to increase the…

  18. Human Relations Education Project. Final Evaluation Report.

    ERIC Educational Resources Information Center

    Buffalo Board of Education, NY.

    This project did the planning and pilot phases of an effort to improve the teaching of human relations in grades K-12 of public and private schools in the Buffalo-Niagara Falls metropolitan area. In the pilot phase, the project furnished on-the-job training for approximately 70 schools. The training was given by teams of human relations…

  19. Evaluation of the Matrix Project. Interchange 77.

    ERIC Educational Resources Information Center

    McIvor, Gill; Moodie, Kristina

    The Matrix Project is a program that has been established in central Scotland with the aim of reducing the risk of offending and anti-social behavior among vulnerable children. The project provides a range of services to children between eight and 11 years of age who are at risk in the local authority areas of Clackmannanshire, Falkirk and…

  20. Evaluation of the Matrix Project. Interchange 77.

    ERIC Educational Resources Information Center

    McIvor, Gill; Moodie, Kristina

    The Matrix Project is a program that has been established in central Scotland with the aim of reducing the risk of offending and anti-social behavior among vulnerable children. The project provides a range of services to children between eight and 11 years of age who are at risk in the local authority areas of Clackmannanshire, Falkirk and…

  1. Project Aprendizaje. 1990-91 Final Evaluation Profile. OREA Report.

    ERIC Educational Resources Information Center

    New York City Board of Education, Brooklyn, NY. Office of Research, Evaluation, and Assessment.

    An evaluation was done of New York City Public Schools' Project Aprendizaje, which served disadvantaged, immigrant, Spanish-speaking high school students at Seward Park High School in Manhattan. The Project enrolled 290 students in grades 9 through 12, 93.1 percent of whom were eligible for the Free Lunch Program. The Project provided students of…

  2. Project T.E.A.C.H.: An Evaluative Study.

    ERIC Educational Resources Information Center

    Howarth, Les

    A survey of 17 graduates of Project T.E.A.C.H. (Teacher Effectiveness and Classroom Handling), an inservice education program offered through the Ontario (Canada) Public School Men Teacher's Association in conjunction with Lesley College, used closed- and open-ended questions to obtain evaluations of the project's effectiveness. Five project areas…

  3. PLATO across the Curriculum: An Evaluation of a Project.

    ERIC Educational Resources Information Center

    Freer, David

    1986-01-01

    A project at the University of Witwatersrand examined the implications of introducing a centrally controlled system of computer-based learning in which 13 university departments utilized PLATO to supplement teaching programs and encourage computer literacy. Department project descriptions and project evaluations (which reported positive student…

  4. Evaluation of the Appalachian Regional Commission's Educational Projects: Final Report.

    ERIC Educational Resources Information Center

    Silverstein, Gary; Bartfai, Nicole; Plishker, Laurie; Snow, Kyle; Frechtling, Joy

    This report presents findings from an evaluation of 84 educational projects funded by the Appalachian Regional Commission (ARC) during the 1990's. Data were collected via document reviews, interviews, a mail survey completed by 78 projects, and eight site visits. Most projects provided services to rural areas or community segments most in need.…

  5. PLATO across the Curriculum: An Evaluation of a Project.

    ERIC Educational Resources Information Center

    Freer, David

    1986-01-01

    A project at the University of Witwatersrand examined the implications of introducing a centrally controlled system of computer-based learning in which 13 university departments utilized PLATO to supplement teaching programs and encourage computer literacy. Department project descriptions and project evaluations (which reported positive student…

  6. Outside Evaluation Report for the Arlington Federal Workplace Literacy Project.

    ERIC Educational Resources Information Center

    Wrigley, Heide Spruck

    The successes and challenges of the Arlington Education and Employment Program (REEP) Workplace Literacy Project in Virginia are described in this evaluation report. REEP's federal Workplace Literacy Project Consortium is operated as a special project within the Department of Adult, Career and Vocational Education of the Arlington Public Schools.…

  7. Special Education Evaluation Project for University Affiliated Facilities. Final Report.

    ERIC Educational Resources Information Center

    Burrello, Leonard C.; And Others

    The final report of the Special Education Evaluation Project for 16 University Affiliated Facilities (UAF) centers provides a chronological review of the project which focused on the training of persons to work with mentally retarded or other handicapped individuals. Outlined are project objectives including the development of descriptors useful…

  8. Social Studies Project Evaluation: Case Study and Recommendations.

    ERIC Educational Resources Information Center

    Napier, John

    1982-01-01

    Describes the development and application of a model for social studies program evaluations. A case study showing how the model's three-step process was used to evaluate the Improving Citizenship Education Project in Fulton County, Georgia is included. (AM)

  9. Social Studies Project Evaluation: Case Study and Recommendations.

    ERIC Educational Resources Information Center

    Napier, John

    1982-01-01

    Describes the development and application of a model for social studies program evaluations. A case study showing how the model's three-step process was used to evaluate the Improving Citizenship Education Project in Fulton County, Georgia is included. (AM)

  10. Service profiling and outcomes benchmarking using the CORE-OM: toward practice-based evidence in the psychological therapies. Clinical Outcomes in Routine Evaluation-Outcome Measures.

    PubMed

    Barkham, M; Margison, F; Leach, C; Lucock, M; Mellor-Clark, J; Evans, C; Benson, L; Connell, J; Audin, K; McGrath, G

    2001-04-01

    To complement the evidence-based practice paradigm, the authors argued for a core outcome measure to provide practice-based evidence for the psychological therapies. Utility requires instruments that are acceptable scientifically, as well as to service users, and a coordinated implementation of the measure at a national level. The development of the Clinical Outcomes in Routine Evaluation-Outcome Measure (CORE-OM) is summarized. Data are presented across 39 secondary-care services (n = 2,710) and within an intensively evaluated single service (n = 1,455). Results suggest that the CORE-OM is a valid and reliable measure for multiple settings and is acceptable to users and clinicians as well as policy makers. Baseline data levels of patient presenting problem severity, including risk, are reported in addition to outcome benchmarks that use the concept of reliable and clinically significant change. Basic quality improvement in outcomes for a single service is considered.

  11. Benchmark simulation models, quo vadis?

    PubMed

    Jeppsson, U; Alex, J; Batstone, D J; Benedetti, L; Comas, J; Copp, J B; Corominas, L; Flores-Alsina, X; Gernaey, K V; Nopens, I; Pons, M-N; Rodríguez-Roda, I; Rosen, C; Steyer, J-P; Vanrolleghem, P A; Volcke, E I P; Vrecko, D

    2013-01-01

    As the work of the IWA Task Group on Benchmarking of Control Strategies for wastewater treatment plants (WWTPs) is coming to an end, it is essential to disseminate the knowledge gained. For this reason, all authors of the IWA Scientific and Technical Report on benchmarking have come together to provide their insights, highlighting areas where knowledge may still be deficient and where new opportunities are emerging, and to propose potential avenues for future development and application of the general benchmarking framework and its associated tools. The paper focuses on the topics of temporal and spatial extension, process modifications within the WWTP, the realism of models, control strategy extensions and the potential for new evaluation tools within the existing benchmark system. We find that there are major opportunities for application within all of these areas, either from existing work already being done within the context of the benchmarking simulation models (BSMs) or applicable work in the wider literature. Of key importance is increasing capability, usability and transparency of the BSM package while avoiding unnecessary complexity.

  12. Closed-Loop Neuromorphic Benchmarks

    PubMed Central

    Stewart, Terrence C.; DeWolf, Travis; Kleinhans, Ashley; Eliasmith, Chris

    2015-01-01

    Evaluating the effectiveness and performance of neuromorphic hardware is difficult. It is even more difficult when the task of interest is a closed-loop task; that is, a task where the output from the neuromorphic hardware affects some environment, which then in turn affects the hardware's future input. However, closed-loop situations are one of the primary potential uses of neuromorphic hardware. To address this, we present a methodology for generating closed-loop benchmarks that makes use of a hybrid of real physical embodiment and a type of “minimal” simulation. Minimal simulation has been shown to lead to robust real-world performance, while still maintaining the practical advantages of simulation, such as making it easy for the same benchmark to be used by many researchers. This method is flexible enough to allow researchers to explicitly modify the benchmarks to identify specific task domains where particular hardware excels. To demonstrate the method, we present a set of novel benchmarks that focus on motor control for an arbitrary system with unknown external forces. Using these benchmarks, we show that an error-driven learning rule can consistently improve motor control performance across a randomly generated family of closed-loop simulations, even when there are up to 15 interacting joints to be controlled. PMID:26696820

  13. Project SEARCH UK--Evaluating Its Employment Outcomes

    ERIC Educational Resources Information Center

    Kaehne, Axel

    2016-01-01

    Background: The study reports the findings of an evaluation of Project SEARCH UK. The programme develops internships for young people with intellectual disabilities who are about to leave school or college. The aim of the evaluation was to investigate at what rate Project SEARCH provided employment opportunities to participants. Methods: The…

  14. Kentucky Migrant Technology Project: External Evaluation Report, 1997-98.

    ERIC Educational Resources Information Center

    Popp, Robert J.

    During its first year of operation (1997-98), the Kentucky Migrant Technology Project successfully implemented its model, used internal and external evaluations to inform improvement of the model, and began plans for expansion into new service areas. This evaluation report is organized around five questions that focus on the project model and its…

  15. Evaluation in Adult Literacy Research. Project ALERT. [Phase I.

    ERIC Educational Resources Information Center

    Ntiri, Daphne Williams, Ed.

    The Adult Literacy and Evaluation Research Team (also known as Project ALERT) was a project conducted by the Detroit Literacy Coalition (DLC) at Wayne State University in 1993-1994 to develop and pilot a user-friendly program model for evaluating literacy operations of community-based organizations throughout Michigan under the provisions of…

  16. The Program Evaluator's Role in Cross-Project Pollination.

    ERIC Educational Resources Information Center

    Yasgur, Bruce J.

    An expanded duties role of the multiple-program evaluator as an integral part of the ongoing decision-making process in all projects served is defended. Assumptions discussed included that need for projects with related objectives to pool resources and avoid duplication of effort and the evaluator's unique ability to provide an objective…

  17. Student Assistance Program Demonstration Project Evaluation. Final Report.

    ERIC Educational Resources Information Center

    Pollard, John A.; Houle, Denise M.

    This document presents the final report on the evaluation of California's model student assistance program (SAP) demonstration projects implemented in five locations across the state from July 1989 through June 1992. The report provides an overall, integrated review of the evaluation of the SAP demonstration projects, summarizes important findings…

  18. Project SEARCH UK--Evaluating Its Employment Outcomes

    ERIC Educational Resources Information Center

    Kaehne, Axel

    2016-01-01

    Background: The study reports the findings of an evaluation of Project SEARCH UK. The programme develops internships for young people with intellectual disabilities who are about to leave school or college. The aim of the evaluation was to investigate at what rate Project SEARCH provided employment opportunities to participants. Methods: The…

  19. Fuzzy Present Value Analysis Model for Evaluating Information System Projects

    SciTech Connect

    Omitaomu, Olufemi A; Badiru, Adedeji B

    2007-01-01

    In this article, the economic evaluation of information system projects using present value is analyzed based on triangular fuzzy numbers. Information system projects usually have numerous uncertainties and several conditions of risk that make their economic evaluation a challenging task. Each year, several information system projects are cancelled before completion as a result of budget overruns at a cost of several billions of dollars to industry. Although engineering economic analysis offers tools and techniques for evaluating risky projects, the tools are not enough to place information system projects on a safe budget/selection track. There is a need for an integrative economic analysis model that will account for the uncertainties in estimating project costs, benefits, and useful lives of uncertain and risky projects. In this study, we propose an approximate method of computing project present value using the concept of fuzzy modeling with special reference to information system projects. This proposed model has the potential of enhancing the project selection process by capturing a better economic picture of the project alternatives. The proposed methodology can also be used for other real-life projects with high degree of uncertainty and risk.

  20. Evaluation of direct-use-project drilling costs

    SciTech Connect

    Dolenc, M.R.; Childs, F.W.; Allman, D.W.; Sanders, R.D.

    1983-01-01

    This study evaluates drilling and completion costs from eleven low-to-moderate temperature geothermal projects carried out under the Program Opportunity Notice (PON) and User-Coupled Confirmation Drilling Programs. Several studies have evaluated geothermal drilling costs, particularly with respect to high-temperature-system drilling costs. This study evaluates drilling costs and individual cost elements for low-to-moderate temperature projects. It considers the effect of drilling depth, rock types, remoteness of location, rig size, and unique operating and subsurface conditions on the total drilling cost. This detailed evaluation should provide the investor in direct-use projects with approximate cost projections by which the economics of such projects can be evaluated.

  1. Authentic e-Learning in a Multicultural Context: Virtual Benchmarking Cases from Five Countries

    ERIC Educational Resources Information Center

    Leppisaari, Irja; Herrington, Jan; Vainio, Leena; Im, Yeonwook

    2013-01-01

    The implementation of authentic learning elements at education institutions in five countries, eight online courses in total, is examined in this paper. The International Virtual Benchmarking Project (2009-2010) applied the elements of authentic learning developed by Herrington and Oliver (2000) as criteria to evaluate authenticity. Twelve…

  2. Authentic e-Learning in a Multicultural Context: Virtual Benchmarking Cases from Five Countries

    ERIC Educational Resources Information Center

    Leppisaari, Irja; Herrington, Jan; Vainio, Leena; Im, Yeonwook

    2013-01-01

    The implementation of authentic learning elements at education institutions in five countries, eight online courses in total, is examined in this paper. The International Virtual Benchmarking Project (2009-2010) applied the elements of authentic learning developed by Herrington and Oliver (2000) as criteria to evaluate authenticity. Twelve…

  3. International benchmarking of specialty hospitals. A series of case studies on comprehensive cancer centres

    PubMed Central

    2010-01-01

    Background Benchmarking is one of the methods used in business that is applied to hospitals to improve the management of their operations. International comparison between hospitals can explain performance differences. As there is a trend towards specialization of hospitals, this study examines the benchmarking process and the success factors of benchmarking in international specialized cancer centres. Methods Three independent international benchmarking studies on operations management in cancer centres were conducted. The first study included three comprehensive cancer centres (CCC), three chemotherapy day units (CDU) were involved in the second study and four radiotherapy departments were included in the final study. Per multiple case study a research protocol was used to structure the benchmarking process. After reviewing the multiple case studies, the resulting description was used to study the research objectives. Results We adapted and evaluated existing benchmarking processes through formalizing stakeholder involvement and verifying the comparability of the partners. We also devised a framework to structure the indicators to produce a coherent indicator set and better improvement suggestions. Evaluating the feasibility of benchmarking as a tool to improve hospital processes led to mixed results. Case study 1 resulted in general recommendations for the organizations involved. In case study 2, the combination of benchmarking and lean management led in one CDU to a 24% increase in bed utilization and a 12% increase in productivity. Three radiotherapy departments of case study 3, were considering implementing the recommendations. Additionally, success factors, such as a well-defined and small project scope, partner selection based on clear criteria, stakeholder involvement, simple and well-structured indicators, analysis of both the process and its results and, adapt the identified better working methods to the own setting, were found. Conclusions The improved

  4. International benchmarking of specialty hospitals. A series of case studies on comprehensive cancer centres.

    PubMed

    van Lent, Wineke A M; de Beer, Relinde D; van Harten, Wim H

    2010-08-31

    Benchmarking is one of the methods used in business that is applied to hospitals to improve the management of their operations. International comparison between hospitals can explain performance differences. As there is a trend towards specialization of hospitals, this study examines the benchmarking process and the success factors of benchmarking in international specialized cancer centres. Three independent international benchmarking studies on operations management in cancer centres were conducted. The first study included three comprehensive cancer centres (CCC), three chemotherapy day units (CDU) were involved in the second study and four radiotherapy departments were included in the final study. Per multiple case study a research protocol was used to structure the benchmarking process. After reviewing the multiple case studies, the resulting description was used to study the research objectives. We adapted and evaluated existing benchmarking processes through formalizing stakeholder involvement and verifying the comparability of the partners. We also devised a framework to structure the indicators to produce a coherent indicator set and better improvement suggestions. Evaluating the feasibility of benchmarking as a tool to improve hospital processes led to mixed results. Case study 1 resulted in general recommendations for the organizations involved. In case study 2, the combination of benchmarking and lean management led in one CDU to a 24% increase in bed utilization and a 12% increase in productivity. Three radiotherapy departments of case study 3, were considering implementing the recommendations.Additionally, success factors, such as a well-defined and small project scope, partner selection based on clear criteria, stakeholder involvement, simple and well-structured indicators, analysis of both the process and its results and, adapt the identified better working methods to the own setting, were found. The improved benchmarking process and the success

  5. Radiation Detection Computational Benchmark Scenarios

    SciTech Connect

    Shaver, Mark W.; Casella, Andrew M.; Wittman, Richard S.; McDonald, Ben S.

    2013-09-24

    Modeling forms an important component of radiation detection development, allowing for testing of new detector designs, evaluation of existing equipment against a wide variety of potential threat sources, and assessing operation performance of radiation detection systems. This can, however, result in large and complex scenarios which are time consuming to model. A variety of approaches to radiation transport modeling exist with complementary strengths and weaknesses for different problems. This variety of approaches, and the development of promising new tools (such as ORNL’s ADVANTG) which combine benefits of multiple approaches, illustrates the need for a means of evaluating or comparing different techniques for radiation detection problems. This report presents a set of 9 benchmark problems for comparing different types of radiation transport calculations, identifying appropriate tools for classes of problems, and testing and guiding the development of new methods. The benchmarks were drawn primarily from existing or previous calculations with a preference for scenarios which include experimental data, or otherwise have results with a high level of confidence, are non-sensitive, and represent problem sets of interest to NA-22. From a technical perspective, the benchmarks were chosen to span a range of difficulty and to include gamma transport, neutron transport, or both and represent different important physical processes and a range of sensitivity to angular or energy fidelity. Following benchmark identification, existing information about geometry, measurements, and previous calculations were assembled. Monte Carlo results (MCNP decks) were reviewed or created and re-run in order to attain accurate computational times and to verify agreement with experimental data, when present. Benchmark information was then conveyed to ORNL in order to guide testing and development of hybrid calculations. The results of those ADVANTG calculations were then sent to PNNL for

  6. What NSF Expects in Project Evaluations for Educational Innovations.

    ERIC Educational Resources Information Center

    Hannah, Judith L.

    1996-01-01

    The National Science Foundation (NSF) sponsors a range of programs to fund innovative approaches to teaching and learning. Focuses on NSF's expectations for project evaluation beginning with a definition of evaluation and a discussion of why evaluation is needed. Also describes planning, formative, and summative evaluation stages and concludes…

  7. Proposed re-evaluation of the 154Eu thermal (n, γ) capture cross-section based on spent fuel benchmarking studies

    DOE PAGES

    Skutnik, Steven E.

    2016-09-22

    154Eu is a nuclide of considerable importance to both non-destructive measurements of used nuclear fuel assembly burnup as well as for calculating the radiation source term for used fuel storage and transportation. But, recent evidence from code validation studies of spent fuel benchmarks have revealed evidence of a systemic bias in predicted 154Eu inventories when using ENDF/B-VII.0 and ENDF/B-VII.1 nuclear data libraries, wherein Eu-154 is consistently over-predicted on the order of 10% or more. Further, this bias is found to correlate with sample burnup, resulting in a larger departure from experimental measurements for higher sample burnups. Here, the bias in Eu-154 is characterized across eleven spent fuel destructive assay benchmarks from five different assemblies. Based on these studies, possible amendments to the ENDF/B-VII.0 and VII.1 evaluations of the 154Eu (n,γ)155Eu are explored. By amending the location of the first resolved resonance for the 154Eu radiative capture cross-section (centered at 0.195 eV in ENDF/B-VII.0 and VII.1) to 0.188 eV and adjusting the neutron capture width proportional tomore » $$\\sqrt1/E$$, the amended cross-section evaluation was found to reduce the bias in predicted 154Eu inventories by approximately 5–7%. And while the amended capture cross-section still results in a residual over-prediction of 154Eu (ranging from 2% to 9%), the effect is substantially attenuated compared with the nominal ENDF/B-VII.0 and VII.1 evaluations.« less

  8. Proposed re-evaluation of the 154Eu thermal (n, γ) capture cross-section based on spent fuel benchmarking studies

    SciTech Connect

    Skutnik, Steven E.

    2016-09-22

    154Eu is a nuclide of considerable importance to both non-destructive measurements of used nuclear fuel assembly burnup as well as for calculating the radiation source term for used fuel storage and transportation. But, recent evidence from code validation studies of spent fuel benchmarks have revealed evidence of a systemic bias in predicted 154Eu inventories when using ENDF/B-VII.0 and ENDF/B-VII.1 nuclear data libraries, wherein Eu-154 is consistently over-predicted on the order of 10% or more. Further, this bias is found to correlate with sample burnup, resulting in a larger departure from experimental measurements for higher sample burnups. Here, the bias in Eu-154 is characterized across eleven spent fuel destructive assay benchmarks from five different assemblies. Based on these studies, possible amendments to the ENDF/B-VII.0 and VII.1 evaluations of the 154Eu (n,γ)155Eu are explored. By amending the location of the first resolved resonance for the 154Eu radiative capture cross-section (centered at 0.195 eV in ENDF/B-VII.0 and VII.1) to 0.188 eV and adjusting the neutron capture width proportional to $\\sqrt1/E$, the amended cross-section evaluation was found to reduce the bias in predicted 154Eu inventories by approximately 5–7%. And while the amended capture cross-section still results in a residual over-prediction of 154Eu (ranging from 2% to 9%), the effect is substantially attenuated compared with the nominal ENDF/B-VII.0 and VII.1 evaluations.

  9. Evaluation of the cytotoxic and genotoxic effects of benchmark multi-walled carbon nanotubes in relation to their physicochemical properties.

    PubMed

    Louro, Henriqueta; Pinhão, Mariana; Santos, Joana; Tavares, Ana; Vital, Nádia; Silva, Maria João

    2016-11-16

    To contribute with scientific evidence to the grouping strategy for the safety assessment of multi-walled carbon nanotubes (MWCNTs), this work describes the investigation of the cytotoxic and genotoxic effects of four benchmark MWCNTs in relation to their physicochemical characteristics, using two types of human respiratory cells. The cytotoxic effects were analysed using the clonogenic assay and replication index determination. A 48h-exposure of cells revealed that NM-401 was the only cytotoxic MWCNT in both cell lines, but after 8-days exposure, the clonogenic assay in A549 cells showed cytotoxic effects for all the tested MWCNTs. Correlation analysis suggested an association between the MWCNTs size in cell culture medium and cytotoxicity. No induction of DNA damage was observed after any MWCNTs in any cell line by the comet assay, while the micronucleus assay revealed that both NM-401 and NM-402 were genotoxic in A549 cells. NM-401 and NM-402 are the two longest MWCNTs analyzed in this work, suggesting that length may be determinant for genotoxicity. No induction of micronuclei was observed in BBEAS-2Beas-2B cell line and the different effect in both cell lines is explained in view of the size-distribution of MWCNTs in the cell culture medium, rather than cell's specificities. Copyright © 2016 Elsevier Ireland Ltd. All rights reserved.

  10. Grey Relational Evaluation on Road Project Delivery Models

    NASA Astrophysics Data System (ADS)

    Lo, Shih-Ching; Chao, Yu; Lee, Gin-Yuan

    2009-08-01

    In this study, four road delivery project models are analyzed by grey relational evaluation. The four models are design-bid-build (DBB), design-build (DB), construction management (CM) and design-build-maintenance (DBM). Evaluating road project delivery models is difficult because the projects differ from road to road, state to state and country to country. Thus, the evaluation data of project delivery systems are poor and lacking. Grey theory is an effective mathematical method, which is a multidisciplinary and generic theory dealing with systems characterized by poor information and/or for which information is lacking. Therefore, grey relational analysis and grey model are employed to compare the efficiency of the four road project delivery models. According to the result, DBM is the best model. DBB is the worst one and DB is better than CM. The results may provide public sectors to employ an adequate model so as to proceed with road construction project.

  11. A guide for mental health clinicians to develop and undertake benchmarking activities.

    PubMed

    Cleary, Michelle; Hunt, Glenn E; Walter, Garry; Tong, Lizabeth

    2010-04-01

    There is a growing expectation for staff to participate in benchmarking activities. If benchmarking projects are to be successful, managers and clinicians need to be aware of the steps involved. In this article, we identify key aspects of benchmarking and consider how clinicians and managers can respond to and meet contemporary requirements for the development of sound benchmarking relationships. Practicalities and issues that must be considered by benchmarking teams are also outlined. Before commencing a benchmarking project, ground rules and benchmarking agreements must be developed and ratified. An understandable benchmarking framework is required: one that is sufficiently robust for clinicians to engage in benchmarking activities and convince others that benchmarking has taken place. There is a need to build the capacity of clinicians in relation to benchmarking.

  12. Nozzle evaluation for Project W-314

    SciTech Connect

    Galbraith, J.D.

    1998-06-19

    Revisions to the waste transfer system piping to be implemented by Project W-314 will eliminate the need to access a majority of interfarm jumper connections associated with specific process pits. Additionally, connections that formerly facilitated waste transfers from the Plutonium-Uranium Extraction (PUREX) Plant are no longer required. This document identified unneeded process pit jumper connections, describes former designated routing, denotes current status (i.e., open or blanked), and recommends appropriate disposition for all. Blanking of identified nozzles should be accomplished by Project W-314 upon installation of jumpers and acceptance by Tank Waste Remediation System (TWRS) Tank Farm Operations.

  13. Evaluation of Project HAPPIER Survey: Illinois.

    ERIC Educational Resources Information Center

    Haenn, Joseph F.

    As part of Project HAPPIER (Health Awareness Patterns Preventing Illnesses and Encouraging Responsibility), a survey was conducted among teachers and other migrant personnel in Illinois to assess the current health needs of migrants. The availability of educational materials was also investigated in the survey in order to ensure that a proposed…

  14. Project RESPECT. Third Year Program Evaluation Report.

    ERIC Educational Resources Information Center

    Kester, Don; Plakos, John; Santos, Will

    In January 1995, John Marshall High School (Los Angeles, California) implemented a 3-year bilingual special alternative instructional program, Redesign of Educational Services Providing Enhanced Computer Technology (Project RESPECT). The federally funded program was to prepare limited-English-proficient (LEP) high school students for higher…

  15. Evaluating the Peruvian Rural Communication Services Project.

    ERIC Educational Resources Information Center

    Mayo, John

    1988-01-01

    Reviews the Peruvian Rural Communication Services (PRCS) Project and outlines selected findings. Topics discussed include a brief description of Peru's economic and social conditions; satellite communication systems; audio teleconferencing; telephone service; planning and administration; research design features; data collection; and project…

  16. Implementing and Evaluating Online Service Learning Projects

    ERIC Educational Resources Information Center

    Helms, Marilyn M.; Rutti, Raina M.; Hervani, Aref Agahei; LaBonte, Joanne; Sarkarat, Sy

    2015-01-01

    As online learning proliferates, professors must adapt traditional projects for an asynchronous environment. Service learning is an effective teaching style fostering interactive learning through integration of classroom activities into communities. While prior studies have documented the appropriateness of service learning in online courses,…

  17. Project Great Start Biennial Evaluation Report.

    ERIC Educational Resources Information Center

    Rudy, Dennis W.

    Project Great Start is designed to provide non-, limited-, and near-native English proficient students with improved, intensified, and increased learning opportunities for accelerated English acquisition and significant academic achievement. It focuses on three groups: students, parents, and school staff. Students and parents benefit from separate…

  18. Process Evaluation of Nebraska's Team Training Project.

    ERIC Educational Resources Information Center

    Scott, David M.; And Others

    1994-01-01

    This article describes a "system approach" training project which utilizes the formation and implementation of localized strategic (action) plans for targeting substance abuse prevention. Participants surveyed in the program reported positive attitudes about the program due to their training and their ability to resist substance abuse…

  19. Implementing and Evaluating Online Service Learning Projects

    ERIC Educational Resources Information Center

    Helms, Marilyn M.; Rutti, Raina M.; Hervani, Aref Agahei; LaBonte, Joanne; Sarkarat, Sy

    2015-01-01

    As online learning proliferates, professors must adapt traditional projects for an asynchronous environment. Service learning is an effective teaching style fostering interactive learning through integration of classroom activities into communities. While prior studies have documented the appropriateness of service learning in online courses,…

  20. Learning with East Aurora Families. Project Evaluation.

    ERIC Educational Resources Information Center

    Bercovitz, Laura

    The Learning with East Aurora Families (LEAF) Project was a 1-year family literacy program developed and implemented by Waubonsee Community College in Sugar Grove, Illinois. It recruited 51 parents and other significant adults of 4- and 5-year-olds enrolled in at-risk programs. Each of the 4-week sessions were divided into 5 components: adult…

  1. Project SCANS Integration. Formative Evaluation Report.

    ERIC Educational Resources Information Center

    Ryan, Ray D.; Pritz, Sandra G.

    Project SCANS (Secretary's Commission on Achieving Necessary Skills) Integration is a 5-year initiative to explore how well all high school instructors would be able to integrate competency-based instruction in designated work-related competencies into their courses and rate students' mastery of the competencies. Three teachers from each of 14…

  2. Evaluating the Peruvian Rural Communication Services Project.

    ERIC Educational Resources Information Center

    Mayo, John

    1988-01-01

    Reviews the Peruvian Rural Communication Services (PRCS) Project and outlines selected findings. Topics discussed include a brief description of Peru's economic and social conditions; satellite communication systems; audio teleconferencing; telephone service; planning and administration; research design features; data collection; and project…

  3. Food Processors Skills Building Project. Evaluation Report.

    ERIC Educational Resources Information Center

    White, Eileen Casey

    The Food Processors Skills Building project was undertaken by four Oregon community colleges, with funds from the Oregon Economic Development Department and 11 local food processing companies, to address basic skills needs in the food processing industry through the development and implementation of an industry-specific curriculum. Based on…

  4. Quality framework proposal for Component Material Evaluation (CME) projects.

    SciTech Connect

    Christensen, Naomi G.; Arfman, John F.; Limary, Siviengxay

    2008-09-01

    This report proposes the first stage of a Quality Framework approach that can be used to evaluate and document Component Material Evaluation (CME) projects. The first stage of the Quality Framework defines two tools that will be used to evaluate a CME project. The first tool is used to decompose a CME project into its essential elements. These elements can then be evaluated for inherent quality by looking at the subelements that impact their level of quality maturity or rigor. Quality Readiness Levels (QRLs) are used to valuate project elements for inherent quality. The Framework provides guidance for the Principal Investigator (PI) and stakeholders for CME project prerequisites that help to ensure the proper level of confidence in the deliverable given its intended use. The Framework also Provides a roadmap that defined when and how the Framework tools should be applied. Use of these tools allow the Principal Investigator (PI) and stakeholders to understand what elements the project will use to execute the project, the inherent quality of the elements, which of those are critical to the project and why, and the risks associated to the project's elements.

  5. How is success or failure in river restoration projects evaluated? Feedback from French restoration projects.

    PubMed

    Morandi, Bertrand; Piégay, Hervé; Lamouroux, Nicolas; Vaudor, Lise

    2014-05-01

    Since the 1990s, French operational managers and scientists have been involved in the environmental restoration of rivers. The European Water Framework Directive (2000) highlights the need for feedback from restoration projects and for evidence-based evaluation of success. Based on 44 French pilot projects that included such an evaluation, the present study includes: 1) an introduction to restoration projects based on their general characteristics 2) a description of evaluation strategies and authorities in charge of their implementation, and 3) a focus on the evaluation of results and the links between these results and evaluation strategies. The results show that: 1) the quality of an evaluation strategy often remains too poor to understand well the link between a restoration project and ecological changes; 2) in many cases, the conclusions drawn are contradictory, making it difficult to determine the success or failure of a restoration project; and 3) the projects with the poorest evaluation strategies generally have the most positive conclusions about the effects of restoration. Recommendations are that evaluation strategies should be designed early in the project planning process and be based on clearly-defined objectives. Copyright © 2014 Elsevier Ltd. All rights reserved.

  6. Decay Data Evaluation Project (DDEP): evaluation of the main 233Pa decay characteristics.

    PubMed

    Chechev, Valery P; Kuzmenko, Nikolay K

    2006-01-01

    The results of a decay data evaluation are presented for 233Pa (beta-) decay to nuclear levels in 233U. These evaluated data have been obtained within the Decay Data Evaluation Project using information published up to 2005.

  7. Criticality Benchmark Analysis of the HTTR Annular Startup Core Configurations

    SciTech Connect

    John D. Bess

    2009-11-01

    One of the high priority benchmarking activities for corroborating the Next Generation Nuclear Plant (NGNP) Project and Very High Temperature Reactor (VHTR) Program is evaluation of Japan's existing High Temperature Engineering Test Reactor (HTTR). The HTTR is a 30 MWt engineering test reactor utilizing graphite moderation, helium coolant, and prismatic TRISO fuel. A large amount of critical reactor physics data is available for validation efforts of High Temperature Gas-cooled Reactors (HTGRs). Previous international reactor physics benchmarking activities provided a collation of mixed results that inaccurately predicted actual experimental performance.1 Reevaluations were performed by the Japanese to reduce the discrepancy between actual and computationally-determined critical configurations.2-3 Current efforts at the Idaho National Laboratory (INL) involve development of reactor physics benchmark models in conjunction with the International Reactor Physics Experiment Evaluation Project (IRPhEP) for use with verification and validation methods in the VHTR Program. Annular cores demonstrate inherent safety characteristics that are of interest in developing future HTGRs.

  8. The Isprs Benchmark on Indoor Modelling

    NASA Astrophysics Data System (ADS)

    Khoshelham, K.; Díaz Vilariño, L.; Peter, M.; Kang, Z.; Acharya, D.

    2017-09-01

    Automated generation of 3D indoor models from point cloud data has been a topic of intensive research in recent years. While results on various datasets have been reported in literature, a comparison of the performance of different methods has not been possible due to the lack of benchmark datasets and a common evaluation framework. The ISPRS benchmark on indoor modelling aims to address this issue by providing a public benchmark dataset and an evaluation framework for performance comparison of indoor modelling methods. In this paper, we present the benchmark dataset comprising several point clouds of indoor environments captured by different sensors. We also discuss the evaluation and comparison of indoor modelling methods based on manually created reference models and appropriate quality evaluation criteria. The benchmark dataset is available for download at: benchmark-on-indoor-modelling.html"target="_blank">http://www2.isprs.org/commissions/comm4/wg5/benchmark-on-indoor-modelling.html.

  9. Pure Dead Brilliant?: Evaluating the Glasgow Story Digitisation Project

    ERIC Educational Resources Information Center

    Anderson, Ian G.

    2007-01-01

    Purpose: The purpose of this paper is to present an evaluation of The Glasgow Story (TGS) digitisation project, funded by the UK's National Lottery's New Opportunities Fund digitisation (NOF-Digi) programme, and a critique of the evaluation process itself. The paper emphasises the need for user impact evaluation and for results to be brought into…

  10. Project Aprendizaje. Final Evaluation Report 1992-93.

    ERIC Educational Resources Information Center

    Clark, Andrew

    This report provides evaluative information regarding the effectiveness of Project Aprendizaje, a New York City program that served 269 Spanish-speaking students of limited English proficiency (LEP). The project promoted parent and community involvement by sponsoring cultural events, such as a large Latin American festival. Students developed…

  11. Project Beacon 1985-86. OEA Evaluation Report.

    ERIC Educational Resources Information Center

    New York City Board of Education, Brooklyn. Office of Educational Assessment.

    This report describes and evaluates the first year of Project Beacon, a bilingual program implemented at four high schools in Queens, New York, in 1985-86. Project Beacon provided instruction in English as a second language, in native language arts, in bilingual academic subjects, and in career-oriented subjects. It reached 570 Spanish-, Chinese-,…

  12. A Program Evaluation Manual for Project Initiators. Final Report.

    ERIC Educational Resources Information Center

    Senf, Gerald; Anderson, David

    Intended for directors of special education projects, the manual provides guidelines for program evaluation. It is explained that the manual developed out of the experiences of the staff of the Leadership Training Institute in Learning Disabilities which provided technical assistance to 43 state projects. The manual's eight major sections focus on…

  13. An Evaluation of Project Gifted 1971-1972.

    ERIC Educational Resources Information Center

    Renzulli, Joseph S.

    Evaluated was Project Gifted, a tri-city (Cranston, East Providence, and Warwick, Rhode Island) program which focused on the training of gifted children in grades 4-6 in the creative thinking process. Project goals were identification of gifted students, development of differential experiences, and development of innovative programs. Cranston's…

  14. Evaluation of Career Education Projects, 1976-1977. Report #7829.

    ERIC Educational Resources Information Center

    Chern, Hermine J.; And Others

    Evaluations of thirty career education projects in the school district of Philadelphia, Pennsylvania are contained in this report. Fifteen of the projects concern classroom or shop instruction, six concern development and/or field testing of curriculum materials, and the remainder involve staff development, installation of shop equipment, job…

  15. Portland Public Schools Project Chrysalis: Year 2 Evaluation Report.

    ERIC Educational Resources Information Center

    Mitchell, Stephanie J.; Gabriel, Roy M.; Hahn, Karen J.; Laws, Katherine E.

    In 1994, the Chrysalis Project in Portland Public Schools received funding to prevent or delay the onset of substance abuse among a special target population: high-risk, female adolescents with a history of childhood abuse. Findings from the evaluation of the project's second year of providing assistance to these students are reported here. During…

  16. An Evaluation of Project Gifted 1971-1972.

    ERIC Educational Resources Information Center

    Renzulli, Joseph S.

    Evaluated was Project Gifted, a tri-city (Cranston, East Providence, and Warwick, Rhode Island) program which focused on the training of gifted children in grades 4-6 in the creative thinking process. Project goals were identification of gifted students, development of differential experiences, and development of innovative programs. Cranston's…

  17. TRIM timber projections: an evaluation based on forest inventory measurements.

    Treesearch

    John R. Mills

    1989-01-01

    Two consecutive timberland inventories collected from permanent plots in the natural pine type in North Carolina were used to evaluate the timber resource inventory model (TRIM). This study compares model predictions with field measurements and examines the effect of inventory data aggregation on the accuracy of projections. Projections were repeated for two geographic...

  18. Childhood Obesity Research Demonstration project: Cross-site evaluation method

    USDA-ARS?s Scientific Manuscript database

    The Childhood Obesity Research Demonstration (CORD) project links public health and primary care interventions in three projects described in detail in accompanying articles in this issue of Childhood Obesity. This article describes a comprehensive evaluation plan to determine the extent to which th...

  19. National Evaluation of the PUSH for Excellence Project. Final Report.

    ERIC Educational Resources Information Center

    Murray, Saundra R.; And Others

    This final report for the PUSH-EXCEL project was designed to provide readers with an idea of the major actors and events that shaped the project's history and implementation, and to describe results of a 3-year evaluation of its impact (Chapter 1). Chapter 2 describes the original impetus for PUSH-EXCEL as arising from the Reverend Jesse Jackson's…

  20. Challenges and Realities: Evaluating a School-Based Service Project.

    ERIC Educational Resources Information Center

    Keir, Scott S.; Millea, Susan

    The Hogg Foundation for Mental Health created the School of the Future (SoF) project to enable selected Texas schools to coordinate and implement school-based social and health services on their campuses and to demonstrate the effectiveness of this method of service delivery by evaluating the project to show the process used and the outcomes that…

  1. Evaluating Quality in Educational Spaces: OECD/CELE Pilot Project

    ERIC Educational Resources Information Center

    von Ahlefeld, Hannah

    2009-01-01

    CELE's International Pilot Project on Evaluating Quality in Educational Spaces aims to assist education authorities, schools and others to maximise the use of and investment in learning environments. This article provides an update on the pilot project, which is currently being implemented in Brazil, Mexico, New Zealand, Portugal and the United…

  2. Evaluation of Career Education Projects, 1976-1977. Report #7829.

    ERIC Educational Resources Information Center

    Chern, Hermine J.; And Others

    Evaluations of thirty career education projects in the school district of Philadelphia, Pennsylvania are contained in this report. Fifteen of the projects concern classroom or shop instruction, six concern development and/or field testing of curriculum materials, and the remainder involve staff development, installation of shop equipment, job…

  3. Project Closeout: Guidance for Final Evaluation of Building America Communities

    SciTech Connect

    Norton, P.; Burch, J.; Hendron, B.

    2008-03-01

    This report presents guidelines for Project Closeout. It is used to determine whether the Building America program is successfully facilitating improved design and practices to achieve energy savings goals in production homes. Its objective is to use energy simulations, targeted utility bill analysis, and feedback from project stakeholders to evaluate the performance of occupied BA communities.

  4. Latin American Literacy Partnership Project. Final Formative Evaluation.

    ERIC Educational Resources Information Center

    Watt, David L. E.

    This final evaluation of the 1991-92 program year of the Latin American literacy Project, designed to foster English language literacy in Spanish-speaking families in Canada, is intended as a formative report, American Literacy Project is intended as a formative report, assessing the changes in the students' language proficiency and the progress…

  5. Automobile Mechanic Training Evaluation Project (AMTEP) Final Report.

    ERIC Educational Resources Information Center

    Losh, Charles

    A project was undertaken to identify, develop, and validate those performance, program, and personal standards judged necessary to operate and evaluate a quality automobile mechanic/technician training program. Included among the project activities were the following: (1) a review of existing literature on performance and program standards; (2)…

  6. An Evaluation of the Favorable Alternate Sites Project. Final Report.

    ERIC Educational Resources Information Center

    Kogan, Deborah; Vencill, Mary

    This final report describes and evaluates the Favorable Alternate Sites Project (FASP), developed in response to the oversettlement of refugees (particularly Southeast Asian refugees) in particular areas of the country. The project's goals were to reduce welfare dependency, increase the ability of FASP refugees to be self-supporting, and reduce…

  7. Evaluating Quality in Educational Spaces: OECD/CELE Pilot Project

    ERIC Educational Resources Information Center

    von Ahlefeld, Hannah

    2009-01-01

    CELE's International Pilot Project on Evaluating Quality in Educational Spaces aims to assist education authorities, schools and others to maximise the use of and investment in learning environments. This article provides an update on the pilot project, which is currently being implemented in Brazil, Mexico, New Zealand, Portugal and the United…

  8. A Project for Research, Development, and Evaluation Training. Final Report.

    ERIC Educational Resources Information Center

    Karr, Chadwick; Porter, Bette C.

    This report summarizes and evaluates the 1969-70 Consortium Research Development (CORD) training and dissemination project conducted by Teaching Research, Oregon State System of Higher Education. The project provided a 2-week summer institute to train selected college and university staff from CORD institutions in the use of individualized,…

  9. Project Familia. Final Evaluation Report, 1992-93. OREA Report.

    ERIC Educational Resources Information Center

    Clarke, Candice

    Project Familia was an Elementary and Secondary Education Act Title VII funded project that, in the year covered by this evaluation, served 41 special education students of limited English proficiency (LEP) from 5 schools, with the participation of 54 parents and 33 siblings. Participating students received English language enrichment and…

  10. Rowland Bilingual/Bicultural Education Project. Evaluation Report.

    ERIC Educational Resources Information Center

    Rowland School District, Rowland Heights, CA.

    A bilingual/bicultural project for kindergarten children in Rowland Heights, California, is described in this evaluation report. The project included 2 kindergarten classes with approximately 30 children in each class and a bilingual staff. Parents and other community members served as educational volunteers and resource persons. Pre-service and…

  11. An Evaluation of the Comp-Lab Project. Final Report.

    ERIC Educational Resources Information Center

    Epes, Mary; And Others

    This final report summarizes an extensive evaluation of the COMP-LAB project, a course which integrates carefully defined classroom instruction with an autotutorial writing laboratory for teaching basic writing and rhetoric, particularly to students impaired by foreign language or nonstandard speech. Information on the project background and the…

  12. Evaluation of 60 continuous quality improvement projects in French hospitals.

    PubMed

    Maguerez, G; Erbault, M; Terra, J L; Maisonneuve, H; Matillon, Y

    2001-04-01

    To evaluate the feasibility of implementing continuous quality improvement (CQI) projects in French health care organizations. The French Ministry of Health issued two calls for CQI projects (in 1995 and 1996). ANAES was commissioned to monitor and evaluate the projects, and to provide advice. ANAES in collaboration with French public hospitals. A jury selected 64 projects from 483 submissions. The first series of projects related to safety issues (e.g. blood transfusions), the second related chiefly to patient management. ANAES instructed project leaders in process analysis (modified four-step FOCUS-PDCA model), convened regular meetings between leaders and performed on-site visits. Objective outcomes: goal achievement, extension of projects to other topics and departments, allocation of resources. Subjective outcomes: changes in attitudes. Statistics were obtained from two questionnaires completed by project leaders. Four projects were discontinued; 82% (49 out of 60) met more than half their objectives. The CQI method was adopted by other departments in 65% and 50% (1st and 2nd series respectively) of cases. Hospital management often chose to provide continued support (81%/88%), offer training (59%/80%), create a CQI unit (62%/73%), and allocate a budget (61%/65%). A positive impact on staff attitudes was noted in over 75% of projects. ANAES' co-ordinated initiative to acquaint a hard core of French public hospitals with CQI proved successful. Identification of the factors for success and of potential hurdles helped pave the way for the national hospital accreditation procedure currently underway.

  13. Benchmarking monthly homogenization algorithms

    NASA Astrophysics Data System (ADS)

    Venema, V. K. C.; Mestre, O.; Aguilar, E.; Auer, I.; Guijarro, J. A.; Domonkos, P.; Vertacnik, G.; Szentimrey, T.; Stepanek, P.; Zahradnicek, P.; Viarre, J.; Müller-Westermeier, G.; Lakatos, M.; Williams, C. N.; Menne, M.; Lindau, R.; Rasol, D.; Rustemeier, E.; Kolokythas, K.; Marinova, T.; Andresen, L.; Acquaotta, F.; Fratianni, S.; Cheval, S.; Klancar, M.; Brunetti, M.; Gruber, C.; Prohom Duran, M.; Likso, T.; Esteban, P.; Brandsma, T.

    2011-08-01

    The COST (European Cooperation in Science and Technology) Action ES0601: Advances in homogenization methods of climate series: an integrated approach (HOME) has executed a blind intercomparison and validation study for monthly homogenization algorithms. Time series of monthly temperature and precipitation were evaluated because of their importance for climate studies and because they represent two important types of statistics (additive and multiplicative). The algorithms were validated against a realistic benchmark dataset. The benchmark contains real inhomogeneous data as well as simulated data with inserted inhomogeneities. Random break-type inhomogeneities were added to the simulated datasets modeled as a Poisson process with normally distributed breakpoint sizes. To approximate real world conditions, breaks were introduced that occur simultaneously in multiple station series within a simulated network of station data. The simulated time series also contained outliers, missing data periods and local station trends. Further, a stochastic nonlinear global (network-wide) trend was added. Participants provided 25 separate homogenized contributions as part of the blind study as well as 22 additional solutions submitted after the details of the imposed inhomogeneities were revealed. These homogenized datasets were assessed by a number of performance metrics including (i) the centered root mean square error relative to the true homogeneous value at various averaging scales, (ii) the error in linear trend estimates and (iii) traditional contingency skill scores. The metrics were computed both using the individual station series as well as the network average regional series. The performance of the contributions depends significantly on the error metric considered. Contingency scores by themselves are not very informative. Although relative homogenization algorithms typically improve the homogeneity of temperature data, only the best ones improve precipitation data

  14. Science Base and Tools for Evaluating Stream Restoration Project Proposals.

    NASA Astrophysics Data System (ADS)

    Cluer, B.; Thorne, C.; Skidmore, P.; Castro, J.; Pess, G.; Beechie, T.; Shea, C.

    2008-12-01

    Stream restoration, stabilization, or enhancement projects typically employ site-specific designs and site- scale habitat improvement projects have become the default solution to many habitat problems and constraints. Such projects are often planned and implemented without thorough consideration of the broader scale problems that may be contributing to habitat degradation, attention to project resiliency to flood events, accounting for possible changes in climate or watershed land use, or ensuring the long term sustainability of the project. To address these issues, NOAA Fisheries and USFWS have collaboratively commissioned research to develop a science document and accompanying tools to support more consistent and comprehensive review of stream management and restoration projects proposals by Service staff responsible for permitting. The science document synthesizes the body of knowledge in fluvial geomorphology and presents it in a way that is accessible to the Services staff biologists, who are not trained experts in this field. Accompanying the science document are two electronic tools: a Project Information Checklist to assist in evaluating whether a proposal includes all the information necessary to allow critical and thorough project evaluation; and a Project Evaluation Tool (in flow chart format) that guides reviewers through the steps necessary to critically evaluate the quality of the information submitted, the goals and objectives of the project, project planning and development, project design, geomorphic-habitat-species relevance, and risks to listed species. Materials for training Services staff and others in the efficient use of the science document and tools have also been developed. The longer term goals of this effort include: enabling consistent and comprehensive reviews that are completed in a timely fashion by regulators; facilitating improved project planning and design by proponents; encouraging projects that are attuned to their watershed

  15. JUPITER PROJECT - JOINT UNIVERSAL PARAMETER IDENTIFICATION AND EVALUATION OF RELIABILITY

    EPA Science Inventory

    The JUPITER (Joint Universal Parameter IdenTification and Evaluation of Reliability) project builds on the technology of two widely used codes for sensitivity analysis, data assessment, calibration, and uncertainty analysis of environmental models: PEST and UCODE.

  16. JUPITER PROJECT - JOINT UNIVERSAL PARAMETER IDENTIFICATION AND EVALUATION OF RELIABILITY

    EPA Science Inventory

    The JUPITER (Joint Universal Parameter IdenTification and Evaluation of Reliability) project builds on the technology of two widely used codes for sensitivity analysis, data assessment, calibration, and uncertainty analysis of environmental models: PEST and UCODE.

  17. 40 CFR 57.604 - Evaluation of projects.

    Code of Federal Regulations, 2010 CFR

    2010-07-01

    ...) PRIMARY NONFERROUS SMELTER ORDERS Research and Development Requirements § 57.604 Evaluation of projects... significant stage of the research and development program, including all relevant information and data...

  18. Evaluation of the Treatment of Diabetic Retinopathy A Research Project

    ERIC Educational Resources Information Center

    Kupfer, Carl

    1973-01-01

    Evaluated is the treatment of diabetic retinopathy (blindness due to ruptured vessels of the retina as a side effect of diabetes), and described is a research project comparing two types of photocoagulation treatment. (DB)

  19. Evaluation of the concrete shield compositions from the 2010 criticality accident alarm system benchmark experiments at the CEA Valduc SILENE facility

    SciTech Connect

    Miller, Thomas Martin; Celik, Cihangir; Dunn, Michael E; Wagner, John C; McMahan, Kimberly L; Authier, Nicolas; Jacquet, Xavier; Rousseau, Guillaume; Wolff, Herve; Savanier, Laurence; Baclet, Nathalie; Lee, Yi-kang; Trama, Jean-Christophe; Masse, Veronique; Gagnier, Emmanuel; Naury, Sylvie; Blanc-Tranchant, Patrick; Hunter, Richard; Kim, Soon; Dulik, George Michael; Reynolds, Kevin H.

    2015-01-01

    In October 2010, a series of benchmark experiments were conducted at the French Commissariat a l'Energie Atomique et aux Energies Alternatives (CEA) Valduc SILENE facility. These experiments were a joint effort between the United States Department of Energy Nuclear Criticality Safety Program and the CEA. The purpose of these experiments was to create three benchmarks for the verification and validation of radiation transport codes and evaluated nuclear data used in the analysis of criticality accident alarm systems. This series of experiments consisted of three single-pulsed experiments with the SILENE reactor. For the first experiment, the reactor was bare (unshielded), whereas in the second and third experiments, it was shielded by lead and polyethylene, respectively. The polyethylene shield of the third experiment had a cadmium liner on its internal and external surfaces, which vertically was located near the fuel region of SILENE. During each experiment, several neutron activation foils and thermoluminescent dosimeters (TLDs) were placed around the reactor. Nearly half of the foils and TLDs had additional high-density magnetite concrete, high-density barite concrete, standard concrete, and/or BoroBond shields. CEA Saclay provided all the concrete, and the US Y-12 National Security Complex provided the BoroBond. Measurement data from the experiments were published at the 2011 International Conference on Nuclear Criticality (ICNC 2011) and the 2013 Nuclear Criticality Safety Division (NCSD 2013) topical meeting. Preliminary computational results for the first experiment were presented in the ICNC 2011 paper, which showed poor agreement between the computational results and the measured values of the foils shielded by concrete. Recently the hydrogen content, boron content, and density of these concrete shields were further investigated within the constraints of the previously available data. New computational results for the first experiment are now available that

  20. Evaluation on Collaborative Satisfaction for Project Management Team in Integrated Project Delivery Mode

    NASA Astrophysics Data System (ADS)

    Zhang, L.; Li, Y.; Wu, Q.

    2013-05-01

    Integrated Project Delivery (IPD) is a newly-developed project delivery approach for construction projects, and the level of collaboration of project management team is crucial to the success of its implementation. Existing research has shown that collaborative satisfaction is one of the key indicators of team collaboration. By reviewing the literature on team collaborative satisfaction and taking into consideration the characteristics of IPD projects, this paper summarizes the factors that influence collaborative satisfaction of IPD project management team. Based on these factors, this research develops a fuzzy linguistic method to effectively evaluate the level of team collaborative satisfaction, in which the authors adopted the 2-tuple linguistic variables and 2-tuple linguistic hybrid average operators to enhance the objectivity and accuracy of the evaluation. The paper demonstrates the practicality and effectiveness of the method through carrying out a case study with the method.

  1. Refining multi-model projections of temperature extremes by evaluation against land-atmosphere coupling diagnostics

    NASA Astrophysics Data System (ADS)

    Sippel, Sebastian; Zscheischler, Jakob; Mahecha, Miguel D.; Orth, Rene; Reichstein, Markus; Vogel, Martha; Seneviratne, Sonia I.

    2017-05-01

    The Earth's land surface and the atmosphere are strongly interlinked through the exchange of energy and matter. This coupled behaviour causes various land-atmosphere feedbacks, and an insufficient understanding of these feedbacks contributes to uncertain global climate model projections. For example, a crucial role of the land surface in exacerbating summer heat waves in midlatitude regions has been identified empirically for high-impact heat waves, but individual climate models differ widely in their respective representation of land-atmosphere coupling. Here, we compile an ensemble of 54 combinations of observations-based temperature (T) and evapotranspiration (ET) benchmarking datasets and investigate coincidences of T anomalies with ET anomalies as a proxy for land-atmosphere interactions during periods of anomalously warm temperatures. First, we demonstrate that a large fraction of state-of-the-art climate models from the Coupled Model Intercomparison Project (CMIP5) archive produces systematically too frequent coincidences of high T anomalies with negative ET anomalies in midlatitude regions during the warm season and in several tropical regions year-round. These coincidences (high T, low ET) are closely related to the representation of temperature variability and extremes across the multi-model ensemble. Second, we derive a land-coupling constraint based on the spread of the T-ET datasets and consequently retain only a subset of CMIP5 models that produce a land-coupling behaviour that is compatible with these benchmark estimates. The constrained multi-model simulations exhibit more realistic temperature extremes of reduced magnitude in present climate in regions where models show substantial spread in T-ET coupling, i.e. biases in the model ensemble are consistently reduced. Also the multi-model simulations for the coming decades display decreased absolute temperature extremes in the constrained ensemble. On the other hand, the differences between projected

  2. Issues in evaluating Superfund cleanup projects

    SciTech Connect

    Yun, J.M.

    1996-12-31

    It is the aim of this paper to highlight issues that have potentially serious, but not immediately obvious, implications in assessing the costs and benefits of a Superfund project. Firstly, a serious error would be to overlook, on the cost-side, the welfare loss due to the excessive litigation (i.e., all relevant legal costs including attorney fees and settlement costs) that a Superfund designation creates. secondly, how the EPA rates the funds to pay for Superfund cleanups must be accounted for--namely, the distortionary effects of the EPA`s taxes on insurance and chemical companies. Harberger`s methodology can be adopted to account for this welfare loss. Finally, on the benefits-side, following Kohlhase and Greenberg and Hughes, housing values can give a valuable second-best assessment of the actual benefits of (local) projects. The appeal of using housing values is that it is a market-based approach. The rationale and ramifications of accounting for the litigation accompanying most EPA cleanups are examined in Section 2. In Section 3, the distortionary effects of the taxes are analyzed. section 4 investigates the use of housing values, near Superfund sites, to estimate the benefits of a cleanup. Finally, in Section 5, the implications and conclusions are summarized. Various illustrative cases will be presented where appropriate.

  3. Benchmarking short sequence mapping tools.

    PubMed

    Hatem, Ayat; Bozdağ, Doruk; Toland, Amanda E; Çatalyürek, Ümit V

    2013-06-07

    The development of next-generation sequencing instruments has led to the generation of millions of short sequences in a single run. The process of aligning these reads to a reference genome is time consuming and demands the development of fast and accurate alignment tools. However, the current proposed tools make different compromises between the accuracy and the speed of mapping. Moreover, many important aspects are overlooked while comparing the performance of a newly developed tool to the state of the art. Therefore, there is a need for an objective evaluation method that covers all the aspects. In this work, we introduce a benchmarking suite to extensively analyze sequencing tools with respect to various aspects and provide an objective comparison. We applied our benchmarking tests on 9 well known mapping tools, namely, Bowtie, Bowtie2, BWA, SOAP2, MAQ, RMAP, GSNAP, Novoalign, and mrsFAST (mrFAST) using synthetic data and real RNA-Seq data. MAQ and RMAP are based on building hash tables for the reads, whereas the remaining tools are based on indexing the reference genome. The benchmarking tests reveal the strengths and weaknesses of each tool. The results show that no single tool outperforms all others in all metrics. However, Bowtie maintained the best throughput for most of the tests while BWA performed better for longer read lengths. The benchmarking tests are not restricted to the mentioned tools and can be further applied to others. The mapping process is still a hard problem that is affected by many factors. In this work, we provided a benchmarking suite that reveals and evaluates the different factors affecting the mapping process. Still, there is no tool that outperforms all of the others in all the tests. Therefore, the end user should clearly specify his needs in order to choose the tool that provides the best results.

  4. Benchmarking short sequence mapping tools

    PubMed Central

    2013-01-01

    Background The development of next-generation sequencing instruments has led to the generation of millions of short sequences in a single run. The process of aligning these reads to a reference genome is time consuming and demands the development of fast and accurate alignment tools. However, the current proposed tools make different compromises between the accuracy and the speed of mapping. Moreover, many important aspects are overlooked while comparing the performance of a newly developed tool to the state of the art. Therefore, there is a need for an objective evaluation method that covers all the aspects. In this work, we introduce a benchmarking suite to extensively analyze sequencing tools with respect to various aspects and provide an objective comparison. Results We applied our benchmarking tests on 9 well known mapping tools, namely, Bowtie, Bowtie2, BWA, SOAP2, MAQ, RMAP, GSNAP, Novoalign, and mrsFAST (mrFAST) using synthetic data and real RNA-Seq data. MAQ and RMAP are based on building hash tables for the reads, whereas the remaining tools are based on indexing the reference genome. The benchmarking tests reveal the strengths and weaknesses of each tool. The results show that no single tool outperforms all others in all metrics. However, Bowtie maintained the best throughput for most of the tests while BWA performed better for longer read lengths. The benchmarking tests are not restricted to the mentioned tools and can be further applied to others. Conclusion The mapping process is still a hard problem that is affected by many factors. In this work, we provided a benchmarking suite that reveals and evaluates the different factors affecting the mapping process. Still, there is no tool that outperforms all of the others in all the tests. Therefore, the end user should clearly specify his needs in order to choose the tool that provides the best results. PMID:23758764

  5. Medical school benchmarking - from tools to programmes.

    PubMed

    Wilkinson, Tim J; Hudson, Judith N; Mccoll, Geoffrey J; Hu, Wendy C Y; Jolly, Brian C; Schuwirth, Lambert W T

    2015-02-01

    Benchmarking among medical schools is essential, but may result in unwanted effects. To apply a conceptual framework to selected benchmarking activities of medical schools. We present an analogy between the effects of assessment on student learning and the effects of benchmarking on medical school educational activities. A framework by which benchmarking can be evaluated was developed and applied to key current benchmarking activities in Australia and New Zealand. The analogy generated a conceptual framework that tested five questions to be considered in relation to benchmarking: what is the purpose? what are the attributes of value? what are the best tools to assess the attributes of value? what happens to the results? and, what is the likely "institutional impact" of the results? If the activities were compared against a blueprint of desirable medical graduate outcomes, notable omissions would emerge. Medical schools should benchmark their performance on a range of educational activities to ensure quality improvement and to assure stakeholders that standards are being met. Although benchmarking potentially has positive benefits, it could also result in perverse incentives with unforeseen and detrimental effects on learning if it is undertaken using only a few selected assessment tools.

  6. PRISMATIC CORE COUPLED TRANSIENT BENCHMARK

    SciTech Connect

    J. Ortensi; M.A. Pope; G. Strydom; R.S. Sen; M.D. DeHart; H.D. Gougar; C. Ellis; A. Baxter; V. Seker; T.J. Downar; K. Vierow; K. Ivanov

    2011-06-01

    The Prismatic Modular Reactor (PMR) is one of the High Temperature Reactor (HTR) design concepts that have existed for some time. Several prismatic units have operated in the world (DRAGON, Fort St. Vrain, Peach Bottom) and one unit is still in operation (HTTR). The deterministic neutronics and thermal-fluids transient analysis tools and methods currently available for the design and analysis of PMRs have lagged behind the state of the art compared to LWR reactor technologies. This has motivated the development of more accurate and efficient tools for the design and safety evaluations of the PMR. In addition to the work invested in new methods, it is essential to develop appropriate benchmarks to verify and validate the new methods in computer codes. The purpose of this benchmark is to establish a well-defined problem, based on a common given set of data, to compare methods and tools in core simulation and thermal hydraulics analysis with a specific focus on transient events. The benchmark-working group is currently seeking OECD/NEA sponsorship. This benchmark is being pursued and is heavily based on the success of the PBMR-400 exercise.

  7. Aquatic Life Benchmarks

    EPA Pesticide Factsheets

    The Aquatic Life Benchmarks is an EPA-developed set of criteria for freshwater species. These benchmarks are based on toxicity values reviewed by EPA and used in the Agency's risk assessments developed as part of the decision-making process for pesticide registration.

  8. Benchmarks in Management Training.

    ERIC Educational Resources Information Center

    Paddock, Susan C.

    1997-01-01

    Data were collected from 12 states with Certified Public Manager training programs to establish benchmarks. The 38 benchmarks were in the following areas: program leadership, stability of administrative/financial support, consistent management philosophy, administrative control, participant selection/support, accessibility, application of…

  9. Benchmarks in Management Training.

    ERIC Educational Resources Information Center

    Paddock, Susan C.

    1997-01-01

    Data were collected from 12 states with Certified Public Manager training programs to establish benchmarks. The 38 benchmarks were in the following areas: program leadership, stability of administrative/financial support, consistent management philosophy, administrative control, participant selection/support, accessibility, application of…

  10. Verification and validation benchmarks.

    SciTech Connect

    Oberkampf, William Louis; Trucano, Timothy Guy

    2007-02-01

    Verification and validation (V&V) are the primary means to assess the accuracy and reliability of computational simulations. V&V methods and procedures have fundamentally improved the credibility of simulations in several high-consequence fields, such as nuclear reactor safety, underground nuclear waste storage, and nuclear weapon safety. Although the terminology is not uniform across engineering disciplines, code verification deals with assessing the reliability of the software coding, and solution verification deals with assessing the numerical accuracy of the solution to a computational model. Validation addresses the physics modeling accuracy of a computational simulation by comparing the computational results with experimental data. Code verification benchmarks and validation benchmarks have been constructed for a number of years in every field of computational simulation. However, no comprehensive guidelines have been proposed for the construction and use of V&V benchmarks. For example, the field of nuclear reactor safety has not focused on code verification benchmarks, but it has placed great emphasis on developing validation benchmarks. Many of these validation benchmarks are closely related to the operations of actual reactors at near-safety-critical conditions, as opposed to being more fundamental-physics benchmarks. This paper presents recommendations for the effective design and use of code verification benchmarks based on manufactured solutions, classical analytical solutions, and highly accurate numerical solutions. In addition, this paper presents recommendations for the design and use of validation benchmarks, highlighting the careful design of building-block experiments, the estimation of experimental measurement uncertainty for both inputs and outputs to the code, validation metrics, and the role of model calibration in validation. It is argued that the understanding of predictive capability of a computational model is built on the level of

  11. The Design of the IGE Evaluation Project Phase IV Comparative Studies. Comparative Study of Phase IV IGE Evaluation Project. Phase IV, Project Paper 80-2.

    ERIC Educational Resources Information Center

    Romberg, Thomas A.; And Others

    This paper outlines the design of two Comparative Studies of Phase IV of the Individually Guided Education (IGE) Evaluation Project. More than 2,000 elementary schools in 25 states use the IGE system. The Evaluation Project was designed to gain a comprehensive view of the system's operation and effectiveness. Phase IV investigated pupil outcomes,…

  12. National Evaluation of Diversion Projects. Executive Summary.

    ERIC Educational Resources Information Center

    Dunford, Franklyn W.; And Others

    In 1976 the Special Emphasis branch of the Office of Juvenile Justice and Delinquency Prevention made $10 million available for the development of 11 diversion programs. A national evaluation of these programs was promoted in the hope of better understanding the viability of diversion as an alternative to traditional practices. The impact of…

  13. Evaluation of EUREKA Project, 1978-1979.

    ERIC Educational Resources Information Center

    Burke, Paul J., Ed.

    An evaluation for 1978-79 was conducted of EUREKA, a career information system in California. Personal visits were made to sixteen EUREKA sites throughout the state, accounting for over 75% of the high schools and agencies with active programs. Both the directors of the programs and counselors were interviewed for their reactions. It was found…

  14. Test Nationally, Benchmark Locally: Using Local DIBELS Benchmarks to Predict Performance on the Pssa

    ERIC Educational Resources Information Center

    Ferchalk, Matthew R.

    2013-01-01

    The Dynamic Indicators of Basic Early Literacy Skills (DIBELS) benchmarks are frequently used to make important decision regarding student performance. More information, however, is needed to understand if the nationally-derived benchmarks created by the DIBELS system provide the most accurate criterion for evaluating reading proficiency. The…

  15. A portfolio evaluation framework for air transportation improvement projects

    NASA Astrophysics Data System (ADS)

    Baik, Hyeoncheol

    This thesis explores the application of portfolio theory to the Air Transportation System (ATS) improvement. The ATS relies on complexly related resources and different stakeholder groups. Moreover, demand for air travel is significantly increasing relative to capacity of air transportation. In this environment, improving the ATS is challenging. Many projects, which are defined as technologies or initiatives, for improvement have been proposed and some have been demonstrated in practice. However, there is no clear understanding of how well these projects work in different conditions nor of how they interact with each other or with existing systems. These limitations make it difficult to develop good project combinations, or portfolios that maximize improvement. To help address this gap, a framework for identifying good portfolios is proposed. The framework can be applied to individual projects or portfolios of projects. Projects or portfolios are evaluated using four different groups of factors (effectiveness, time-to-implement, scope of applicability, and stakeholder impacts). Portfolios are also evaluated in terms of interaction-determining factors (prerequisites, co-requisites, limiting factors, and amplifying factors) because, while a given project might work well in isolation, interdependencies between projects or with existing systems could result in lower overall performance in combination. Ways to communicate a portfolio to decision makers are also introduced. The framework is unique because (1) it allows using a variety of available data, and (2) it covers diverse benefit metrics. For demonstrating the framework, an application to ground delay management projects serves as a case study. The portfolio evaluation approach introduced in this thesis can aid decision makers and researchers at universities and aviation agencies such as Federal Aviation Administration (FAA), National Aeronautics and Space Administration (NASA), and Department of Defense (DoD), in

  16. Helical Screw Expander Evaluation Project. Final report

    SciTech Connect

    McKay, R.

    1982-03-01

    A functional 1-MW geothermal electric power plant that featured a helical screw expander was produced and then tested in Utah in 1978 to 1979 with a demonstrated average performance of approximately 45% machine efficiency over a wide range of test conditions in noncondensing operation on two-phase geothermal fluids. The Project also produced a computer-equipped data system, an instrumentation and control van, and a 1000-kW variable load bank, all integrated into a test array designed for operation at a variety of remote test sites. Additional testing was performed in Mexico in 1980 under a cooperative test program using the same test array, and machine efficiency was measured at 62% maximum with the rotors partially coated with scale, compared with approximately 54% maximum in Utah with uncoated rotors, confirming the importance of scale deposits within the machine on performance. Data are presented for the Utah testing and for the noncondensing phases of the testing in Mexico. Test time logged was 437 hours during the Utah tests and 1101 hours during the Mexico tests.

  17. A performance benchmark test for geodynamo simulations

    NASA Astrophysics Data System (ADS)

    Matsui, H.; Heien, E. M.

    2013-12-01

    In the last ten years, a number of numerical dynamo models have successfully represented basic characteristics of the geomagnetic field. As new models and numerical methods continue to be developed, it is important to update and extend benchmarks for testing these models. The first dynamo benchmark of Christensen et al. (2001) was applied to models based on spherical harmonic expansion methods. However, only a few groups have reported results of the dynamo benchmark using local methods (Harder and Hansen, 2005; Matsui and Okuda, 2005; Chan et al., 2007) because of the difficulty treating magnetic boundary conditions based on the local methods. On the other hand, spherical harmonics expansion methods perform poorly on massively parallel computers because global data communications are required for the spherical harmonics expansions to evaluate nonlinear terms. We perform benchmark tests to asses various numerical methods for the next generation of geodynamo simulations. The purpose of this benchmark test is to assess numerical geodynamo models on a massively parallel computational platform. To compare among many numerical methods as possible, we consider the model with the insulated magnetic boundary by Christensen et al. (2001) and with the pseudo vacuum magnetic boundary, because the pseudo vacuum boundaries are implemented easier by using the local method than the magnetic insulated boundaries. In the present study, we consider two kinds of benchmarks, so-called accuracy benchmark and performance benchmark. In the accuracy benchmark, we compare the dynamo models by using modest Ekman and Rayleigh numbers proposed by Christensen et. al. (2001). We investigate a required spatial resolution for each dynamo code to obtain less than 1% difference from the suggested solution of the benchmark test using the two magnetic boundary conditions. In the performance benchmark, we investigate computational performance under the same computational environment. We perform these

  18. Restaurant Energy Use Benchmarking Guideline

    SciTech Connect

    Hedrick, R.; Smith, V.; Field, K.

    2011-07-01

    A significant operational challenge for food service operators is defining energy use benchmark metrics to compare against the performance of individual stores. Without metrics, multiunit operators and managers have difficulty identifying which stores in their portfolios require extra attention to bring their energy performance in line with expectations. This report presents a method whereby multiunit operators may use their own utility data to create suitable metrics for evaluating their operations.

  19. A framework for benchmarking land models

    SciTech Connect

    Luo, Yiqi; Randerson, J.; Abramowitz, G.; Bacour, C.; Blyth, E.; Carvalhais, N.; Ciais, Philippe; Dalmonech, D.; Fisher, J.B.; Fisher, R.; Friedlingstein, P.; Hibbard, Kathleen A.; Hoffman, F. M.; Huntzinger, Deborah; Jones, C.; Koven, C.; Lawrence, David M.; Li, D.J.; Mahecha, M.; Niu, S.L.; Norby, Richard J.; Piao, S.L.; Qi, X.; Peylin, P.; Prentice, I.C.; Riley, William; Reichstein, M.; Schwalm, C.; Wang, Y.; Xia, J. Y.; Zaehle, S.; Zhou, X. H.

    2012-10-09

    Land models, which have been developed by the modeling community in the past few decades to predict future states of ecosystems and climate, have to be critically evaluated for their performance skills of simulating ecosystem responses and feedback to climate change. Benchmarking is an emerging procedure to measure performance of models against a set of defined standards. This paper proposes a benchmarking framework for evaluation of land model performances and, meanwhile, highlights major challenges at this infant stage of benchmark analysis. The framework includes (1) targeted aspects of model performance to be evaluated, (2) a set of benchmarks as defined references to test model performance, (3) metrics to measure and compare performance skills among models so as to identify model strengths and deficiencies, and (4) model improvement. Land models are required to simulate exchange of water, energy, carbon and sometimes other trace gases between the atmosphere and land surface, and should be evaluated for their simulations of biophysical processes, biogeochemical cycles, and vegetation dynamics in response to climate change across broad temporal and spatial scales. Thus, one major challenge is to select and define a limited number of benchmarks to effectively evaluate land model performance. The second challenge is to develop metrics of measuring mismatches between models and benchmarks. The metrics may include (1) a priori thresholds of acceptable model performance and (2) a scoring system to combine data–model mismatches for various processes at different temporal and spatial scales. The benchmark analyses should identify clues of weak model performance to guide future development, thus enabling improved predictions of future states of ecosystems and climate. The near-future research effort should be on development of a set of widely acceptable benchmarks that can be used to objectively, effectively, and reliably evaluate fundamental properties of land models

  20. A framework for benchmarking land models

    SciTech Connect

    Luo, Yiqi; Randerson, James T.; Hoffman, Forrest; Norby, Richard J

    2012-01-01

    Land models, which have been developed by the modeling community in the past few decades to predict future states of ecosystems and climate, have to be critically evaluated for their performance skills of simulating ecosystem responses and feedback to climate change. Benchmarking is an emerging procedure to measure performance of models against a set of defined standards. This paper proposes a benchmarking framework for evaluation of land model performances and, meanwhile, highlights major challenges at this infant stage of benchmark analysis. The framework includes (1) targeted aspects of model performance to be evaluated, (2) a set of benchmarks as defined references to test model performance, (3) metrics to measure and compare performance skills among models so as to identify model strengths and deficiencies, and (4) model improvement. Land models are required to simulate exchange of water, energy, carbon and sometimes other trace gases between the atmosphere and land surface, and should be evaluated for their simulations of biophysical processes, biogeochemical cycles, and vegetation dynamics in response to climate change across broad temporal and spatial scales. Thus, one major challenge is to select and define a limited number of benchmarks to effectively evaluate land model performance. The second challenge is to develop metrics of measuring mismatches between models and benchmarks. The metrics may include (1) a priori thresholds of acceptable model performance and (2) a scoring system to combine data model mismatches for various processes at different temporal and spatial scales. The benchmark analyses should identify clues of weak model performance to guide future development, thus enabling improved predictions of future states of ecosystems and climate. The near-future research effort should be on development of a set of widely acceptable benchmarks that can be used to objectively, effectively, and reliably evaluate fundamental properties of land models

  1. Iterative Knowledge-Based Scoring Functions Derived from Rigid and Flexible Decoy Structures: Evaluation with the 2013 and 2014 CSAR Benchmarks.

    PubMed

    Yan, Chengfei; Grinter, Sam Z; Merideth, Benjamin Ryan; Ma, Zhiwei; Zou, Xiaoqin

    2016-06-27

    In this study, we developed two iterative knowledge-based scoring functions, ITScore_pdbbind(rigid) and ITScore_pdbbind(flex), using rigid decoy structures and flexible decoy structures, respectively, that were generated from the protein-ligand complexes in the refined set of PDBbind 2012. These two scoring functions were evaluated using the 2013 and 2014 CSAR benchmarks. The results were compared with the results of two other scoring functions, the Vina scoring function and ITScore, the scoring function that we previously developed from rigid decoy structures for a smaller set of protein-ligand complexes. A graph-based method was developed to evaluate the root-mean-square deviation between two conformations of the same ligand with different atom names and orders due to different file preparations, and the program is freely available. Our study showed that the two new scoring functions developed from the larger training set yielded significantly improved performance in binding mode predictions. For binding affinity predictions, all four scoring functions showed protein-dependent performance. We suggest the development of protein-family-dependent scoring functions for accurate binding affinity prediction.

  2. Preview: Evaluation of the 1973-1974 Bilingual/Bicultural Project. Formative Evaluation Report.

    ERIC Educational Resources Information Center

    Ligon, Glynn; And Others

    The formative report provided the Austin Independent School District personnel with information useful for planning the remaining activities for the 1973-74 Bilingual/Bicultural Project and the activities for the 1974-75 Project. Emphasis was on what had been done to evaluate the 1973-74 Project, the data which was or would be available for the…

  3. Preview: Evaluation of the 1973-1974 Bilingual/Bicultural Project. Formative Evaluation Report.

    ERIC Educational Resources Information Center

    Ligon, Glynn; And Others

    The formative report provided the Austin Independent School District personnel with information useful for planning the remaining activities for the 1973-74 Bilingual/Bicultural Project and the activities for the 1974-75 Project. Emphasis was on what had been done to evaluate the 1973-74 Project, the data which was or would be available for the…

  4. Benchmarking and the laboratory

    PubMed Central

    Galloway, M; Nadin, L

    2001-01-01

    This article describes how benchmarking can be used to assess laboratory performance. Two benchmarking schemes are reviewed, the Clinical Benchmarking Company's Pathology Report and the College of American Pathologists' Q-Probes scheme. The Clinical Benchmarking Company's Pathology Report is undertaken by staff based in the clinical management unit, Keele University with appropriate input from the professional organisations within pathology. Five annual reports have now been completed. Each report is a detailed analysis of 10 areas of laboratory performance. In this review, particular attention is focused on the areas of quality, productivity, variation in clinical practice, skill mix, and working hours. The Q-Probes scheme is part of the College of American Pathologists programme in studies of quality assurance. The Q-Probes scheme and its applicability to pathology in the UK is illustrated by reviewing two recent Q-Probe studies: routine outpatient test turnaround time and outpatient test order accuracy. The Q-Probes scheme is somewhat limited by the small number of UK laboratories that have participated. In conclusion, as a result of the government's policy in the UK, benchmarking is here to stay. Benchmarking schemes described in this article are one way in which pathologists can demonstrate that they are providing a cost effective and high quality service. Key Words: benchmarking • pathology PMID:11477112

  5. Evaluation Toolkit: A Tailored Approach to Evaluation for Parenting Projects.

    ERIC Educational Resources Information Center

    Shaw, Catherine

    This toolkit presents a collection of accessible guidelines, measures, and tools to guide and implement evaluation of parenting education and support interventions. Designed primarily for people who are new to evaluation, it contains additional advice and guidance for those with a higher level of understanding or knowledge and it may also be…

  6. Factors Common to High-Utilization Evaluations. Evaluation Productivity Project.

    ERIC Educational Resources Information Center

    Alkin, Marvin; And Others

    This paper reports on the factors that characterize high-utilization evaluations. It is based on materials submitted to an American Educational Research Association (AERA) Division H competition for outstanding examples of evaluation utilization. The paper is organized into three sections. The first section outlines the background of the study:…

  7. Rasch Model Analysis on the Effectiveness of Early Evaluation Questions as a Benchmark for New Students Ability

    ERIC Educational Resources Information Center

    Arsad, Norhana; Kamal, Noorfazila; Ayob, Afida; Sarbani, Nizaroyani; Tsuey, Chong Sheau; Misran, Norbahiah; Husain, Hafizah

    2013-01-01

    This paper discusses the effectiveness of the early evaluation questions conducted to determine the academic ability of the new students in the Department of Electrical, Electronics and Systems Engineering. Questions designed are knowledge based--on what the students have learned during their pre-university level. The results show students have…

  8. Summative Evaluation of the Manukau Family Literacy Project, 2004

    ERIC Educational Resources Information Center

    Benseman, John Robert; Sutton, Alison Joy

    2005-01-01

    This report covers a summative evaluation of a family literacy project in Auckland, New Zealand. The evaluation covered 70 adults and their children over a two year period. Outcomes for the program included literacy skill gains for both adults and children, increased levels of self-confidence and self-efficacy, greater parental involvement in…

  9. Evaluating Injury Prevention Programs: The Oklahoma City Smoke Alarm Project.

    ERIC Educational Resources Information Center

    Mallonee, Sue

    2000-01-01

    Illustrates how evaluating the Oklahoma City Smoke Alarm Project increased its success in reducing residential fire-related injuries and deaths. The program distributed and tested smoke alarms in residential dwellings and offered educational materials on fire prevention and safety. Evaluation provided sound data on program processes and outcomes,…

  10. 40 CFR 57.604 - Evaluation of projects.

    Code of Federal Regulations, 2014 CFR

    2014-07-01

    ...) PRIMARY NONFERROUS SMELTER ORDERS Research and Development Requirements § 57.604 Evaluation of projects. The research and development proposal shall include a provision for the employment of a qualified independent engineering firm to prepare written reports at least annually which evaluate each...

  11. Corrections Education Evaluation System Project. Site Visit Report.

    ERIC Educational Resources Information Center

    Nelson, Orville; And Others

    Site visits to five correctional institutions in Wisconsin were conducted as part of the development of an evaluation model for the competency-based vocational education (CBVE) project for the Wisconsin Correctional System. The evaluators' perceptions of the CBVE system are presented with recommendations for improvement. Site visits were conducted…

  12. Participatory Evaluation with Youth Leads to Community Action Project

    ERIC Educational Resources Information Center

    Ashton, Carolyn; Arnold, Mary E.; Wells, Elissa E.

    2010-01-01

    4-H has long emphasized the importance of civic engagement and community service for positive youth development. One pathway to this ideal is youth action research and evaluation. This article demonstrates how participatory youth research and evaluation can lead to the successful implementation of community action projects. It describes the…

  13. The ASCD Healthy School Communities Project: Formative Evaluation Results

    ERIC Educational Resources Information Center

    Valois, Robert F.; Lewallen, Theresa C.; Slade, Sean; Tasco, Adriane N.

    2015-01-01

    Purpose: The purpose of this paper is to report the formative evaluation results from the Association for Supervision and Curriculum Development Healthy School Communities (HSC) pilot project. Design/methodology/approach: This study utilized 11 HSC pilot sites in the USA (eight sites) and Canada (three sites). The evaluation question was…

  14. The ASCD Healthy School Communities Project: Formative Evaluation Results

    ERIC Educational Resources Information Center

    Valois, Robert F.; Lewallen, Theresa C.; Slade, Sean; Tasco, Adriane N.

    2015-01-01

    Purpose: The purpose of this paper is to report the formative evaluation results from the Association for Supervision and Curriculum Development Healthy School Communities (HSC) pilot project. Design/methodology/approach: This study utilized 11 HSC pilot sites in the USA (eight sites) and Canada (three sites). The evaluation question was…

  15. CENDL project, the chinese evaluated nuclear data library

    NASA Astrophysics Data System (ADS)

    Ge, Zhigang; Wu, Haicheng; Chen, Guochang; Xu, Ruirui

    2017-09-01

    The status of Chinese Evaluated Nuclear Data Library (CENDL) and the relevant CENDL project are introduced in this paper. Recently, a new version CENDL-3.2b0 was being prepared on the basis of the previous CENDL-3.1. The data in the light and actinide nuclide regions are updated from CENDL-3.1, and the new evaluations and calculations are performed mainly around structure and fission product nuclide regions. Covariance was also evaluated for structure and actinide nuclides. At the same time, the methodologies are systematically developed to fulfil the requirements of evaluations for CENDL-3.2b0. The updated nuclear reaction models for light and middle-heavy nuclides, non-model dependent nuclear data evaluation, covariance evaluation approaches, systematics, and integral validation system of nuclear data are incorporated in present CENDL project. The future developments are also planned.

  16. Comparison of the Mortality Probability Admission Model III, National Quality Forum, and Acute Physiology and Chronic Health Evaluation IV hospital mortality models: implications for national benchmarking*.

    PubMed

    Kramer, Andrew A; Higgins, Thomas L; Zimmerman, Jack E

    2014-03-01

    Physiology and Chronic Health Evaluation IVa had better accuracy within patient subgroups and for specific admission diagnoses. Acute Physiology and Chronic Health Evaluation IVa offered the best discrimination and calibration on a large common dataset and excluded fewer patients than Mortality Probability Admission Model III or ICU Outcomes Model/National Quality Forum. The choice of ICU performance benchmarks should be based on a comparison of model accuracy using data for identical patients.

  17. Evaluation in Cross-Cultural Contexts: Proposing a Framework for International Education and Training Project Evaluations.

    ERIC Educational Resources Information Center

    bin Yahya, Ismail; And Others

    This paper focuses on the need for increased sensitivity and responsiveness in international education and training project evaluations, particularly those in Third World countries. A conceptual-theoretical framework for designing and developing models appropriate for evaluating education and training projects in non-Western cultures is presented.…

  18. TWODANT benchmark. Progress report

    SciTech Connect

    Lee, Sung

    1994-01-11

    TWODANT (Two-Dimensional, Diffusion-Accelerated, Neutral-Particle Transport) code has been benchmarked against 6 critical experiments (Jezebel plutonium critical assembly) and their k effective values compared with those of KENO and MCNP codes.

  19. Benchmarking TENDL-2012

    NASA Astrophysics Data System (ADS)

    van der Marck, S. C.; Koning, A. J.; Rochman, D. A.

    2014-04-01

    The new release of the TENDL nuclear data library, TENDL-2012, was tested by performing many benchmark calculations. Close to 2000 criticality safety benchmark cases were used, as well as many benchmark shielding cases. All the runs could be compared with similar runs based on the nuclear data libraries ENDF/B-VII.1, JENDL-4.0, and JEFF-3.1.1 respectively. The results are that many of the criticality safety results obtained with TENDL-2012 are close to the ones for the other libraries. In particular the results for the thermal spectrum cases with LEU fuel are good. Nevertheless, there is a fair amount of cases for which the TENDL-2012 results are not as good as the other libraries. Especially a number of fast spectrum cases with reflectors are not well described. The results for the shielding benchmarks are mostly similar to the ones for the other libraries. Some isolated cases with differences are identified.

  20. Diagnostic Algorithm Benchmarking

    NASA Technical Reports Server (NTRS)

    Poll, Scott

    2011-01-01

    A poster for the NASA Aviation Safety Program Annual Technical Meeting. It describes empirical benchmarking on diagnostic algorithms using data from the ADAPT Electrical Power System testbed and a diagnostic software framework.

  1. Final report : PATTON Alliance gazetteer evaluation project.

    SciTech Connect

    Bleakly, Denise Rae

    2007-08-01

    In 2005 the National Ground Intelligence Center (NGIC) proposed that the PATTON Alliance provide assistance in evaluating and obtaining the Integrated Gazetteer Database (IGDB), developed for the Naval Space Warfare Command Research group (SPAWAR) under Advance Research and Development Activity (ARDA) funds by MITRE Inc., fielded to the text-based search tool GeoLocator, currently in use by NGIC. We met with the developers of GeoLocator and identified their requirements for a better gazetteer. We then validated those requirements by reviewing the technical literature, meeting with other members of the intelligence community (IC), and talking with both the United States Geologic Survey (USGS) and the National Geospatial Intelligence Agency (NGA), the authoritative sources for official geographic name information. We thus identified 12 high-level requirements from users and the broader intelligence community. The IGDB satisfies many of these requirements. We identified gaps and proposed ways of closing these gaps. Three important needs have not been addressed but are critical future needs for the broader intelligence community. These needs include standardization of gazetteer data, a web feature service for gazetteer information that is maintained by NGA and USGS but accessible to users, and a common forum that brings together IC stakeholders and federal agency representatives to provide input to these activities over the next several years. Establishing a robust gazetteer web feature service that is available to all IC users may go a long way toward resolving the gazetteer needs within the IC. Without a common forum to provide input and feedback, community adoption may take significantly longer than anticipated with resulting risks to the war fighter.

  2. Evaluation of the El Dorado micellar-polymer demonstration project

    SciTech Connect

    Vanhorn, L.E.

    1983-01-01

    The El Dorado Micellar-Polymer Demonstration Project has been a cooperative venture between Cities Service Co. and the U.S. Department of Energy. The objective of the project was to determine if it was technically and economically feasible to produce commercial volumes of oil using a micellar-polymer process in the El Dorado field. The project was designed to allow a side-by-side comparison of 2 distinctly different micellar-polymer processes in the same field in order that the associated benefits and problems of each could be determined. These are described and evaluated.

  3. Evaluating success of mobile health projects in the developing world.

    PubMed

    Ginige, J Anupama; Maeder, Anthony J; Long, Vanessa

    2014-01-01

    Many mobile health (mHealth) projects, typically deploying pilot or small scale implementations, have been undertaken in developing world settings and reported with a widely varying range of claims being made on their effectiveness and benefits. As a result, there is little evidence for which aspects of such projects lead to successful outcomes. This paper describes a literature review of papers from PubMed undertaken to identify strong contributions to execution and evaluation of mHealth projects in developing world settings, and suggests a template for classifying the main success factors to assist with collating evidence in the future.

  4. Virtual decoy sets for molecular docking benchmarks.

    PubMed

    Wallach, Izhar; Lilien, Ryan

    2011-02-28

    Virtual docking algorithms are often evaluated on their ability to separate active ligands from decoy molecules. The current state-of-the-art benchmark, the Directory of Useful Decoys (DUD), minimizes bias by including decoys from a library of synthetically feasible molecules that are physically similar yet chemically dissimilar to the active ligands. We show that by ignoring synthetic feasibility, we can compile a benchmark that is comparable to the DUD and less biased with respect to physical similarity.

  5. Collected notes from the Benchmarks and Metrics Workshop

    NASA Technical Reports Server (NTRS)

    Drummond, Mark E.; Kaelbling, Leslie P.; Rosenschein, Stanley J.

    1991-01-01

    In recent years there has been a proliferation of proposals in the artificial intelligence (AI) literature for integrated agent architectures. Each architecture offers an approach to the general problem of constructing an integrated agent. Unfortunately, the ways in which one architecture might be considered better than another are not always clear. There has been a growing realization that many of the positive and negative aspects of an architecture become apparent only when experimental evaluation is performed and that to progress as a discipline, we must develop rigorous experimental methods. In addition to the intrinsic intellectual interest of experimentation, rigorous performance evaluation of systems is also a crucial practical concern to our research sponsors. DARPA, NASA, and AFOSR (among others) are actively searching for better ways of experimentally evaluating alternative approaches to building intelligent agents. One tool for experimental evaluation involves testing systems on benchmark tasks in order to assess their relative performance. As part of a joint DARPA and NASA funded project, NASA-Ames and Teleos Research are carrying out a research effort to establish a set of benchmark tasks and evaluation metrics by which the performance of agent architectures may be determined. As part of this project, we held a workshop on Benchmarks and Metrics at the NASA Ames Research Center on June 25, 1990. The objective of the workshop was to foster early discussion on this important topic. We did not achieve a consensus, nor did we expect to. Collected here is some of the information that was exchanged at the workshop. Given here is an outline of the workshop, a list of the participants, notes taken on the white-board during open discussions, position papers/notes from some participants, and copies of slides used in the presentations.

  6. Correlational effect size benchmarks.

    PubMed

    Bosco, Frank A; Aguinis, Herman; Singh, Kulraj; Field, James G; Pierce, Charles A

    2015-03-01

    Effect size information is essential for the scientific enterprise and plays an increasingly central role in the scientific process. We extracted 147,328 correlations and developed a hierarchical taxonomy of variables reported in Journal of Applied Psychology and Personnel Psychology from 1980 to 2010 to produce empirical effect size benchmarks at the omnibus level, for 20 common research domains, and for an even finer grained level of generality. Results indicate that the usual interpretation and classification of effect sizes as small, medium, and large bear almost no resemblance to findings in the field, because distributions of effect sizes exhibit tertile partitions at values approximately one-half to one-third those intuited by Cohen (1988). Our results offer information that can be used for research planning and design purposes, such as producing better informed non-nil hypotheses and estimating statistical power and planning sample size accordingly. We also offer information useful for understanding the relative importance of the effect sizes found in a particular study in relationship to others and which research domains have advanced more or less, given that larger effect sizes indicate a better understanding of a phenomenon. Also, our study offers information about research domains for which the investigation of moderating effects may be more fruitful and provide information that is likely to facilitate the implementation of Bayesian analysis. Finally, our study offers information that practitioners can use to evaluate the relative effectiveness of various types of interventions. PsycINFO Database Record (c) 2015 APA, all rights reserved.

  7. A client/server database system for project evaluation

    SciTech Connect

    Brule, M.R.; Fair, W.B.; Jiang, J.; Sanvido, R.D.

    1994-12-31

    PETS (Project Evaluation Tool Set) is a networked client/server system that provides a full set of decision-support tools for evaluating the business potential of onshore and offshore development projects. This distributed workgroup computing system combines and streamlines preliminary design, routine cost estimation, economic evaluation, and risk analysis for conceptual developments as well as for ongoing projects and operations. A flexible and extendible client/server integration framework links in-house and third-party software applications with a database and an expert-system knowledgebase, and, where appropriate, links the applications among themselves. The capability and richness of inexpensive commercial operating systems and off-the-shelf applications have made building a client/server system like PETS possible in a relatively short time and at low cost. We will discuss the object-oriented design of the PETS system, detail its capabilities, and outline the methods used to integrate applications from other domains.

  8. Childhood Obesity Research Demonstration Project: Cross-Site Evaluation Methods

    PubMed Central

    Lee, Rebecca E.; Mehta, Paras; Thompson, Debbe; Bhargava, Alok; Carlson, Coleen; Kao, Dennis; Layne, Charles S.; Ledoux, Tracey; O'Connor, Teresia; Rifai, Hanadi; Gulley, Lauren; Hallett, Allen M.; Kudia, Ousswa; Joseph, Sitara; Modelska, Maria; Ortega, Dana; Parker, Nathan; Stevens, Andria

    2015-01-01

    Abstract Introduction: The Childhood Obesity Research Demonstration (CORD) project links public health and primary care interventions in three projects described in detail in accompanying articles in this issue of Childhood Obesity. This article describes a comprehensive evaluation plan to determine the extent to which the CORD model is associated with changes in behavior, body weight, BMI, quality of life, and healthcare satisfaction in children 2–12 years of age. Design/Methods: The CORD Evaluation Center (EC-CORD) will analyze the pooled data from three independent demonstration projects that each integrate public health and primary care childhood obesity interventions. An extensive set of common measures at the family, facility, and community levels were defined by consensus among the CORD projects and EC-CORD. Process evaluation will assess reach, dose delivered, and fidelity of intervention components. Impact evaluation will use a mixed linear models approach to account for heterogeneity among project-site populations and interventions. Sustainability evaluation will assess the potential for replicability, continuation of benefits beyond the funding period, institutionalization of the intervention activities, and community capacity to support ongoing program delivery. Finally, cost analyses will assess how much benefit can potentially be gained per dollar invested in programs based on the CORD model. Conclusions: The keys to combining and analyzing data across multiple projects include the CORD model framework and common measures for the behavioral and health outcomes along with important covariates at the individual, setting, and community levels. The overall objective of the comprehensive evaluation will develop evidence-based recommendations for replicating and disseminating community-wide, integrated public health and primary care programs based on the CORD model. PMID:25679060

  9. Developing integrated benchmarks for DOE performance measurement

    SciTech Connect

    Barancik, J.I.; Kramer, C.F.; Thode, Jr. H.C.

    1992-09-30

    The objectives of this task were to describe and evaluate selected existing sources of information on occupational safety and health with emphasis on hazard and exposure assessment, abatement, training, reporting, and control identifying for exposure and outcome in preparation for developing DOE performance benchmarks. Existing resources and methodologies were assessed for their potential use as practical performance benchmarks. Strengths and limitations of current data resources were identified. Guidelines were outlined for developing new or improved performance factors, which then could become the basis for selecting performance benchmarks. Data bases for non-DOE comparison populations were identified so that DOE performance could be assessed relative to non-DOE occupational and industrial groups. Systems approaches were described which can be used to link hazards and exposure, event occurrence, and adverse outcome factors, as needed to generate valid, reliable, and predictive performance benchmarks. Data bases were identified which contain information relevant to one or more performance assessment categories . A list of 72 potential performance benchmarks was prepared to illustrate the kinds of information that can be produced through a benchmark development program. Current information resources which may be used to develop potential performance benchmarks are limited. There is need to develop an occupational safety and health information and data system in DOE, which is capable of incorporating demonstrated and documented performance benchmarks prior to, or concurrent with the development of hardware and software. A key to the success of this systems approach is rigorous development and demonstration of performance benchmark equivalents to users of such data before system hardware and software commitments are institutionalized.

  10. Developing integrated benchmarks for DOE performance measurement

    SciTech Connect

    Barancik, J.I.; Kramer, C.F.; Thode, Jr. H.C.

    1992-09-30

    The objectives of this task were to describe and evaluate selected existing sources of information on occupational safety and health with emphasis on hazard and exposure assessment, abatement, training, reporting, and control identifying for exposure and outcome in preparation for developing DOE performance benchmarks. Existing resources and methodologies were assessed for their potential use as practical performance benchmarks. Strengths and limitations of current data resources were identified. Guidelines were outlined for developing new or improved performance factors, which then could become the basis for selecting performance benchmarks. Data bases for non-DOE comparison populations were identified so that DOE performance could be assessed relative to non-DOE occupational and industrial groups. Systems approaches were described which can be used to link hazards and exposure, event occurrence, and adverse outcome factors, as needed to generate valid, reliable, and predictive performance benchmarks. Data bases were identified which contain information relevant to one or more performance assessment categories . A list of 72 potential performance benchmarks was prepared to illustrate the kinds of information that can be produced through a benchmark development program. Current information resources which may be used to develop potential performance benchmarks are limited. There is need to develop an occupational safety and health information and data system in DOE, which is capable of incorporating demonstrated and documented performance benchmarks prior to, or concurrent with the development of hardware and software. A key to the success of this systems approach is rigorous development and demonstration of performance benchmark equivalents to users of such data before system hardware and software commitments are institutionalized.

  11. A benchmark server using high resolution protein structure data, and benchmark results for membrane helix predictions.

    PubMed

    Rath, Emma M; Tessier, Dominique; Campbell, Alexander A; Lee, Hong Ching; Werner, Tim; Salam, Noeris K; Lee, Lawrence K; Church, W Bret

    2013-03-27

    Helical membrane proteins are vital for the interaction of cells with their environment. Predicting the location of membrane helices in protein amino acid sequences provides substantial understanding of their structure and function and identifies membrane proteins in sequenced genomes. Currently there is no comprehensive benchmark tool for evaluating prediction methods, and there is no publication comparing all available prediction tools. Current benchmark literature is outdated, as recently determined membrane protein structures are not included. Current literature is also limited to global assessments, as specialised benchmarks for predicting specific classes of membrane proteins were not previously carried out. We present a benchmark server at http://sydney.edu.au/pharmacy/sbio/software/TMH_benchmark.shtml that uses recent high resolution protein structural data to provide a comprehensive assessment of the accuracy of existing membrane helix prediction methods. The server further allows a user to compare uploaded predictions generated by novel methods, permitting the comparison of these novel methods against all existing methods compared by the server. Benchmark metrics include sensitivity and specificity of predictions for membrane helix location and orientation, and many others. The server allows for customised evaluations such as assessing prediction method performances for specific helical membrane protein subtypes.We report results for custom benchmarks which illustrate how the server may be used for specialised benchmarks. Which prediction method is the best performing method depends on which measure is being benchmarked. The OCTOPUS membrane helix prediction method is consistently one of the highest performing methods across all measures in the benchmarks that we performed. The benchmark server allows general and specialised assessment of existing and novel membrane helix prediction methods. Users can employ this benchmark server to determine the most

  12. Intermediate evaluation of USAID/Cairo energy policy planning project

    SciTech Connect

    Wilbanks, T.J.; Wright, S.B. ); Barron, W.F. ); Kamel, A.M. ); Santiago, H.T. )

    1992-01-01

    Three years ago, a team from the Oak Ridge National Laboratory and the Oak Ridge Associated Universities, supplemented by an expert from the US Department of Energy and a senior Egyptian energy professional, carried out what was termed an intermediate evaluation'' of a major energy policy project in Egypt. Supported by USAID/Cairo, the project had concentrated on developing and strengthening an Organization for Energy Planning (OEP) within the Government of India, and it was actually scheduled to end less than a year after this evaluation. The evaluation was submitted to USAID/Cairo and circulated elsewhere in the US Agency for International Development and the Government of Egypt as an internal report. Over the next several years, the USAID energy planning project ended and the functions performed by OEP were merged with planning capabilities in the electric power sector. Now that the major issues addressed by the evaluation report have been resolved, we are making it available to a broader audience as a contribution to the general literature on development project evaluation and institution-building.

  13. Intermediate evaluation of USAID/Cairo energy policy planning project

    SciTech Connect

    Wilbanks, T.J.; Wright, S.B.; Barron, W.F.; Kamel, A.M.; Santiago, H.T.

    1992-09-01

    Three years ago, a team from the Oak Ridge National Laboratory and the Oak Ridge Associated Universities, supplemented by an expert from the US Department of Energy and a senior Egyptian energy professional, carried out what was termed an ``intermediate evaluation`` of a major energy policy project in Egypt. Supported by USAID/Cairo, the project had concentrated on developing and strengthening an Organization for Energy Planning (OEP) within the Government of India, and it was actually scheduled to end less than a year after this evaluation. The evaluation was submitted to USAID/Cairo and circulated elsewhere in the US Agency for International Development and the Government of Egypt as an internal report. Over the next several years, the USAID energy planning project ended and the functions performed by OEP were merged with planning capabilities in the electric power sector. Now that the major issues addressed by the evaluation report have been resolved, we are making it available to a broader audience as a contribution to the general literature on development project evaluation and institution-building.

  14. Evaluation and comparison of benchmark QSAR models to predict a relevant REACH endpoint: The bioconcentration factor (BCF).

    PubMed

    Gissi, Andrea; Lombardo, Anna; Roncaglioni, Alessandra; Gadaleta, Domenico; Mangiatordi, Giuseppe Felice; Nicolotti, Orazio; Benfenati, Emilio

    2015-02-01

    The bioconcentration factor (BCF) is an important bioaccumulation hazard assessment metric in many regulatory contexts. Its assessment is required by the REACH regulation (Registration, Evaluation, Authorization and Restriction of Chemicals) and by CLP (Classification, Labeling and Packaging). We challenged nine well-known and widely used BCF QSAR models against 851 compounds stored in an ad-hoc created database. The goodness of the regression analysis was assessed by considering the determination coefficient (R(2)) and the Root Mean Square Error (RMSE); Cooper's statistics and Matthew's Correlation Coefficient (MCC) were calculated for all the thresholds relevant for regulatory purposes (i.e. 100L/kg for Chemical Safety Assessment; 500L/kg for Classification and Labeling; 2000 and 5000L/kg for Persistent, Bioaccumulative and Toxic (PBT) and very Persistent, very Bioaccumulative (vPvB) assessment) to assess the classification, with particular attention to the models' ability to control the occurrence of false negatives. As a first step, statistical analysis was performed for the predictions of the entire dataset; R(2)>0.70 was obtained using CORAL, T.E.S.T. and EPISuite Arnot-Gobas models. As classifiers, ACD and logP-based equations were the best in terms of sensitivity, ranging from 0.75 to 0.94. External compound predictions were carried out for the models that had their own training sets. CORAL model returned the best performance (R(2)ext=0.59), followed by the EPISuite Meylan model (R(2)ext=0.58). The latter gave also the highest sensitivity on external compounds with values from 0.55 to 0.85, depending on the thresholds. Statistics were also compiled for compounds falling into the models Applicability Domain (AD), giving better performances. In this respect, VEGA CAESAR was the best model in terms of regression (R(2)=0.94) and classification (average sensitivity>0.80). This model also showed the best regression (R(2)=0.85) and sensitivity (average>0.70) for

  15. An Economic Evaluation Framework for Assessing Renewable Energy Projects

    SciTech Connect

    Omitaomu, Olufemi A; Badiru, Adedeji B

    2012-01-01

    It is becoming increasingly imperative to integrate renewable energy, such as solar and wind, into electricity generation due to increased regulations on air and water pollution and a sociopolitical desire to develop more clean energy sources. This increased spotlight on renewable energy requires evaluating competing projects using either conventional economic analysis techniques or other economics-based models and approaches in order to select a subset of the projects to be funded. Even then, there are reasons to suspect that techniques applied to renewable energy projects may result in decisions that will reject viable projects due to the use of a limited number of quantifiable and tangible attributes about the projects. This paper presents a framework for economic evaluation of renewable energy projects. The framework is based on a systems approach in which the processes within the entire network of the system, from generation to consumption, are accounted for. Furthermore, the framework uses the concept of fuzzy system to calculate the value of information under conditions of uncertainty.

  16. BENCHMARK DOSE TECHNICAL GUIDANCE DOCUMENT ...

    EPA Pesticide Factsheets

    The U.S. EPA conducts risk assessments for an array of health effects that may result from exposure to environmental agents, and that require an analysis of the relationship between exposure and health-related outcomes. The dose-response assessment is essentially a two-step process, the first being the definition of a point of departure (POD), and the second extrapolation from the POD to low environmentally-relevant exposure levels. The benchmark dose (BMD) approach provides a more quantitative alternative to the first step in the dose-response assessment than the current NOAEL/LOAEL process for noncancer health effects, and is similar to that for determining the POD proposed for cancer endpoints. As the Agency moves toward harmonization of approaches for human health risk assessment, the dichotomy between cancer and noncancer health effects is being replaced by consideration of mode of action and whether the effects of concern are likely to be linear or nonlinear at low doses. Thus, the purpose of this project is to provide guidance for the Agency and the outside community on the application of the BMD approach in determining the POD for all types of health effects data, whether a linear or nonlinear low dose extrapolation is used. A guidance document is being developed under the auspices of EPA's Risk Assessment Forum. The purpose of this project is to provide guidance for the Agency and the outside community on the application of the benchmark dose (BMD) appr

  17. Encouraging Strong Family Relationships. Policy Matters: Setting and Measuring Benchmarks for State Policies. A Discussion Paper for the "Policy Matters" Project

    ERIC Educational Resources Information Center

    Anyabwile, Thabiti

    2004-01-01

    "Policy Matters" is an initiative of the Center for the Study of Social Policy. The "Policy Matters" project is designed to develop and make available coherent, comprehensive information regarding the strength and adequacy of state policies affecting children, families, and communities. The project seeks to establish consensus among policy experts…

  18. Protein-Protein Docking Benchmark Version 3.0

    PubMed Central

    Hwang, Howook; Pierce, Brian; Mintseris, Julian; Janin, Joël; Weng, Zhiping

    2009-01-01

    We present version 3.0 of our publicly available protein-protein docking benchmark. This update includes 40 new test cases, representing a 48% increase from Benchmark 2.0. For all of the new cases, the crystal structures of both binding partners are available. As with Benchmark 2.0, SCOP1 (Structural Classification of Proteins) was used to remove redundant test cases. The 124 unbound-unbound test cases in Benchmark 3.0 are classified into 88 rigid-body cases, 19 medium difficulty cases, and 17 difficult cases, based on the degree of conformational change at the interface upon complex formation. In addition to providing the community with more test cases for evaluating docking methods, the expansion of Benchmark 3.0 will facilitate the development of new algorithms that require a large number of training examples. Benchmark 3.0 is available to the public at http://zlab.bu.edu/benchmark. PMID:18491384

  19. Using fish population models in hydro project evaluation

    SciTech Connect

    Power, M.V.; McKinley, R.S.

    1997-04-01

    Technical details of the use of population-level fisheries models in evaluating the environmental impacts of hydroelectric projects are described. Population models are grouped into four types, and evaluated in terms of usefulness of results, prediction uncertainty, and data requirements. The four types of models identified are stock-recruitment, Leslie matrix, life-cycle, and individual-based. Each model is discussed in some detail, with sample results and comparisons between models.

  20. [Evaluation of 12 pilot projects to improve outpatient palliative care].

    PubMed

    Schmidt-Wolf, G; Elsner, F; Lindena, G; Hilgers, R-D; Heussen, N; Rolke, R; Ostgathe, C; Radbruch, L

    2013-12-01

    With a priority programme the German Cancer Aid supported the development of quality-assured outpatient palliative care to cover the whole country. The 12 regional pilot projects funded with the aim to improve outpatient palliative care in different models and different frameworks were concurrently monitored and evaluated. The supported projects, starting and ending individually, documented all patients who were cared for using HOPE (Hospice and palliative care evaluation) and MIDOS (Minimal documentation system for palliative patients). Total data were analyzed for 3239 patients decriptively. In addition to the quantitative data the experiences of the projects were recorded in a number of workshops (2008, 2009, 2010, and 2012). In particular, the experiences reported in the final meeting in July 2012 were considered for this article as well as the final reports for the German Cancer Aid. In the quantitative evaluation 85.6% of 3239 palliative care patients had a cancer diagnosis. In all model projects the goal of a network with close cooperation of primary providers, social support, and outpatient and inpatient specialist services has been achieved. For all projects, the initial financing of the German Cancer Aid was extremely important, because contracts with health insurance funds were negotiated slowly, and could then be built on the experiences with the projects. The participants of the project-completion meeting emphasized the need to carry out a market analysis before starting palliative care organizations considering the different regional structures and target groups of patients. Education, training and continuing education programs contribute significantly to the network. A reliably funded coordination center/case management across all institutions is extremely important. © Georg Thieme Verlag KG Stuttgart · New York.

  1. Final Report on the Evaluation of Project Upswing's First Year.

    ERIC Educational Resources Information Center

    Plantec, P.; And Others

    This technical report describes the evaluation of the first year of Project Upswing, a 2-year experimental study to determine the potential contribution of volunteers in helping young children overcome learning difficulties. The three large groups of first grade children involved received tutoring either from specially trained volunteers,…

  2. Project GET SET, 1985-1986. OEA Evaluation Report.

    ERIC Educational Resources Information Center

    New York City Board of Education, Brooklyn. Office of Educational Assessment.

    This report evaluates the achievement of Project GET SET, a Bronx, New York, program designed to provide supportive and peer-tutorial services to Hispanic junior high school students for the following purposes: (1) to reinforce English language skills, especially reading and writing; (2) to reinforce native language arts skills; (3) to offer…

  3. Education North Evaluation Project. The Second Annual Report.

    ERIC Educational Resources Information Center

    Ingram, E. J.; McIntosh, R. G.

    The report and evaluation of Education North (a project designed to encourage parents, community members, and teachers in small, isolated, primarily Native and Metis communities in northern Alberta to work together to meet community educational needs) is comprised of three parts. Part One presents an update of Education North activities and…

  4. Service Learning in Medical Education: Project Description and Evaluation

    ERIC Educational Resources Information Center

    Borges, Nicole J.; Hartung, Paul J.

    2007-01-01

    Although medical education has long recognized the importance of community service, most medical schools have not formally nor fully incorporated service learning into their curricula. To address this problem, we describe the initial design, development, implementation, and evaluation of a service-learning project within a first-year medical…

  5. 40 CFR 57.604 - Evaluation of projects.

    Code of Federal Regulations, 2011 CFR

    2011-07-01

    ... 57.604 Protection of Environment ENVIRONMENTAL PROTECTION AGENCY (CONTINUED) AIR PROGRAMS (CONTINUED) PRIMARY NONFERROUS SMELTER ORDERS Research and Development Requirements § 57.604 Evaluation of projects. The research and development proposal shall include a provision for the employment of a qualified...

  6. 40 CFR 57.604 - Evaluation of projects.

    Code of Federal Regulations, 2012 CFR

    2012-07-01

    ... 57.604 Protection of Environment ENVIRONMENTAL PROTECTION AGENCY (CONTINUED) AIR PROGRAMS (CONTINUED) PRIMARY NONFERROUS SMELTER ORDERS Research and Development Requirements § 57.604 Evaluation of projects. The research and development proposal shall include a provision for the employment of a qualified...

  7. Developing and Evaluating a Cardiovascular Risk Reduction Project.

    ERIC Educational Resources Information Center

    Brownson, Ross C.; Mayer, Jeffrey P.; Dusseault, Patricia; Dabney, Sue; Wright, Kathleen; Jackson-Thompson, Jeannette; Malone, Bernard; Goodman, Robert

    1997-01-01

    Describes the development and baseline evaluation data from the Ozark Heart Health Project, a community-based cardiovascular disease risk reduction program in rural Missouri that targeted smoking, physical inactivity, and poor diet. Several Ozark counties participated in either intervention or control groups, and researchers conducted surveillance…

  8. ESEA Title I Projects Evaluation Report 1967, Volume I.

    ERIC Educational Resources Information Center

    Pittsburgh Public Schools, PA.

    Reports of Pittsburgh's 1967 ESEA Title I projects are presented in two volumes. The 17 reports in Volume I, which adhere to the procedures established in an evaluation model, are of programs in communication skills, camping, vocational education, music, standard English, social development, revised class organization, remedial reading by means of…

  9. In-depth Evaluation of the Associated Schools Project.

    ERIC Educational Resources Information Center

    Churchill, Stacy; Omari, Issa

    1980-01-01

    Describes methods and conclusions of an in-depth evaluation of the UNESCO Associated Schools Project for International Understanding. The report includes suggestions for improving course content, teaching methods, and instructional materials. Improvements in program quality, international coordination, information dissemination, and expansion into…

  10. Resource Sharing and Public Education. Project Evaluation Report.

    ERIC Educational Resources Information Center

    Leton, Donald A.

    Two exemplary projects developed to maximize student options through the sharing of facilities, programs, and options were evaluated. The Remedial-Developmental Program developed a communications network between McKinley and Roosevelt High Schools and Kapiolani Community College to share student information and use it for individual student…

  11. Niagara Falls HEW 309 Project 1974-1975: Evaluation Report.

    ERIC Educational Resources Information Center

    Skeen, Elois M.

    The document reports an outside evaluation of a Niagara Falls Adult Basic Education Program special project entitled "Identification of Preferred Cognitive Styles and Matching Adult Reading Program Alternatives for the 0-4 Grade Levels." It was concerned with (1) research, training in cognitive style mapping, and development of a survey…

  12. Developing and Evaluating a Cardiovascular Risk Reduction Project.

    ERIC Educational Resources Information Center

    Brownson, Ross C.; Mayer, Jeffrey P.; Dusseault, Patricia; Dabney, Sue; Wright, Kathleen; Jackson-Thompson, Jeannette; Malone, Bernard; Goodman, Robert

    1997-01-01

    Describes the development and baseline evaluation data from the Ozark Heart Health Project, a community-based cardiovascular disease risk reduction program in rural Missouri that targeted smoking, physical inactivity, and poor diet. Several Ozark counties participated in either intervention or control groups, and researchers conducted surveillance…

  13. Evaluation of Project TREC: Teaching Respect for Every Culture.

    ERIC Educational Resources Information Center

    Mitchell, Stephanie

    The purpose of Teaching Respect for Every Culture (TREC) was to ensure that racial/ethnic, gender, disability, and other circumstances did not bar student access to alcohol/drug education, prevention, and intervention services. This report describes the implementation and evaluation of the TREC Project. Five objectives of TREC were to: (1)…

  14. Instruments and Scoring Guide of the Experiential Education Evaluation Project.

    ERIC Educational Resources Information Center

    Conrad, Dan; Hedin, Diane

    As a result of the Experiential Education Evaluation Project the publication identifies instruments used to measure and assess experiential learning programs. The following information is given for each instrument: rationale for its inclusion in the study; precise issues or outcomes designed to measure, validity and reliability data; and…

  15. Project "Freestyle": Ad Hoc: Fast-Turn-Around Evaluation.

    ERIC Educational Resources Information Center

    Smith, Karen

    Project "Freestyle" involved the development of prototypical television materials and a comic book intended to combat sex-role stereotyping in career-related attitudes of nine to twelve-year-old children. At various times during the early developmental stages of "Freestyle" materials, "ad hoc fast-turn-around" formative evaluations were conducted.…

  16. Evaluation of the Universal Design for Learning Projects

    ERIC Educational Resources Information Center

    Cooper-Martin, Elizabeth; Wolanin, Natalie

    2014-01-01

    The Office of Shared Accountability evaluated the "Universal Design for Learning" (UDL) projects during spring 2013. UDL is an instructional framework that seeks to give all students equal opportunities to learn, by providing multiple means of representation, of action and expression, and of engagement. To inform future implementation…

  17. Native Speaker Program: Evaluation of a Pilot Project.

    ERIC Educational Resources Information Center

    Westcott, D.B.

    1968-01-01

    With the idea of providing a valid conditioning and testing situation by which the student's self-confidence and effectiveness in communicating in a foreign language could be improved, this study evaluates a pilot project designed to identify and measure the effects of having a native speaker in the classroom for a period of time as opposed to a…

  18. Expedited Permanency Planning: Evaluation of the Kentucky Adoptions Opportunities Project.

    ERIC Educational Resources Information Center

    Martin, Mavin H.; Barbee, Anita P.; Antle, Becky F.; Sar, Bibhuti

    2002-01-01

    Presents evaluation findings of a 3-year Kentucky Adoptions Opportunities Project. Notes that a majority of children had one or both parents coping with multiple risk factors including mental illness, substance abuse, mental retardation, or family violence. Discusses major barriers to permanency, as well as policy and practice implications in the…

  19. Project Achieve Evaluation Report: Year One, 2001-2002.

    ERIC Educational Resources Information Center

    Speas, Carol

    This report is an evaluation of the pilot year of Project Achieve, a major local instructional initiative at six elementary schools and two middle schools in the Wake County Public School System (WCPSS), North Carolina, that was designed to help reach the WCPSS goal of 95% of students at or above grade level. Participating schools had a higher…

  20. Orthographic Projection. Courseware Evaluation for Vocational and Technical Education.

    ERIC Educational Resources Information Center

    Turner, Gordon; And Others

    This courseware evaluation rates the Orthographic Projection program developed by Hobar Publications. (The program--not contained in this document--uses computer graphics to present abstract visual concepts such as points, lines, and planes.) Part A describes the program in terms of subject area and hardware requirements (Apple II), indicates its…