Science.gov

Sample records for 10th grade benchmark

  1. Predicting 10th Grade FCAT Success. Research Brief. Volume 0401

    ERIC Educational Resources Information Center

    Froman, Terry; Bayne, Joseph

    2004-01-01

    Florida law requires that students achieve a passing score on the Grade 10 Florida Comprehensive Assessment Test (FCAT) to qualify for a standard high school diploma (Section 1008.22(3)(c)5, Florida Statutes). Students who were administered the Grade 10 FCAT for the first time during the 2002 administrations or later must earn a developmental…

  2. An Early Warning System: Predicting 10th Grade FCAT Success from 6th Grade FCAT Performance. Research Brief. Volume 0711

    ERIC Educational Resources Information Center

    Froman, Terry; Brown, Shelly; Lapadula, Maria

    2008-01-01

    This Research Brief presents a method for predicting 10th grade Florida Comprehensive Assessment Test (FCAT) success from 6th grade FCAT performance. A simple equation provides the most probable single score prediction, and give-or-take error margins define high and low probability zones for expected 10th grade scores. In addition, a double-entry…

  3. Indiana's Academic Standards: 10th Grade English/Language Arts.

    ERIC Educational Resources Information Center

    Indiana State Dept. of Education, Indianapolis.

    This booklet of academic standards spells out what students should know and be able to do in Grade 10 English/Language Arts. The booklet gives examples to help students understand what is required to meet the standards and provides a list of 10 things parents can do to help their child get a good education. It outlines the following seven

  4. Cigarette Use in 6th Through 10th Grade: The Sarasota County Demonstration Project

    ERIC Educational Resources Information Center

    Zapata, Lauren B.; Forthofer, Melinda S.; Eaton, Danice K.; Brown, Kelli McCormack; Bryant, Carol A.; Reynolds, Sherri T.; McDermot, Robert J.

    2004-01-01

    Objectives: To identify factors associated with cigarette smoking in the 6 th -grade through 10 th -grade youth population of Sarasota County, Florida. Methods: A closed-ended, quantitative survey was completed by 2004 youth and used to extract population-specific data on the correlates of cigarette use. Results: A range of factors influence

  5. The School Compositional Effect of Single Parenthood on 10th-Grade Achievement.

    ERIC Educational Resources Information Center

    Pong, Suet-ling

    1998-01-01

    Uses data from the National Education Longitudinal Study (NELS) to reveal a detrimental contextual effect on 10th-grade mathematics and reading achievement associated with attending a school with a high concentration of children from single-parent homes. Maintains this effect is evident even when individual demographic characteristics and family

  6. Predicting 3rd Grade and 10th Grade FCAT Success for 2006-07. Research Brief. Volume 0601

    ERIC Educational Resources Information Center

    Froman, Terry; Rubiera, Vilma

    2006-01-01

    For the past few years the Florida School Code has set the Florida Comprehensive Assessment Test (FCAT) performance requirements for promotion of 3rd graders and graduation for 10th graders. Grade 3 students who do not score at level 2 or higher on the FCAT SSS Reading must be retained unless exempted for special circumstances. Grade 10 students…

  7. Science Grades K-4 Benchmarks.

    ERIC Educational Resources Information Center

    Arkansas State Dept. of Education, Little Rock.

    This document explains the Arkansas science benchmarks for grades K-4 which were developed to provide guidance to science teachers. The content standards for three strands--physical science systems, life science systems, and Earth science/space science systems--were identified based on grade levels. (YDS)

  8. Tobacco use among 10th grade students in Istanbul and related variables.

    PubMed

    Evren, Cuneyt; Evren, Bilge; Bozkurt, Muge

    2014-04-01

    Aim of this study was to determine prevalence of cigarette smoking and hookah use among 10th grade students in Istanbul, Turkey, and to compare sociodemographic, psychological and behavioral variables according to frequency of tobacco use. Cross-sectional online self-report survey conducted in 45 schools from the 15 districts in Istanbul/Turkey. The questionnaire included sections about demographic data, family characteristics, school life, psychological symptoms and use of substances including tobacco, hookah, alcohol, marijuana, volatiles, heroin, cocaine, non-prescribed legal tranquillizers (benzodiazepines, alprazolam etc.) and illegal tranquillizers (flunitrazepam). The analyses were conducted based on the 4957 subjects. Trial at least once in life is observed as 45.4% for hookah use and as 24.4% for cigarette use. Risk of hookah and cigarette use was significantly higher in male students than in female students. Frequency of tobacco use is related with various sociodemographic, psychological and behavioral variables. Our data also shows that using tobacco and alcohol increases the risk of all the other substances use and these effects are interrelated. The data suggest that there is a link between tobacco use and substance use, psychological, behavioral and social factors. There is also a strong association between tobacco use and suicidal behavior as well as self-mutilative, impulsive, hyperactive, delinquent, aggressive and behavioral problems. The illumination of these relationships may be relevant in prevention and management of tobacco use as well as important problems, such as substance use, impulsivity, hyperactivity, delinquent, aggressive self-mutilative and suicidal behavior among 10th grade students in Istanbul. PMID:24655632

  9. The Effect of Case-Based Instruction on 10th Grade Students' Understanding of Gas Concepts

    ERIC Educational Resources Information Center

    Yalinkaya, Eylem; Boz, Yezdan

    2015-01-01

    The main purpose of the present study was to investigate the effect of case-based instruction on remedying 10th grade students' alternative conceptions related to gas concepts. 128 tenth grade students from two high schools participated in this study. In each school, one of the classes was randomly assigned as the experimental group and the

  10. Investigating the Effects of a DNA Fingerprinting Workshop on 10th Grade Students' Self Efficacy and Attitudes toward Science.

    ERIC Educational Resources Information Center

    Sonmez, Duygu; Simcox, Amanda

    The purpose of this study was investigate the effects of a DNA Fingerprinting Workshop on 10th grade students' self efficacy and attitudes toward science. The content of the workshop based on high school science curriculum and includes multimedia instruction, laboratory experiment and participation of undergraduate students as mentors. N=93

  11. Examining General and Specific Factors in the Dimensionality of Oral Language and Reading in 4th-10th Grades

    ERIC Educational Resources Information Center

    Foorman, Barbara R.; Koon, Sharon; Petscher, Yaacov; Mitchell, Alison; Truckenmiller, Adrea

    2015-01-01

    The objective of this study was to explore dimensions of oral language and reading and their influence on reading comprehension in a relatively understudied population--adolescent readers in 4th through 10th grades. The current study employed latent variable modeling of decoding fluency, vocabulary, syntax, and reading comprehension so as to

  12. Investigating the Effects of a DNA Fingerprinting Workshop on 10th Grade Students' Self Efficacy and Attitudes toward Science.

    ERIC Educational Resources Information Center

    Sonmez, Duygu; Simcox, Amanda

    The purpose of this study was investigate the effects of a DNA Fingerprinting Workshop on 10th grade students' self efficacy and attitudes toward science. The content of the workshop based on high school science curriculum and includes multimedia instruction, laboratory experiment and participation of undergraduate students as mentors. N=93…

  13. Influence of V-Diagrams on 10th Grade Turkish Students' Achievement in the Subject of Mechanical Waves

    ERIC Educational Resources Information Center

    Tekes, Hanife; Gonen, Selahattin

    2012-01-01

    The purpose of the present study was to examine how the use of V-diagrams one of the learning techniques used in laboratory studies in experiments conducted regarding the 10th grade lesson unit of "waves" influenced students' achievements. In the study, a quasi-experimental design with a pretest and posttest control group was used. The study was…

  14. Context Effects in Norwegian 10th-Grade Students' Reports on Learning Strategies Using the Cross-Curricular Competencies Instrument

    ERIC Educational Resources Information Center

    Samuelstuen, Marit S.; Braten, Ivar; Valas, Harald

    2007-01-01

    This study examined contextual effects in Norwegian 10th-grade students' reports on cognitive and metacognitive strategies when using the Cross-Curricular Competencies (CCC) instrument. The CCC was first administered as the final task in a long (three-hour) testing session in the Organisation for Economic Co-operation and Development Program for

  15. Examining General and Specific Factors in the Dimensionality of Oral Language and Reading in 4th-10th Grades

    ERIC Educational Resources Information Center

    Foorman, Barbara R.; Koon, Sharon; Petscher, Yaacov; Mitchell, Alison; Truckenmiller, Adrea

    2015-01-01

    The objective of this study was to explore dimensions of oral language and reading and their influence on reading comprehension in a relatively understudied population--adolescent readers in 4th through 10th grades. The current study employed latent variable modeling of decoding fluency, vocabulary, syntax, and reading comprehension so as to…

  16. Progression in Complexity: Contextualizing Sustainable Marine Resources Management in a 10th Grade Classroom

    NASA Astrophysics Data System (ADS)

    Bravo-Torija, Beatriz; Jiménez-Aleixandre, María-Pilar

    2012-01-01

    Sustainable management of marine resources raises great challenges. Working with this socio-scientific issue in the classroom requires students to apply complex models about energy flow and trophic pyramids in order to understand that food chains represent transfer of energy, to construct meanings for sustainable resources management through discourse, and to connect them to actions and decisions in a real-life context. In this paper we examine the process of elaboration of plans for resources management in a marine ecosystem by 10th grade students (15-16 year) in the context of solving an authentic task. A complete class ( N = 14) worked in a sequence about ecosystems. Working in small groups, the students made models of energy flow and trophic pyramids, and used them to solve the problem of feeding a small community for a long time. Data collection included videotaping and audiotaping of all of the sessions, and collecting the students' written productions. The research objective is to examine the process of designing a plan for sustainable resources management in terms of the discursive moves of the students across stages in contextualizing practices, or different degrees of complexity (Jiménez-Aleixandre & Reigosa International Journal of Science Education, 14(1): 51-61 2006), understood as transformations from theoretical statements to decisions about the plan. The analysis of students' discursive moves shows how the groups progressed through stages of connecting different models, between them and with the context, in order to solve the task. The challenges related to taking this sustainability issue to the classroom are discussed.

  17. Human papillomavirus vaccine uptake, knowledge and attitude among 10th grade students in Berlin, Germany, 2010

    PubMed Central

    Stöcker, Petra; Dehnert, Manuel; Schuster, Melanie; Wichmann, Ole; Deleré, Yvonne

    2013-01-01

    Purpose: Since March 2007, the Standing Committee on Vaccination (STIKO) recommends HPV vaccination for all 12–17 y-old females in Germany. In the absence of an immunization register, we aimed at assessing HPV-vaccination coverage and knowledge among students in Berlin, the largest city in Germany, to identify factors influencing HPV-vaccine uptake. Results: Between September and December 2010, 442 students completed the questionnaire (mean age 15.1; range 14–19). In total 281/442 (63.6%) students specified HPV correctly as a sexually transmitted infection. Of 238 participating girls, 161 (67.6%) provided their vaccination records. Among these, 66 (41.0%) had received the recommended three HPV-vaccine doses. Reasons for being HPV-unvaccinated were reported by 65 girls: Dissuasion from parents (40.2%), dissuasion from their physician (18.5%), and concerns about side-effects (30.8%) (multiple choices possible). The odds of being vaccinated increased with age (Odds Ratio (OR) 2.19, 95% Confidence Interval (CI) 1.16, 4.15) and decreased with negative attitude toward vaccinations (OR = 0.33, 95%CI 0.13, 0.84). Methods: Self-administered questionnaires were distributed to 10th grade school students in 14 participating schools in Berlin to assess socio-demographic characteristics, knowledge, and statements on vaccinations. Vaccination records were reviewed. Multivariable statistical methods were applied to identify independent predictors for HPV-vaccine uptake among female participants. Conclusions: HPV-vaccine uptake was low among school girls in Berlin. Both, physicians and parents were influential regarding their HPV-vaccination decision even though personal perceptions played an important role as well. School programs could be beneficial to improve knowledge related to HPV and vaccines, and to offer low-barrier access to HPV vaccination. PMID:22995838

  18. The Earlier the Better? Taking the AP in 10th Grade. Research Report No. 2012-10

    ERIC Educational Resources Information Center

    Rodriguez, Awilda; McKillip, Mary E. M.; Niu, Sunny X.

    2013-01-01

    In this report, the authors examine the impact of scoring a 1 or 2 on an AP Exam in 10th grade on later AP Exam participation and performance. As access to AP courses increases within and across schools, a growing number of students are taking AP courses and exams in the earlier grades of high school. Using a matched sample of AP and no-AP

  19. Comparing Overexcitabilities of Gifted and Non-Gifted 10th Grade Students in Turkey

    ERIC Educational Resources Information Center

    Yakmaci-Guzel, Buket; Akarsu, Fusun

    2006-01-01

    The study compares overexcitability scores of Turkish 10th graders who are grouped in terms of their intellectual abilities, motivation, creativity and leadership as well as gender. 711 students who were administered Raven Advanced Progressive Matrices Test (APM) were divided into three intellectual ability categories. From this pool, 105 subjects

  20. Changes in Educational Expectations between 10th and 12th Grades across Cohorts

    ERIC Educational Resources Information Center

    Park, Sueuk; Wells, Ryan; Bills, David

    2015-01-01

    The mean levels of educational expectations of American high school students have increased over the past generation; individual educational expectations change as students mature. Using the National Education Longitudinal Study and the Education Longitudinal Study, we examined simultaneously the changes in individuals' expectations from 10th to…

  1. Mental health problems in the 10th grade and non-completion of upper secondary school: the mediating role of grades in a population-based longitudinal study

    PubMed Central

    2014-01-01

    Background School drop-out is a problem all over the world with adverse life-course consequences. The aim of this paper is to study how internalising and externalising problems in the 10th grade are associated with non-completion of upper secondary school, and to examine the mediating role of grade points in the 10th grade across general academic and vocational tracks in upper secondary school. We also study the impact of health behaviour. Methods Population-based health surveys were linked with Norwegian registries on education and sociodemographic factors (n?=?10 931). Mental health was assessed by the self-report Strengths and Difficulties Questionnaire. Logistic regression was used to analyse the relations between mental health and health behaviour in 10th grade and non-completion of upper secondary school. The mediating effect of grade points was studied by causal mediation analysis. Results Adolescents not completing upper secondary school reported more externalising problems and girls more internalising problems in the 10th grade, after adjustments. Smoking and physical inactivity increased the odds of non-completion of upper secondary school. Causal mediation analyses showed that a reduction in externalising problems of 10 percentage points led to lower rates of non-completion of 45 percentage points, and about three-quarters of this total effect was mediated by grades. For internalising problems the total effect was significant only for girls (1 percentage point), and the mediated effect of grades was about 30%. The effect of mental health problems on school dropout was mainly the same in both vocational and general tracks. Conclusions Assuming a causal relationship from mental health problems to school performance, this study suggests that externalising problems impair educational attainment. A reduction of such problems may improve school performance, reduce school drop-out and reduce the adverse life-course consequences. PMID:24406098

  2. Water: How Good is Good Enough? Teacher's Guide. Science Module (9th-10th Grade Chemistry).

    ERIC Educational Resources Information Center

    Georgia Univ., Athens. Coll. of Education.

    This is a teacher's guide for a module designed to integrate environmental education into ninth- and tenth-grade chemistry classes. The module, pilot tested in Gwinnett County, Georgia in classes of students, many of whom had learning disabilities, emphasizes activity learning and considerable review. The module is divided into four parts. Part

  3. Scope, Sequence & Coordination: 9th and 10th Grade Science. Supplementary Evaluation Report, December, 1997.

    ERIC Educational Resources Information Center

    Lawrenz, Frances; Huffman, Douglas

    The purpose of this report is to synthesize two other evaluation reports ascertaining the effectiveness of the Scope, Sequence & Coordination (SS&C) project on ninth and tenth grade student achievement of the National Science Education Standards. The overall evaluation utilized a time lag, post-test-only design in which the perceptions and

  4. Science Grades 5-8 Benchmarks.

    ERIC Educational Resources Information Center

    Arkansas State Dept. of Education, Little Rock.

    This document explains the Arkansas science benchmarks for grades 5-8 which were developed to provide guidance to science teachers. The content standards for three strands--physical science systems, life science systems, and Earth science/space science systems--were identified based on grade levels. (YDS)

  5. Predicting 3rd Grade and 10th Grade FCAT Success for 2007-08. Research Brief. Volume 0702

    ERIC Educational Resources Information Center

    Froman, Terry; Rubiera, Vilma

    2008-01-01

    For the past few years the Florida School Code has set the Florida Comprehensive Assessment Test (FCAT) performance requirements for promotion of 3rd graders and graduation for 10 graders. Grade 3 students who do not score at level 2 or higher on the FCAT SSS Reading must be retained unless exempted for special circumstances. Grade 10 students…

  6. Investigating the intrinsic and extrinsic work values of 10th grade students in science-oriented charter schools

    NASA Astrophysics Data System (ADS)

    Ozer, Ozgur

    The purpose of this study was to investigate to what extent gender, achievement level, and income level predict the intrinsic and extrinsic work values of 10th grade students. The study explored whether group differences were good predictors of scores in work values. The research was a descriptive, cross-sectional study conducted on 131 10th graders who attended science-oriented charter schools. Students took Super's Work Values Instrument, a Likert-type test that links to 15 work values, which can be categorized as intrinsic and extrinsic values (Super, 1970). Multiple regression analysis was employed as the main analysis followed by ANCOVA. Multiple regression analysis results indicated that there is evidence that 8.9% of the variance in intrinsic work values and 10.2% of the variance in extrinsic work values can be explained by the independent variables ( p < .05). Achievement Level and Income Level may help predict intrinsic work value scores; Achievement Level may also help predict extrinsic work values. Achievement Level was the covariate in ANCOVA. Results indicated that males (M = .174) in this sample have a higher mean of extrinsic work values than that of females (M = -.279). However, there was no statistically significant difference between the intrinsic work values by gender. One possible interpretation of this might be school choice; students in these science-oriented charter schools may have higher intrinsic work values regardless of gender. Results indicated that there was no statistically significant difference among the means of extrinsic work values by income level (p < .05). However, free lunch students (M = .268) have a higher mean of intrinsic work values than that of paid lunch students ( M = -.279). A possible interpretation of this might be that lower income students benefit greatly from the intrinsic work values in overcoming obstacles. Further research is needed in each of these areas. The study produced statistically significant results with little practical significance. Students, parents, teachers, and counselors may still be advised to consider the work value orientations of students during the career choice process.

  7. Gender and Ethnic Differences in Smoking, Drinking and Illicit Drug Use among American 8th, 10th and 12th Grade Students, 1976-2000.

    ERIC Educational Resources Information Center

    Wallace, John M., Jr.; Bachman, Jerald G.; O'Malley, Patrick M.; Schulenberg, John E.; Cooper, Shauna M.; Johnston, Lloyd D.

    2003-01-01

    Paper examines ethnic differences in licit and illicit drug use among American 8th, 10th, and 12th grade students, with a particular focus on girls. Across ethnic groups, drug use is highest among Native American girls and lowest among black and Asian American girls. Trend data suggest that girls' and boys' drug use patterns are converging.

  8. Perceptions of 9th and 10th Grade Students on How Their Environment, Cognition, and Behavior Motivate Them in Algebra and Geometry Courses

    ERIC Educational Resources Information Center

    Harootunian, Alen

    2012-01-01

    In this study, relationships were examined between students' perception of their cognition, behavior, environment, and motivation. The purpose of the research study was to explore the extent to which 9th and 10th grade students' perception of environment, cognition, and behavior can predict their motivation in Algebra and Geometry

  9. The Basic Program of Vocational Agriculture in Louisiana. Ag I and Ag II (9th and 10th Grades). Volume I. Bulletin 1690-I.

    ERIC Educational Resources Information Center

    Louisiana State Dept. of Education, Baton Rouge. Div. of Vocational Education.

    This document is the first volume of a state curriculum guide on vocational agriculture for use in the 9th and 10th grades in Louisiana. Three instructional areas are profiled in this volume: orientation to vocational agriculture, agricultural leadership, and soil science. The three units of the orientation area cover introducing beginning…

  10. Mental and somatic health complaints associated with school bullying between 10th and 12th grade students; results from cross sectional studies in Oslo, Norway

    PubMed Central

    2009-01-01

    Background Bullying is a widespread and serious problem that might influence both mental and psychical well being as well as school performance and social life. The aim of this study was to describe the prevalence of bullying, mental health problems and psychical complaints among 10th and 12th grade students and to analyze the association between bullying, mental health problems and muscle and skeletal complaints. Methods Two cross sectional studies of adolescents living in Oslo, Norway the first conducted in 2001 among 10th grade students (15/16 years old) and the second in 2004 among 12th grade students (18/19 years old). Both surveys were based on self report, were mostly school based and had almost identical questionnaires. There were around 3700 participants in both surveys, but the participation rate was lower in the latter survey (88 versus 80%). The Hopkins Symptoms Check List (HSCL-10) and the Strength and Difficulties Questionnaire (SDQ) were used to measure mental health problems. Results Bullying is decreasing both among boys and girls while the prevalence of internalized mental health problems are increasing from 10th to 12th grade. For muscle and skeletal pain there is a diverging trend between boys and girls, with an increase among girls and a decrease among boys. The highest Odds Ratios, as a measure for the association between bullying, mental health problems and pain, were found for internalized mental health problems at both 10th and 12th grade both for boys and girls. Conclusion Both internalized and externalized mental health problems together with pain seem to be associated with bullying irrespective of school type and gender. PMID:19309497

  11. Trends in Substance Use Among 6th- to 10th-Grade Students From 1998 to 2010: Findings From a National Probability Study

    PubMed Central

    Brooks-Russell, Ashley; Farhat, Tilda; Haynie, Denise; Simons-Morton, Bruce

    2016-01-01

    Of the handful of national studies tracking trends in adolescent substance use in the United States, only the Health Behavior in School-Aged Children (HBSC) study collects data from 6th through 10th graders. The purpose of this study was to examine trends from 1998 to 2010 (four time points) in the prevalence of tobacco, alcohol, and marijuana use among 6th through 10th graders. Differences in trends by grade, gender, and race/ethnicity were examined for each substance use behavior, with a primary focus on trends for sixth and seventh graders. Overall, there were significant declines in tobacco, alcohol, and marijuana use from 1998 to 2010. The declines were largest for the younger grades, which suggest promise for future declines among high school students as these cohorts age into high school. PMID:26949282

  12. Teachers' and Students' Perceptions Regarding Technology-Assisted Instruction in 10th-Grade Mathematics Classrooms

    ERIC Educational Resources Information Center

    Brown, Martha Jean Ware

    2012-01-01

    Tenth grade students in a high school in the southern United States were not meeting minimum mathematics competencies. Educators, families, and students needed more information on ways to enhance mathematics understanding and performance. The purpose of this qualitative case study was to examine students' and teachers' perceptions…

  13. Growth: How Much is Too Much? Student Book. Science Module (9th-10th Grade Biology). Revised Edition.

    ERIC Educational Resources Information Center

    Georgia Univ., Athens. Coll. of Education.

    This learning module is designed to integrate environmental education into ninth- and tenth-grade chemistry classes. This module and a companion social studies module were pilot tested in Gwinnett County, Georgia in 1975-76. The module is divided into four parts. Part one provides a broad overview of unit content and proposes questions to

  14. Teachers' and Students' Perceptions Regarding Technology-Assisted Instruction in 10th-Grade Mathematics Classrooms

    ERIC Educational Resources Information Center

    Brown, Martha Jean Ware

    2012-01-01

    Tenth grade students in a high school in the southern United States were not meeting minimum mathematics competencies. Educators, families, and students needed more information on ways to enhance mathematics understanding and performance. The purpose of this qualitative case study was to examine students' and teachers' perceptions

  15. Assessing the Judicious Use of the "Language" of Certain Types of Graphs by 10th Grade Biology Pupils.

    ERIC Educational Resources Information Center

    Dreyfus, Amos; Mazouz, Yossef

    1992-01-01

    Assesses the ability of tenth grade students (n=364) to acquire meanings of graphs that are frequently used in their biology textbooks. Indicates that the main source of failure to process information equally well from tables and graphs was not a lack of basic analytical skills but rather a lack of understanding of the relationship between…

  16. Water: How Good is Good Enough? Teacher's Guide. Social Studies Module (9th-10th Grade Social Studies).

    ERIC Educational Resources Information Center

    Georgia Univ., Athens. Coll. of Education.

    This teacher's guide is for an environmental education module to integrate topics of water quality in ninth- and tenth-grade social studies classes. This module was pilot tested in Gwinnett County, Georgia in 1975-76. Included in the guide are overall objectives, the module sequence, an introduction, a suggested teaching sequence, a word review

  17. A Typology of Chemistry Classroom Environments: Exploring the Relationships between 10th Grade Students' Perceptions, Attitudes and Gender

    ERIC Educational Resources Information Center

    Giallousi, M.; Gialamas, V.; Pavlatou, E. A.

    2013-01-01

    The present study was the first in Greece in which educational effectiveness theory constituted a knowledge base for investigating the impact of chemistry classroom environment in 10 Grade students' enjoyment of class. An interpretive heuristic schema was developed and utilised in order to incorporate two factors of teacher behaviour at

  18. Water: How Good is Good Enough? Student Book. Science Module (9th-10th Grade Chemistry). Revised Edition.

    ERIC Educational Resources Information Center

    Georgia Univ., Athens. Coll. of Education.

    This learning module is designed to integrate environmental education into ninth- and tenth-grade chemistry classes. This module and a companion social studies module were pilot tested in Gwinnett County, Georgia in classes of students, many of whom had learning disabilities. It emphasizes activity learning. The module is divided into four parts.

  19. Water: How Good is Good Enough? Student Book. Social Studies Module (9th-10th Grade Social Studies).

    ERIC Educational Resources Information Center

    Georgia Univ., Athens. Coll. of Education.

    This is an environmental education module for integrating topics of water quality in ninth- and tenth-grade social studies classes. The module was pilot tested in Gwinnett County, Georgia in 1975-76. The module sequence is divided into four parts. The first part provides an introductory episode to stimulate student interest in an environmental

  20. Growth: How Much is Too Much? Student Book. Science Module (9th-10th Grade Biology). Revised Edition.

    ERIC Educational Resources Information Center

    Georgia Univ., Athens. Coll. of Education.

    This learning module is designed to integrate environmental education into ninth- and tenth-grade chemistry classes. This module and a companion social studies module were pilot tested in Gwinnett County, Georgia in 1975-76. The module is divided into four parts. Part one provides a broad overview of unit content and proposes questions to…

  1. Growth: How Much is Too Much? Student Book. Social Studies Module (9th-10th Grade Social Studies). Revised Edition.

    ERIC Educational Resources Information Center

    Georgia Univ., Athens. Coll. of Education.

    This learning module is designed to integrate environmental education into ninth- and tenth-grade social studies courses. The module and a parallel module designed for chemistry classes were pilot tested in Gwinnett County, Georgia in 1975-76. The module is divided into four parts. The first part alerts students to the serious problems that growth

  2. Growth: How Much is Too Much? Teacher's Guide. Social Studies Module (9th-10th Grade Social Studies).

    ERIC Educational Resources Information Center

    Georgia Univ., Athens. Coll. of Education.

    This is the teacher's guide for a learning module designed to integrate environmental education into ninth- and tenth-grade social studies classes. This module and a parallel module designed for chemistry classes were pilot tested in Gwinnett County, Georgia in 1975-76. The module is divided into four parts. The first part alerts students to the

  3. Mountain Dew[R] or Mountain Don't?: A Pilot Investigation of Caffeine Use Parameters and Relations to Depression and Anxiety Symptoms in 5th- and 10th-Grade Students

    ERIC Educational Resources Information Center

    Luebbe, Aaron M.; Bell, Debora J.

    2009-01-01

    Background: Caffeine, the only licit psychoactive drug available to minors, may have a harmful impact on students' health and adjustment, yet little is known about its use or effects on students, especially from a developmental perspective. Caffeine use in 5th- and 10th-grade students was examined in a cross-sectional design, and relations and…

  4. Mountain Dew[R] or Mountain Don't?: A Pilot Investigation of Caffeine Use Parameters and Relations to Depression and Anxiety Symptoms in 5th- and 10th-Grade Students

    ERIC Educational Resources Information Center

    Luebbe, Aaron M.; Bell, Debora J.

    2009-01-01

    Background: Caffeine, the only licit psychoactive drug available to minors, may have a harmful impact on students' health and adjustment, yet little is known about its use or effects on students, especially from a developmental perspective. Caffeine use in 5th- and 10th-grade students was examined in a cross-sectional design, and relations and

  5. The impact of high-stakes, state-mandated student performance assessment on 10th grade English, mathematics, and science teachers' instructional practices

    NASA Astrophysics Data System (ADS)

    Vogler, Kenneth E.

    The purpose of this study was to determine if the public release of student results on high-stakes, state-mandated performance assessments influence instructional practices, and if so in what manner. The research focused on changes in teachers' instructional practices and factors that may have influenced such changes since the public release of high-stakes, state-mandated student performance assessment scores. The data for this study were obtained from a 54-question survey instrument given to a stratified random sample of teachers teaching at least one section of 10th grade English, mathematics, or science in an academic public high school within Massachusetts. Two hundred and fifty-seven (257) teachers, or 62% of the total sample, completed the survey instrument. An analysis of the data found that teachers are making changes in their instructional practices. The data show notable increases in the use of open-response questions, creative/critical thinking questions, problem-solving activities, use of rubrics or scoring guides, writing assignments, and inquiry/investigation. Teachers also have decreased the use of multiple-choice and true-false questions, textbook-based assignments, and lecturing. Also, the data show that teachers felt that changes made in their instructional practices were most influenced by an "interest in helping my students attain MCAS assessment scores that will allow them to graduate high school" and by an "interest in helping my school improve student (MCAS) assessment scores," Finally, mathematics teachers and teachers with 13--19 years of experience report making significantly more changes than did others. It may be interpreted from the data that the use of state-mandated student performance assessments and the high-stakes attached to this type of testing program contributed to changes in teachers' instructional practices. The changes in teachers' instructional practices have included increases in the use of instructional practices deemed by educational researchers as the "best practices."

  6. Effect of cooperative learning strategies on student verbal interactions and achievement during conceptual change instruction in 10th grade general science

    NASA Astrophysics Data System (ADS)

    Lonning, Robert A.

    This study evaluated the effects of cooperative learning on students' verbal interaction patterns and achievement in a conceptual change instructional model in secondary science. Current conceptual change instructional models recognize the importance of student-student verbal interactions, but lack specific strategies to encourage these interactions. Cooperative learning may provide the necessary strategies. Two sections of low-ability 10th-grade students were designated the experimental and control groups. Students in both sections received identical content instruction on the particle model of matter using conceptual change teaching strategies. Students worked in teacher-assigned small groups on in-class assignments. The experimental section used cooperative learning strategies involving instruction in collaborative skills and group evaluation of assignments. The control section received no collaborative skills training and students were evaluated individually on group work. Gains on achievement were assessed using pre- and posttreatment administrations of an investigator-designed short-answer essay test. The assessment strategies used in this study represent an attempt to measure conceptual change. Achievement was related to students' ability to correctly use appropriate scientific explanations of events and phenomena and to discard use of naive conceptions. Verbal interaction patterns of students working in groups were recorded on videotape and analyzed using an investigator-designed verbal interaction scheme. The targeted verbalizations used in the interaction scheme were derived from the social learning theories of Piaget and Vygotsky. It was found that students using cooperative learning strategies showed greater achievement gains as defined above and made greater use of specific verbal patterns believed to be related to increased learning. The results of the study demonstrated that cooperative learning strategies enhance conceptual change instruction. More research is needed to identify the specific variables mediating the effects of cooperative learning strategies on conceptual change learning. The methods employed in this study may provide some of the tools for this research.

  7. Affective decision-making deficits, linked to a dysfunctional ventromedial prefrontal cortex, revealed in 10th-grade Chinese adolescent smokers

    PubMed Central

    Xiao, Lin; Bechara, Antoine; Cen, Steven; Grenard, Jerry L.; Stacy, Alan W.; Gallaher, Peggy; Wei, Yonglan; Jia, Yong; Johnson, C. Anderson

    2008-01-01

    This study addressed the question of whether poor decision making would be associated with adolescent past 7-day smoking. We conducted a cross-sectional study of 208 10th-grade adolescents in Chengdu City, China. We used the Iowa Gambling Task (IGT) to assess decision-making, and the Self-ordered Pointing Task (SOPT) to assess working memory capacity. Paper and pencil questionnaires assessed the school academic performance (SAP) and smoking variables. The results showed that a significantly higher proportion of past 7-day smokers (91.7%) were susceptible to future smoking and cigarette offers from best friends compared to other levels of smokers (never, ever and past 30-day smokers). Consistent with these behavioral data, the neuropsychological assessments revealed that relative to never smokers, past 7-day adolescent smokers (but not ever smokers or past 30-day smokers) demonstrated significantly lower scores on the IGT. Moreover, a higher proportion of past 7-day smokers (91.7%) performed poorly (no more than an overall net score of 10) on the IGT than nonsmokers and irregular (ever or past 30-day) smokers (about 65.3%). There were no differences on working memory performance for smokers (at any level) compared to never smokers after adjusting for school-type. In addition, logistic regression showed that the IGT significantly predicted past 7-day smoking after controlling for the working memory, school academic performance and demographic variables. These results suggest that poor affective decision making might predispose some adolescents to smoking in the future or in the social situations where their peers are smoking. Intervention targeting affective decision making might hold promise for reducing adolescents’ risks for substance use. PMID:18584472

  8. Trends in Substance Use among 6th-to 10th-Grade Students from 1998 to 2010: Findings from a National Probability Study

    ERIC Educational Resources Information Center

    Brooks-Russell, Ashley; Farhat, Tilda; Haynie, Denise; Simons-Morton, Bruce

    2014-01-01

    Of the handful of national studies tracking trends in adolescent substance use in the United States, only the Health Behavior in School-Aged Children (HBSC) study collects data from 6th through 10th graders. The purpose of this study was to examine trends from 1998 to 2010 (four time points) in the prevalence of tobacco, alcohol, and marijuana use

  9. Trends in Substance Use among 6th-to 10th-Grade Students from 1998 to 2010: Findings from a National Probability Study

    ERIC Educational Resources Information Center

    Brooks-Russell, Ashley; Farhat, Tilda; Haynie, Denise; Simons-Morton, Bruce

    2014-01-01

    Of the handful of national studies tracking trends in adolescent substance use in the United States, only the Health Behavior in School-Aged Children (HBSC) study collects data from 6th through 10th graders. The purpose of this study was to examine trends from 1998 to 2010 (four time points) in the prevalence of tobacco, alcohol, and marijuana use…

  10. REPORT FOR COMMERCIAL GRADE NICKEL CHARACTERIZATION AND BENCHMARKING

    SciTech Connect

    2012-12-20

    Oak Ridge Associated Universities (ORAU), under the Oak Ridge Institute for Science and Education (ORISE) contract, has completed the collection, sample analysis, and review of analytical results to benchmark the concentrations of gross alpha-emitting radionuclides, gross beta-emitting radionuclides, and technetium-99 in commercial grade nickel. This report presents methods, change management, observations, and statistical analysis of materials procured from sellers representing nine countries on four continents. The data suggest there is a low probability of detecting alpha- and beta-emitting radionuclides in commercial nickel. Technetium-99 was not detected in any samples, thus suggesting it is not present in commercial nickel.

  11. The Impact of Internet Virtual Physics Laboratory Instruction on the Achievement in Physics, Science Process Skills and Computer Attitudes of 10th-Grade Students

    ERIC Educational Resources Information Center

    Yang, Kun-Yuan; Heh, Jia-Sheng

    2007-01-01

    The purpose of this study was to investigate and compare the impact of Internet Virtual Physics Laboratory (IVPL) instruction with traditional laboratory instruction in physics academic achievement, performance of science process skills, and computer attitudes of tenth grade students. One-hundred and fifty students from four classes at one private

  12. The Basic Program of Vocational Agriculture in Louisiana. Ag I and Ag II (9th and 10th Grades). Volume III. Bulletin 1690-III.

    ERIC Educational Resources Information Center

    Louisiana State Dept. of Education, Baton Rouge. Div. of Vocational Education.

    This curriculum guide, the third volume of the series, outlines the basic program of vocational agriculture for Louisiana students in the ninth and tenth grades. Covered in the five units on plant science are growth processes of plants, cultural practices for plants, insects affecting plants, seed and plant selection, and diseases that affect…

  13. World History Fair and Exposition. 10th Grade Lesson. Schools of California Online Resources for Education (SCORE): Connecting California's Classrooms to the World.

    ERIC Educational Resources Information Center

    Davis, Cheryl

    This tenth grade lesson plan asks student teams to research and analyze the history, the challenges, and the future of contemporary nations and share their discoveries in a poster presentation at a technology pavilion. The lesson plan states that each team, as a participating nation, is required to include three themes in their countries' pavilion

  14. Trends in Bullying, Physical Fighting, and Weapon Carrying Among 6th- Through 10th-Grade Students From 1998 to 2010: Findings From a National Study

    PubMed Central

    Brooks-Russell, Ashley; Wang, Jing; Iannotti, Ronald J.

    2014-01-01

    Objectives. We examined trends from 1998 to 2010 in bullying, bullying victimization, physical fighting, and weapon carrying and variations by gender, grade level, and race/ethnicity among US adolescents. Methods. The Health Behavior in School-Aged Children surveys of nationally representative samples of students in grades 6 through 10 were completed in 1998 (n = 15 686), 2002 (n = 14 818), 2006 (n = 9229), and 2010 (n = 10 926). We assessed frequency of bullying behaviors, physical fighting, and weapon carrying as well as weapon type and subtypes of bullying. We conducted logistic regression analyses, accounting for the complex sampling design, to identify trends and variations by demographic factors. Results. Bullying perpetration, bullying victimization, and physical fighting declined from 1998 to 2010. Weapon carrying increased for White students only. Declines in bullying perpetration and victimization were greater for boys than for girls. Declines in bullying perpetration and physical fighting were greater for middle-school students than for high-school students. Conclusions. Declines in most violent behaviors are encouraging; however, lack of decline in weapon carrying merits further attention. PMID:24825213

  15. The Impact of Internet Virtual Physics Laboratory Instruction on the Achievement in Physics, Science Process Skills and Computer Attitudes of 10th-Grade Students

    NASA Astrophysics Data System (ADS)

    Yang, Kun-Yuan; Heh, Jia-Sheng

    2007-10-01

    The purpose of this study was to investigate and compare the impact of Internet Virtual Physics Laboratory (IVPL) instruction with traditional laboratory instruction in physics academic achievement, performance of science process skills, and computer attitudes of tenth grade students. One-hundred and fifty students from four classes at one private senior high school in Taoyuan Country, Taiwan, R.O.C. were sampled. All four classes contained 75 students who were equally divided into an experimental group and a control group. The pre-test results indicated that the students' entry-level physics academic achievement, science process skills, and computer attitudes were equal for both groups. On the post-test, the experimental group achieved significantly higher mean scores in physics academic achievement and science process skills. There was no significant difference in computer attitudes between the groups. We concluded that the IVPL had potential to help tenth graders improve their physics academic achievement and science process skills.

  16. Rates of Substance Use of American Indian Students in 8th, 10th, and 12th Grades Living on or Near Reservations: Update, 20092012

    PubMed Central

    Harness, Susan D.; Swaim, Randall C.; Beauvais, Fred

    2014-01-01

    Objectives Understanding the similarities and differences between substance use rates for American Indian (AI) young people and young people nationally can better inform prevention and treatment efforts. We compared substance use rates for a large sample of AI students living on or near reservations for the years 20092012 with national prevalence rates from Monitoring the Future (MTF). Methods We identified and sampled schools on or near AI reservations by region; 1,399 students in sampled schools were administered the American Drug and Alcohol Survey. We computed lifetime, annual, and last-month prevalence measures by grade and compared them with MTF results for the same time period. Results Prevalence rates for AI students were significantly higher than national rates for nearly all substances, especially for 8th graders. Rates of marijuana use were very high, with lifetime use higher than 50% for all grade groups. Other findings of interest included higher binge drinking rates and OxyContin use for AI students. Conclusions The results from this study demonstrate that adolescent substance use is still a major problem among reservation-based AI adolescent students, especially 8th graders, where prevalence rates were sometimes dramatically higher than MTF rates. Given the high rates of substance use-related problems on reservations, such as academic failure, delinquency, violent criminal behavior, suicidality, and alcohol-related mortality, the costs to members of this population and to society will continue to be much too high until a comprehensive understanding of the root causes of substance use are established. PMID:24587550

  17. Effects of a novel science curriculum versus traditional science curriculum on problem solving skills and attitudes for 10th grade students

    NASA Astrophysics Data System (ADS)

    Gauchat, Carrie

    This study utilized both quantitative and qualitative methods in investigating how a novel science curriculum, geared towards the 21 st century student, affected skills and attitudes towards science for tenth grade students. The quantitative portion of the study was a quasi-experimental design since random groups were not possible. This portion of the study used a pretest/posttest design to measure any improvement in science skills, and a Likert scale survey to measure any improvements in students' attitudes. Statistical tests revealed no significant differences between students who received the novel curriculum versus those students who received a traditional curriculum. Both groups showed significant improvements in all skill areas. Qualitatively, the researcher used informal teacher interviews and student surveys to identify the most relevant and effective curriculum components for the 21st century student. The findings suggest that the task of creating a meaningful and relevant curriculum based on the necessary skills of this century is not an easy task. There is much more work to be done in this area, but according to the qualitative findings integrated design and student technology are promising.

  18. Teachers' Perceptions of the Effectiveness of Benchmark Assessment Data to Predict Student Math Grades

    ERIC Educational Resources Information Center

    Lewis, Lawanna M.

    2010-01-01

    The purpose of this correlational quantitative study was to examine the extent to which teachers perceive the use of benchmark assessment data as effective; the extent to which the time spent teaching mathematics is associated with students' mathematics grades, and the extent to which the results of math benchmark assessment influence teachers'

  19. Analysis of Weapons-Grade MOX VVER-1000 Benchmarks with HELIOS and KENO

    SciTech Connect

    Dourougie, C.; Emmett, M.B.; Gehin, J.C.; Lillie, R.A.

    1999-07-01

    Calculations of computational benchmark problems for the disposition of weapons-grade mixed-oxide fuel in VVER-1000 reactors have been performed with the HELIOS fuel assembly analysis code. The benchmarks cover pin cell, single fuel assembly, and multiassembly structures with several different fuel types, moderator densities, and boron content for operational and off-normal conditions. Fuel depletion is performed to a burnup of 60 MWd/kg. The detailed results have been tabulated for multiplication factors, macroscopic and microscopic cross sections, reaction rates, fuel isotopics, and fission distributions. Additional results are given for a preliminary analysis of a multiassembly variant using the KENO Monte Carlo code with cross sections from HELIOS.

  20. Sample Grade Level Benchmarks, Grades K-4, Based on the 1998 Arkansas State Mathematics Framework.

    ERIC Educational Resources Information Center

    Arkansas State Dept. of Education, Little Rock.

    This document presents the application and use of mathematics learning proposed by the Arkansas curriculum frameworks for grades K-4. The standards are presented in chart form and organized into five strands: (1) number sense, properties, and operations; (2) geometry and spatial sense; (3) measurement; (4) data analysis, statistics, and…

  1. Sample Grade Level Benchmarks, Grades 5-8, Based on the 1998 Arkansas State Mathematics Framework.

    ERIC Educational Resources Information Center

    Arkansas State Dept. of Education, Little Rock.

    This document presents the application and use of mathematics learning proposed by the Arkansas curriculum frameworks for grades 5-8. The standards are presented in chart form and organized into five strands: (1) number sense, properties, and operations; (2) geometry and spatial sense; (3) measurement; (4) data analysis, statistics, and…

  2. Columbines 10th Anniversary Finds Lessons Learned

    ERIC Educational Resources Information Center

    Trump, Kenneth S.

    2009-01-01

    When school administrators hear that the 10th anniversary of the Columbine High School attack is approaching, most shake their heads in disbelief. They are amazed that 10 years have passed since this watershed event, which changed the landscape of K-12 school safety. In this article, the author reflects on the lessons learned from the Columbine

  3. The Relationship between Mid-Year Benchmark and End-of-Grade Assessments: 2010-11. D&A Report No. 11.24

    ERIC Educational Resources Information Center

    McMillen, Brad

    2012-01-01

    This analysis was conducted to examine the relationship between Wake Count Public Schools' students' performance on mid-year benchmark assessments and End-of-Grade (EOG) tests in both reading and math in grades 3-5 in school year 2010-11. Strong positive correlations were found between the mid-year benchmark assessment and EOG test scores for each…

  4. Analyses of Weapons-Grade MOX VVER-1000 Neutronics Benchmarks: Pin-Cell Calculations with SCALE/SAS2H

    SciTech Connect

    Ellis, R.J.

    2001-01-11

    A series of unit pin-cell benchmark problems have been analyzed related to irradiation of mixed oxide fuel in VVER-1000s (water-water energetic reactors). One-dimensional, discrete-ordinates eigenvalue calculations of these benchmarks were performed at ORNL using the SAS2H control sequence module of the SCALE-4.3 computational code system, as part of the Fissile Materials Disposition Program (FMDP) of the US DOE. Calculations were also performed using the SCALE module CSAS to confirm the results. The 238 neutron energy group SCALE nuclear data library 238GROUPNDF5 (based on ENDF/B-V) was used for all calculations. The VVER-1000 pin-cell benchmark cases modeled with SAS2H included zero-burnup calculations for eight fuel material variants (from LEU UO{sub 2} to weapons-grade MOX) at five different reactor states, and three fuel depletion cases up to high burnup. Results of the SAS2H analyses of the VVER-1000 neutronics benchmarks are presented in this report. Good general agreement was obtained between the SAS2H results, the ORNL results using HELIOS-1.4 with ENDF/B-VI nuclear data, and the results from several Russian benchmark studies using the codes TVS-M, MCU-RFFI/A, and WIMS-ABBN. This SAS2H benchmark study is useful for the verification of HELIOS calculations, the HELIOS code being the principal computational tool at ORNL for physics studies of assembly design for weapons-grade plutonium disposition in Russian reactors.

  5. PREFACE: 10th Joint Conference on Chemistry

    NASA Astrophysics Data System (ADS)

    2016-02-01

    The 10th Joint Conference on Chemistry is an international conference organized by 4 chemistry departments of 4 universities in central Java, Indonesia. The universities are Sebelas Maret University, Diponegoro University, Semarang State University and Soedirman University. The venue was at Solo, Indonesia, at September 8-9, 2015. The total conference participants are 133 including the invited speakers. The conference emphasized the multidisciplinary chemical issue and impact of today's sustainable chemistry which covering the following topics: • Material innovation for sustainable goals • Development of renewable and sustainable energy based on chemistry • New drug design, experimental and theoretical methods • Green synthesis and characterization of material (from molecule to functionalized materials) • Catalysis as core technology in industry • Natural product isolation and optimization

  6. 10th Arnual Great Moonbuggy Race

    NASA Technical Reports Server (NTRS)

    2003-01-01

    Students from across the United States and as far away as Puerto Rico came to Huntsville, Alabama for the 10th annual Great Moonbuggy Race at the U.S. Space Rocket Center. Sixty-eight teams, representing high schools and colleges from all over the United States, and Puerto Rico, raced human powered vehicles over a lunar-like terrain. Vehicles powered by two team members, one male and one female, raced one at a time over a half-mile obstacle course of simulated moonscape terrain. The competition is inspired by development, some 30 years ago, of the Lunar Roving Vehicle (LRV), a program managed by the Marshall Space Flight Center. The LRV team had to design a compact, lightweight, all-terrain vehicle that could be transported to the Moon in the small Apollo spacecraft. The Great Moonbuggy Race challenges students to design and build a human powered vehicle so they will learn how to deal with real-world engineering problems similar to those faced by the actual NASA LRV team. In this photograph, racers from C-1 High School in Lafayette County, Missouri, get ready to tackle the course. The team pedaled its way to victory over 29 other teams to take first place honors. It was the second year in a row a team from the school has placed first in the high school division. (NASA/MSFC)

  7. 10th Arnual Great Moonbuggy Race

    NASA Technical Reports Server (NTRS)

    2003-01-01

    Students from across the United States and as far away as Puerto Rico came to Huntsville, Alabama for the 10th annual Great Moonbuggy Race at the U.S. Space Rocket Center. Sixty-eight teams, representing high schools and colleges from all over the United States, and Puerto Rico, raced human powered vehicles over a lunar-like terrain. Vehicles powered by two team members, one male and one female, raced one at a time over a half-mile obstacle course of simulated moonscape terrain. The competition is inspired by development, some 30 years ago, of the Lunar Roving Vehicle (LRV), a program managed by the Marshall Space Flight Center. The LRV team had to design a compact, lightweight, all-terrain vehicle that could be transported to the Moon in the small Apollo spacecraft. The Great Moonbuggy Race challenges students to design and build a human powered vehicle so they will learn how to deal with real-world engineering problems similar to those faced by the actual NASA LRV team. In this photograph, Team No. 1 from North Dakota State University in Fargo conquers one of several obstacles on their way to victory. The team captured first place honors in the college level competition.

  8. PREFACE: 10th International LISA Symposium

    NASA Astrophysics Data System (ADS)

    Ciani, Giacomo; Conklin, John W.; Mueller, Guido

    2015-05-01

    The LISA Symposia have become a mainstay of the gravitational wave community. Held every two years, they are the prime opportunity for our community to discuss the exciting science, technology, mission designs, and progress of the Laser Interferometer Space Antenna. The 8th LISA symposium, held at Stanford University in the summer of 2010 was the largest symposium so far and was dominated by progress and hopes that the LISA mission will soon excel following the expected launch of the LISA pathfinder (LPF), no later than 2012, and the expected prioritization by the Decadal survey which was released 6 weeks later. The following years were challenging. Although the Decadal survey ranked LISA very high, NASA's budget issues, mostly due to the cost increase of the James Webb Space Telescope, and continued delays in LPF put too much stress on the LISA project and it officially ended in 2011. The LISA International Science Team (LIST), the core group of LISA scientists and technologists, was dissolved and the community in the U.S. was struggling to maintain cohesion. In the wake of these events, ESA started a new selection process for their next three large missions, L1, L2, and L3, and the European LISA team developed the New Gravitational wave Observatory (NGO), an evolved LISA concept, as an ESA only L1 candidate. A few weeks before the 9th LISA Symposium, held in Paris in May 2012, ESA announced its decision to select JUICE, a planetary mission to Jupiter and its moons, as its next large science mission (L1). Despite having the highest ranked science case, NGO was not selected due to further delays in LPF and the general feeling outside the GW community that the technology is perhaps too challenging to be pulled off in time for the L1 launch in 2022. Many U.S. members of the LISA community cancelled their travel plans and the mood at that symposium ranged from resignation to defiance. Hope for a somewhat timely launch of a LISA-like mission rested upon L2, the next large mission in Europe, and a potential comprehensive technology development program followed by a number one selection in the 2020 Decadal Survey in the U.S. The selection of L2 was combined with the selection of L3 and the newly formed eLISA consortium submitted an updated NGO concept under the name eLISA, or Evolved LISA, to the competition. It was widely believed that the launch date of 2028 for L2, would be seen by the selection committee as providing sufficient time to retire any remaining technological risks for LISA. However, the committee selected the 'Hot and Energetic Universe', an X-ray mission, as the science theme for L2 and the 'Gravitational Universe', the eLISA science theme, for L3. Although very disappointed, it was not a surprising decision. LPF did experience further delays just prior to and during the selection process, which may have influenced the decision. The strong technology program in the U.S. never materialized because WFIRST, the highest priority large mission in the 2010 Decadal following JWST, not only moved ahead but was also up-scoped significantly. The L3 selection, the WFIRST schedule, and the missing comprehensive technology development in the U.S. will make a launch of a GW mission in the 2020s very difficult. Although many in the LISA community, including ourselves, did not want to accept this harsh reality, this was the situation just prior to the 10th LISA symposium. However, despite all of this, the LISA team is now hopeful! In May of 2014 the LISA community gathered at the University of Florida in Gainesville to discuss progress in both the science and technology of LISA. The most notable plenary and contributed sessions included updates on the progress of LISA Pathfinder, which remains on track for launch in the second half of 2015(!), the science of LISA which ranges from super-massive black hole mergers and cosmology to the study of compact binaries within our own galaxy, and updates from other programs that share some of LISA's science or technology. Plenary talks from the pulsar timing and ground-based laser interferometer groups told of the reasonable expectation of gravitational wave detection within the next 4 to 8 years. We also heard about the GRACE Follow-on mission, which will demonstrate a precision laser ranging system in space in 2017, using technology that is somewhat similar to that of LISA. Presentations on the Large Synoptic Survey Telescope, Athena, the Cherenkov Telescope Array, and WFIRST provided data on the landscape in which LISA will live in the 2030s. Beyond the 10th symposium there is much to look forward to. There is high-expectation that LISA Pathfinder will launch in 2015, prior to the 11th symposium in Zürich, which, for the first time, will be dedicated to the results of LPF and not its preparation. Ground-based gravitational wave observatories, especially Advanced LIGO, are rapidly approaching their required sensitivities and could make the first direct detection before the 12th LISA symposium. Advanced VIRGO and KAGRA, and the pulsar timing community are also hopeful that they will reach the required sensitivity within this decade or shortly thereafter. These events will dramatically improve the perception of gravitational wave science by the broader astronomy and astrophysics communities. The U.S. LISA team is also embolden by the announcement that NASA is now planning to join ESA in the gravitational wave L3 mission as a junior partner and will begin funding a technology development program to support this partnership. A space-based gravitational wave mission is inevitable. At the time of the 10th LISA Symposium, it was not clear if the gradient of LISA's trajectory was perceived as positive or negative. But in hindsight, 2014 will hopefully be seen as a time when LISA regained some of the ground recently lost and began accelerating towards launch.

  9. EDITORIAL: 10th anniversary of attosecond pulses 10th anniversary of attosecond pulses

    NASA Astrophysics Data System (ADS)

    Kienberger, Reinhard; Chang, Zenghu; Nam, Chang Hee

    2012-04-01

    Attosecond science is an emerging field prompted by the development of attosecond light sources during the last decade. Since shorter light sources can diagnose faster phenomena, the generation of very short light pulses is critical in ultrafast measurements. In 2001 two groups at Commissariat l'Energie Atomique (CEA), Saclay, and Technische Universitt Wien (TUW), Vienna, succeeded in proving the production of attosecond pulse trains and isolated attosecond pulses using high harmonic generation processes, respectively. For the temporal characterization, cross correlation techniques such as RABITT and FROG CRAB that make use of the photoionization of atoms by harmonic and infrared pulses have been demonstrated, in addition to autocorrelation techniques. Femtosecond laser technology has progressed so well that carrier envelope phase (CEP)-stabilized few cycle lasers are now widely used in the production of attosecond pulses. For the generation of isolated attosecond pulses the high-harmonic generation processes have been controlled: the amplitude gating method has been powerful with CEP-stabilized sub-two-cycle laser pulses and the temporal gating methods, including polarization gating and double optical gating, have been effective in accommodating multi-cycle laser pulses. In achieving near transform-limited attosecond pulses the compensation of the inherent attosecond chirp by material dispersion has been useful. The attosecond pulse duration has reached sub-100 attoseconds both for attosecond pulse trains and isolated attosecond pulses, and it is expected that attosecond pulses shorter than one atomic unit (24 as) will be obtained soon. With attosecond light sources, attosecond science has become an active research field. Ultrafast phenomena in atoms and molecules have been looked at in detail, especially in the time domain. The photoionization processes of atoms and molecules occurring in sub-optical cycles have been analyzed, and ultrafast characteristic times in atoms and molecules, such as Auger decay time and autoionization lifetime, have been measured directly as compared to indirect spectroscopic measurements normally done using synchrotron light sources. The reconstruction of molecular orbital wave functions has been demonstrated by developing the molecular tomography method. Ultrafast phenomena in condensed matter and in nanostructures have been tackled also. The successful development of attosecond light sources has thus opened up a variety of new research activities in ultrafast optical science; it will be continued and accelerated further in coming years with intensive research investments by more groups joining the field of attosecond science. In this special issue celebrating the 10th year of attosecond pulse generation 6 review articles and 16 regular articles are included. Although it does not cover all active research areas, we sincerely hope it gives a glimpse of active research activities in attosecond science throughout the world.

  10. Factors Related to Alcohol Use among 6th through 10th Graders: The Sarasota County Demonstration Project

    ERIC Educational Resources Information Center

    Eaton, Danice K.; Forthofer, Melinda S.; Zapata, Lauren B.; Brown, Kelli R. McCormack; Bryant, Carol A.; Reynolds, Sherri T.; McDermott, Robert J.

    2004-01-01

    Alcohol consumption by youth can produce negative health outcomes. This study identified correlates of lifetime alcohol use, recent alcohol use, and binge drinking among youth in sixth through 10th grade (n = 2,004) in Sarasota County, Fla. Results from a closed-ended, quantitative survey acknowledged a range of personal, social and environmental

  11. NADEOSA 10th Anniversary Conference Formal Evaluation Report

    ERIC Educational Resources Information Center

    South African Journal of Higher Education, 2007

    2007-01-01

    NADEOSA held its 10th anniversary conference during August 2006 on the theme "Celebrating 10 years of NADEOSA. Exploring the role of ICTs in addressing educational needs: Identifying the myths and miracles." The conference evaluation strategy consisted of a number of consciously planned processes, a dialogical and reflective account which had been

  12. Beyond Discipline: From Compliance to Community. 10th Anniversary Edition

    ERIC Educational Resources Information Center

    Kohn, Alfie

    2006-01-01

    In this 10th anniversary edition of an Association for Supervision and Curriculum Development (ASCD) best seller, the author reflects on his revolutionary ideas in the context of today's emphasis on school accountability and high-stakes testing. The author relates how his innovative approach--where teachers learn to work with students, rather than…

  13. Beyond Discipline: From Compliance to Community. 10th Anniversary Edition

    ERIC Educational Resources Information Center

    Kohn, Alfie

    2006-01-01

    In this 10th anniversary edition of an Association for Supervision and Curriculum Development (ASCD) best seller, the author reflects on his revolutionary ideas in the context of today's emphasis on school accountability and high-stakes testing. The author relates how his innovative approach--where teachers learn to work with students, rather than

  14. Byzantine psychosomatic medicine (10th- 15th century).

    PubMed

    Eftychiadis, A C

    1999-01-01

    Original elements of the psychosomatic medicine are examined by the most important byzantine physicians and medico-philosophers during the 10th -15th centuries. These topics concern the psycosomatic unity of the human personality, the psychosomatic disturbances, diseases and interactions, organic diseases, which cause psychical disorders, psychical pathological reactions, which result in somatic diseases, the psychology of the depth of the soul, the psychosomatic pathogenetic reasons of psychiatric and neurological diseases and suicide, the influence of witchcraft on psychosomatic affections, maniac and demoniac patients. The psychosomatic treatment has a holistic preventive and curative character and encloses sanitary and dietary measures, physiotherapy, curative bathing, strong purgation, pharmaceutical preparations proportional to the disease, religious disposition, psychoanalysis and psychotherapy with dialogue and the contribution of the divine factor. The late byzantine medical science contributed mainly to the progress of the psychosomatic medicine and therapeutics. The saint woman physician Hermione (1st -2nd cent.) is considered as the protectress of psychosomatic medicine. PMID:11624574

  15. Prevalence and Patterns of Polysubstance Use in a Nationally Representative Sample of 10th Graders in the United States

    PubMed Central

    Conway, Kevin P.; Vullo, Genevieve C.; Nichter, Brandon; Wang, Jing; Compton, Wilson M.; Iannotti, Ronald J.; Simons-Morton, Bruce

    2013-01-01

    Purpose The current study examines the prevalence and demographic correlates of self-reported substance use and identifies subgroups of polysubstance users among a cohort of US 10th grade students. Methods A nationally representative school-based cohort of US 10th grade students completed the NEXT Generation Health Study baseline survey in the spring of 2010 (N = 2,524). Results Past-year use of marijuana was most common among illicit drugs (26%), followed by misuse of medication (9%) and use of other illicit drugs (8%). During the past month, alcohol use was reported by more than one-third (35%), binge drinking by 27%, and cigarette smoking by 19%. Results further show that substance use varied somewhat by demographic characteristics. Results from the latent class analysis of polysubstance use indicated a 4-class solution as the best-fitting model; Class 1 (59%) included the ‘non-user’ group; Class 2 (23%) comprised the ‘predominant alcohol user’ group; Class 3 (11%) formed the ‘predominant marijuana user’ group; Class 4 (8%) was characterized as the ‘predominant polysubstance user’ group. Somatic and depressive symptoms varied significantly by class membership, with predominant polysubstance users reporting elevated levels of somatic and depressive symptoms. Conclusions The findings from this national study of 10th-grade students indicate high rates of substance and polysubstance use. The high level of depressive and somatic symptoms among polysubstance users indicates the need for mental health screening and referral. PMID:23465320

  16. The 10th Annual Bioassays and Bioanalytical Method Development Conference.

    PubMed

    Ma, Mark; Tudan, Christopher; Koltchev, Dolly

    2015-01-01

    The 10th Annual Bioassays and Bioanalytical Method Development Conference was hosted in Boston, MA, USA on 20-22 October 2014. This meeting brought together scientists from the biopharmaceutical and life sciences industries, the regulatory agency and academia to share and discuss current trends in cell-based assays and bioanalysis, challenges and ideas for the future of the bioassays and bioanalytical method development. The experiences associated with new and innovative technologies were evaluated as well as their impact on the current bioassays methodologies and bioanalysis workflow, including quality, feasibility, outsourcing strategies and challenges, productivity and compliance. Several presentations were also provided by members of the US FDA, sharing both scientific and regulatory paradigms including a most recent update on the position of the FDA with specific aspects of the draft Bioanalytical Method Validation guidance following its review of the industry's responses. The meeting was jointly coincided with the 15th Annual Immunogenicity for Biotherapeutics meeting, allowing for attendees to also familiarize themselves with new and emerging approaches to overcome the effect of immunogenicity, in addition to investigative strategies. PMID:25826138

  17. The Comparative Toxicogenomics Database's 10th year anniversary: update 2015

    PubMed Central

    Davis, Allan Peter; Grondin, Cynthia J.; Lennon-Hopkins, Kelley; Saraceni-Richards, Cynthia; Sciaky, Daniela; King, Benjamin L.; Wiegers, Thomas C.; Mattingly, Carolyn J.

    2015-01-01

    Ten years ago, the Comparative Toxicogenomics Database (CTD; http://ctdbase.org/) was developed out of a need to formalize, harmonize and centralize the information on numerous genes and proteins responding to environmental toxic agents across diverse species. CTD's initial approach was to facilitate comparisons of nucleotide and protein sequences of toxicologically significant genes by curating these sequences and electronically annotating them with chemical terms from their associated references. Since then, however, CTD has vastly expanded its scope to robustly represent a triad of chemicalgene, chemicaldisease and genedisease interactions that are manually curated from the scientific literature by professional biocurators using controlled vocabularies, ontologies and structured notation. Today, CTD includes 24 million toxicogenomic connections relating chemicals/drugs, genes/proteins, diseases, taxa, phenotypes, Gene Ontology annotations, pathways and interaction modules. In this 10th year anniversary update, we outline the evolution of CTD, including our increased data content, newPathway View visualization tool, enhanced curation practices, pilot chemicalphenotype results and impending exposure data set. The prototype database originally described in our first report has transformed into a sophisticated resource used actively today to help scientists develop and test hypotheses about the etiologies of environmentally influenced diseases. PMID:25326323

  18. Supracostal Approach for PCNL: Is 10th and 11th Intercostal Space Safe According to Clavien Classification System?

    PubMed Central

    Kara, Cengiz; De?irmenci, Tansu; Kozacioglu, Zafer; Gunlusoy, Bulent; Koras, Omer; Minareci, Suleyman

    2014-01-01

    The purpose of this study was to evaluate the success and morbidity of percutaneous nephrolithotomy (PCNL) performed through the 11th and 10th intercostal space. Between March 2005 and February 2012, 612 patients underwent PCNL, 243 of whom had a supracostal access. The interspace between the 11th and 12th rib was used in 204 cases (group 1) and between the 10th and 11th interspaces in 39 cases (group 2). PCNL was performed using standard supracostal technique in all patients. The operative time, success rate, hospital stay, and complications according to the modified Clavien classification were compared between group 1 and group 2. The stone-free rate was 86.8% in group 1 and 84.6% in group 2 after one session of PCNL. Auxiliary procedures consisting of ureterorenoscopy (URS) and shock wave lithotripsy (SWL) were required in 5 and 7 patients, respectively, in group 1; and in 1 patient each in group 2 . After the auxiliary procedures, stone-free rates increased to 92.6% in group 1 and 89.7% in group 2. A total of 74 (30.4%) complications were documented in the 2 groups according to modified Clavien classification. Grade-I complications were recorded in 20 (8.2%), grade-II in 38 (15.6%), grade-IIIa in 13 (5.3%), and grade-IIIb in 2 (0.8%) patients; grade-IVa was recorded in 1 (0.4%) patient. There were no grade-IVb or grade-V complications. Overall complication rate was 30.9% in group 1 and 28.2% in group 2. Supracostal PCNL in selected cases is effective and safe with acceptable complications. The modified Clavien system provides a standardized grading system for complications of PCNL. PMID:25437600

  19. PREFACE: ISEC 2005: The 10th International Superconductive Electronics Conference

    NASA Astrophysics Data System (ADS)

    Rogalla, Horst

    2006-05-01

    The 10th International Superconductive Electronics Conference took place in Noordwijkerhout in the Netherlands, 5-9 September 2005, not far from the birthplace of superconductivity in Leiden nearly 100 years ago. There have been many reasons to celebrate the 10th ISEC: not only was it the 20th anniversary, but also the achievements since the first conference in Tokyo in 1987 are tremendous. We have seen whole new groups of superconductive materials come into play, such as oxide superconductors with maximum Tc in excess of 100 K, carbon nanotubes, as well as the realization of new digital concepts from saturation logic to the ultra-fast RSFQ-logic. We have learned that superconductors not only show s-wave symmetries in the spatial arrangement of the order parameter, but also that d-wave dependence in oxide superconductors is now well accepted and can even be successfully applied to digital circuits. We are now used to operating SQUIDs in liquid nitrogen; fT sensitivity of SQUID magnetometers is not surprising anymore and can even be reached with oxide-superconductor based SQUIDs. Even frequency discriminating wide-band single photon detection with superconductive devices, and Josephson voltage standards with tens of thousands of junctions, nowadays belong to the daily life of advanced laboratories. ISEC has played a very important role in this development. The first conferences were held in 1987 and 1989 in Tokyo, and subsequently took place in Glasgow (UK), Boulder (USA), Nagoya (Japan), Berlin (Germany), Berkeley (USA), Osaka (Japan), Sydney (Australia), and in 2005 for the first time in the Netherlands. These conferences have provided platforms for the presentation of the research and development results of this community and for the vivid discussion of achievements and strategies for the further development of superconductive electronics. The 10th conference has played a very important role in this context. The results in laboratories show great potential and the first demonstrations underline this potential both in digital as well as analogue applications. But a broad application of this technology in industry and non-superconductive research is only slowly starting. An intensification of the contact between possible users from industry and research on the one hand, and from basic and applied research on the other, is needed, and ISEC is an excellent forum to stimulate this discussion. With the special session `Superconductive Electronics in Space' and the special focus on new devices and technologies like π-junctions and π-shift devices, we took a first step in this direction and we hope that it has helped to stimulate the further development of superconductive electronics. When we first declared our willingness to organize ISEC in the year 2005, Gerrit Gerritsma (from my group in Twente) and I intended to share the chair and the responsibility. After the regrettable death of Gerrit in 2002, there was some doubt if we should continue with the preparations for this conference. However, with the help of co-workers in Twente and colleagues from all over Europe and the rest of the world, organizing the conference became an enjoyable task. The programme committee consisted primarily of colleagues from the European Society `Fluxonics', of working groups of the European Thematic Network SCENET, and of the European Science Foundation Program `PiShift'. The comments and guidelines of the International Advisory Board were a great help and their advice set the style of the conference. Without the sponsorship of a number of organizations it would not have been possible to organize the conference: The European Science Foundation (ESF) supported this conference and its special topics on advanced superconductive devices within its `PiShift' programme. Furthermore, we are grateful to the European Society `Fluxonics', the European Thematic Network SCENET and the European Space Agency ESA/ESTEC for their support. Last but not least, the financial and organizational contribution by the University of Twente is gratefully acknowledged. A limited number of conference contributions have been selected, after peer reviewing, for publication in this special issue of Superconductor Science and Technology. A special word of thanks goes to Arnold Silver for writing his banquet speech, Paul Seidel for his conference summary, and Gordon Donaldson for his editorial work in realizing this issue.

  20. Interests of 5th through 10th Grade Students toward Human Biology

    ERIC Educational Resources Information Center

    Erten, Sinan

    2008-01-01

    This study investigated the middle and high school students' interests towards the subjects of human biology, specifically, "Human Health and Nutrition" and "Human Body and Organs." The study also investigated sources of their interests and factors that impact their interests, namely people that they interact and courses that they take about

  1. Using Diagrams versus Text for Spaced Restudy: Effects on Learning in 10th Grade Biology Classes

    ERIC Educational Resources Information Center

    Bergey, Bradley W.; Cromley, Jennifer G.; Kirchgessner, Mandy L.; Newcombe, Nora S.

    2015-01-01

    Background and Aim: Spaced restudy has been typically tested with written learning materials, but restudy with visual representations in actual classrooms is under-researched. We compared the effects of two spaced restudy interventions: A Diagram-Based Restudy (DBR) warm-up condition and a business-as-usual Text-Based Restudy (TBR) warm-up…

  2. School Composition and Context Factors that Moderate and Predict 10th-Grade Science Proficiency

    ERIC Educational Resources Information Center

    Hogrebe, Mark C.; Tate, William F., IV

    2010-01-01

    Background: Performance in high school science is a critical indicator of science literacy and regional competitiveness. Factors that influence science proficiency have been studied using national databases, but these do not answer all questions about variable relationships at the state level. School context factors and opportunities to learn

  3. Interests of 5th through 10th Grade Students Regarding Enviromental Protection Issues

    ERIC Educational Resources Information Center

    Erten, Sinan

    2015-01-01

    This study investigates the extent of interest among middle and high school students in environmental protection issues along with the sources of their interests and factors that impact their interests, namely people with whom they interact and courses that they take related to the environment, science and technology. In addition, it is confirmed…

  4. Using Diagrams versus Text for Spaced Restudy: Effects on Learning in 10th Grade Biology Classes

    ERIC Educational Resources Information Center

    Bergey, Bradley W.; Cromley, Jennifer G.; Kirchgessner, Mandy L.; Newcombe, Nora S.

    2015-01-01

    Background and Aim: Spaced restudy has been typically tested with written learning materials, but restudy with visual representations in actual classrooms is under-researched. We compared the effects of two spaced restudy interventions: A Diagram-Based Restudy (DBR) warm-up condition and a business-as-usual Text-Based Restudy (TBR) warm-up

  5. General Shop Competencies in Vocational Agriculture for 9th and 10th Grade Classes.

    ERIC Educational Resources Information Center

    Novotny, Ronald; And Others

    The document presents unit plans which offer list of experiences and competencies to be learned for general shop occupations in vocational agriculture. The units include: (1) arc welding, (2) oxy-acetylene welding, (3) flat concrete, (4) concrete block, (5) lumber patterns and wood building materials, (6) metal fasteners, (7) wood adhesives, (8)

  6. Progression in Complexity: Contextualizing Sustainable Marine Resources Management in a 10th Grade Classroom

    ERIC Educational Resources Information Center

    Bravo-Torija, Beatriz; Jimenez-Aleixandre, Maria-Pilar

    2012-01-01

    Sustainable management of marine resources raises great challenges. Working with this socio-scientific issue in the classroom requires students to apply complex models about energy flow and trophic pyramids in order to understand that food chains represent transfer of energy, to construct meanings for sustainable resources management through

  7. Who's Skipping School: Characteristics of Truants in 8th and 10th Grade

    ERIC Educational Resources Information Center

    Henry, Kimberly L.

    2007-01-01

    Background: Truancy is a serious concern that affects most school districts in the United States; however, we do not have accurate estimates of the prevalence of truancy due to inconsistent tracking and reporting practices of schools. As a result, our best current estimates of the national state of truancy may be from self-reported data. In this…

  8. Does STES-Oriented Science Education Promote 10th-Grade Students' Decision-Making Capability?

    ERIC Educational Resources Information Center

    Levy Nahum, Tami; Ben-Chaim, David; Azaiza, Ibtesam; Herskovitz, Orit; Zoller, Uri

    2010-01-01

    Today's society is continuously coping with sustainability-related complex issues in the Science-Technology-Environment-Society (STES) interfaces. In those contexts, the need and relevance of the development of students' higher-order cognitive skills (HOCS) such as question-asking, critical-thinking, problem-solving and decision-making

  9. Vocational Agriculture II Curriculum Guide, 10th Grade. Kansas Vocational Agriculture Education: Basic Core Curriculum II.

    ERIC Educational Resources Information Center

    Pittsburg State Univ., KS. Kansas Vocational Curriculum and Research Center.

    This basic core curriculum for vocational agriculture education contains 35 units of instruction in five content areas: agricultural chemicals (1 unit), leadership (2 units), farm management (5 units), plant and soil science (10 units), animal science (8 units), and farm mechanics (9 units). Each unit follows a typical format that includes…

  10. Vocational Agriculture II Curriculum Guide, 10th Grade. Kansas Vocational Agriculture Education: Basic Core Curriculum II.

    ERIC Educational Resources Information Center

    Pittsburg State Univ., KS. Kansas Vocational Curriculum and Research Center.

    This basic core curriculum for vocational agriculture education contains 35 units of instruction in five content areas: agricultural chemicals (1 unit), leadership (2 units), farm management (5 units), plant and soil science (10 units), animal science (8 units), and farm mechanics (9 units). Each unit follows a typical format that includes

  11. An analysis of women's ways of knowing in a 10th grade integrated science classroom

    NASA Astrophysics Data System (ADS)

    Kochheiser, Karen Lynn

    All students can learn science, but how they learn science may differ. This study is about learning science and its relationship to gender. Women need to develop and establish connections with the objects that they are learning and be able to establish a voice in a science classroom. Unfortunately, traditional science classrooms still view science as a male domain and tend to discourage women from pursuing higher levels of science or science related careers. The ways that women learn science are a very complex set of interactions. In order to describe these interactions, this study explored how women's ways of knowing are represented in a high school science classroom. Nine women from an enriched integrated biology and earth science class contributed to this study. The women contributed to this study by participating in individual and group interviews, questionnaires, journals, observations and participant review of the interviews. The ways that these women learn science were described in terms of Belenky, Clinchy, Goldberger, and Tarule's Women's Ways of Knowing: The Development of Self, Voice, and Mind (1997). The women's ways of learning in this classroom tended to be situational with the women fitting different categories of knowing depending on the situation. Most of the women demonstrated periods of time where they wanted to be heard or tried to establish a voice in the classroom. The study helps to provide a theory for how women make choices in their learning of science and the struggle to be successful in a male dominated discipline. The women participating in this study gained an awareness of how they learn science and how that can be used to make them even more successful in the classroom. The awareness of how women learn science will also be of great benefit to other teachers and educators as the work for science reform continues to make science a 'science for all'.

  12. Interests of 5th through 10th Grade Students toward Human Biology

    ERIC Educational Resources Information Center

    Erten, Sinan

    2008-01-01

    This study investigated the middle and high school students' interests towards the subjects of human biology, specifically, "Human Health and Nutrition" and "Human Body and Organs." The study also investigated sources of their interests and factors that impact their interests, namely people that they interact and courses that they take about…

  13. Interests of 5th through 10th Grade Students Regarding Enviromental Protection Issues

    ERIC Educational Resources Information Center

    Erten, Sinan

    2015-01-01

    This study investigates the extent of interest among middle and high school students in environmental protection issues along with the sources of their interests and factors that impact their interests, namely people with whom they interact and courses that they take related to the environment, science and technology. In addition, it is confirmed

  14. Progression in Complexity: Contextualizing Sustainable Marine Resources Management in a 10th Grade Classroom

    ERIC Educational Resources Information Center

    Bravo-Torija, Beatriz; Jimenez-Aleixandre, Maria-Pilar

    2012-01-01

    Sustainable management of marine resources raises great challenges. Working with this socio-scientific issue in the classroom requires students to apply complex models about energy flow and trophic pyramids in order to understand that food chains represent transfer of energy, to construct meanings for sustainable resources management through…

  15. Research and Education: The Foundations for Rehabilitation Service Delivery--10th Annual National Rehabilitation Educators Conference April 6th-10th, 2010

    ERIC Educational Resources Information Center

    Chou, Chih Chin

    2010-01-01

    The theme of the 10th annual National Rehabilitation Educators conference emphasized research and teaching ideals in the areas of clinical supervision, evidence-based practice in rehabilitation, rehabilitation counseling process, effective rehabilitation counseling training strategies, accreditation and licensure, rehabilitation ethics, and

  16. 16. NORTHEAST CORNER VIEW OF 10TH AND 11TH FLOOR WINDOWS. ...

    Library of Congress Historic Buildings Survey, Historic Engineering Record, Historic Landscapes Survey

    16. NORTHEAST CORNER VIEW OF 10TH AND 11TH FLOOR WINDOWS. CORNER SHOWS THE DIAGONALLY FLUTED SPIRAL DESIGN OF THE RELIEF COLUMN. - Pacific Telephone & Telegraph Company Building, 1519 Franklin Street, Oakland, Alameda County, CA

  17. 1. Historic American Buildings Survey Joseph Hill, Photographer August 10th, ...

    Library of Congress Historic Buildings Survey, Historic Engineering Record, Historic Landscapes Survey

    1. Historic American Buildings Survey Joseph Hill, Photographer August 10th, 1936 (Copied from small photo taken by survey members) OLD APARTMENT HOUSE - Jansonist Colony, Old Apartment House, Main Street, Bishop Hill, Henry County, IL

  18. Predictors of Student Performance in Grades 7 and 8 Mathematics: The Correlation between Benchmark Tests and Performance on the Texas Assessment of Knowledge and Skills (TAKS) Math Tests

    ERIC Educational Resources Information Center

    Allen, Timothy Dale

    2012-01-01

    School districts throughout Texas have used archived Texas Assessment of Knowledge and Skills (TAKS) tests as a benchmark to predict student performance on future TAKS tests without substantial quantitative evidence that these types of benchmark tests are valid predictors of student performance. The purpose of this quantitative correlational study

  19. How Many Letters Should Preschoolers in Public Programs Know? The Diagnostic Efficiency of Various Preschool Letter-Naming Benchmarks for Predicting First-Grade Literacy Achievement

    ERIC Educational Resources Information Center

    Piasta, Shayne B.; Petscher, Yaacov; Justice, Laura M.

    2012-01-01

    Review of current federal and state standards indicates little consensus or empirical justification regarding appropriate goals, often referred to as benchmarks, for preschool letter-name learning. The present study investigated the diagnostic efficiency of various letter-naming benchmarks using a longitudinal database of 371 children who attended

  20. Alberta K-12 ESL Proficiency Benchmarks

    ERIC Educational Resources Information Center

    Salmon, Kathy; Ettrich, Mike

    2012-01-01

    The Alberta K-12 ESL Proficiency Benchmarks are organized by division: kindergarten, grades 1-3, grades 4-6, grades 7-9, and grades 10-12. They are descriptors of language proficiency in listening, speaking, reading, and writing. The descriptors are arranged in a continuum of seven language competences across five proficiency levels. Several

  1. County Data Book, 2000: Kentucky Kids Count. 10th Annual Edition.

    ERIC Educational Resources Information Center

    Albright, Danielle; Hall, Douglas; Mellick, Donna; Miller, Debra; Town, Jackie

    This 10th annual Kids Count data book reports on trends in the well-being of Kentucky's children. The statistical portrait is based on indicators in the areas of well being, child risk factors, and demography. The indicators are as follows: (1) healthy births, including birth weights and prenatal care; (2) maternal risk characteristics, including

  2. County Data Book, 2000: Kentucky Kids Count. 10th Annual Edition.

    ERIC Educational Resources Information Center

    Albright, Danielle; Hall, Douglas; Mellick, Donna; Miller, Debra; Town, Jackie

    This 10th annual Kids Count data book reports on trends in the well-being of Kentucky's children. The statistical portrait is based on indicators in the areas of well being, child risk factors, and demography. The indicators are as follows: (1) healthy births, including birth weights and prenatal care; (2) maternal risk characteristics, including…

  3. A PAIR OF 10TH CAVALRY AMBULANCES, PARKED NEXT TO ONE ...

    Library of Congress Historic Buildings Survey, Historic Engineering Record, Historic Landscapes Survey

    A PAIR OF 10TH CAVALRY AMBULANCES, PARKED NEXT TO ONE OF THE STABLE LABELED "M.D. 10." PHOTOGRAPH TAKEN CIRCA 1918 (FORT HUACHUCA HISTORICAL MUSEUM, PHOTOGRAPH 1918.00.00.135, PHOTOGRAPHER UNIDENTIFIED, CREATED BY AND PROPERTY OF THE UNITED STATES ARMY) - Fort Huachuca, Cavalry Stables, Clarkson Road, Sierra Vista, Cochise County, AZ

  4. MedlinePlus en español marks its 10th anniversary

    MedlinePLUS

    ... nlm.nih.gov/medlineplus/spanishanniversary.html MedlinePlus en español Marks its 10 th Anniversary To use the ... on this page, please enable JavaScript. MedlinePlus en español ( http://medlineplus.gov/espanol ) launched in September 2002. ...

  5. Making a Difference: Education at the 10th International Conference on Zebrafish Development and Genetics

    PubMed Central

    Liang, Jennifer O.; Pickart, Michael A.; Pierret, Chris; Tomasciewicz, Henry G.

    2012-01-01

    Abstract Scientists, educators, and students met at the 10th International Conference on Zebrafish Development and Genetics during the 2-day Education Workshop, chaired by Dr. Jennifer Liang and supported in part by the Genetics Society of America. The goal of the workshop was to share expertise, to discuss the challenges faced when using zebrafish in the classroom, and to articulate goals for expanding the impact of zebrafish in education. PMID:23244686

  6. 14. CLOSEUP VIEW OF THE 10TH AND 11TH FLOOR WINDOWS. ...

    Library of Congress Historic Buildings Survey, Historic Engineering Record, Historic Landscapes Survey

    14. CLOSE-UP VIEW OF THE 10TH AND 11TH FLOOR WINDOWS. WINDOWS HAVE WHITE TERRA COTTA SILLS, HEADS AND MULLIONS. ARCHES ARE OF TERRA COTTA INCLUDING ORNAMENTATION ABOVE THE 11TH FLOOR WINDOWS. CIRCULAR ORNAMENTATIONS BETWEEN ARCHES ARE TERRA COTTA PAINTED IN BRONZE COLOR. LOUVERS ON THE WINDOWS ARE NOT PART OF THE ORIGINAL DESIGN. THIS IS THE FRONT ELEVATION. - Pacific Telephone & Telegraph Company Building, 1519 Franklin Street, Oakland, Alameda County, CA

  7. From the corner of N. 10th St. and W. O'Neill ...

    Library of Congress Historic Buildings Survey, Historic Engineering Record, Historic Landscapes Survey

    From the corner of N. 10th St. and W. O'Neill Ave. Looking west. Housing # 157-162 are on the right, building 156 is straight ahead, and buildings 153, 152, 116, and 115 are to the left. The golf course is directly west of these buildings. - Fitzsimons General Hospital, Bounded by East Colfax to south, Peoria Street to west, Denver City/County & Adams County Line to north, & U.S. Route 255 to east, Aurora, Adams County, CO

  8. The Relationship of Grade Span in 9th Grade to Math Achievement in High School

    ERIC Educational Resources Information Center

    West, John; Miller, Mary Lou; Myers, Jim; Norton, Timothy

    2015-01-01

    Purpose, Scope, and Method of Study: The purpose of this study was to determine if a correlation exists between grade span for ninth grade and gains in math achievement test scores in 10th grade and 12th grade. A quantitative, longitudinal, correlational research design was employed to investigate the research questions. The population was high

  9. Final Report 10th Conference on the Intersections of Particle and Nuclear Physics

    SciTech Connect

    Marshak, Marvin L.

    2013-11-03

    The 10th Conference on the Intersections of Particle and Nuclear Physics was held in LaJolla, California on May 26 to May 31, 2009. The Conference Proceedings are published by the American Institute of Physics in Volume 1182 of the AIP Conference Proceedings (ISBN: 978-0-7354-0723-7). The Proceedings include papers from each of the Conference Presenters and a detailed schedule of talks at the Conference. The Table of Contents of the Conference Proceedings is available at http://scitation.aip.org/content/aip/proceeding/aipcp/1182. Support by the U.S. Department of Energy and by DOE Laboratories was essential to the success of the Conference.

  10. From the corner of E. Mccloskey Ave. and N. 10th ...

    Library of Congress Historic Buildings Survey, Historic Engineering Record, Historic Landscapes Survey

    From the corner of E. Mccloskey Ave. and N. 10th St., looking west with building 135 (gas station) on the left. Beyond it is building 119 and to the right of 119 is the gable end of the north side of 120. Beyond and perpendicular to building 120 are 118 and 117. - Fitzsimons General Hospital, Bounded by East Colfax to south, Peoria Street to west, Denver City/County & Adams County Line to north, & U.S. Route 255 to east, Aurora, Adams County, CO

  11. Under Construction: Benchmark Assessments and Common Core Math Implementation in Grades K-8. Formative Evaluation Cycle Report for the Math in Common Initiative, Volume 1

    ERIC Educational Resources Information Center

    Flaherty, John, Jr.; Sobolew-Shubin, Alexandria; Heredia, Alberto; Chen-Gaddini, Min; Klarin, Becca; Finkelstein, Neal D.

    2014-01-01

    Math in Common (MiC) is a five-year initiative that supports a formal network of 10 California school districts as they implement the Common Core State Standards in mathematics (CCSS-M) across grades K-8. As the MiC initiative moves into its second year, one of the central activities that each of the districts is undergoing to support CCSS

  12. Students Left Behind: Measuring 10th to 12th Grade Student Persistence Rates in Texas High Schools

    PubMed Central

    Domina, Thurston; Ghosh-Dastidar, Bonnie; Tienda, Marta

    2012-01-01

    The No Child Left Behind Act requires states to publish high school graduation rates for public schools and the U.S. Department of Education is currently considering a mandate to standardize high school graduation rate reporting. However, no consensus exists among researchers or policy-makers about how to measure high school graduation rates. In this paper, we use longitudinal data tracking a cohort of students at 82 Texas public high schools to assess the accuracy and precision of three widely-used high school graduation rate measures: Texass official graduation rates, and two competing estimates based on publicly available enrollment data from the Common Core of Data. Our analyses show that these widely-used approaches yield inaccurate and highly imprecise estimates of high school graduation and persistence rates. We propose several guidelines for using existing graduation and persistence rate data and argue that a national effort to track students as they progress through high school is essential to reconcile conflicting estimates. PMID:23077375

  13. Students Left Behind: Measuring 10(th) to 12(th) Grade Student Persistence Rates in Texas High Schools.

    PubMed

    Domina, Thurston; Ghosh-Dastidar, Bonnie; Tienda, Marta

    2010-06-01

    The No Child Left Behind Act requires states to publish high school graduation rates for public schools and the U.S. Department of Education is currently considering a mandate to standardize high school graduation rate reporting. However, no consensus exists among researchers or policy-makers about how to measure high school graduation rates. In this paper, we use longitudinal data tracking a cohort of students at 82 Texas public high schools to assess the accuracy and precision of three widely-used high school graduation rate measures: Texas's official graduation rates, and two competing estimates based on publicly available enrollment data from the Common Core of Data. Our analyses show that these widely-used approaches yield inaccurate and highly imprecise estimates of high school graduation and persistence rates. We propose several guidelines for using existing graduation and persistence rate data and argue that a national effort to track students as they progress through high school is essential to reconcile conflicting estimates. PMID:23077375

  14. Students Left behind: Measuring 10th to 12th Grade Student Persistence Rates in Texas High Schools

    ERIC Educational Resources Information Center

    Domina, Thurston; Ghosh-Dastidar, Bonnie; Tienda, Marta

    2010-01-01

    The No Child Left Behind Act requires states to publish high school graduation rates for public schools; the U.S. Department of Education is currently considering a mandate to standardize high school graduation rate reporting. However, no consensus exists among researchers or policymakers about how to measure high school graduation rates. We use

  15. The Insertion of Local Wisdom into Instructional Materials of Bahasa Indonesia for 10th Grade Students in Senior High School

    ERIC Educational Resources Information Center

    Anggraini, Purwati; Kusniarti, Tuti

    2015-01-01

    This current study aimed at investigating Bahasa Indonesia textbooks with regards to local wisdom issues. The preliminary study was utilized as the basis for developing instructional materials of Bahasa Indonesia that are rich of characters. Bahasa Indonesia instructional materials containing local wisdoms not only equip students with broad…

  16. Successes with Reversing the Negative Student Attitudes Developed in Typical Biology Classes for 8th and 10th Grade Students

    ERIC Educational Resources Information Center

    Hacieminoglu, Esme; Ali, Mohamed Moustafa; Oztas, Fulya; Yager, Robert E.

    2016-01-01

    The purpose of this study is to compare changes in attitudes of students about their study of biology in the classes thought by five biology teachers who experienced an Iowa Chautauqua workshop with and two non-Chautauqua teachers who had no experience with any professional development program. The results indicated that there are significant

  17. A Learning Progression for Deepening Students' Understandings of Modern Genetics across the 5th-10th Grades

    ERIC Educational Resources Information Center

    Duncan, Ravit Golan; Rogat, Aaron D.; Yarden, Anat

    2009-01-01

    Over the past several decades, there has been a tremendous growth in our understanding of genetic phenomena and the intricate and complicated mechanisms that mediate genetic effects. Given the complexity of content in modern genetics and the inadequacy of current instructional methods and materials it seems that a more coherent and extensive

  18. Successes with Reversing the Negative Student Attitudes Developed in Typical Biology Classes for 8th and 10th Grade Students

    ERIC Educational Resources Information Center

    Hacieminoglu, Esme; Ali, Mohamed Moustafa; Oztas, Fulya; Yager, Robert E.

    2016-01-01

    The purpose of this study is to compare changes in attitudes of students about their study of biology in the classes thought by five biology teachers who experienced an Iowa Chautauqua workshop with and two non-Chautauqua teachers who had no experience with any professional development program. The results indicated that there are significant…

  19. Exploring the Relationship between Virtual Learning Environment Preference, Use, and Learning Outcomes in 10th Grade Earth Science Students

    ERIC Educational Resources Information Center

    Lin, Ming-Chao; Tutwiler, M. Shane; Chang, Chun-Yen

    2011-01-01

    This study investigated the relationship between the use of a three-dimensional Virtual Reality Learning Environment for Field Trip (3DVLE[subscript (ft)]) system and the achievement levels of senior high school earth science students. The 3DVLE[subscript (ft)] system was presented in two separate formats: Teacher Demonstrated Based and Student

  20. A Cross-Analysis of the Mathematics Teacher's Activity: An Example in a French 10th-Grade Class

    ERIC Educational Resources Information Center

    Robert, Aline; Rogalski, Janine

    2005-01-01

    The purpose of this paper is to contribute to the debate about how to tackle the issue of "the teacher in the teaching/learning process", and to propose a methodology for analysing the teacher's activity in the classroom, based on concepts used in the fields of the didactics of mathematics as well as in cognitive ergonomics. This methodology

  1. A Cross-Analysis of the Mathematics Teacher's Activity: An Example in a French 10th-Grade Class

    ERIC Educational Resources Information Center

    Robert, Aline; Rogalski, Janine

    2005-01-01

    The purpose of this paper is to contribute to the debate about how to tackle the issue of "the teacher in the teaching/learning process", and to propose a methodology for analysing the teacher's activity in the classroom, based on concepts used in the fields of the didactics of mathematics as well as in cognitive ergonomics. This methodology…

  2. Space Commerce 1994 Forum: The 10th National Space Symposium. Proceedings report

    NASA Technical Reports Server (NTRS)

    Lipskin, Beth Ann (Editor); Patterson, Sara (Editor); Aragon, Larry (Editor); Brescia, David A. (Editor); Flannery, Jack (Editor); Mossey, Roberty (Editor); Regan, Christopher (Editor); Steeby, Kurt (Editor); Suhr, Stacy (Editor); Zimkas, Chuck (Editor)

    1994-01-01

    The theme of the 10th National Space Symposium was 'New Windows of Opportunity'. These proceedings cover the following: Business Trends in High Tech Commercialization; How to Succeed in Space Technology Business -- Making Dollars and Sense; Obstacles and Opportunities to Success in Technology Commercialization NASA's Commercial Technology Mission -- a New Way of Doing Business: Policy and Practices; Field Center Practices; Practices in Action -- A New Way: Implementation and Business Opportunities; Space Commerce Review; Windows of Opportunity; the International Space Station; Space Support Forum; Spacelift Update; Competitive Launch Capabilities; Supporting Life on Planet Earth; National Security Space Issues; NASA in the Balance; Earth and Space Observations -- Did We Have Cousins on Mars?; NASA: A New Vision for Science; and Space Technology Hall of Fame.

  3. 10th World IHEA and ECHE Joint Congress: health economics in the age of longevity.

    PubMed

    Jakovljevic, Mihajlo B; Getzen, Thomas E; Torbica, Aleksandra; Anegawa, Tomofumi

    2014-12-01

    The 10th consecutive World Health Economics conference was organized jointly by International Health Economics Association and European Conference on Health Economics Association and took place at The Trinity College, Dublin, Ireland in July 2014. It has attracted broad participation from the global professional community devoted to health economics teaching,research and policy applications. It has provided a forum for lively discussion on hot contemporary issues such as health expenditure projections, reimbursement regulations,health technology assessment, universal insurance coverage, demand and supply of hospital services, prosperity diseases, population aging and many others. The high-profile debate fostered by this meeting is likely to inspire further methodological advances worldwide and spreading of evidence-based policy practice from OECD towards emerging markets. PMID:25301000

  4. New directions in research: report from the 10th International Conference on AIDS.

    PubMed Central

    Berger, P B

    1995-01-01

    Research findings presented at the 10th International Conference on AIDS, held in Yokohama, Japan, in August 1994, indicate that few advances have been made in standard antiretroviral therapy for HIV infection. The perinatal administration of AZT (zidovudine) was reported to reduce transmission of HIV from mother to child, and its use in combination with acyclovir appears to improve survival among patients with advanced disease. Other research has focused on asymptomatic patients with long-standing HIV infection. Their survival may be related to the activity of cell antiviral factor, a cytokine produced by CD8+ cells. In gene therapy research, one approach involved the genetic alteration of target cells to enable them to render the virus harmless. A second approach consisted of enhancing the function of CD8+ cells to allow them to compensate for dysfunctional CD4+ cells. The author believes that gene therapy may offer the greatest hope of an effective treatment for HIV infection. PMID:7780908

  5. Space Commerce 1994 Forum: The 10th National Space Symposium. Proceedings report

    NASA Astrophysics Data System (ADS)

    Lipskin, Beth Ann; Patterson, Sara; Aragon, Larry; Brescia, David A.; Flannery, Jack; Mossey, Roberty; Regan, Christopher; Steeby, Kurt; Suhr, Stacy; Zimkas, Chuck

    1994-04-01

    The theme of the 10th National Space Symposium was 'New Windows of Opportunity'. These proceedings cover the following: Business Trends in High Tech Commercialization; How to Succeed in Space Technology Business -- Making Dollars and Sense; Obstacles and Opportunities to Success in Technology Commercialization NASA's Commercial Technology Mission -- a New Way of Doing Business: Policy and Practices; Field Center Practices; Practices in Action -- A New Way: Implementation and Business Opportunities; Space Commerce Review; Windows of Opportunity; the International Space Station; Space Support Forum; Spacelift Update; Competitive Launch Capabilities; Supporting Life on Planet Earth; National Security Space Issues; NASA in the Balance; Earth and Space Observations -- Did We Have Cousins on Mars?; NASA: A New Vision for Science; and Space Technology Hall of Fame.

  6. Mathematics Content Standards Benchmarks and Performance Standards

    ERIC Educational Resources Information Center

    New Mexico Public Education Department, 2008

    2008-01-01

    New Mexico Mathematics Content Standards, Benchmarks, and Performance Standards identify what students should know and be able to do across all grade levels, forming a spiraling framework in the sense that many skills, once introduced, develop over time. While the Performance Standards are set forth at grade-specific levels, they do not exist as

  7. Sequenced Benchmarks for K-8 Science.

    ERIC Educational Resources Information Center

    Kendall, John S.; DeFrees, Keri L.; Richardson, Amy

    This document describes science benchmarks for grades K-8 in Earth and Space Science, Life Science, and Physical Science. Each subject area is divided into topics followed by a short content description and grade level information. Source documents for this paper included science content guides from California, Ohio, South Carolina, and South…

  8. Updating the ACT College Readiness Benchmarks. ACT Research Report Series 2013 (6)

    ERIC Educational Resources Information Center

    Allen, Jeff

    2013-01-01

    The ACT College Readiness Benchmarks are the ACT College Readiness Assessment scores associated with a 50% chance of earning a B or higher grade in typical first-year credit-bearing college courses. The Benchmarks also correspond to an approximate 75% chance of earning a C or higher grade in these courses. There are four Benchmarks, corresponding

  9. FOREWORD: 10th Anglo-French Physical Acoustics Conference (AFPAC 2011)

    NASA Astrophysics Data System (ADS)

    Lhmery, Alain; Saffari, Nader

    2012-03-01

    The Anglo-French Physical Acoustics Conference (AFPAC) had its 10th annual meeting in Villa Clythia, Frjus, France, from 19-21 January 2011. This series of meetings is a collaboration between the Physical Acoustics Group (PAG) of the Institute of Physics and the Groupe d'Acoustique Physique, Sous-marine et UltraSonore (GAPSUS) of the Socit Franaise d'Acoustique. The conference has its loyal supporters whom we wish to thank. It is their loyalty that has made this conference a success. AFPAC alternates between the UK and France and its format has been designed to ensure that it remains a friendly meeting of very high scientific quality, offering a broad spectrum of subjects, welcoming young researchers and PhD students and giving them the opportunity to give their first presentations in an 'international' conference, but with limited pressure. For the third consecutive year AFPAC is followed by the publication of its proceedings in the form of 18 peer-reviewed papers, which cover the most recent research developments in the field of Physical Acoustics in the UK and France. Alain Lhmery CEA, France Nader Saffari UCL, United Kingdom

  10. The 8th-10 th January 2009 snowfalls: a case of Mediterranean warm advection event

    NASA Astrophysics Data System (ADS)

    Aguado, F.; Ayensa, E.; Barriga, M.; Del Hoyo, J.; Fernández, A.; Garrido, N.; Martín, A.; Martín, F.; Roa, I. Martínez, A.; Pascual, R.

    2009-09-01

    From 8 th to 10 th of January 2009, significant snowfalls were reported in many areas of the Iberian Peninsula and the Balearic Islands. This relevant event was very important from the meteorological and social impact point of views. The snow affected many zones, especially the regions of Madrid, Castilla & León and Castilla-La Mancha (Spanish central plateau) with the persistence and thickness of solid precipitation. Up to twenty-five centimetres of snow were reported in some places. On 9th of January the snowfalls caused great social and media impact due to the fact that they took place in the early hours in the Madrid metropolitan areas, affecting both air traffic and land transport. The "Madrid-Barajas" airport was closed and the city was collapsed during several hours. A study of this situation appears in the poster. The snowstorm was characterized by the previous irruption of an European continental polar air mass, that subsequently interacted with a wet and warm air mass of Mediterranean origin, all preceded by low level easterly flows. This type of snowfall is called "warm advection". These winter situations are very efficient from precipitation point of view, generating significant snowfalls and affecting a lot of areas.

  11. Rituximab-combination chemotherapy achieves a 10th cycle of remission for Burkitt's lymphoma.

    PubMed

    Umeda, Katsutsugu; Fujino, Hisanori; Saida, Satoshi; Kato, Itaru; Hiramatsu, Hidefumi; Yamada, Tomomi; Hori, Toshinori; Adachi, Souichi; Heike, Toshio; Watanabe, Ken-ichiro

    2015-04-01

    A 14-year-old girl with multiple intra-abdominal tumors was diagnosed with stage III Burkitt's lymphoma. She achieved complete remission after multi-drug chemotherapy, but she relapsed after six courses. Autologous peripheral blood stem cells (PBSC) or allogeneic PBSC harvested from an HLA-identical sibling were insufficient, and her family did not agree to bone marrow collection from the sibling. Although the patient relapsed nine times (the relapses involved intra-abdominal organs or bone) during the following 4 years 7 months, treatment with rituximab monotherapy or in combination with ifosphamide, carboplastin, and etoposide, or local irradiation (33.8-40.0 Gy) to treat the bone metastases, proved effective, resulting in complete or partial remission. At the time of writing, the patient was in a 10th cycle of remission lasting 1 year 6 months and had not required transplantation. Thus, a chemotherapy regimen including rituximab might be effective for Burkitt's lymphoma in patients experiencing multiple relapse. PMID:25868956

  12. Benchmarking in Student Affairs.

    ERIC Educational Resources Information Center

    Mosier, Robert E.; Schwarzmueller, Gary J.

    2002-01-01

    Discusses the use of benchmarking in student affairs, focusing on issues related to student housing. Provides examples of how benchmarking has influenced administrative practice at many institutions. (EV)

  13. 10th annual meeting of the Safety Pharmacology Society: an overview.

    PubMed

    Cavero, Icilio

    2011-03-01

    The 10th annual meeting of the Safety Pharmacology (SP) Society covered numerous topics of educational and practical research interest. Biopolymers - the theme of the keynote address - were presented as essential components of medical devices, diagnostic tools, biosensors, human tissue engineering and pharmaceutical formulations for optimized drug delivery. Toxicology and SP investigators - the topic of the Distinguished Service Award Lecture - were encouraged to collaborate in the development of SP technologies and protocols applicable to toxicology studies. Pharmaceutical companies, originally organizations bearing all risks for developing their portfolios, are increasingly moving towards fully integrated networks which outsource core activities (including SP studies) to large contract research organizations. Future nonclinical data are now expected to be of such high quality and predictability power that they may obviate the need for certain expensive and time-consuming clinical investigations. In this context, SP is called upon to extend its risk assessment purview to areas which currently are not systematically covered, such as drug-induced QRS interval prolongation, negative emotions and feelings (e.g., depression), and minor chronic cardiovascular and metabolic changes (e.g., as produced by drugs for type 2 diabetes) which can be responsible for delayed morbidity and mortality. The recently approved ICH S9 guidance relaxes the traditional regulatory SP package in order to accelerate the clinical access to anticancer drugs for patients with advanced malignancies. The novel FDA 'Animal Rule' guidance proposes that for clinical candidates with well-understood toxicities, marketing approval may be granted exclusively on efficacy data generated in animal studies as human clinical investigations for these types of drugs are either unfeasible or unethical. In conclusion, the core messages of this meeting are that SP should consistently operate according to the 'fit-for-purpose' principle and gradually integrate new mechanism-oriented safety paradigms into the traditional ones for ensuring more effectively the safety of drugs for any population of patients in need. PMID:21314442

  14. Eruption Mechanism of the 10th Century Eruption in Baitoushan Volcano, China/North Korea

    NASA Astrophysics Data System (ADS)

    Shimano, T.; Miyamoto, T.; Nakagawa, M.; Ban, M.; Maeno, F.; Nishimoto, J.; Jien, X.; Taniguchi, H.

    2005-12-01

    Baitoushan volcano, China/North Korea, is one of the most active volcanoes in Northeastern Asia, and the 10th century eruption was the most voluminous eruption in the world in recent 2000 years. The sequence of the eruption reconstructed recently consists mainly of 6 units of deposits (Miyamoto et al., 2004); plinian airfall (unit B), large pyroclastic flow (unit C), plinian airfall with some intra- plinian pyroclastic flows (unit D), sub-plinian airfall (unit E), and large pyroclastic flow (unit F) with base surge (unit G) in ascending order. The magma erupted during steady eruption in earlier phase was comendite (unit B-C; Phase 1), whereas the magma during fluctuating eruptions in later phase is characterized by trachyte to trachyandesite with various amount of comendite (unit D-G; Phase 2). The wide variety in composition and occurrence of banded pumices strongly indicate mixing or mingling of the two magmas just prior to or during the eruption. The initial water contents had been determined for comendite by melt inclusion analyses (ca. 5.2 wt.%; Horn and Schmincke, 2000). Although the initial water content of the trachytic magma has not been correctly determined yet, the reported water contents of trachytic melt inclusions are lower (3-4 wt.%) than those of comenditic melt (Horn and Schmincke, 2000). We investigated juvenile materials of the eruption sequentially in terms of vesicularity, H2O content in matrix glass and textural characteristics. The vesicularity of pumices are generally high (>0.75) for all units. The residual water contents of the comenditic pumices during Phase 1 are relatively uniform (1.6 wt.%), whereas those of the trachytic scoria during Phase 2 and gray pumices during Phase 1 are low (ca. 0.7-1.3 wt.%). These facts may indicate that the difference in the initial water content, rather than the ascent mechanism of magma, controls the steadiness or fluctuation in eruption styles and the mass flux during the eruption.

  15. Report: Combustion Byproducts and Their Health Effects: Summary of the 10th International Congress.

    PubMed

    Dellinger, Barry; D'Alessio, Antonio; D'Anna, Andrea; Ciajolo, Anna; Gullett, Brian; Henry, Heather; Keener, Mel; Lighty, Joann; Lomnicki, Slawomir; Lucas, Donald; Oberdörster, Günter; Pitea, Demetrio; Suk, William; Sarofim, Adel; Smith, Kirk R; Stoeger, Tobias; Tolbert, Paige; Wyzga, Ron; Zimmermann, Ralf

    2008-10-01

    The 10th International Congress on Combustion Byproducts and their Health Effects was held in Ischia, Italy, from June 17-20, 2007. It is sponsored by the US NIEHS, NSF, Coalition for Responsible Waste Incineration (CRWI), and Electric Power Research Institute (EPRI). The congress focused on: the origin, characterization, and health impacts of combustion-generated fine and ultrafine particles; emissions of mercury and dioxins, and the development/application of novel analytical/diagnostic tools. The consensus of the discussion was that particle-associated organics, metals, and persistent free radicals (PFRs) produced by combustion sources are the likely source of the observed health impacts of airborne PM rather than simple physical irritation of the particles. Ultrafine particle-induced oxidative stress is a likely progenitor of the observed health impacts, but important biological and chemical details and possible catalytic cycles remain unresolved. Other key conclusions were: (1) In urban settings, 70% of airborne fine particles are a result of combustion emissions and 50% are due to primary emissions from combustion sources, (2) In addition to soot, combustion produces one, possibly two, classes of nanoparticles with mean diameters of ~10 nm and ~1 nm. (3) The most common metrics used to describe particle toxicity, viz. surface area, sulfate concentration, total carbon, and organic carbon, cannot fully explain observed health impacts, (4) Metals contained in combustion-generated ultrafine and fine particles mediate formation of toxic air pollutants such as PCDD/F and PFRs. (5) The combination of metal-containing nanoparticles, organic carbon compounds, and PFRs can lead to a cycle generating oxidative stress in exposed organisms. PMID:22476005

  16. Mathematics 10th Year. Curriculum Bulletin, 1964-65 Series, Number 8.

    ERIC Educational Resources Information Center

    New York City Board of Education, Brooklyn, NY. Bureau of Curriculum Development.

    This bulletin suggests teaching procedures for a tenth-grade geometry course. An introduction states general objectives and discusses the nature of inductive and deductive thinking. Lesson outlines follow which give a statement of objectives and a plan for development. Major lesson groupings are congruent triangles, parallel lines, quadrilaterals,

  17. Carcass traits, cut yields, and compositional end points in high-lean-yielding pork carcasses: effects of 10th rib backfat and loin eye area.

    PubMed

    Pringle, T D; Williams, S E

    2001-01-01

    Pork carcasses (n = 133) were used to investigate the influence of carcass fatness and muscling on composition and yields of pork primal and subprimal cuts fabricated to varying levels of s.c. fat. Carcasses were selected from commercial packing plants in the southeastern United States, using a 3 x 3 factorial arrangement with three levels of 10th rib backfat depth (< 2.03, 2.03 to 2.54, and > 2.54 cm) and three levels of loin eye area (LEA; < 35.5, 35.5 to 41.9, and > 41.9 cm2). Sides from the selected carcasses were shipped to the University of Georgia for carcass data collection by trained USDA-AMS and University of Georgia personnel and fabrication. Sides were fabricated to four lean cuts (picnic shoulder, Boston butt, loin, and ham) and the skinned belly. The four lean cuts were further fabricated into boneless cuts with s.c. fat trim levels of 0.64, 0.32, and 0 cm. The percentages of four lean cuts, boneless cuts (four lean cuts plus skinned, trimmed belly) at 0.64, 0.32, and 0 cm s.c. fat, fat-free lean, and total fat were calculated. Data were analyzed using a least squares fixed effects model, with the main effects of 10th rib backfat and LEA and their interaction. Fatness and muscling traits increased (P < 0.05) as 10th rib backfat and LEA category increased, respectively. However, fat depth measures were not affected greatly by LEA category, nor were muscling measures greatly affected by backfat category. The percentage yield of cuts decreased (P < 0.05) as backfat category increased. Cut yields from the picnic shoulder, Boston butt, and belly were not affected (P > 0.05) by LEA category, whereas the yield of boneless loin and ham increased (P < 0.05) as LEA category increased. Compositionally, the percentage of four lean cuts, boneless cuts at varying trim levels, and fat-free lean decreased incrementally (P < 0.05) as backfat depth increased, whereas parentage total fat and USDA grade increased (P < 0.05) as backfat depth increased. As LEA increased, percentage boneless cuts trimmed to 0.32 and 0 cm s.c. fat and fat-free lean increased and total fat decreased; however, the difference was only significant in the smallest LEA category. Collectively, these data show that decreased carcass fatness plays a greater role in increasing primal and subprimal cut yields and carcass composition than muscling even in lean, heavily muscled carcasses. PMID:11204691

  18. Risk Communication and Public Education in Edmonton, Alberta, Canada on the 10th Anniversary of the "Black Friday" Tornado

    ERIC Educational Resources Information Center

    Blanchard-Boehm, R. Denise; Cook, M. Jeffrey

    2004-01-01

    In July 1997, on the 10th anniversary of the great "Black Friday" Tornado, city officials of Edmonton, the print and broadcast media, agencies dealing in emergency management, and the national weather organisation recounted stories of the 1987, F5 tornado that struck Edmonton on a holiday weekend. The information campaign also presented

  19. Graduate Students Lend Their Voices: Reflections on the 10th Seminar in Health and Environmental Education Research

    ERIC Educational Resources Information Center

    Russell, Joshua; White, Peta; Fook, Tanya Chung Tiam; Kayira, Jean; Muller, Susanne; Oakley, Jan

    2010-01-01

    Graduate students were invited by their faculty advisors to attend the 10th Seminar in Health and Environmental Education Research. Afterward, they were encouraged to comment on their experiences, involvement, and positioning. Two main authors developed survey questions and retrieved, analyzed, and synthesized the responses of four other graduate

  20. Proceedings of the International Conference on Mobile Learning 2014. (10th, Madrid, Spain, February 28-March 2, 2014)

    ERIC Educational Resources Information Center

    Sánchez, Inmaculada Arnedillo, Ed.; Isaías, Pedro, Ed.

    2014-01-01

    These proceedings contain the papers of the 10th International Conference on Mobile Learning 2014, which was organised by the International Association for Development of the Information Society, in Madrid, Spain, February 28-March 2, 2014. The Mobile Learning 2014 International Conference seeks to provide a forum for the presentation and…

  1. Risk Communication and Public Education in Edmonton, Alberta, Canada on the 10th Anniversary of the "Black Friday" Tornado

    ERIC Educational Resources Information Center

    Blanchard-Boehm, R. Denise; Cook, M. Jeffrey

    2004-01-01

    In July 1997, on the 10th anniversary of the great "Black Friday" Tornado, city officials of Edmonton, the print and broadcast media, agencies dealing in emergency management, and the national weather organisation recounted stories of the 1987, F5 tornado that struck Edmonton on a holiday weekend. The information campaign also presented…

  2. Selected Papers from the International Conference on College Teaching and Learning (10th, Jacksonville, Florida, April 1999).

    ERIC Educational Resources Information Center

    Chambers, Jack A., Ed.

    These 20 papers were selected from those presented at the 10th International Conference on College Teaching and Learning. Papers have the following titles and authors: (1) "Case It! A Project to Integrate Collaborative Case-Based Learning into International Undergraduate Biology Curricula" (Bergland, Klyczek, Lundeberg, Mogen, Johnson); (2) "The

  3. The NAS parallel benchmarks

    NASA Technical Reports Server (NTRS)

    Bailey, David (Editor); Barton, John (Editor); Lasinski, Thomas (Editor); Simon, Horst (Editor)

    1993-01-01

    A new set of benchmarks was developed for the performance evaluation of highly parallel supercomputers. These benchmarks consist of a set of kernels, the 'Parallel Kernels,' and a simulated application benchmark. Together they mimic the computation and data movement characteristics of large scale computational fluid dynamics (CFD) applications. The principal distinguishing feature of these benchmarks is their 'pencil and paper' specification - all details of these benchmarks are specified only algorithmically. In this way many of the difficulties associated with conventional benchmarking approaches on highly parallel systems are avoided.

  4. Fortified Settlements of the 9th and 10th Centuries ad in Central Europe: Structure, Function and Symbolism

    PubMed Central

    Herold, Hajnalka

    2012-01-01

    THE STRUCTURE, FUNCTION(S) and symbolism of early medieval (9th10th centuries ad) fortified settlements from central Europe, in particular todays Austria, Hungary, Czech Republic and Slovakia, are examined in this paper. It offers an overview of the current state of research together with new insights based on analysis of the site of Gars-Thunau in Lower Austria. Special emphasis is given to the position of the fortified sites in the landscape, to the elements of the built environment and their spatial organisation, as well as to graves within the fortified area. The region under study was situated on the SE border of the Carolingian (and later the Ottonian) Empire, with some of the discussed sites lying in the territory of the Great Moravian Empire in the 9th and 10th centuries. These sites can therefore provide important comparative data for researchers working in other parts of the Carolingian Empire and neighbouring regions. PMID:23564981

  5. Benchmarking: The New Tool.

    ERIC Educational Resources Information Center

    Stralser, Steven

    1995-01-01

    This article suggests that benchmarking, the process of comparing one's own operation with the very best, can be used to make improvements in colleges and universities. Six steps are outlined: determining what to benchmark, forming a team, discovering who to benchmark, collecting and analyzing data, using the data to redesign one's own operation,

  6. Benchmarking for Higher Education.

    ERIC Educational Resources Information Center

    Jackson, Norman, Ed.; Lund, Helen, Ed.

    The chapters in this collection explore the concept of benchmarking as it is being used and developed in higher education (HE). Case studies and reviews show how universities in the United Kingdom are using benchmarking to aid in self-regulation and self-improvement. The chapters are: (1) "Introduction to Benchmarking" (Norman Jackson and Helen

  7. Benchmarking for Higher Education.

    ERIC Educational Resources Information Center

    Jackson, Norman, Ed.; Lund, Helen, Ed.

    The chapters in this collection explore the concept of benchmarking as it is being used and developed in higher education (HE). Case studies and reviews show how universities in the United Kingdom are using benchmarking to aid in self-regulation and self-improvement. The chapters are: (1) "Introduction to Benchmarking" (Norman Jackson and Helen…

  8. Evaluating the Ability of Drama-Based Instruction To Influence the Socialization of Tenth Grade English Students Labeled as "Low Ability."

    ERIC Educational Resources Information Center

    Danielson, Trygve R.

    A study investigated the effect of drama-based instruction on the learning of social skills by student labeled as "low ability" in a 10th-grade required English class in a rural high school. Two separate classes of "low ability" 10th-grade English students in Janesville, Wisconsin, were presented with social skills training utilizing role playing,…

  9. [A modified retroperitoneal approach to the kidney in patients with a highly deformed thorax: obtaining a wide operative field through subperiosteal resection of the 10th, 11th and 12th ribs].

    PubMed

    Satoh, Yuji; Kanou, Takehiro; Takagi, Norito; Tokuda, Yuji; Uozumi, Jiro; Masaki, Zenjiro

    2005-07-01

    We herein report a technique which facilitates a retroperitoneal approach to the kidney in cases of highly deformed thorax due to kyphoscoliosis. The operation consists of a lumbar oblique incision with removal of the 11th rib, combined with the additional removal of the 12th and 10th ribs. Resection of the upper two ribs was performed subperiosteally, leaving the periosteum of the deep side untouched. However, the deep side periosteum of the 12th rib was incised caudal from the pleural margin in order to facilitate exposure of the diaphragm. The retroperitoneal space was entered through the tip of the 11th rib bed. The diaphragm was incised dorso-medially at a level 1 cm caudal from the lower margin of the pleura, to an extent necessary to enable the pleura together with the cranial diaphragm to be manoeuvred in an upward direction. Two cases with renal tuberculosis associated with high-grade kyphosis and one case with staghorn calculi accompanied with lordosis were operated on utilizing this technique. In the former two cases, the thoracic cage was in direct contact with the iliac bone and there was practically no space between the rib border and the iliac crest. This was also true of the third case, but the grade of deformity was not as extensive as in the former two cases. Removal of the 10th, 11th and 12th ribs could be achieved without injuring the pleura and a satisfactorily large operating field could thus be developed which enabled a simple nephrectomy to be performed without difficulty. The characteristic feature of the described approach is that resection of the 10th and 11th ribs is simply to facilitate manoevrability of the wound margin, without going through the rib bed. The technique could be advantageous in selected cases where there is a highly deformed thorax. PMID:16083038

  10. Carpenter, Tractors and Microbes for the Development of Logical-Mathematical Thinking--The Way 10th Graders and Pre-Service Teachers Solve Thinking Challenges

    ERIC Educational Resources Information Center

    Gazit, Avikam

    2012-01-01

    The objective of this case study was to investigate the ability of 10th graders and pre-service teachers to solve logical-mathematical thinking challenges. The challenges do not require mathematical knowledge beyond that of primary school but rather an informed use of the problem representation. The percentage of correct answers given by the 10th

  11. Operationalizing the Rubric: The Effect of Benchmark Selection on the Assessed Quality of Writing.

    ERIC Educational Resources Information Center

    Popp, Sharon E. Osborn; Ryan, Joseph M.; Thompson, Marilyn S.; Behrens, John T.

    The purposes of this study were to investigate the role of benchmark writing samples in direct assessment of writing and to examine the consequences of differential benchmark selection with a common writing rubric. The influences of discourse and grade level were also examined within the context of differential benchmark selection. Raters scored

  12. The Nature and Predictive Validity of a Benchmark Assessment Program in an American Indian School District

    ERIC Educational Resources Information Center

    Payne, Beverly J. R.

    2013-01-01

    This mixed methods study explored the nature of a benchmark assessment program and how well the benchmark assessments predicted End-of-Grade (EOG) and End-of-Course (EOC) test scores in an American Indian school district. Five major themes were identified and used to develop a Dimensions of Benchmark Assessment Program Effectiveness model:

  13. Is the 10th and 11th Intercostal Space a Safe Approach for Percutaneous Nephrostomy and Nephrolithotomy?

    SciTech Connect

    Muzrakchi, Ahmed Al; Szmigielski, W. Omar, Ahmed J.S.; Younes, Nagy M.

    2004-09-15

    The aim of this study was to determine the rate of complications in percutaneous nephrostomy (PCN) and nephrolithotomy (PCNL) performed through the 11th and 10th intercostal spaces using our monitoring technique and to discuss the safety of the procedure. Out of 398 PCNs and PCNLs carried out during a 3-year period, 56 patients had 57 such procedures performed using an intercostal approach. The 11th intercostal route was used in 42 and the 10th in 15 cases. One patient had two separate nephrostomies performed through the 10th and 11th intercostal spaces. The technique utilizes bi-planar fluoroscopy with a combination of a conventional angiographic machine to provide anterior-posterior fluoroscopy and a C-arm mobile fluoroscopy machine to give a lateral view, displayed on two separate monitors. None of the patients had clinically significant thoracic or abdominal complications. Two patients had minor chest complications. Only one developed changes (plate atelectasis, elevation of the hemi-diaphragm) directly related to the nephrostomy (2%). The second patient had bilateral plate atelectasis and unilateral congestive lung changes after PCNL. These changes were not necessarily related to the procedure but rather to general anesthesia during nephrolithotomy. The authors consider PCN or PCNL through the intercostal approach a safe procedure with a negligible complication rate, provided that it is performed under bi-planar fluoroscopy, which allows determination of the skin entry point just below the level of pleural reflection and provides three-dimensional monitoring of advancement of the puncturing needle toward the target entry point.

  14. New insights into mechanisms of glomerular injury and repair from the 10th International Podocyte Conference 2014.

    PubMed

    Grahammer, Florian; Benzing, Thomas; Huber, Tobias B

    2015-05-01

    Glomerular kidney diseases are a major cause of end-stage renal disease (ESRD). Thus, it comes as wonderful news that glomerular research is advancing at a remarkable pace. Researchers from around the world met at the 10th International Podocyte Conference in Freiburg, Germany, to discuss the latest developments and findings in this innovative field of kidney research. The meeting highlighted the tremendous progress in our understanding of podocyte-related disorders and promised a rapid transfer of this knowledge into novel treatment options for proteinuric kidney diseases. PMID:25715118

  15. Analysis and test for space shuttle propellant dynamics (1/10th scale model test results). Volume 1: Technical discussion

    NASA Technical Reports Server (NTRS)

    Berry, R. L.; Tegart, J. R.; Demchak, L. J.

    1979-01-01

    Space shuttle propellant dynamics during ET/Orbiter separation in the RTLS (return to launch site) mission abort sequence were investigated in a test program conducted in the NASA KC-135 "Zero G" aircraft using a 1/10th-scale model of the ET LOX Tank. Low-g parabolas were flown from which thirty tests were selected for evaluation. Data on the nature of low-g propellant reorientation in the ET LOX tank, and measurements of the forces exerted on the tank by the moving propellent will provide a basis for correlation with an analytical model of the slosh phenomenon.

  16. Benchmarks for target tracking

    NASA Astrophysics Data System (ADS)

    Dunham, Darin T.; West, Philip D.

    2011-09-01

    The term benchmark originates from the chiseled horizontal marks that surveyors made, into which an angle-iron could be placed to bracket ("bench") a leveling rod, thus ensuring that the leveling rod can be repositioned in exactly the same place in the future. A benchmark in computer terms is the result of running a computer program, or a set of programs, in order to assess the relative performance of an object by running a number of standard tests and trials against it. This paper will discuss the history of simulation benchmarks that are being used by multiple branches of the military and agencies of the US government. These benchmarks range from missile defense applications to chemical biological situations. Typically, a benchmark is used with Monte Carlo runs in order to tease out how algorithms deal with variability and the range of possible inputs. We will also describe problems that can be solved by a benchmark.

  17. MCNP neutron benchmarks

    SciTech Connect

    Hendricks, J.S.; Whalen, D.J.; Cardon, D.A.; Uhle, J.L.

    1991-10-08

    Over 50 neutron benchmark calculations have recently been completed as part of an ongoing program to validate the MCNP Monte Carlo radiation transport code. The new and significant aspects of this work are as follows: These calculations are the first attempt at a validation program for MCNP and the first official benchmarking of version 4 of the code. We believe the chosen set of benchmarks is a comprehensive set that may be useful for benchmarking other radiation transport codes and data libraries. These calculations provide insight into how well neutron transport calculations can be expected to model a wide variety of problems.

  18. Predicting Long-Term College Success through Degree Completion Using ACT[R] Composite Score, ACT Benchmarks, and High School Grade Point Average. ACT Research Report Series, 2012 (5)

    ERIC Educational Resources Information Center

    Radunzel, Justine; Noble, Julie

    2012-01-01

    This study compared the effectiveness of ACT[R] Composite score and high school grade point average (HSGPA) for predicting long-term college success. Outcomes included annual progress towards a degree (based on cumulative credit-bearing hours earned), degree completion, and cumulative grade point average (GPA) at 150% of normal time to degree

  19. BENCHMARK DOSE SOFTWARE (BMDS)

    EPA Science Inventory

    EPA has announced the latest update to the Benchmark Dose Software (BMDS) tool which is used to facilitate the application of benchmark dose (BMD) methods to EPA hazardous pollutant risk assessments. This latest version (1.4.1b) contains seventeen (17) different models that ar...

  20. USGS Yosemite Benchmark Streamgage

    USGS Multimedia Gallery

    In this image, the USGS Hydrologic Benchmark Streamgage on the Merced River in Yosemite Valley can be seen. Hydrologic Benchmark streamgages are those at which no human development exists upstream of the streamgage. This streamgage is a cooperation between the National Park Service, National Oceanic...

  1. Thermal Performance Benchmarking (Presentation)

    SciTech Connect

    Moreno, G.

    2014-11-01

    This project will benchmark the thermal characteristics of automotive power electronics and electric motor thermal management systems. Recent vehicle systems will be benchmarked to establish baseline metrics, evaluate advantages and disadvantages of different thermal management systems, and identify areas of improvement to advance the state-of-the-art.

  2. Verification and validation benchmarks.

    SciTech Connect

    Oberkampf, William Louis; Trucano, Timothy Guy

    2007-02-01

    Verification and validation (V&V) are the primary means to assess the accuracy and reliability of computational simulations. V&V methods and procedures have fundamentally improved the credibility of simulations in several high-consequence fields, such as nuclear reactor safety, underground nuclear waste storage, and nuclear weapon safety. Although the terminology is not uniform across engineering disciplines, code verification deals with assessing the reliability of the software coding, and solution verification deals with assessing the numerical accuracy of the solution to a computational model. Validation addresses the physics modeling accuracy of a computational simulation by comparing the computational results with experimental data. Code verification benchmarks and validation benchmarks have been constructed for a number of years in every field of computational simulation. However, no comprehensive guidelines have been proposed for the construction and use of V&V benchmarks. For example, the field of nuclear reactor safety has not focused on code verification benchmarks, but it has placed great emphasis on developing validation benchmarks. Many of these validation benchmarks are closely related to the operations of actual reactors at near-safety-critical conditions, as opposed to being more fundamental-physics benchmarks. This paper presents recommendations for the effective design and use of code verification benchmarks based on manufactured solutions, classical analytical solutions, and highly accurate numerical solutions. In addition, this paper presents recommendations for the design and use of validation benchmarks, highlighting the careful design of building-block experiments, the estimation of experimental measurement uncertainty for both inputs and outputs to the code, validation metrics, and the role of model calibration in validation. It is argued that the understanding of predictive capability of a computational model is built on the level of achievement in V&V activities, how closely related the V&V benchmarks are to the actual application of interest, and the quantification of uncertainties related to the application of interest.

  3. The Interpretations and Applications of Boethius's Introduction to the Arithmetic II,1 at the End of the 10th Century

    NASA Astrophysics Data System (ADS)

    Otisk, Marek

    This paper deals with comments and glosses to the first chapter of the second book of Boethius's Introduction to Arithmetic from the last quarter of the 10th century. Those texts were written by Gerbert of Aurillac (Scholium ad Boethii Arithmeticam Institutionem l. II, c. 1), Abbo of Fleury (commentary on the Calculus by Victorius of Aquitaine, the so-called De numero, mensura et pondere), Notker of Liège (De superparticularibus) and by the anonymous author (De arithmetica Boetii). The main aim of this paper is to show that Boethius's statements about the converting numerical sequences to equality from this work could be interpreted minimally in two different ways. This paper discussed also the application of this topic in other liberal arts (like astronomy, music, grammar etc.) and in playing game called rithmomachia, the medieval philosophers' game.

  4. A performance geodynamo benchmark

    NASA Astrophysics Data System (ADS)

    Matsui, H.; Heien, E. M.

    2014-12-01

    In the last ten years, a number of numerical dynamo models have successfully represented basic characteristics of the geomagnetic field. However, to approach the parameters regime of the Earth's outer core, we need massively parallel computational environment for extremely large spatial resolutions. Local methods are expected to be more suitable for massively parallel computation because the local methods needs less data communication than the spherical harmonics expansion, but only a few groups have reported results of the dynamo benchmark using local methods (Harder and Hansen, 2005; Matsui and Okuda, 2005; Chan et al., 2007) because of the difficulty treating magnetic boundary conditions based on the local methods. On the other hand, some numerical dynamo models using spherical harmonics expansion has performed successfully with thousands of processes. We perform benchmark tests to asses various numerical methods to asses the next generation of geodynamo simulations. The purpose of the present benchmark test is to assess numerical geodynamo models on a massively parallel computational platform. To compare among many numerical methods as possible, we consider the model with the insulated magnetic boundary by Christensen et al. (2001) and with the pseudo vacuum magnetic boundary, because the pseudo vacuum boundaries are implemented easier by using the local method than the magnetic insulated boundaries. In the present study, we consider two kinds of benchmarks, so-called accuracy benchmark and performance benchmark. In the present study, we will report the results of the performance benchmark. We perform the participated dynamo models under the same computational environment (XSEDE TACC Stampede), and investigate computational performance. To simplify the problem, we choose the same model and parameter regime as the accuracy benchmark test, but perform the simulations with much finer spatial resolutions as possible to investigate computational capability (e.g. parallel scalability and capability, maximum spatial resolutions)under the closer condition to the Earth's outer core. We compare the results of the performance benchmark tests by various codes and discuss characteristics of the simulation methods for geodynamo problems.

  5. Toxicological Benchmarks for Wildlife

    SciTech Connect

    Sample, B.E. Opresko, D.M. Suter, G.W.

    1993-01-01

    Ecological risks of environmental contaminants are evaluated by using a two-tiered process. In the first tier, a screening assessment is performed where concentrations of contaminants in the environment are compared to no observed adverse effects level (NOAEL)-based toxicological benchmarks. These benchmarks represent concentrations of chemicals (i.e., concentrations presumed to be nonhazardous to the biota) in environmental media (water, sediment, soil, food, etc.). While exceedance of these benchmarks does not indicate any particular level or type of risk, concentrations below the benchmarks should not result in significant effects. In practice, when contaminant concentrations in food or water resources are less than these toxicological benchmarks, the contaminants may be excluded from further consideration. However, if the concentration of a contaminant exceeds a benchmark, that contaminant should be retained as a contaminant of potential concern (COPC) and investigated further. The second tier in ecological risk assessment, the baseline ecological risk assessment, may use toxicological benchmarks as part of a weight-of-evidence approach (Suter 1993). Under this approach, based toxicological benchmarks are one of several lines of evidence used to support or refute the presence of ecological effects. Other sources of evidence include media toxicity tests, surveys of biota (abundance and diversity), measures of contaminant body burdens, and biomarkers. This report presents NOAEL- and lowest observed adverse effects level (LOAEL)-based toxicological benchmarks for assessment of effects of 85 chemicals on 9 representative mammalian wildlife species (short-tailed shrew, little brown bat, meadow vole, white-footed mouse, cottontail rabbit, mink, red fox, and whitetail deer) or 11 avian wildlife species (American robin, rough-winged swallow, American woodcock, wild turkey, belted kingfisher, great blue heron, barred owl, barn owl, Cooper's hawk, and red-tailed hawk, osprey) (scientific names for both the mammalian and avian species are presented in Appendix B). [In this document, NOAEL refers to both dose (mg contaminant per kg animal body weight per day) and concentration (mg contaminant per kg of food or L of drinking water)]. The 20 wildlife species were chosen because they are widely distributed and provide a representative range of body sizes and diets. The chemicals are some of those that occur at U.S. Department of Energy (DOE) waste sites. The NOAEL-based benchmarks presented in this report represent values believed to be nonhazardous for the listed wildlife species; LOAEL-based benchmarks represent threshold levels at which adverse effects are likely to become evident. These benchmarks consider contaminant exposure through oral ingestion of contaminated media only. Exposure through inhalation and/or direct dermal exposure are not considered in this report.

  6. KRITZ-2 Experimental Benchmark Analysis

    SciTech Connect

    Pavlovichev, A.M.

    2001-09-28

    The KRITZ-2 experiment has been adopted by the OECD/NEA Task Force on Reactor-Based Plutonium Disposition for use as a benchmark exercise. The KRITZ-2 experiment consists of three different core configurations (one with near-weapons-grade MOX) with critical conditions a 20 C and 245 C. The KRITZ-2 experiment has calculated the MCU-REA code, which is a continuous energy Monte Carlo code system developed at the Russian Research Center--Kurchatov Institute and is used extensively in the Fissile Materials Disposition Program. The calculated results for k{sub eff} and fission rate distributions are compared with the experimental data and results of other codes. The results are in good agreement with the experimental values.

  7. TWODANT benchmark. Progress report

    SciTech Connect

    Lee, Sung

    1994-01-11

    TWODANT (Two-Dimensional, Diffusion-Accelerated, Neutral-Particle Transport) code has been benchmarked against 6 critical experiments (Jezebel plutonium critical assembly) and their k effective values compared with those of KENO and MCNP codes.

  8. Diagnostic Algorithm Benchmarking

    NASA Technical Reports Server (NTRS)

    Poll, Scott

    2011-01-01

    A poster for the NASA Aviation Safety Program Annual Technical Meeting. It describes empirical benchmarking on diagnostic algorithms using data from the ADAPT Electrical Power System testbed and a diagnostic software framework.

  9. Starlink Benchmarking Utility

    NASA Astrophysics Data System (ADS)

    Gledhill, T. M.; Clayton, C. A.

    The Starlink Benchmarking Utility provides a set of tools for investigating the performance of computer systems running astronomy data reduction software. This manual is intended for Starlink Site Managers and describes how to install and use the package.

  10. BENCHMARKING SUSTAINABILITY ENGINEERING EDUCATION

    EPA Science Inventory

    The goals of this project are to develop and apply a methodology for benchmarking curricula in sustainability engineering and to identify individuals active in sustainability engineering education.

  11. Differential Effects on Student Demographic Groups of Using ACT® College Readiness Assessment Composite Score, Act Benchmarks, and High School Grade Point Average for Predicting Long-Term College Success through Degree Completion. ACT Research Report Series, 2013 (5)

    ERIC Educational Resources Information Center

    Radunzel, Justine; Noble, Julie

    2013-01-01

    In this study, we evaluated the differential effects on racial/ethnic, family income, and gender groups of using ACT® College Readiness Assessment Composite score and high school grade point average (HSGPA) for predicting long-term college success. Outcomes included annual progress towards a degree (based on cumulative credit-bearing hours…

  12. XAFS study of copper and silver nanoparticles in glazes of medieval middle-east lustreware (10th-13th century)

    NASA Astrophysics Data System (ADS)

    Padovani, S.; Puzzovio, D.; Sada, C.; Mazzoldi, P.; Borgia, I.; Sgamellotti, A.; Brunetti, B. G.; Cartechini, L.; D'Acapito, F.; Maurizio, C.; Shokoui, F.; Oliaiy, P.; Rahighi, J.; Lamehi-Rachti, M.; Pantos, E.

    2006-06-01

    It has recently been shown that lustre decoration of medieval and Renaissance pottery consists of silver and copper nanoparticles dispersed in the glassy matrix of the ceramic glaze. Here the findings of an X-ray absorption fine structure (XAFS) study on lustred glazes of shards belonging to 10th and 13rd century pottery from the National Museum of Iran are reported. Absorption spectra in the visible range have been also measured in order to investigate the relations between colour and glaze composition. Gold colour is mainly due to Ag nanoparticles, though Ag+, Cu+ and Cu2+ ions can be also dispersed within the glassy matrix, with different ratios. Red colour is mainly due to Cu nanoparticles, although some Ag nanoparticles, Ag+ and Cu+ ions can be present. The achievement of metallic Cu and the absence of Cu2+ indicate a higher reduction of copper in red lustre. These findings are in substantial agreement with previous results on Italian Renaissance pottery. In spite of the large heterogeneity of cases, the presence of copper and silver ions in the glaze confirms that lustre formation is mediated by a copper- and silver-alkali ion exchange, followed by nucleation and growth of metal nanoparticles.

  13. Report on the 10th anniversary of international drug discovery science and technology conference, 8 - 10 november 2012, nanjing, china.

    PubMed

    Everett, Jeremy R

    2013-03-01

    The 10th Anniversary of International Drug Discovery Science and Technology (IDDST) Conference was held in Nanjing, China from 8 to 10 November 2012. The conference ran in parallel with the 2nd Annual Symposium of Drug Delivery Systems. Over 400 delegates from both conferences came together for the Opening Ceremony and Keynote Addresses but otherwise pursued separate paths in the huge facilities of the Nanjing International Expo Centre. The IDDST was arranged into 19 separate Chapters covering drug discovery biology, target validation, chemistry, rational drug design, pharmacology and toxicology, drug screening technology, 'omics' technologies, analytical, automation and enabling technologies, informatics, stem cells and regenerative medicine, bioprocessing, generics, biosimilars and biologicals and seven disease areas: cancer, CNS, respiratory and inflammation, autoimmune, emerging infectious, bone and orphan diseases. There were also two sessions of a 'Bench to Bedside to Business' Program and a Chinese Scientist programme. In each period of the IDDST conference, up to seven sessions were running in parallel. This Meeting Highlight samples just a fraction of the content of this large meeting. The talks included have as a link, the use of new approaches to drug discovery. Many other excellent talks could have been highlighted and the author has necessarily had to be selective. PMID:23339328

  14. An Experiment on Effects of Redundant Audio in Computer Based Instruction on Achievement, Attitude, and Learning Time in 10th Grade Math.

    ERIC Educational Resources Information Center

    Rehaag, Darlene M.; Szabo, Michael

    The effects of the inclusion of matched redundant digital audio on achievement, time spent in learning, and attitude toward computer-based instruction (CBI) delivered mathematics were studied with 82 high school students. Differential effects on students of varying entry learning mathematics performance were also investigated. Subjects were

  15. Nebraska Vocational Agribusiness Curriculum for City Schools. Career Opportunities in Agribusiness. Basic Skill in Agribusiness. A Curriculum Guide. 10th Grade.

    ERIC Educational Resources Information Center

    Nebraska Univ., Lincoln. Dept. of Agricultural Education.

    Designed for use with high school sophomores, this agribusiness curriculum for city schools contains thirty-one units of instruction in the areas of career opportunities in agribusiness and vocational agribusiness skills. Among the units included are (1) Career Selection, (2) Parliamentary Procedure and Public Speaking, (3) Career Opportunities in

  16. A Comparison Study of AVID and GEAR UP 10th-Grade Students in Two High Schools in the Rio Grande Valley of Texas

    ERIC Educational Resources Information Center

    Watt, Karen M.; Huerta, Jeffery; Lozano, Aliber

    2007-01-01

    This study examines 4 groups of high school students enrolled in 2 college preparatory programs, AVID and GEAR UP. Differences in student educational aspirations, expectations and anticipations, knowledge of college entrance requirements, knowledge of financial aid, and academic achievement in mathematics were examined. Adelman's (1999)

  17. The Enlightenment Music Contract. 10th Grade Lesson. Schools of California Online Resources for Education (SCORE): Connecting California's Classrooms to the World.

    ERIC Educational Resources Information Center

    Kelly, Freda

    The "philosophes" of the Enlightenment Period were a group of free (different) thinkers who offered commentary on societal issues. Often, they were like one of today's social commentators suggesting reforms for the political system. Since the United States during the era of the Revolutionary War was seeking reform of what they considered English

  18. PNNL Information Technology Benchmarking

    SciTech Connect

    DD Hostetler

    1999-09-08

    Benchmarking is a methodology for searching out industry best practices that lead to superior performance. It is exchanging information, not just with any organization, but with organizations known to be the best within PNNL, in industry, or in dissimilar industries with equivalent functions. It is used as a continuous improvement tool for business and technical processes, products, and services. Information technology--comprising all computer and electronic communication products and services--underpins the development and/or delivery of many PNNL products and services. This document describes the Pacific Northwest National Laboratory's (PNNL's) approach to information technology (IT) benchmarking. The purpose is to engage other organizations in the collaborative process of benchmarking in order to improve the value of IT services provided to customers. TM document's intended audience consists of other US Department of Energy (DOE) national laboratories and their IT staff. Although the individual participants must define the scope of collaborative benchmarking, an outline of IT service areas for possible benchmarking is described.

  19. Shielding Benchmark Computational Analysis

    SciTech Connect

    Hunter, H.T.; Slater, C.O.; Holland, L.B.; Tracz, G.; Marshall, W.J.; Parsons, J.L.

    2000-09-17

    Over the past several decades, nuclear science has relied on experimental research to verify and validate information about shielding nuclear radiation for a variety of applications. These benchmarks are compared with results from computer code models and are useful for the development of more accurate cross-section libraries, computer code development of radiation transport modeling, and building accurate tests for miniature shielding mockups of new nuclear facilities. When documenting measurements, one must describe many parts of the experimental results to allow a complete computational analysis. Both old and new benchmark experiments, by any definition, must provide a sound basis for modeling more complex geometries required for quality assurance and cost savings in nuclear project development. Benchmarks may involve one or many materials and thicknesses, types of sources, and measurement techniques. In this paper the benchmark experiments of varying complexity are chosen to study the transport properties of some popular materials and thicknesses. These were analyzed using three-dimensional (3-D) models and continuous energy libraries of MCNP4B2, a Monte Carlo code developed at Los Alamos National Laboratory, New Mexico. A shielding benchmark library provided the experimental data and allowed a wide range of choices for source, geometry, and measurement data. The experimental data had often been used in previous analyses by reputable groups such as the Cross Section Evaluation Working Group (CSEWG) and the Organization for Economic Cooperation and Development/Nuclear Energy Agency Nuclear Science Committee (OECD/NEANSC).

  20. The FTIO Benchmark

    NASA Technical Reports Server (NTRS)

    Fagerstrom, Frederick C.; Kuszmaul, Christopher L.; Woo, Alex C. (Technical Monitor)

    1999-01-01

    We introduce a new benchmark for measuring the performance of parallel input/ouput. This benchmark has flexible initialization. size. and scaling properties that allows it to satisfy seven criteria for practical parallel I/O benchmarks. We obtained performance results while running on the a SGI Origin2OOO computer with various numbers of processors: with 4 processors. the performance was 68.9 Mflop/s with 0.52 of the time spent on I/O, with 8 processors the performance was 139.3 Mflop/s with 0.50 of the time spent on I/O, with 16 processors the performance was 173.6 Mflop/s with 0.43 of the time spent on I/O. and with 32 processors the performance was 259.1 Mflop/s with 0.47 of the time spent on I/O.

  1. Translational benchmark risk analysis

    PubMed Central

    Piegorsch, Walter W.

    2010-01-01

    Translational development – in the sense of translating a mature methodology from one area of application to another, evolving area – is discussed for the use of benchmark doses in quantitative risk assessment. Illustrations are presented with traditional applications of the benchmark paradigm in biology and toxicology, and also with risk endpoints that differ from traditional toxicological archetypes. It is seen that the benchmark approach can apply to a diverse spectrum of risk management settings. This suggests a promising future for this important risk-analytic tool. Extensions of the method to a wider variety of applications represent a significant opportunity for enhancing environmental, biomedical, industrial, and socio-economic risk assessments. PMID:20953283

  2. Accelerated randomized benchmarking

    NASA Astrophysics Data System (ADS)

    Granade, Christopher; Ferrie, Christopher; Cory, D. G.

    2015-01-01

    Quantum information processing offers promising advances for a wide range of fields and applications, provided that we can efficiently assess the performance of the control applied in candidate systems. That is, we must be able to determine whether we have implemented a desired gate, and refine accordingly. Randomized benchmarking reduces the difficulty of this task by exploiting symmetries in quantum operations. Here, we bound the resources required for benchmarking and show that, with prior information, we can achieve several orders of magnitude better accuracy than in traditional approaches to benchmarking. Moreover, by building on state-of-the-art classical algorithms, we reach these accuracies with near-optimal resources. Our approach requires an order of magnitude less data to achieve the same accuracies and to provide online estimates of the errors in the reported fidelities. We also show that our approach is useful for physical devices by comparing to simulations.

  3. What Do 2nd and 10th Graders Have in Common? Worms and Technology: Using Technology to Collaborate across Boundaries

    ERIC Educational Resources Information Center

    Culver, Patti; Culbert, Angie; McEntyre, Judy; Clifton, Patrick; Herring, Donna F.; Notar, Charles E.

    2009-01-01

    The article is about the collaboration between two classrooms that enabled a second grade class to participate in a high school biology class. Through the use of modern video conferencing equipment, Mrs. Culbert, with the help of the Dalton State College Educational Technology Training Center (ETTC), set up a live, two way video and audio feed of…

  4. NAS Parallel Benchmarks Results

    NASA Technical Reports Server (NTRS)

    Subhash, Saini; Bailey, David H.; Lasinski, T. A. (Technical Monitor)

    1995-01-01

    The NAS Parallel Benchmarks (NPB) were developed in 1991 at NASA Ames Research Center to study the performance of parallel supercomputers. The eight benchmark problems are specified in a pencil and paper fashion i.e. the complete details of the problem to be solved are given in a technical document, and except for a few restrictions, benchmarkers are free to select the language constructs and implementation techniques best suited for a particular system. In this paper, we present new NPB performance results for the following systems: (a) Parallel-Vector Processors: Cray C90, Cray T'90 and Fujitsu VPP500; (b) Highly Parallel Processors: Cray T3D, IBM SP2 and IBM SP-TN2 (Thin Nodes 2); (c) Symmetric Multiprocessing Processors: Convex Exemplar SPP1000, Cray J90, DEC Alpha Server 8400 5/300, and SGI Power Challenge XL. We also present sustained performance per dollar for Class B LU, SP and BT benchmarks. We also mention NAS future plans of NPB.

  5. Bryce Canyon Benchmark

    USGS Multimedia Gallery

    A USGS elevation benchmark in Bryce Canyon National Park. Bryce Canyon is a unique sandstone formation in southern Utah. It is home to a large number of hoodoos, which are oddly shaped pillars of rock that formed due to different erosion rates for the dolomite that caps them and the sandstone that ...

  6. Benchmarking the World's Best

    ERIC Educational Resources Information Center

    Tucker, Marc S.

    2012-01-01

    A century ago, the United States was a world leader in industrial benchmarking. However, after World War II, once no one could compete with the U.S., it became complacent. Many industrialized countries now have higher student achievement and more equitable and efficient education systems. A higher proportion of young people in their workforces

  7. Benchmarks: WICHE Region 2012

    ERIC Educational Resources Information Center

    Western Interstate Commission for Higher Education, 2013

    2013-01-01

    Benchmarks: WICHE Region 2012 presents information on the West's progress in improving access to, success in, and financing of higher education. The information is updated annually to monitor change over time and encourage its use as a tool for informed discussion in policy and education communities. To establish a general context for the

  8. Comparison of five benchmarks

    SciTech Connect

    Huss, J. E.; Pennline, J. A.

    1987-02-01

    Five benchmark programs were obtained and run on the NASA Lewis CRAY X-MP/24. A comparison was made between the programs codes and between the methods for calculating performance figures. Several multitasking jobs were run to gain experience in how parallel performance is measured.

  9. Monte Carlo Benchmark

    Energy Science and Technology Software Center (ESTSC)

    2010-10-20

    The "Monte Carlo Benchmark" (MCB) is intended to model the computatiional performance of Monte Carlo algorithms on parallel architectures. It models the solution of a simple heuristic transport equation using a Monte Carlo technique. The MCB employs typical features of Monte Carlo algorithms such as particle creation, particle tracking, tallying particle information, and particle destruction. Particles are also traded among processors using MPI calls.

  10. Advances in Classification Research. Volume 10. Proceedings of the ASIS SIG/CR Classification Research Workshop (10th, Washington, DC, November 1-5, 1999). ASIST Monograph Series.

    ERIC Educational Resources Information Center

    Albrechtsen, Hanne, Ed.; Mai, Jens-Erik, Ed.

    This volume is a compilation of the papers presented at the 10th ASIS (American Society for Information Science) workshop on classification research. Major themes include the social and cultural informatics of classification and coding systems, subject access and indexing theory, genre analysis and the agency of documents in the ordering of

  11. Creating Cultures of Peace: Pedagogical Thought and Practice. Selected Papers from the 10th Triennial World Conference (September 10-15, 2001, Madrid, Spain)

    ERIC Educational Resources Information Center

    Benton, Jean E., Ed.; Swami, Piyush, Ed.

    2007-01-01

    The 10th Triennial World Conference of the World Council for Curriculum and Instruction (WCCI) was held September 10-15, 2001 in Madrid, Spain. The theme of the conference was "Cultures of Peace." Thirty-four papers and presentations are divided into nine sections. Part I, Tributes to the Founders of WCCI, includes: (1) Tribute to Alice Miel

  12. Carpenter, Tractors and Microbes for the Development of Logical-Mathematical Thinking--The Way 10th Graders and Pre-Service Teachers Solve Thinking Challenges

    ERIC Educational Resources Information Center

    Gazit, Avikam

    2012-01-01

    The objective of this case study was to investigate the ability of 10th graders and pre-service teachers to solve logical-mathematical thinking challenges. The challenges do not require mathematical knowledge beyond that of primary school but rather an informed use of the problem representation. The percentage of correct answers given by the 10th…

  13. Creating Cultures of Peace: Pedagogical Thought and Practice. Selected Papers from the 10th Triennial World Conference (September 10-15, 2001, Madrid, Spain)

    ERIC Educational Resources Information Center

    Benton, Jean E., Ed.; Swami, Piyush, Ed.

    2007-01-01

    The 10th Triennial World Conference of the World Council for Curriculum and Instruction (WCCI) was held September 10-15, 2001 in Madrid, Spain. The theme of the conference was "Cultures of Peace." Thirty-four papers and presentations are divided into nine sections. Part I, Tributes to the Founders of WCCI, includes: (1) Tribute to Alice Miel…

  14. Eighth Grade Level Science Sample Curriculum.

    ERIC Educational Resources Information Center

    Arkansas State Dept. of Education, Little Rock.

    This document presents a sample of the Arkansas science curriculum and identifies the content standards for physical science systems, life science systems, and Earth science/space science systems for eighth grade students. Each content standard is explained and includes student learning expectations, eighth grade benchmarks, assessments, and

  15. Fifth Grade Level Science Sample Curriculum.

    ERIC Educational Resources Information Center

    Arkansas State Dept. of Education, Little Rock.

    This document presents a sample of the Arkansas science curriculum and identifies the content standards for physical science systems, life science systems, and Earth science/space science systems for fifth grade students. Each content standard is explained and includes student learning expectations, fifth grade benchmarks, assessments, and

  16. First Grade Level Science Sample Curriculum.

    ERIC Educational Resources Information Center

    Arkansas State Dept. of Education, Little Rock.

    This document presents a sample of the Arkansas science curriculum and identifies the content standards for physical science systems, life science systems, and Earth science/space science systems for first grade students. Each content standard is explained and includes student learning expectations, first grade benchmarks, assessments, and

  17. Fourth Grade Level Science Sample Curriculum.

    ERIC Educational Resources Information Center

    Arkansas State Dept. of Education, Little Rock.

    This document presents a sample of the Arkansas science curriculum and identifies the content standards for physical science systems, life science systems, and Earth science/space science systems for fourth grade students. Each content standard is explained and includes student learning expectations, fourth grade benchmarks, assessments, and

  18. Seventh Grade Level Science Sample Curriculum.

    ERIC Educational Resources Information Center

    Arkansas State Dept. of Education, Little Rock.

    This document presents a sample of the Arkansas science curriculum and identifies the content standards for physical science systems, life science systems, and Earth science/space science systems for seventh grade students. Each content standard is explained and includes student learning expectations, seventh grade benchmarks, assessments, and

  19. Second Grade Level Science Sample Curriculum.

    ERIC Educational Resources Information Center

    Arkansas State Dept. of Education, Little Rock.

    This document presents a sample of the Arkansas science curriculum and identifies the content standards for physical science systems, life science systems, and Earth science/space science systems for second grade students. Each content standard is explained and includes student learning expectations, second grade benchmarks, assessments, and

  20. Third Grade Level Science Sample Curriculum.

    ERIC Educational Resources Information Center

    Arkansas State Dept. of Education, Little Rock.

    This document presents a sample of the Arkansas science curriculum and identifies the content standards for physical science systems, life science systems, and Earth science/space science systems for third grade students. Each content standard is explained and includes student learning expectations, third grade benchmarks, assessments, and

  1. Sixth Grade Level Science Sample Curriculum.

    ERIC Educational Resources Information Center

    Arkansas State Dept. of Education, Little Rock.

    This document presents a sample of the Arkansas science curriculum and identifies the content standards for physical science systems, life science systems, and Earth science/space science systems for sixth grade students. Each content standard is explained and includes student learning expectations, sixth grade benchmarks, assessments, and

  2. 21st Century Curriculum: Does Auto-Grading Writing Actually Work?

    ERIC Educational Resources Information Center

    T.H.E. Journal, 2013

    2013-01-01

    The West Virginia Department of Education's auto grading initiative dates back to 2004--a time when school districts were making their first forays into automation. The Charleston based WVDE had instituted a statewide writing assessment in 1984 for students in fourth, seventh, and 10th grades and was looking to expand that program without

  3. 21st Century Curriculum: Does Auto-Grading Writing Actually Work?

    ERIC Educational Resources Information Center

    T.H.E. Journal, 2013

    2013-01-01

    The West Virginia Department of Education's auto grading initiative dates back to 2004--a time when school districts were making their first forays into automation. The Charleston based WVDE had instituted a statewide writing assessment in 1984 for students in fourth, seventh, and 10th grades and was looking to expand that program without…

  4. Social Studies Performance Standards. North Dakota Standards and Benchmarks.

    ERIC Educational Resources Information Center

    North Dakota State Dept. of Public Instruction, Bismarck.

    This document is a companion to the North Dakota Social Studies Standards (December 2000), and provides descriptions of levels of student performance and achievement in relation to the state standards and benchmarks. It is designed for use by all educators in grades K-12. Components of the document include: content standards (statement that

  5. Radiography benchmark 2014

    SciTech Connect

    Jaenisch, G.-R. Deresch, A. Bellon, C.; Schumm, A.; Lucet-Sanchez, F.; Guerin, P.

    2015-03-31

    The purpose of the 2014 WFNDEC RT benchmark study was to compare predictions of various models of radiographic techniques, in particular those that predict the contribution of scattered radiation. All calculations were carried out for homogenous materials and a mono-energetic X-ray point source in the energy range between 100 keV and 10 MeV. The calculations were to include the best physics approach available considering electron binding effects. Secondary effects like X-ray fluorescence and bremsstrahlung production were to be taken into account if possible. The problem to be considered had two parts. Part I examined the spectrum and the spatial distribution of radiation behind a single iron plate. Part II considered two equally sized plates, made of iron and aluminum respectively, only evaluating the spatial distribution. Here we present the results of above benchmark study, comparing them to MCNP as the assumed reference model. The possible origins of the observed deviations are discussed.

  6. Radiography benchmark 2014

    NASA Astrophysics Data System (ADS)

    Jaenisch, G.-R.; Deresch, A.; Bellon, C.; Schumm, A.; Lucet-Sanchez, F.; Guerin, P.

    2015-03-01

    The purpose of the 2014 WFNDEC RT benchmark study was to compare predictions of various models of radiographic techniques, in particular those that predict the contribution of scattered radiation. All calculations were carried out for homogenous materials and a mono-energetic X-ray point source in the energy range between 100 keV and 10 MeV. The calculations were to include the best physics approach available considering electron binding effects. Secondary effects like X-ray fluorescence and bremsstrahlung production were to be taken into account if possible. The problem to be considered had two parts. Part I examined the spectrum and the spatial distribution of radiation behind a single iron plate. Part II considered two equally sized plates, made of iron and aluminum respectively, only evaluating the spatial distribution. Here we present the results of above benchmark study, comparing them to MCNP as the assumed reference model. The possible origins of the observed deviations are discussed.

  7. Benchmark Airport Charges

    NASA Technical Reports Server (NTRS)

    deWit, A.; Cohn, N.

    1999-01-01

    The Netherlands Directorate General of Civil Aviation (DGCA) commissioned Hague Consulting Group (HCG) to complete a benchmark study of airport charges at twenty eight airports in Europe and around the world, based on 1996 charges. This study followed previous DGCA research on the topic but included more airports in much more detail. The main purpose of this new benchmark study was to provide insight into the levels and types of airport charges worldwide and into recent changes in airport charge policy and structure, This paper describes the 1996 analysis. It is intended that this work be repeated every year in order to follow developing trends and provide the most up-to-date information possible.

  8. MPI Multicore Linktest Benchmark

    Energy Science and Technology Software Center (ESTSC)

    2008-01-25

    The MPI Multicore Linktest (LinkTest) measures the aggregate bandwidth from/to a multicore node in a parallel system. It allows the user to specify a variety of different node layout and communication routine variations and reports the maximal observed bandwidth across all specified options. In particular, this benchmark is able to vary the number of tasks on the root node and thereby allows users to study the impact of multicore architectures on MPI communication performance.

  9. The NAS Parallel Benchmarks

    SciTech Connect

    Bailey, David H.

    2009-11-15

    The NAS Parallel Benchmarks (NPB) are a suite of parallel computer performance benchmarks. They were originally developed at the NASA Ames Research Center in 1991 to assess high-end parallel supercomputers. Although they are no longer used as widely as they once were for comparing high-end system performance, they continue to be studied and analyzed a great deal in the high-performance computing community. The acronym 'NAS' originally stood for the Numerical Aeronautical Simulation Program at NASA Ames. The name of this organization was subsequently changed to the Numerical Aerospace Simulation Program, and more recently to the NASA Advanced Supercomputing Center, although the acronym remains 'NAS.' The developers of the original NPB suite were David H. Bailey, Eric Barszcz, John Barton, David Browning, Russell Carter, LeoDagum, Rod Fatoohi, Samuel Fineberg, Paul Frederickson, Thomas Lasinski, Rob Schreiber, Horst Simon, V. Venkatakrishnan and Sisira Weeratunga. The original NAS Parallel Benchmarks consisted of eight individual benchmark problems, each of which focused on some aspect of scientific computing. The principal focus was in computational aerophysics, although most of these benchmarks have much broader relevance, since in a much larger sense they are typical of many real-world scientific computing applications. The NPB suite grew out of the need for a more rational procedure to select new supercomputers for acquisition by NASA. The emergence of commercially available highly parallel computer systems in the late 1980s offered an attractive alternative to parallel vector supercomputers that had been the mainstay of high-end scientific computing. However, the introduction of highly parallel systems was accompanied by a regrettable level of hype, not only on the part of the commercial vendors but even, in some cases, by scientists using the systems. As a result, it was difficult to discern whether the new systems offered any fundamental performance advantage over vector supercomputers, and, if so, which of the parallel offerings would be most useful in real-world scientific computation. In part to draw attention to some of the performance reporting abuses prevalent at the time, the present author wrote a humorous essay 'Twelve Ways to Fool the Masses,' which described in a light-hearted way a number of the questionable ways in which both vendor marketing people and scientists were inflating and distorting their performance results. All of this underscored the need for an objective and scientifically defensible measure to compare performance on these systems.

  10. Self-benchmarking Guide for Data Centers: Metrics, Benchmarks, Actions

    SciTech Connect

    Mathew, Paul; Ganguly, Srirupa; Greenberg, Steve; Sartor, Dale

    2009-07-13

    This guide describes energy efficiency metrics and benchmarks that can be used to track the performance of and identify potential opportunities to reduce energy use in data centers. This guide is primarily intended for personnel who have responsibility for managing energy use in existing data centers - including facilities managers, energy managers, and their engineering consultants. Additionally, data center designers may also use the metrics and benchmarks described in this guide for goal-setting in new construction or major renovation. This guide provides the following information: (1) A step-by-step outline of the benchmarking process. (2) A set of performance metrics for the whole building as well as individual systems. For each metric, the guide provides a definition, performance benchmarks, and potential actions that can be inferred from evaluating this metric. (3) A list and descriptions of the data required for computing the metrics. This guide is complemented by spreadsheet templates for data collection and for computing the benchmarking metrics. This guide builds on prior data center benchmarking studies supported by the California Energy Commission. Much of the benchmarking data are drawn from the LBNL data center benchmarking database that was developed from these studies. Additional benchmark data were obtained from engineering experts including facility designers and energy managers. This guide also builds on recent research supported by the U.S. Department of Energy's Save Energy Now program.

  11. Algebraic Multigrid Benchmark

    Energy Science and Technology Software Center (ESTSC)

    2013-05-06

    AMG2013 is a parallel algebraic multigrid solver for linear systems arising from problems on unstructured grids. It has been derived directly from the Boomer AMG solver in the hypre library, a large linear solvers library that is being developed in the Center for Applied Scientific Computing (CASC) at LLNL. The driver provided in the benchmark can build various test problems. The default problem is a Laplace type problem on an unstructured domain with various jumpsmore » and an anisotropy in one part.« less

  12. Benchmarking Query Execution Robustness

    NASA Astrophysics Data System (ADS)

    Wiener, Janet L.; Kuno, Harumi; Graefe, Goetz

    Benchmarks that focus on running queries on a well-tuned database system ignore a long-standing problem: adverse runtime conditions can cause database system performance to vary widely and unexpectedly. When the query execution engine does not exhibit resilience to these adverse conditions, addressing the resultant performance problems can contribute significantly to the total cost of ownership for a database system in over-provisioning, lost efficiency, and increased human administrative costs. For example, focused human effort may be needed to manually invoke workload management actions or fine-tune the optimization of specific queries.

  13. Benchmarking multimedia performance

    NASA Astrophysics Data System (ADS)

    Zandi, Ahmad; Sudharsanan, Subramania I.

    1998-03-01

    With the introduction of faster processors and special instruction sets tailored to multimedia, a number of exciting applications are now feasible on the desktops. Among these is the DVD playback consisting, among other things, of MPEG-2 video and Dolby digital audio or MPEG-2 audio. Other multimedia applications such as video conferencing and speech recognition are also becoming popular on computer systems. In view of this tremendous interest in multimedia, a group of major computer companies have formed, Multimedia Benchmarks Committee as part of Standard Performance Evaluation Corp. to address the performance issues of multimedia applications. The approach is multi-tiered with three tiers of fidelity from minimal to full compliant. In each case the fidelity of the bitstream reconstruction as well as quality of the video or audio output are measured and the system is classified accordingly. At the next step the performance of the system is measured. In many multimedia applications such as the DVD playback the application needs to be run at a specific rate. In this case the measurement of the excess processing power, makes all the difference. All these make a system level, application based, multimedia benchmark very challenging. Several ideas and methodologies for each aspect of the problems will be presented and analyzed.

  14. Core Benchmarks Descriptions

    SciTech Connect

    Pavlovichev, A.M.

    2001-05-24

    Actual regulations while designing of new fuel cycles for nuclear power installations comprise a calculational justification to be performed by certified computer codes. It guarantees that obtained calculational results will be within the limits of declared uncertainties that are indicated in a certificate issued by Gosatomnadzor of Russian Federation (GAN) and concerning a corresponding computer code. A formal justification of declared uncertainties is the comparison of calculational results obtained by a commercial code with the results of experiments or of calculational tests that are calculated with an uncertainty defined by certified precision codes of MCU type or of other one. The actual level of international cooperation provides an enlarging of the bank of experimental and calculational benchmarks acceptable for a certification of commercial codes that are being used for a design of fuel loadings with MOX fuel. In particular, the work is practically finished on the forming of calculational benchmarks list for a certification of code TVS-M as applied to MOX fuel assembly calculations. The results on these activities are presented.

  15. Self-benchmarking Guide for Laboratory Buildings: Metrics, Benchmarks, Actions

    SciTech Connect

    Mathew, Paul; Greenberg, Steve; Sartor, Dale

    2009-07-13

    This guide describes energy efficiency metrics and benchmarks that can be used to track the performance of and identify potential opportunities to reduce energy use in laboratory buildings. This guide is primarily intended for personnel who have responsibility for managing energy use in existing laboratory facilities - including facilities managers, energy managers, and their engineering consultants. Additionally, laboratory planners and designers may also use the metrics and benchmarks described in this guide for goal-setting in new construction or major renovation. This guide provides the following information: (1) A step-by-step outline of the benchmarking process. (2) A set of performance metrics for the whole building as well as individual systems. For each metric, the guide provides a definition, performance benchmarks, and potential actions that can be inferred from evaluating this metric. (3) A list and descriptions of the data required for computing the metrics. This guide is complemented by spreadsheet templates for data collection and for computing the benchmarking metrics. This guide builds on prior research supported by the national Laboratories for the 21st Century (Labs21) program, supported by the U.S. Department of Energy and the U.S. Environmental Protection Agency. Much of the benchmarking data are drawn from the Labs21 benchmarking database and technical guides. Additional benchmark data were obtained from engineering experts including laboratory designers and energy managers.

  16. Self-benchmarking Guide for Cleanrooms: Metrics, Benchmarks, Actions

    SciTech Connect

    Mathew, Paul; Sartor, Dale; Tschudi, William

    2009-07-13

    This guide describes energy efficiency metrics and benchmarks that can be used to track the performance of and identify potential opportunities to reduce energy use in laboratory buildings. This guide is primarily intended for personnel who have responsibility for managing energy use in existing laboratory facilities - including facilities managers, energy managers, and their engineering consultants. Additionally, laboratory planners and designers may also use the metrics and benchmarks described in this guide for goal-setting in new construction or major renovation. This guide provides the following information: (1) A step-by-step outline of the benchmarking process. (2) A set of performance metrics for the whole building as well as individual systems. For each metric, the guide provides a definition, performance benchmarks, and potential actions that can be inferred from evaluating this metric. (3) A list and descriptions of the data required for computing the metrics. This guide is complemented by spreadsheet templates for data collection and for computing the benchmarking metrics. This guide builds on prior research supported by the national Laboratories for the 21st Century (Labs21) program, supported by the U.S. Department of Energy and the U.S. Environmental Protection Agency. Much of the benchmarking data are drawn from the Labs21 benchmarking database and technical guides. Additional benchmark data were obtained from engineering experts including laboratory designers and energy managers.

  17. A Benchmarking Model. Benchmarking Quality Performance in Vocational Technical Education.

    ERIC Educational Resources Information Center

    Losh, Charles

    The Skills Standards Projects have provided further emphasis on the need for benchmarking U.S. vocational-technical education (VTE) against international competition. Benchmarking is an ongoing systematic process designed to identify, as quantitatively as possible, those practices that produce world class performance. Metrics are those things that

  18. Piping benchmark problems

    SciTech Connect

    Bezler, P.; Hartzman, M.

    1981-01-01

    A set of benchmark problems and solutions have been developed for verifying the adequacy of computer programs used for dynamic analysis and design of nuclear piping systems by the Response Spectrum Method. The problems range from simple to complex configurations which are assumed to experience linear elastic behavior. The dynamic loading is represented by uniform support motion, assumed to be induced by seismic excitation in three spatial directions. The solutions consist of frequencies, participation factors, nodal displacement components and internal force and moment components. Solutions to associated anchor point motion static problems are not included. A summary description of each problem including a description of the input parameters, pertinent output results and the reasons for problem selection are included. The problem set is presently being used by the Mechanical Branch of the Nuclear Regulatory Commission to confirm the adequacy of the computer programs used by applicants for nuclear power plant licenses. 9 refs., 14 figs., 2 tabs.

  19. Cleanroom energy benchmarking results

    SciTech Connect

    Tschudi, William; Xu, Tengfang

    2001-09-01

    A utility market transformation project studied energy use and identified energy efficiency opportunities in cleanroom HVAC design and operation for fourteen cleanrooms. This paper presents the results of this work and relevant observations. Cleanroom owners and operators know that cleanrooms are energy intensive but have little information to compare their cleanroom's performance over time, or to others. Direct comparison of energy performance by traditional means, such as watts/ft{sup 2}, is not a good indicator with the wide range of industrial processes and cleanliness levels occurring in cleanrooms. In this project, metrics allow direct comparison of the efficiency of HVAC systems and components. Energy and flow measurements were taken to determine actual HVAC system energy efficiency. The results confirm a wide variation in operating efficiency and they identify other non-energy operating problems. Improvement opportunities were identified at each of the benchmarked facilities. Analysis of the best performing systems and components is summarized, as are areas for additional investigation.

  20. Benchmarking foreign electronics technologies

    SciTech Connect

    Bostian, C.W.; Hodges, D.A.; Leachman, R.C.; Sheridan, T.B.; Tsang, W.T.; White, R.M.

    1994-12-01

    This report has been drafted in response to a request from the Japanese Technology Evaluation Center`s (JTEC) Panel on Benchmarking Select Technologies. Since April 1991, the Competitive Semiconductor Manufacturing (CSM) Program at the University of California at Berkeley has been engaged in a detailed study of quality, productivity, and competitiveness in semiconductor manufacturing worldwide. The program is a joint activity of the College of Engineering, the Haas School of Business, and the Berkeley Roundtable on the International Economy, under sponsorship of the Alfred P. Sloan Foundation, and with the cooperation of semiconductor producers from Asia, Europe and the United States. Professors David A. Hodges and Robert C. Leachman are the project`s Co-Directors. The present report for JTEC is primarily based on data and analysis drawn from that continuing program. The CSM program is being conducted by faculty, graduate students and research staff from UC Berkeley`s Schools of Engineering and Business, and Department of Economics. Many of the participating firms are represented on the program`s Industry Advisory Board. The Board played an important role in defining the research agenda. A pilot study was conducted in 1991 with the cooperation of three semiconductor plants. The research plan and survey documents were thereby refined. The main phase of the CSM benchmarking study began in mid-1992 and will continue at least through 1997. reports are presented on the manufacture of integrated circuits; data storage; wireless technology; human-machine interfaces; and optoelectronics. Selected papers are indexed separately for inclusion in the Energy Science and Technology Database.

  1. Former Yosemite Hydrologic Benchmark Streamgage

    USGS Multimedia Gallery

    In this image, the former USGS Hydrologic Benchmark Streamgage on the Merced River in Yosemite Valley can be seen. Hydrologic Benchmark streamgages are those at which no human development exists upstream of the streamgage. The streamgage station has been in existence since 1915, and was replaced by ...

  2. Internal Benchmarking for Institutional Effectiveness

    ERIC Educational Resources Information Center

    Ronco, Sharron L.

    2012-01-01

    Internal benchmarking is an established practice in business and industry for identifying best in-house practices and disseminating the knowledge about those practices to other groups in the organization. Internal benchmarking can be done with structures, processes, outcomes, or even individuals. In colleges or universities with multicampuses or a

  3. Applications of Integral Benchmark Data

    SciTech Connect

    Giuseppe Palmiotti; Teruhiko Kugo; Fitz Trumble; Albert C. Kahler; Dale Lancaster

    2014-10-09

    The International Reactor Physics Experiment Evaluation Project (IRPhEP) and the International Criticality Safety Benchmark Evaluation Project (ICSBEP) provide evaluated integral benchmark data that may be used for validation of reactor physics / nuclear criticality safety analytical methods and data, nuclear data testing, advanced modeling and simulation, and safety analysis licensing activities. The handbooks produced by these programs are used in over 30 countries. Five example applications are presented in this paper: (1) Use of IRPhEP Data in Uncertainty Analyses and Cross Section Adjustment, (2) Uncertainty Evaluation Methods for Reactor Core Design at JAEA Using Reactor Physics Experimental Data, (3) Application of Benchmarking Data to a Broad Range of Criticality Safety Problems, (4) Cross Section Data Testing with ICSBEP Benchmarks, and (5) Use of the International Handbook of Evaluated Reactor Physics Benchmark Experiments to Support the Power Industry.

  4. FireHose Streaming Benchmarks

    SciTech Connect

    2015-01-27

    The FireHose Streaming Benchmarks are a suite of stream-processing benchmarks defined to enable comparison of streaming software and hardware, both quantitatively vis-a-vis the rate at which they can process data, and qualitatively by judging the effort involved to implement and run the benchmarks. Each benchmark has two parts. The first is a generator which produces and outputs datums at a high rate in a specific format. The second is an analytic which reads the stream of datums and is required to perform a well-defined calculation on the collection of datums, typically to find anomalous datums that have been created in the stream by the generator. The FireHose suite provides code for the generators, sample code for the analytics (which users are free to re-implement in their own custom frameworks), and a precise definition of each benchmark calculation.

  5. FireHose Streaming Benchmarks

    Energy Science and Technology Software Center (ESTSC)

    2015-01-27

    The FireHose Streaming Benchmarks are a suite of stream-processing benchmarks defined to enable comparison of streaming software and hardware, both quantitatively vis-a-vis the rate at which they can process data, and qualitatively by judging the effort involved to implement and run the benchmarks. Each benchmark has two parts. The first is a generator which produces and outputs datums at a high rate in a specific format. The second is an analytic which reads the streammore » of datums and is required to perform a well-defined calculation on the collection of datums, typically to find anomalous datums that have been created in the stream by the generator. The FireHose suite provides code for the generators, sample code for the analytics (which users are free to re-implement in their own custom frameworks), and a precise definition of each benchmark calculation.« less

  6. Change and Continuity in Student Achievement from Grades 3 to 5: A Policy Dilemma

    ERIC Educational Resources Information Center

    McCaslin, Mary; Burross, Heidi Legg; Good, Thomas L.

    2005-01-01

    In this article we examine student performance on mandated tests in grades 3, 4, and 5 in one state. We focus on this interval, which w e term "the fourth grade window," based on our hypothesis that students in grade four are particularly vulnerable to decrements in achievement. The national focus on the third grade as the critical benchmark in

  7. [Congresses of the Croatian Medical Association regarding unpublished proceedings of the 10th congress in Zadar on September 25-28, 1996].

    PubMed

    Drazanci?, Ante

    2011-01-01

    The first annual meeting of Croatian physicians, with characteristics of a congress, was held in 1899 at the 25th anniversary of the Croatian Medical Association. From 1954 to 1996, during almost 60 years of existence of the Croatian Medical Association, ten congresses of the Association were held. The congresses were during the development of modern medicine devoted to different medical questions, including some problems of national pathology, of the structure and restructuring of health care. The work and the content of congresses were published in the proceedings except for the 8th Congress in 1987 and the 10th in 1996. By reading main lectures, invited lectures and free papers the knowledge of that period can be gained. Many papers are even today actual, even today it could be learned from them. With more details, using published proceedings the 9th congress and the 10th congress are described on the basis preserved program, of a brief report in home journal and ample preserved correspondence. The national medical congres dedicated to technology advancement and to numerous problems of national pathology may be actual even today. They could help to solve many problems of health care, contribute to its improvement and convey consensus on its further development. PMID:22329299

  8. Benchmarking in the sunshine state

    SciTech Connect

    Varella, F.

    1996-09-01

    Comparably sized municipal electric utilities in Florida have assembled a group to benchmark key performance areas. There are many benchmarking or performance measurement groups striving to identify best business practices in an effort to keep companies flourishing in these tough times. As the work forces shrink and workloads increase, these tools are critical for managerial analysis. Over the pat couple of years, Fort Pierce Utilities Authority in Florida has been involved in the initial benchmarking process of the American Public Power Association, the results of which are documented in the June 1994 publication entitled Benchmarking for Electric Utilities: Report of APP`s Benchmarking Task Force. This report provided the foundation to establish a benchmarking group in Florida that would be relevant to and characteristic of the public power industry in the state. Using the expertise of a consultant, R.W. Beck of Orlando, and the Florida Municipal Power Agency, the group pursued the idea of developing a database from Florida municipal electric utilities of similar size (annual sales over 100 gWh hours, but less than 1,000 gWh) for the purpose of establishing performance benchmarks. The results of this study are discussed.

  9. Correlational effect size benchmarks.

    PubMed

    Bosco, Frank A; Aguinis, Herman; Singh, Kulraj; Field, James G; Pierce, Charles A

    2015-03-01

    Effect size information is essential for the scientific enterprise and plays an increasingly central role in the scientific process. We extracted 147,328 correlations and developed a hierarchical taxonomy of variables reported in Journal of Applied Psychology and Personnel Psychology from 1980 to 2010 to produce empirical effect size benchmarks at the omnibus level, for 20 common research domains, and for an even finer grained level of generality. Results indicate that the usual interpretation and classification of effect sizes as small, medium, and large bear almost no resemblance to findings in the field, because distributions of effect sizes exhibit tertile partitions at values approximately one-half to one-third those intuited by Cohen (1988). Our results offer information that can be used for research planning and design purposes, such as producing better informed non-nil hypotheses and estimating statistical power and planning sample size accordingly. We also offer information useful for understanding the relative importance of the effect sizes found in a particular study in relationship to others and which research domains have advanced more or less, given that larger effect sizes indicate a better understanding of a phenomenon. Also, our study offers information about research domains for which the investigation of moderating effects may be more fruitful and provide information that is likely to facilitate the implementation of Bayesian analysis. Finally, our study offers information that practitioners can use to evaluate the relative effectiveness of various types of interventions. PMID:25314367

  10. Virtual machine performance benchmarking.

    PubMed

    Langer, Steve G; French, Todd

    2011-10-01

    The attractions of virtual computing are many: reduced costs, reduced resources and simplified maintenance. Any one of these would be compelling for a medical imaging professional attempting to support a complex practice on limited resources in an era of ever tightened reimbursement. In particular, the ability to run multiple operating systems optimized for different tasks (computational image processing on Linux versus office tasks on Microsoft operating systems) on a single physical machine is compelling. However, there are also potential drawbacks. High performance requirements need to be carefully considered if they are to be executed in an environment where the running software has to execute through multiple layers of device drivers before reaching the real disk or network interface. Our lab has attempted to gain insight into the impact of virtualization on performance by benchmarking the following metrics on both physical and virtual platforms: local memory and disk bandwidth, network bandwidth, and integer and floating point performance. The virtual performance metrics are compared to baseline performance on "bare metal." The results are complex, and indeed somewhat surprising. PMID:21207096

  11. Benchmarking massively parallel architectures

    SciTech Connect

    Lubeck, O.; Moore, J.; Simmons, M.; Wasserman, H.

    1993-01-01

    The purpose of this paper is to summarize some initial experiences related to measuring the performance of massively parallel processors (MPPs) at Los Alamos National Laboratory (LANL). Actually, the range of MPP architectures the authors have used is rather limited, being confined mostly to the Thinking Machines Corporation (TMC) Connection Machine CM-2 and CM-5. Some very preliminary work has been carried out on the Kendall Square KSR-1, and efforts related to other machines, such as the Intel Paragon and the soon-to-be-released CRAY T3D are planned. This paper will concentrate more on methodology rather than discuss specific architectural strengths and weaknesses; the latter is expected to be the subject of future reports. MPP benchmarking is a field in critical need of structure and definition. As the authors have stated previously, such machines have enormous potential, and there is certainly a dire need for orders of magnitude computational power over current supercomputers. However, performance reports for MPPs must emphasize actual sustainable performance from real applications in a careful, responsible manner. Such has not always been the case. A recent paper has described in some detail, the problem of potentially misleading performance reporting in the parallel scientific computing field. Thus, in this paper, the authors briefly offer a few general ideas on MPP performance analysis.

  12. Issues in Benchmark Metric Selection

    NASA Astrophysics Data System (ADS)

    Crolotte, Alain

    It is true that a metric can influence a benchmark but will esoteric metrics create more problems than they will solve? We answer this question affirmatively by examining the case of the TPC-D metric which used the much debated geometric mean for the single-stream test. We will show how a simple choice influenced the benchmark and its conduct and, to some extent, DBMS development. After examining other alternatives our conclusion is that the real measure for a decision-support benchmark is the arithmetic mean.

  13. Phase-covariant quantum benchmarks

    SciTech Connect

    Calsamiglia, J.; Aspachs, M.; Munoz-Tapia, R.; Bagan, E.

    2009-05-15

    We give a quantum benchmark for teleportation and quantum storage experiments suited for pure and mixed test states. The benchmark is based on the average fidelity over a family of phase-covariant states and certifies that an experiment cannot be emulated by a classical setup, i.e., by a measure-and-prepare scheme. We give an analytical solution for qubits, which shows important differences with standard state estimation approach, and compute the value of the benchmark for coherent and squeezed states, both pure and mixed.

  14. California commercial building energy benchmarking

    SciTech Connect

    Kinney, Satkartar; Piette, Mary Ann

    2003-07-01

    Building energy benchmarking is the comparison of whole-building energy use relative to a set of similar buildings. It provides a useful starting point for individual energy audits and for targeting buildings for energy-saving measures in multiple-site audits. Benchmarking is of interest and practical use to a number of groups. Energy service companies and performance contractors communicate energy savings potential with ''typical'' and ''best-practice'' benchmarks while control companies and utilities can provide direct tracking of energy use and combine data from multiple buildings. Benchmarking is also useful in the design stage of a new building or retrofit to determine if a design is relatively efficient. Energy managers and building owners have an ongoing interest in comparing energy performance to others. Large corporations, schools, and government agencies with numerous facilities also use benchmarking methods to compare their buildings to each other. The primary goal of Task 2.1.1 Web-based Benchmarking was the development of a web-based benchmarking tool, dubbed Cal-Arch, for benchmarking energy use in California commercial buildings. While there were several other benchmarking tools available to California consumers prior to the development of Cal-Arch, there were none that were based solely on California data. Most available benchmarking information, including the Energy Star performance rating, were developed using DOE's Commercial Building Energy Consumption Survey (CBECS), which does not provide state-level data. Each database and tool has advantages as well as limitations, such as the number of buildings and the coverage by type, climate regions and end uses. There is considerable commercial interest in benchmarking because it provides an inexpensive method of screening buildings for tune-ups and retrofits. However, private companies who collect and manage consumption data are concerned that the identities of building owners might be revealed and hence are reluctant to share their data. The California Commercial End Use Survey (CEUS), the primary source of data for Cal-Arch, is a unique source of information on commercial buildings in California. It has not been made public; however, it was made available by CEC to LBNL for the purpose of developing a public benchmarking tool.

  15. Benchmark script and evaluation criteria

    SciTech Connect

    Not Available

    1980-10-31

    The Department of Energy's Office of Energy Emergency Management Information System requires time-sharing and remote-job-entry services for their state Energy Emergency Management Information System (EEMIS-S). The Office is soliciting time-sharing services through use of GSA's Teleprocessing Service Program - Multiple Award Schedule Contracts (TSP-MASC) process. The functional mandatory and desirable requirements that candidate vendors should provide are described in the Functional Requirements Document (DOE/EIA/11581-T1). This document, DOE/EIA/11581-T2, describes the Benchmark Scenario and Vendor Evaluation Criteria. Section 1.0 describes the Benchmark, the Benchmark Evaluation, and the Workload Definition. The Vendor Evaluation Criteria and mandatory and desirable hardware and service checklists are described in Section 2.0. To provide compliance with these requirements, a benchmark test will be performed using the vendor's system.

  16. Benchmarking hypercube hardware and software

    NASA Technical Reports Server (NTRS)

    Grunwald, Dirk C.; Reed, Daniel A.

    1986-01-01

    It was long a truism in computer systems design that balanced systems achieve the best performance. Message passing parallel processors are no different. To quantify the balance of a hypercube design, an experimental methodology was developed and the associated suite of benchmarks was applied to several existing hypercubes. The benchmark suite includes tests of both processor speed in the absence of internode communication and message transmission speed as a function of communication patterns.

  17. Data-Intensive Benchmarking Suite

    Energy Science and Technology Software Center (ESTSC)

    2008-11-26

    The Data-Intensive Benchmark Suite is a set of programs written for the study of data-or storage-intensive science and engineering problems, The benchmark sets cover: general graph searching (basic and Hadoop Map/Reduce breadth-first search), genome sequence searching, HTTP request classification (basic and Hadoop Map/Reduce), low-level data communication, and storage device micro-beachmarking

  18. Research on computer systems benchmarking

    NASA Technical Reports Server (NTRS)

    Smith, Alan Jay (Principal Investigator)

    1996-01-01

    This grant addresses the topic of research on computer systems benchmarking and is more generally concerned with performance issues in computer systems. This report reviews work in those areas during the period of NASA support under this grant. The bulk of the work performed concerned benchmarking and analysis of CPUs, compilers, caches, and benchmark programs. The first part of this work concerned the issue of benchmark performance prediction. A new approach to benchmarking and machine characterization was reported, using a machine characterizer that measures the performance of a given system in terms of a Fortran abstract machine. Another report focused on analyzing compiler performance. The performance impact of optimization in the context of our methodology for CPU performance characterization was based on the abstract machine model. Benchmark programs are analyzed in another paper. A machine-independent model of program execution was developed to characterize both machine performance and program execution. By merging these machine and program characterizations, execution time can be estimated for arbitrary machine/program combinations. The work was continued into the domain of parallel and vector machines, including the issue of caches in vector processors and multiprocessors. All of the afore-mentioned accomplishments are more specifically summarized in this report, as well as those smaller in magnitude supported by this grant.

  19. Benchmarks for GADRAS performance validation.

    SciTech Connect

    Mattingly, John K.; Mitchell, Dean James; Rhykerd, Charles L., Jr.

    2009-09-01

    The performance of the Gamma Detector Response and Analysis Software (GADRAS) was validated by comparing GADRAS model results to experimental measurements for a series of benchmark sources. Sources for the benchmark include a plutonium metal sphere, bare and shielded in polyethylene, plutonium oxide in cans, a highly enriched uranium sphere, bare and shielded in polyethylene, a depleted uranium shell and spheres, and a natural uranium sphere. The benchmark experimental data were previously acquired and consist of careful collection of background and calibration source spectra along with the source spectra. The calibration data were fit with GADRAS to determine response functions for the detector in each experiment. A one-dimensional model (pie chart) was constructed for each source based on the dimensions of the benchmark source. The GADRAS code made a forward calculation from each model to predict the radiation spectrum for the detector used in the benchmark experiment. The comparisons between the GADRAS calculation and the experimental measurements are excellent, validating that GADRAS can correctly predict the radiation spectra for these well-defined benchmark sources.

  20. Implementation of Benchmarking Transportation Logistics Practices and Future Benchmarking Organizations

    SciTech Connect

    Thrower, A.W.; Patric, J.; Keister, M.

    2008-07-01

    The purpose of the Office of Civilian Radioactive Waste Management's (OCRWM) Logistics Benchmarking Project is to identify established government and industry practices for the safe transportation of hazardous materials which can serve as a yardstick for design and operation of OCRWM's national transportation system for shipping spent nuclear fuel and high-level radioactive waste to the proposed repository at Yucca Mountain, Nevada. The project will present logistics and transportation practices and develop implementation recommendations for adaptation by the national transportation system. This paper will describe the process used to perform the initial benchmarking study, highlight interim findings, and explain how these findings are being implemented. It will also provide an overview of the next phase of benchmarking studies. The benchmarking effort will remain a high-priority activity throughout the planning and operational phases of the transportation system. The initial phase of the project focused on government transportation programs to identify those practices which are most clearly applicable to OCRWM. These Federal programs have decades of safe transportation experience, strive for excellence in operations, and implement effective stakeholder involvement, all of which parallel OCRWM's transportation mission and vision. The initial benchmarking project focused on four business processes that are critical to OCRWM's mission success, and can be incorporated into OCRWM planning and preparation in the near term. The processes examined were: transportation business model, contract management/out-sourcing, stakeholder relations, and contingency planning. More recently, OCRWM examined logistics operations of AREVA NC's Business Unit Logistics in France. The next phase of benchmarking will focus on integrated domestic and international commercial radioactive logistic operations. The prospective companies represent large scale shippers and have vast experience in safely and efficiently shipping spent nuclear fuel and other radioactive materials. Additional business processes may be examined in this phase. The findings of these benchmarking efforts will help determine the organizational structure and requirements of the national transportation system. (authors)

  1. Variables Distinguishing Between Effective and Ineffective Writers in the Tenth Grade.

    ERIC Educational Resources Information Center

    Donelson, Kenneth L.

    1967-01-01

    To provide data for examining the variables that distinguish effective writers from ineffective ones, 1821 mid-west 10th-grade students wrote three themes at 3-week intervals. From a first reading of these papers, the sets of three themes from 432 pupils, which were judged strong or weak in effective writing, were submitted to four judges for

  2. [Hygienic assessment of lifestyle and health status in 10th-11th-form pupils directed to have a higher medical education].

    PubMed

    Timoshenko, K T

    2008-01-01

    Ninety-seven pupils from the 10th-to-11th classes formed on a competitive basis for intensive education, for forming motivation for future medical profession were examined using a set of psychophysiological tests that could evaluate the central nervous and cardiovascular systems, psychophysiological adaptation, task performance, and personality traits. The vast majority of the examinees were found to follow the hygienic recommendation of the day regimen, which corresponded to the principles of healthy lifestyle. In 99% of the pupils, mental capacity was rated as fair (66%) and high (33%), as evidenced by psychophysiological testing. Fifty-six per cent of the examinees were observed to have mental adaptive disorders that might reflect age-related psychological immaturity in them at the completing stage of schooling. PMID:19097437

  3. Current Use of Biomarkers in Acute Kidney Injury: Report and Summary of Recommendations from the 10th Acute Dialysis Quality Initiative Consensus Conference

    PubMed Central

    Murray, Patrick T.; Mehta, Ravindra L.; Shaw, Andrew; Ronco, Claudio; Endre, Zoltan; Kellum, John A.; Chawla, Lakhmir; Cruz, Dinna; Ince, Can; Okusa, Mark

    2013-01-01

    Over the last decade there has been considerable progress in the discovery and development of biomarkers of kidney disease, and several have now been evaluated in different clinical settings. While there is a growing literature on the performance of various biomarkers in clinical studies, there is limited information on how these biomarkers would be utilized by clinicians to manage patients with acute kidney injury (AKI). Recognizing this gap in knowledge, we convened the 10th Acute Dialysis Quality Initiative (ADQI) meeting to review the literature on biomarkers in AKI and their application in clinical practice. We asked an international group of experts to assess four broad areas for biomarker utilization for AKI: risk assessment, diagnosis and staging; differential diagnosis; prognosis and management and novel physiological techniques including imaging. This article provides a summary of the key findings and recommendations of the group, to equip clinicians to effectively use biomarkers in AKI. PMID:24107851

  4. Incorporating Ninth-Grade PSAT/NMSQT Scores into AP Potential Predictions for AP European History and AP World History. Statistical Report 2014-1

    ERIC Educational Resources Information Center

    Zhang, Xiuyuan; Patel, Priyank; Ewing, Maureen

    2015-01-01

    Historically, AP Potential correlations and expectancy tables have been based on 10th-and 11th-grade PSAT/NMSQT examinees and 11th-and 12th-grade AP examinees for all subjects (Zhang, Patel, & Ewing,2014; Ewing, Camara, & Millsap, 2006; Camara & Millsap, 1998). However, a large number of students take AP European History and AP

  5. Grade Inflation

    ERIC Educational Resources Information Center

    Hendrickson, Hildegard R.

    1976-01-01

    It is hypothesized that the average level of undergraduate grades has been increasing due to a change in grading practices rather than other fundamental factors. Printed sources were searched for the various reasons that have been cited and possible solutions are offered. (LBH)

  6. Closed-Loop Neuromorphic Benchmarks.

    PubMed

    Stewart, Terrence C; DeWolf, Travis; Kleinhans, Ashley; Eliasmith, Chris

    2015-01-01

    Evaluating the effectiveness and performance of neuromorphic hardware is difficult. It is even more difficult when the task of interest is a closed-loop task; that is, a task where the output from the neuromorphic hardware affects some environment, which then in turn affects the hardware's future input. However, closed-loop situations are one of the primary potential uses of neuromorphic hardware. To address this, we present a methodology for generating closed-loop benchmarks that makes use of a hybrid of real physical embodiment and a type of "minimal" simulation. Minimal simulation has been shown to lead to robust real-world performance, while still maintaining the practical advantages of simulation, such as making it easy for the same benchmark to be used by many researchers. This method is flexible enough to allow researchers to explicitly modify the benchmarks to identify specific task domains where particular hardware excels. To demonstrate the method, we present a set of novel benchmarks that focus on motor control for an arbitrary system with unknown external forces. Using these benchmarks, we show that an error-driven learning rule can consistently improve motor control performance across a randomly generated family of closed-loop simulations, even when there are up to 15 interacting joints to be controlled. PMID:26696820

  7. Closed-Loop Neuromorphic Benchmarks

    PubMed Central

    Stewart, Terrence C.; DeWolf, Travis; Kleinhans, Ashley; Eliasmith, Chris

    2015-01-01

    Evaluating the effectiveness and performance of neuromorphic hardware is difficult. It is even more difficult when the task of interest is a closed-loop task; that is, a task where the output from the neuromorphic hardware affects some environment, which then in turn affects the hardware's future input. However, closed-loop situations are one of the primary potential uses of neuromorphic hardware. To address this, we present a methodology for generating closed-loop benchmarks that makes use of a hybrid of real physical embodiment and a type of “minimal” simulation. Minimal simulation has been shown to lead to robust real-world performance, while still maintaining the practical advantages of simulation, such as making it easy for the same benchmark to be used by many researchers. This method is flexible enough to allow researchers to explicitly modify the benchmarks to identify specific task domains where particular hardware excels. To demonstrate the method, we present a set of novel benchmarks that focus on motor control for an arbitrary system with unknown external forces. Using these benchmarks, we show that an error-driven learning rule can consistently improve motor control performance across a randomly generated family of closed-loop simulations, even when there are up to 15 interacting joints to be controlled. PMID:26696820

  8. Randomized Benchmarking of Multiqubit Gates

    NASA Astrophysics Data System (ADS)

    Gaebler, J. P.; Meier, A. M.; Tan, T. R.; Bowler, R.; Lin, Y.; Hanneke, D.; Jost, J. D.; Home, J. P.; Knill, E.; Leibfried, D.; Wineland, D. J.

    2012-06-01

    We describe an extension of single-qubit gate randomized benchmarking that measures the error of multiqubit gates in a quantum information processor. This platform-independent protocol evaluates the performance of Clifford unitaries, which form a basis of fault-tolerant quantum computing. We implemented the benchmarking protocol with trapped ions and found an error per random two-qubit Clifford unitary of 0.1620.008, thus setting the first benchmark for such unitaries. By implementing a second set of sequences with an extra two-qubit phase gate inserted after each step, we extracted an error per phase gate of 0.0690.017. We conducted these experiments with transported, sympathetically cooled ions in a multizone Paul trapa system that can in principle be scaled to larger numbers of ions.

  9. FLOWTRAN-TF code benchmarking

    SciTech Connect

    Flach, G.P.

    1990-12-01

    FLOWTRAN-TF is a two-component (air-water), two-phase thermal-hydraulics code designed for performing accident analyses of SRS reactor fuel assemblies during the Emergency Cooling System (ECS) phase of a Double Ended Guillotine Break (DEGB) Loss Of Coolant Accident (LOCA). A description of the code is given by Flach et al. (1990). This report provides benchmarking results for the version of FLOWTRAN-TF used to compute the Recommended K-Reactor Restart ECS Power Limit (Smith et al., 1990a; 1990b). Individual constitutive relations are benchmarked in Sections 2 through 5 while in Sections 6 and 7 integral code benchmarking results are presented. An overall assessment of FLOWTRAN-TF for its intended use in computing the ECS power limit completes the document.

  10. DOE Commercial Building Benchmark Models

    SciTech Connect

    Torcellini, P.; Deru, M.; Griffith, B.; Benne, K.; Halverson, M.; Winiarski, D.; Crawley, D.

    2008-01-01

    The excitement surrounding the drive to build and renovate commercial buildings to achieve exemplary and even 'net zero performance,' coupled with the realization that complex systems engineering is usually required to achieve such levels, has led to a broader use of computer energy simulations. To provide a consistent baseline of comparison and save time conducting such simulations, the U.S. Department of Energy (DOE) - through three of its national laboratories - has developed a set of standard benchmark building models for new and existing buildings. These models represent a complete revision of the DOE benchmark buildings originally developed in 2006. The shapes, thermal zoning, and operation of the models are more indicative of real buildings than in the previous versions. DOE has developed 15 benchmark buildings that represent most of the commercial building stock, across 16 locations (representing all U.S. climate zones) and with three vintages (new, pre-1980, and post-1980 construction). This paper will provide an executive summary overview of these benchmark buildings, and how they can save building analysts valuable time. Fully documented and implemented to use with the EnergyPlus energy simulation program, the benchmark models are publicly available and new versions will be created to maintain compatibility with new releases of EnergyPlus. The benchmark buildings will form the basis for research on specific building technologies, energy code development, appliance standards, and measurement of progress toward DOE energy goals. Having a common starting point allows us to better share and compare research results and move forward to make more energy efficient buildings.

  11. Radiation Detection Computational Benchmark Scenarios

    SciTech Connect

    Shaver, Mark W.; Casella, Andrew M.; Wittman, Richard S.; McDonald, Ben S.

    2013-09-24

    Modeling forms an important component of radiation detection development, allowing for testing of new detector designs, evaluation of existing equipment against a wide variety of potential threat sources, and assessing operation performance of radiation detection systems. This can, however, result in large and complex scenarios which are time consuming to model. A variety of approaches to radiation transport modeling exist with complementary strengths and weaknesses for different problems. This variety of approaches, and the development of promising new tools (such as ORNL’s ADVANTG) which combine benefits of multiple approaches, illustrates the need for a means of evaluating or comparing different techniques for radiation detection problems. This report presents a set of 9 benchmark problems for comparing different types of radiation transport calculations, identifying appropriate tools for classes of problems, and testing and guiding the development of new methods. The benchmarks were drawn primarily from existing or previous calculations with a preference for scenarios which include experimental data, or otherwise have results with a high level of confidence, are non-sensitive, and represent problem sets of interest to NA-22. From a technical perspective, the benchmarks were chosen to span a range of difficulty and to include gamma transport, neutron transport, or both and represent different important physical processes and a range of sensitivity to angular or energy fidelity. Following benchmark identification, existing information about geometry, measurements, and previous calculations were assembled. Monte Carlo results (MCNP decks) were reviewed or created and re-run in order to attain accurate computational times and to verify agreement with experimental data, when present. Benchmark information was then conveyed to ORNL in order to guide testing and development of hybrid calculations. The results of those ADVANTG calculations were then sent to PNNL for compilation. This is a report describing the details of the selected Benchmarks and results from various transport codes.

  12. Benchmark of a Cubieboard cluster

    NASA Astrophysics Data System (ADS)

    Schnepf, M. J.; Gudu, D.; Rische, B.; Fischer, M.; Jung, C.; Hardt, M.

    2015-12-01

    We built a cluster of ARM-based Cubieboards2 which has a SATA interface to connect a harddrive. This cluster was set up as a storage system using Ceph and as a compute cluster for high energy physics analyses. To study the performance in these applications, we ran two benchmarks on this cluster. We also checked the energy efficiency of the cluster using the preseted benchmarks. Performance and energy efficency of our cluster were compared with a network-attached storage (NAS), and with a desktop PC.

  13. Grading Meningioma

    PubMed Central

    Okuchi, Sachi; Okada, Tomohisa; Yamamoto, Akira; Kanagaki, Mitsunori; Fushimi, Yasutaka; Okada, Tsutomu; Yamauchi, Moritaka; Kataoka, Masako; Arakawa, Yoshiki; Takahashi, Jun C.; Minamiguchi, Sachiko; Miyamoto, Susumu; Togashi, Kaori

    2015-01-01

    Abstract The purpose was to compare capability of fluorine-18 fluorodeoxyglucose (FDG)-PET and thallium-201 (Tl)-SPECT for grading meningioma. This retrospective study was conducted as a case-control study under approval by the institutional review board. In the hospital information system, 67 patients (22 men and 45 women) who had both FDG-PET and Tl-SPECT preoperative examinations were found with histopathologic diagnosis of meningioma. The maximum FDG uptake values of the tumors were measured, and they were standardized to the whole body (SUVmax) and normalized as gray matter ratio (SUVRmax). Mean and maximum Tl uptake ratios (TURmean and TURmax, respectively) of the tumors were measured and normalized as ratios to those of the contralateral normal brain. Receiver-operating characteristic curve analyses of the 4 indexes were conducted for differentiation between low- and high-grade meningiomas, and areas under the curves (AUCs) were compared. Correlation coefficients were calculated between these indexes and Ki-67. Fifty-six meningiomas were classified as grade I (low grade), and 11 were grade II or III (high grade). In all 4 indexes, a significant difference was observed between low- and high-grade meningiomas (P?grades comparable to FDG-PET. PMID:25674763

  14. Real-Time Benchmark Suite

    Energy Science and Technology Software Center (ESTSC)

    1992-01-17

    This software provides a portable benchmark suite for real time kernels. It tests the performance of many of the system calls, as well as the interrupt response time and task response time to interrupts. These numbers provide a baseline for comparing various real-time kernels and hardware platforms.

  15. Benchmark Lisp And Ada Programs

    NASA Technical Reports Server (NTRS)

    Davis, Gloria; Galant, David; Lim, Raymond; Stutz, John; Gibson, J.; Raghavan, B.; Cheesema, P.; Taylor, W.

    1992-01-01

    Suite of nonparallel benchmark programs, ELAPSE, designed for three tests: comparing efficiency of computer processing via Lisp vs. Ada; comparing efficiencies of several computers processing via Lisp; or comparing several computers processing via Ada. Tests efficiency which computer executes routines in each language. Available for computer equipped with validated Ada compiler and/or Common Lisp system.

  16. PyMPI Dynamic Benchmark

    Energy Science and Technology Software Center (ESTSC)

    2007-02-16

    Pynamic is a benchmark designed to test a system's ability to handle the Dynamic Linking and Loading (DLL) requirements of Python-based scientific applications. This benchmark is developed to add a workload to our testing environment, a workload that represents a newly emerging class of DLL behaviors. Pynamic buildins on pyMPI, and MPI extension to Python C-extension dummy codes and a glue layer that facilitates linking and loading of the generated dynamic modules into the resultingmore » pyMPI. Pynamic is configurable, enabling modeling the static properties of a specific code as described in section 5. It does not, however, model any significant computationss of the target and hence, it is not subjected to the same level of control as the target code. In fact, HPC computer vendors and tool developers will be encouraged to add it to their tesitn suite once the code release is completed. an ability to produce and run this benchmark is an effective test for valifating the capability of a compiler and linker/loader as well as an OS kernel and other runtime system of HPC computer vendors. In addition, the benchmark is designed as a test case for stressing code development tools. Though Python has recently gained popularity in the HPC community, it heavy DLL operations have hindered certain HPC code development tools, notably parallel debuggers, from performing optimally.« less

  17. Processor Emulator with Benchmark Applications

    SciTech Connect

    Lloyd, G. Scott; Pearce, Roger; Gokhale, Maya

    2015-11-13

    A processor emulator and a suite of benchmark applications have been developed to assist in characterizing the performance of data-centric workloads on current and future computer architectures. Some of the applications have been collected from other open source projects. For more details on the emulator and an example of its usage, see reference [1].

  18. A comparison of five benchmarks

    NASA Technical Reports Server (NTRS)

    Huss, Janice E.; Pennline, James A.

    1987-01-01

    Five benchmark programs were obtained and run on the NASA Lewis CRAY X-MP/24. A comparison was made between the programs codes and between the methods for calculating performance figures. Several multitasking jobs were run to gain experience in how parallel performance is measured.

  19. PRISMATIC CORE COUPLED TRANSIENT BENCHMARK

    SciTech Connect

    J. Ortensi; M.A. Pope; G. Strydom; R.S. Sen; M.D. DeHart; H.D. Gougar; C. Ellis; A. Baxter; V. Seker; T.J. Downar; K. Vierow; K. Ivanov

    2011-06-01

    The Prismatic Modular Reactor (PMR) is one of the High Temperature Reactor (HTR) design concepts that have existed for some time. Several prismatic units have operated in the world (DRAGON, Fort St. Vrain, Peach Bottom) and one unit is still in operation (HTTR). The deterministic neutronics and thermal-fluids transient analysis tools and methods currently available for the design and analysis of PMRs have lagged behind the state of the art compared to LWR reactor technologies. This has motivated the development of more accurate and efficient tools for the design and safety evaluations of the PMR. In addition to the work invested in new methods, it is essential to develop appropriate benchmarks to verify and validate the new methods in computer codes. The purpose of this benchmark is to establish a well-defined problem, based on a common given set of data, to compare methods and tools in core simulation and thermal hydraulics analysis with a specific focus on transient events. The benchmark-working group is currently seeking OECD/NEA sponsorship. This benchmark is being pursued and is heavily based on the success of the PBMR-400 exercise.

  20. Benchmarks for industrial energy efficiency

    SciTech Connect

    Amarnath, K.R.; Kumana, J.D.; Shah, J.V.

    1996-12-31

    What are the standards for improving energy efficiency for industries such as petroleum refining, chemicals, and glass manufacture? How can different industries in emerging markets and developing accelerate the pace of improvements? This paper discusses several case studies and experiences relating to this subject emphasizing the use of energy efficiency benchmarks. Two important benchmarks are discussed. The first is based on a track record of outstanding performers in the related industry segment; the second benchmark is based on site specific factors. Using energy use reduction targets or benchmarks, projects have been implemented in Mexico, Poland, India, Venezuela, Brazil, China, Thailand, Malaysia, Republic of South Africa and Russia. Improvements identified through these projects include a variety of recommendations. The use of oxy-fuel and electric furnaces in the glass industry in Poland; reconfiguration of process heat recovery systems for refineries in China, Malaysia, and Russia; recycling and reuse of process wastewater in Republic of South Africa; cogeneration plant in Venezuela. The paper will discuss three case studies of efforts undertaken in emerging market countries to improve energy efficiency.

  1. Austin Community College Benchmarking Update.

    ERIC Educational Resources Information Center

    Austin Community Coll., TX. Office of Institutional Effectiveness.

    Austin Community College contracted with MGT of America, Inc. in spring 1999 to develop a peer and benchmark (best) practices analysis on key indicators. These indicators were updated in spring 2002 using data from eight Texas community colleges and four non-Texas institutions that represent large, comprehensive, urban community colleges, similar

  2. Genetic counseling/consultation in South-East Asia: a report from the workshop at the 10th Asia pacific conference on human genetics.

    PubMed

    Zayts, Olga; Sarangi, Srikant; Thong, Meow-Keong; Chung, Brian Hon-yin; Lo, Ivan Fao-man; Kan, Anita Sik-yau; Lee, Juliana Mei-har; Padilla, Carmencita David; Cutiongco-de la Paz, Eva Maria; Faradz, Sultana M H; Wasant, Pornswan

    2013-12-01

    This paper reports on the workshop 'Genetic Counseling/Consultations in South-East Asia' at the 10(th) Asia Pacific Conference on Human Genetics in Kuala Lumpur, Malaysia, in December 2012. The workshop brought together professionals and language/communication scholars from South-East Asia, and the UK. The workshop aimed at addressing culture- and context-specific genetic counseling/consultation practices in South-East Asia. As a way of contextualizing genetic counseling/consultation in South-East Asia, we first offer an overview of communication-oriented research generally, drawing attention to consultation and counseling as part of a communicative continuum with distinctive interactional features. We then provide examples of genetic counseling/consultation research in Hong Kong. As other countries in South-East Asia have not yet embarked on communication-oriented empirical research, we report on the current practices of genetic counseling/consultation in these countries in order to identify similarities and differences as well as key obstacles that could be addressed through future research. Three issues emerged as 'problematic': language, religion and culture. We suggest that communication-oriented research can provide a starting point for evidence-based reflections on how to incorporate a counseling mentality in genetic consultation. To conclude, we discuss the need for creating a platform for targeted training of genetic counselors based on communication-oriented research findings. PMID:24048708

  3. The complexity and challenges of the International Classification of Diseases, Ninth Revision, Clinical Modification to International Classification of Diseases, 10th Revision, Clinical Modification transition in EDs.

    PubMed

    Krive, Jacob; Patel, Mahatkumar; Gehm, Lisa; Mackey, Mark; Kulstad, Erik; Li, Jianrong John; Lussier, Yves A; Boyd, Andrew D

    2015-05-01

    Beginning October 2015, the Center for Medicare and Medicaid Services will require medical providers to use the vastly expanded International Classification of Diseases, 10th Revision, Clinical Modification (ICD-10-CM) system. Despite wide availability of information and mapping tools for the next generation of the ICD classification system, some of the challenges associated with transition from ICD-9-CM to ICD-10-CM are not well understood. To quantify the challenges faced by emergency physicians, we analyzed a subset of a 2010 Illinois Medicaid database of emergency department ICD-9-CM codes, seeking to determine the accuracy of existing mapping tools in order to better prepare emergency physicians for the change to the expanded ICD-10-CM system. We found that 27% of 1830 codes represented convoluted multidirectional mappings. We then analyzed the convoluted transitions and found that 8% of total visit encounters (23% of the convoluted transitions) were clinically incorrect. The ambiguity and inaccuracy of these mappings may impact the workflow associated with the translation process and affect the potential mapping between ICD codes and Current Procedural Codes, which determine physician reimbursement. PMID:25863652

  4. Imaging in the Age of Precision Medicine: Summary of the Proceedings of the 10th Biannual Symposium of the International Society for Strategic Studies in Radiology.

    PubMed

    Herold, Christian J; Lewin, Jonathan S; Wibmer, Andreas G; Thrall, James H; Krestin, Gabriel P; Dixon, Adrian K; Schoenberg, Stefan O; Geckle, Rena J; Muellner, Ada; Hricak, Hedvig

    2016-04-01

    During the past decade, with its breakthroughs in systems biology, precision medicine (PM) has emerged as a novel health-care paradigm. Challenging reductionism and broad-based approaches in medicine, PM is an approach for disease treatment and prevention that takes into account individual variability in genes, environment, and lifestyle. It involves integrating information from multiple sources in a holistic manner to achieve a definitive diagnosis, focused treatment, and adequate response assessment. Biomedical imaging and imaging-guided interventions, which provide multiparametric morphologic and functional information and enable focused, minimally invasive treatments, are key elements in the infrastructure needed for PM. The emerging discipline of radiogenomics, which links genotypic information to phenotypic disease manifestations at imaging, should also greatly contribute to patient-tailored care. Because of the growing volume and complexity of imaging data, decision-support algorithms will be required to help physicians apply the most essential patient data for optimal management. These innovations will challenge traditional concepts of health care and business models. Reimbursement policies and quality assurance measures will have to be reconsidered and adapted. In their 10th biannual symposium, which was held in August 2013, the members of the International Society for Strategic Studies in Radiology discussed the opportunities and challenges arising for the imaging community with the transition to PM. This article summarizes the discussions and central messages of the symposium. (©) RSNA, 2015. PMID:26465058

  5. An INTEGRAL view of the high-energy sky (the first 10 years) - 9th INTEGRAL Workshop and celebration of the 10th anniversary of the launch

    NASA Astrophysics Data System (ADS)

    The 9th INTEGRAL workshop "An INTEGRAL view of the high-energy sky (the first 10 years)" took place from 15 to 19 October 2012 in Paris, Bibliothque Nationale de France (Bibliothque Franois Mitterrand). The workshop was sponsored by ESA, CNES and other French and European Institutions. During this week, and in particular on 17 October 2012, we celebrated the 10th anniversary of the launch of the INTEGRAL mission. The main goal of this workshop was to present and to discuss (via invited and contributed talks and posters) latest results obtained in the field of high-energy astrophysics using the International Gamma-Ray Astrophysics Laboratory INTEGRAL, as well as results from observations from other ground- and space-based high-energy observatories and from associated multi-wavelength campaigns. Contributions to the workshop covered the following scientific topics: - X-ray binaries (IGR sources, black holes, neutron stars, white dwarfs) - Isolated neutron stars (gamma-ray pulsars, magnetars) - Nucleo-synthesis (SNe, Novae, SNRs, ISM) and gamma-ray lines (511 keV) - Galactic diffuse continuum emission (including Galactic Ridge) - Massive black holes in AGNs, elliptical galaxies, nucleus of the Galaxy - Sky surveys, source populations and unidentified gamma-ray sources - Cosmic background radiation - Gamma-ray bursts - Coordinated observations with other ground- and space-based observatories - Science data processing and analysis (posters only) - Future instruments and missions (posters only)

  6. Evaluation of elemental status of ancient human bone samples from Northeastern Hungary dated to the 10th century AD by XRF

    NASA Astrophysics Data System (ADS)

    János, I.; Szathmáry, L.; Nádas, E.; Béni, A.; Dinya, Z.; Máthé, E.

    2011-11-01

    The present study is a multielemental analysis of bone samples belonging to skeletal individuals originating from two contemporaneous (10th century AD) cemeteries (Tiszavasvári Nagy-Gyepáros and Nagycserkesz-Nádasibokor sites) in Northeastern Hungary, using the XRF analytical technique. Emitted X-rays were detected in order to determine the elemental composition of bones and to appreciate the possible influence of the burial environment on the elemental content of the human skeletal remains. Lumbar vertebral bodies were used for analysis. Applying the ED(P)XRF technique concentration of the following elements were determined: P, Ca, K, Na, Mg, Al, Cl, Mn, Fe, Zn, Br and Sr. The results indicated post mortem mineral exchange between the burial environment (soil) and bones (e.g. the enhanced levels of Fe and Mn) and referred to diagenetic alteration processes during burials. However, other elements such as Zn, Sr and Br seemed to be accumulated during the past life. On the basis of statistical analysis, clear separation could not be observed between the two excavation sites in their bone elemental concentrations which denoted similar diagenetic influences, environmental conditions. The enhanced levels of Sr might be connected with the past dietary habits, especially consumption of plant food.

  7. Benchmarking: Achieving the best in class

    SciTech Connect

    Kaemmerer, L

    1996-05-01

    Oftentimes, people find the process of organizational benchmarking an onerous task, or, because they do not fully understand the nature of the process, end up with results that are less than stellar. This paper presents the challenges of benchmarking and reasons why benchmarking can benefit an organization in today`s economy.

  8. Testing (Validating?) Cross Sections with ICSBEP Benchmarks

    SciTech Connect

    Kahler, Albert C. III

    2012-06-28

    We discuss how to use critical benchmarks from the International Handbook of Evaluated Criticality Safety Benchmark Experiments to determine the applicability of specific cross sections to the end-user's problem of interest. Particular attention is paid to making sure the selected suite of benchmarks includes the user's range of applicability (ROA).

  9. The Nature of Grading.

    ERIC Educational Resources Information Center

    Ornstein, Allan C.

    1989-01-01

    Discusses grading in elementary and secondary schools, focusing on the following topics: combining and weighting data, grading criteria, absolute grade standards, relative grade standards, contracting for grades, mastery and continuous progress grading, grading for effort or improvement, report cards, and procedures for grading students. (SR)

  10. Effect of a Counseling Intervention Program on Tenth Grade Students' Attendance, Discipline Referrals, and Academic Achievement

    ERIC Educational Resources Information Center

    Austin, Dorothy Deona Martin

    2013-01-01

    Poor student achievement, high discipline referrals, and student absenteeism were issues in a rural school with a population of approximately 400 students. The purpose of this study was to examine the influence of Effective Teens training on the attendance, discipline referrals, and academic achievement of 10th grade students. The theoretical

  11. Positive predictive values of the International Classification of Disease, 10th edition diagnoses codes for diverticular disease in the Danish National Registry of Patients

    PubMed Central

    Erichsen, Rune; Strate, Lisa; Srensen, Henrik Toft; Baron, John A

    2010-01-01

    Objective: To investigate the accuracy of diagnostic coding for diverticular disease in the Danish National Registry of Patients (NRP). Study design and setting: At Aalborg Hospital, Denmark, with a catchment area of 640,000 inhabitants, we identified 100 patients recorded in the NRP with a diagnosis of diverticular disease (International Classification of Disease codes, 10th revision [ICD-10] K572K579) during the 19992008 period. We assessed the positive predictive value (PPV) as a measure of the accuracy of discharge codes for diverticular disease using information from discharge abstracts and outpatient notes as the reference standard. Results: Of the 100 patients coded with diverticular disease, 49 had complicated diverticular disease, whereas 51 had uncomplicated diverticulosis. For the overall diagnosis of diverticular disease (K57), the PPV was 0.98 (95% confidence intervals [CIs]: 0.93, 0.99). For the more detailed subgroups of diagnosis indicating the presence or absence of complications (K573K579) the PPVs ranged from 0.67 (95% CI: 0.09, 0.99) to 0.92 (95% CI: 0.52, 1.00). The diagnosis codes did not allow accurate identification of uncomplicated disease or any specific complication. However, the combined ICD-10 codes K572, K574, and K578 had a PPV of 0.91 (95% CI: 0.71, 0.99) for any complication. Conclusion: The diagnosis codes in the NRP can be used to identify patients with diverticular disease in general; however, they do not accurately discern patients with uncomplicated diverticulosis or with specific diverticular complications. PMID:21694857

  12. IBC’s 23rd Annual Antibody Engineering, 10th Annual Antibody Therapeutics International Conferences and the 2012 Annual Meeting of The Antibody Society

    PubMed Central

    Klöhn, Peter-Christian; Wuellner, Ulrich; Zizlsperger, Nora; Zhou, Yu; Tavares, Daniel; Berger, Sven; Zettlitz, Kirstin A.; Proetzel, Gabriele; Yong, May; Begent, Richard H.J.; Reichert, Janice M

    2013-01-01

    The 23rd Annual Antibody Engineering, 10th Annual Antibody Therapeutics international conferences, and the 2012 Annual Meeting of The Antibody Society, organized by IBC Life Sciences with contributions from The Antibody Society and two Scientific Advisory Boards, were held December 3–6, 2012 in San Diego, CA. The meeting drew over 800 participants who attended sessions on a wide variety of topics relevant to antibody research and development. As a prelude to the main events, a pre-conference workshop held on December 2, 2012 focused on intellectual property issues that impact antibody engineering. The Antibody Engineering Conference was composed of six sessions held December 3–5, 2012: (1) From Receptor Biology to Therapy; (2) Antibodies in a Complex Environment; (3) Antibody Targeted CNS Therapy: Beyond the Blood Brain Barrier; (4) Deep Sequencing in B Cell Biology and Antibody Libraries; (5) Systems Medicine in the Development of Antibody Therapies/Systematic Validation of Novel Antibody Targets; and (6) Antibody Activity and Animal Models. The Antibody Therapeutics conference comprised four sessions held December 4–5, 2012: (1) Clinical and Preclinical Updates of Antibody-Drug Conjugates; (2) Multifunctional Antibodies and Antibody Combinations: Clinical Focus; (3) Development Status of Immunomodulatory Therapeutic Antibodies; and (4) Modulating the Half-Life of Antibody Therapeutics. The Antibody Society’s special session on applications for recording and sharing data based on GIATE was held on December 5, 2012, and the conferences concluded with two combined sessions on December 5–6, 2012: (1) Development Status of Early Stage Therapeutic Antibodies; and (2) Immunomodulatory Antibodies for Cancer Therapy. PMID:23575266

  13. Geothermal Heat Pump Benchmarking Report

    SciTech Connect

    1997-01-17

    A benchmarking study was conducted on behalf of the Department of Energy to determine the critical factors in successful utility geothermal heat pump programs. A Successful program is one that has achieved significant market penetration. Successfully marketing geothermal heat pumps has presented some major challenges to the utility industry. However, select utilities have developed programs that generate significant GHP sales. This benchmarking study concludes that there are three factors critical to the success of utility GHP marking programs: (1) Top management marketing commitment; (2) An understanding of the fundamentals of marketing and business development; and (3) An aggressive competitive posture. To generate significant GHP sales, competitive market forces must by used. However, because utilities have functioned only in a regulated arena, these companies and their leaders are unschooled in competitive business practices. Therefore, a lack of experience coupled with an intrinsically non-competitive culture yields an industry environment that impedes the generation of significant GHP sales in many, but not all, utilities.

  14. Benchmarking neuromorphic systems with Nengo

    PubMed Central

    Bekolay, Trevor; Stewart, Terrence C.; Eliasmith, Chris

    2015-01-01

    Nengo is a software package for designing and simulating large-scale neural models. Nengo is architected such that the same Nengo model can be simulated on any of several Nengo backends with few to no modifications. Backends translate a model to specific platforms, which include GPUs and neuromorphic hardware. Nengo also contains a large test suite that can be run with any backend and focuses primarily on functional performance. We propose that Nengo's large test suite can be used to benchmark neuromorphic hardware's functional performance and simulation speed in an efficient, unbiased, and future-proof manner. We implement four benchmark models and show that Nengo can collect metrics across five different backends that identify situations in which some backends perform more accurately or quickly. PMID:26539076

  15. Benchmarking neuromorphic systems with Nengo.

    PubMed

    Bekolay, Trevor; Stewart, Terrence C; Eliasmith, Chris

    2015-01-01

    Nengo is a software package for designing and simulating large-scale neural models. Nengo is architected such that the same Nengo model can be simulated on any of several Nengo backends with few to no modifications. Backends translate a model to specific platforms, which include GPUs and neuromorphic hardware. Nengo also contains a large test suite that can be run with any backend and focuses primarily on functional performance. We propose that Nengo's large test suite can be used to benchmark neuromorphic hardware's functional performance and simulation speed in an efficient, unbiased, and future-proof manner. We implement four benchmark models and show that Nengo can collect metrics across five different backends that identify situations in which some backends perform more accurately or quickly. PMID:26539076

  16. MPI Multicore Torus Communication Benchmark

    Energy Science and Technology Software Center (ESTSC)

    2008-02-05

    The MPI Multicore Torus Communications Benchmark (TorusTest) measues the aggegate bandwidth across all six links from/to any multicore node in a logical torus. It can run in wo modi: using a static or a random mapping of tasks to torus locations. The former can be used to achieve optimal mappings and aggregate bandwidths that can be achieved with varying node mappings.

  17. Restaurant Energy Use Benchmarking Guideline

    SciTech Connect

    Hedrick, R.; Smith, V.; Field, K.

    2011-07-01

    A significant operational challenge for food service operators is defining energy use benchmark metrics to compare against the performance of individual stores. Without metrics, multiunit operators and managers have difficulty identifying which stores in their portfolios require extra attention to bring their energy performance in line with expectations. This report presents a method whereby multiunit operators may use their own utility data to create suitable metrics for evaluating their operations.

  18. HS06 Benchmark for an ARM Server

    NASA Astrophysics Data System (ADS)

    Kluth, Stefan

    2014-06-01

    We benchmarked an ARM cortex-A9 based server system with a four-core CPU running at 1.1 GHz. The system used Ubuntu 12.04 as operating system and the HEPSPEC 2006 (HS06) benchmarking suite was compiled natively with gcc-4.4 on the system. The benchmark was run for various settings of the relevant gcc compiler options. We did not find significant influence from the compiler options on the benchmark result. The final HS06 benchmark result is 10.4.

  19. RISKIND verification and benchmark comparisons

    SciTech Connect

    Biwer, B.M.; Arnish, J.J.; Chen, S.Y.; Kamboj, S.

    1997-08-01

    This report presents verification calculations and benchmark comparisons for RISKIND, a computer code designed to estimate potential radiological consequences and health risks to individuals and the population from exposures associated with the transportation of spent nuclear fuel and other radioactive materials. Spreadsheet calculations were performed to verify the proper operation of the major options and calculational steps in RISKIND. The program is unique in that it combines a variety of well-established models into a comprehensive treatment for assessing risks from the transportation of radioactive materials. Benchmark comparisons with other validated codes that incorporate similar models were also performed. For instance, the external gamma and neutron dose rate curves for a shipping package estimated by RISKIND were compared with those estimated by using the RADTRAN 4 code and NUREG-0170 methodology. Atmospheric dispersion of released material and dose estimates from the GENII and CAP88-PC codes. Verification results have shown the program to be performing its intended function correctly. The benchmark results indicate that the predictions made by RISKIND are within acceptable limits when compared with predictions from similar existing models.

  20. Salient Object Detection: A Benchmark

    NASA Astrophysics Data System (ADS)

    Borji, Ali; Cheng, Ming-Ming; Jiang, Huaizu; Li, Jia

    2015-12-01

    We extensively compare, qualitatively and quantitatively, 40 state-of-the-art models (28 salient object detection, 10 fixation prediction, 1 objectness, and 1 baseline) over 6 challenging datasets for the purpose of benchmarking salient object detection and segmentation methods. From the results obtained so far, our evaluation shows a consistent rapid progress over the last few years in terms of both accuracy and running time. The top contenders in this benchmark significantly outperform the models identified as the best in the previous benchmark conducted just two years ago. We find that the models designed specifically for salient object detection generally work better than models in closely related areas, which in turn provides a precise definition and suggests an appropriate treatment of this problem that distinguishes it from other problems. In particular, we analyze the influences of center bias and scene complexity in model performance, which, along with the hard cases for state-of-the-art models, provide useful hints towards constructing more challenging large scale datasets and better saliency models. Finally, we propose probable solutions for tackling several open problems such as evaluation scores and dataset bias, which also suggest future research directions in the rapidly-growing field of salient object detection.

  1. The Federal Forecasters Conference--1999. Papers and Proceedings (10th, Washington, DC, June 24, 1999) and Selected Papers from the International Symposium on Forecasting (19th, Washington, DC, June 27-30, 1999).

    ERIC Educational Resources Information Center

    Gerald, Debra E., Ed.

    The 10th Federal Forecasters Conference provided a forum where 127 forecasters from different federal agencies and other organizations met to discuss various aspects of the conference's theme, "Forecasting in the New Millennium," that could be applied in the United States. A keynote address, "Procedures for Auditing Federal Forecasts" by J. Scott

  2. Standing adult human phantoms based on 10th, 50th and 90th mass and height percentiles of male and female Caucasian populations

    NASA Astrophysics Data System (ADS)

    Cassola, V. F.; Milian, F. M.; Kramer, R.; de Oliveira Lira, C. A. B.; Khoury, H. J.

    2011-07-01

    Computational anthropomorphic human phantoms are useful tools developed for the calculation of absorbed or equivalent dose to radiosensitive organs and tissues of the human body. The problem is, however, that, strictly speaking, the results can be applied only to a person who has the same anatomy as the phantom, while for a person with different body mass and/or standing height the data could be wrong. In order to improve this situation for many areas in radiological protection, this study developed 18 anthropometric standing adult human phantoms, nine models per gender, as a function of the 10th, 50th and 90th mass and height percentiles of Caucasian populations. The anthropometric target parameters for body mass, standing height and other body measures were extracted from PeopleSize, a well-known software package used in the area of ergonomics. The phantoms were developed based on the assumption of a constant body-mass index for a given mass percentile and for different heights. For a given height, increase or decrease of body mass was considered to reflect mainly the change of subcutaneous adipose tissue mass, i.e. that organ masses were not changed. Organ mass scaling as a function of height was based on information extracted from autopsy data. The methods used here were compared with those used in other studies, anatomically as well as dosimetrically. For external exposure, the results show that equivalent dose decreases with increasing body mass for organs and tissues located below the subcutaneous adipose tissue layer, such as liver, colon, stomach, etc, while for organs located at the surface, such as breasts, testes and skin, the equivalent dose increases or remains constant with increasing body mass due to weak attenuation and more scatter radiation caused by the increasing adipose tissue mass. Changes of standing height have little influence on the equivalent dose to organs and tissues from external exposure. Specific absorbed fractions (SAFs) have also been calculated with the 18 anthropometric phantoms. The results show that SAFs decrease with increasing height and increase with increasing body mass. The calculated data suggest that changes of the body mass may have a significant effect on equivalent doses, primarily for external exposure to organs and tissue located below the adipose tissue layer, while for superficial organs, for changes of height and for internal exposures the effects on equivalent dose are small to moderate.

  3. IBC’s 23rd Antibody Engineering and 10th Antibody Therapeutics Conferences and the Annual Meeting of The Antibody Society

    PubMed Central

    Marquardt, John; Begent, Richard H.J.; Chester, Kerry; Huston, James S.; Bradbury, Andrew; Scott, Jamie K.; Thorpe, Philip E.; Veldman, Trudi; Reichert, Janice M.; Weiner, Louis M.

    2012-01-01

    Now in its 23rd and 10th years, respectively, the Antibody Engineering and Antibody Therapeutics conferences are the Annual Meeting of The Antibody Society. The scientific program covers the full spectrum of challenges in antibody research and development from basic science through clinical development. In this preview of the conferences, the chairs provide their thoughts on sessions that will allow participants to track emerging trends in (1) the development of next-generation immunomodulatory antibodies; (2) the complexity of the environment in which antibodies must function; (3) antibody-targeted central nervous system (CNS) therapies that cross the blood brain barrier; (4) the extension of antibody half-life for improved efficacy and pharmacokinetics (PK)/pharmacodynamics (PD); and (5) the application of next generation DNA sequencing to accelerate antibody research. A pre-conference workshop on Sunday, December 2, 2012 will update participants on recent intellectual property (IP) law changes that affect antibody research, including biosimilar legislation, the America Invents Act and recent court cases. Keynote presentations will be given by Andreas Plückthun (University of Zürich), who will speak on engineering receptor ligands with powerful cellular responses; Gregory Friberg (Amgen Inc.), who will provide clinical updates of bispecific antibodies; James D. Marks (University of California, San Francisco), who will discuss a systems approach to generating tumor targeting antibodies; Dario Neri (Swiss Federal Institute of Technology Zürich), who will speak about delivering immune modulators at the sites of disease; William M. Pardridge (University of California, Los Angeles), who will discuss delivery across the blood-brain barrier; and Peter Senter (Seattle Genetics, Inc.), who will present his vision for the future of antibody-drug conjugates. For more information on these meetings or to register to attend, please visit www.IBCLifeSciences.com/AntibodyEng or call 800-390-4078. Members of The Antibody Society and mAbs journal subscribers receive a 20% discount for meeting registration. To obtain this discount, email kdostie@ibcusa.com. mAbs is the official therapeutics journal of The Antibody Society and offers a discounted subscription to Society members for $49. PMID:23007482

  4. Fault modeling of the Mw 7.0 shallow intra-slab strike-slip earthquake occurred on 2011 July 10th using near-field tsunami record

    NASA Astrophysics Data System (ADS)

    Kubota, T.; Hino, R.; Iinuma, T.

    2014-12-01

    On 2011 July 10th, an earthquake of Mw 7.0 occurred in the shallow part of the Pacific slab beneath the large coseismic slip area of the 2011 Tohoku-Oki earthquake. This event has a strike-slip focal mechanism with steep dipping nodal planes. Near the epicenter, aftershocks determined by OBS deployment formed clear two orthogonal lineaments with identical strikes of the focal mechanism solution, suggesting that the aftershock activity occurred along the two conjugate faults. The strikes of these faults were almost parallel to the direction of the magnetic lineations and the fracture zones of the incoming Pacific plate, suggesting that the earthquake was the re-rupture of congenital fractures under the extensional stress induced by the Tohoku-Oki earthquake. It is of great interest to know the down-dip size of the source fault not only to understand the mechanical nature of the slab but also the post-2011 stress state. Coseismic seafloor deformation and tsunami associated with the earthquake were observed by ocean bottom pressure gauges deployed within ~ 100 km from the epicenter. We estimated the finite fault model of this event to discuss the rupture properties of the earthquake. We sought the source model assuming a rectangular fault with a uniform slip assuming the strike of the fault to be one of those of two nodal planes of the focal mechanism. The two preferable source models corresponding to the two nodal planes explained the observed data equally well. For either model, the depth of the downdip end exceeds 40 km below the plate boundary, meaning the fault widths (down-dip size) were much larger than the depth extent of the aftershock distribution (~ 15 km). We sought another source model assuming the simultaneous rupture of the conjugate faults and found that the width of the fault model was more consistent with the aftershock distribution than the single rupture plane models. The 2011 intraslab strike-slip earthquake might be a compound rupture of the spreading-related fracture and the transform fault like fracture within the subducting Pacific slab.

  5. NASA Software Engineering Benchmarking Effort

    NASA Technical Reports Server (NTRS)

    Godfrey, Sally; Rarick, Heather

    2012-01-01

    Benchmarking was very interesting and provided a wealth of information (1) We did see potential solutions to some of our "top 10" issues (2) We have an assessment of where NASA stands with relation to other aerospace/defense groups We formed new contacts and potential collaborations (1) Several organizations sent us examples of their templates, processes (2) Many of the organizations were interested in future collaboration: sharing of training, metrics, Capability Maturity Model Integration (CMMI) appraisers, instructors, etc. We received feedback from some of our contractors/ partners (1) Desires to participate in our training; provide feedback on procedures (2) Welcomed opportunity to provide feedback on working with NASA

  6. Benchmarking for Excellence and the Nursing Process

    NASA Technical Reports Server (NTRS)

    Sleboda, Claire

    1999-01-01

    Nursing is a service profession. The services provided are essential to life and welfare. Therefore, setting the benchmark for high quality care is fundamental. Exploring the definition of a benchmark value will help to determine a best practice approach. A benchmark is the descriptive statement of a desired level of performance against which quality can be judged. It must be sufficiently well understood by managers and personnel in order that it may serve as a standard against which to measure value.

  7. Computational Chemistry Comparison and Benchmark Database

    National Institute of Standards and Technology Data Gateway

    SRD 101 NIST Computational Chemistry Comparison and Benchmark Database (Web, free access)   The NIST Computational Chemistry Comparison and Benchmark Database is a collection of experimental and ab initio thermochemical properties for a selected set of molecules. The goals are to provide a benchmark set of molecules for the evaluation of ab initio computational methods and allow the comparison between different ab initio computational methods for the prediction of thermochemical properties.

  8. Pynamic: the Python Dynamic Benchmark

    SciTech Connect

    Lee, G L; Ahn, D H; de Supinksi, B R; Gyllenhaal, J C; Miller, P J

    2007-07-10

    Python is widely used in scientific computing to facilitate application development and to support features such as computational steering. Making full use of some of Python's popular features, which improve programmer productivity, leads to applications that access extremely high numbers of dynamically linked libraries (DLLs). As a result, some important Python-based applications severely stress a system's dynamic linking and loading capabilities and also cause significant difficulties for most development environment tools, such as debuggers. Furthermore, using the Python paradigm for large scale MPI-based applications can create significant file IO and further stress tools and operating systems. In this paper, we present Pynamic, the first benchmark program to support configurable emulation of a wide-range of the DLL usage of Python-based applications for large scale systems. Pynamic has already accurately reproduced system software and tool issues encountered by important large Python-based scientific applications on our supercomputers. Pynamic provided insight for our system software and tool vendors, and our application developers, into the impact of several design decisions. As we describe the Pynamic benchmark, we will highlight some of the issues discovered in our large scale system software and tools using Pynamic.

  9. TRENDS: Compendium of Benchmark Objects

    NASA Astrophysics Data System (ADS)

    Gonzales, Erica J.; Crepp, Justin R.; Bechter, Eric; Johnson, John A.; Montet, Benjamin T.; Howard, Andrew; Marcy, Geoffrey W.; Isaacson, Howard T.

    2016-01-01

    The physical properties of faint stellar and substellar objects are highly uncertain. For example, the masses of brown dwarfs are usually inferred using theoretical models, which are age dependent and have yet to be properly tested. With the goal of identifying new benchmark objects through observations with NIRC2 at Keck, we have carried out a comprehensive adaptive-optics survey as part of the TRENDS (TaRgetting bENchmark-objects with Doppler Spectroscopy) high-contrast imaging program. TRENDS targets nearby (d < 100 pc), Sun-like stars showing long-term radial velocity accelerations. We present the discovery of 28 confirmed, co-moving companions as well as 19 strong candidate companions to F-, G-, and K-stars with well-determined parallaxes and metallicities. Benchmark objects of this nature lend themselves to a three dimensional orbit determination that will ultimately yield a precise dynamical mass. Unambiguous mass measurements of very low mass companions, which straddle the hydrogen-burning boundary, will allow our compendium of objects to serve as excellent testbeds to substantiate theoretical evolutionary and atmospheric models in regimes where they currently breakdown (low temperature, low mass, and old age).

  10. Benchmarking steganographic and steganalysis techniques

    NASA Astrophysics Data System (ADS)

    Kharrazi, Mehdi; Sencar, Husrev T.; Memon, Nasir

    2005-03-01

    There have been a number of steganography embedding techniques proposed over the past few years. In turn the development of these techniques have led to an increased interest in steganalysis techniques. More specifically Universal steganalysis techniques have become more attractive since they work independently of the embedding technique. In this work, our goal is to compare a number of universal steganalysis techniques proposed in the literature which include techniques based on binary similarity measures, wavelet coefficients' statistics, and DCT based image features. These universal steganalysis techniques are tested against a number of well know embedding techniques, including Outguess, F5, Model based, and perturbed quantization. Our experiments are done using a large dataset of JPEG images, obtained by randomly crawling a set of publicly available websites. The image dataset is categorized with respect to the size and quality. We benchmark embedding rate versus detectability performances of several widely used embedding as well as universal steganalysis techniques. Furthermore, we provide a framework for benchmarking future techniques.

  11. Selecting Students for Pre-Algebra: Examination of the Relative Utility of the Anchorage Pre-Algebra Screening Tests and the State of Alaska Standards Based Benchmark 2 Mathematics Study. An Examination of Consequential Validity and Recommendation.

    ERIC Educational Resources Information Center

    Fenton, Ray

    This study examined the relative efficacy of the Anchorage (Alaska) Pre-Algebra Test and the State of Alaska Benchmark in 2 Math examination as tools used in the process of recommending grade 6 students for grade 7 Pre-Algebra placement. The consequential validity of the tests is explored in the context of class placements and grades earned. The

  12. Method and system for benchmarking computers

    DOEpatents

    Gustafson, John L. (Ames, IA)

    1993-09-14

    A testing system and method for benchmarking computer systems. The system includes a store containing a scalable set of tasks to be performed to produce a solution in ever-increasing degrees of resolution as a larger number of the tasks are performed. A timing and control module allots to each computer a fixed benchmarking interval in which to perform the stored tasks. Means are provided for determining, after completion of the benchmarking interval, the degree of progress through the scalable set of tasks and for producing a benchmarking rating relating to the degree of progress for each computer.

  13. Characterizing universal gate sets via dihedral benchmarking

    NASA Astrophysics Data System (ADS)

    Carignan-Dugas, Arnaud; Wallman, Joel J.; Emerson, Joseph

    2015-12-01

    We describe a practical experimental protocol for robustly characterizing the error rates of non-Clifford gates associated with dihedral groups, including small single-qubit rotations. Our dihedral benchmarking protocol is a generalization of randomized benchmarking that relaxes the usual unitary 2-design condition. Combining this protocol with existing randomized benchmarking schemes enables practical universal gate sets for quantum information processing to be characterized in a way that is robust against state-preparation and measurement errors. In particular, our protocol enables direct benchmarking of the ? /8 gate even under the gate-dependent error model that is expected in leading approaches to fault-tolerant quantum computation.

  14. Mock Tribunal in Action: Mock International Criminal Tribunal for the Former Yugoslavia. 10th Grade Lesson. Schools of California Online Resources for Education (SCORE): Connecting California's Classrooms to the World.

    ERIC Educational Resources Information Center

    Fix, Terrance

    In this lesson, students role-play as members of the International Criminal Tribunal for the former Yugoslavia that will bring to trial "Persons Responsible for Serious Violations of International Humanitarian Law." Students represent the following groups: International Criminal Tribunal; Prosecution; Defense; Serbians; Croatians; Bosnian Muslims;

  15. Kauffman Teen Survey. An Annual Report on Teen Health Behaviors: Use of Alcohol, Tobacco, and Other Drugs among 8th-, 10th-, and 12th-Grade Students in Greater Kansas City, 1991-92 to 2000-01.

    ERIC Educational Resources Information Center

    Ewing Marion Kauffman Foundation, Kansas City, MO.

    The Ewing Marion Kauffman Foundation began surveying Kansas City area teens during the 1984-85 school year. The Kauffman Teen Survey now addresses two sets of issues for teens. Teen Health Behaviors, addressed in this report, have been a focus of the survey since its inception. The report focuses on teen use of alcohol, tobacco, and other drugs in

  16. The Zoo, Benchmarks & You: How To Reach the Oregon State Benchmarks with Zoo Resources.

    ERIC Educational Resources Information Center

    2002

    This document aligns Oregon state educational benchmarks and standards with Oregon Zoo resources. Benchmark areas examined include English, mathematics, science, social studies, and career and life roles. Brief descriptions of the programs offered by the zoo are presented. (SOE)

  17. Improving Grading Consistency through Grade Lift Reporting

    ERIC Educational Resources Information Center

    Millet, Ido

    2010-01-01

    We define Grade Lift as the difference between average class grade and average cumulative class GPA. This metric provides an assessment of how lenient the grading was for a given course. In 2006, we started providing faculty members individualized Grade Lift reports reflecting their position relative to an anonymously plotted school-wide

  18. HPC Analytics Support. Requirements for Uncertainty Quantification Benchmarks

    SciTech Connect

    Paulson, Patrick R.; Purohit, Sumit; Rodriguez, Luke R.

    2015-05-01

    This report outlines techniques for extending benchmark generation products so they support uncertainty quantification by benchmarked systems. We describe how uncertainty quantification requirements can be presented to candidate analytical tools supporting SPARQL. We describe benchmark data sets for evaluating uncertainty quantification, as well as an approach for using our benchmark generator to produce data sets for generating benchmark data sets.

  19. COG validation: SINBAD Benchmark Problems

    SciTech Connect

    Lent, E M; Sale, K E; Buck, R M; Descalle, M

    2004-02-23

    We validated COG, a 3D Monte Carlo radiation transport code, against experimental data and MNCP4C simulations from the Shielding Integral Benchmark Archive Database (SINBAD) compiled by RSICC. We modeled three experiments: the Osaka Nickel and Aluminum sphere experiments conducted at the OKTAVIAN facility, and the liquid oxygen experiment conducted at the FNS facility. COG results are in good agreement with experimental data and generally within a few % of MCNP results. There are several possible sources of discrepancy between MCNP and COG results: (1) the cross-section database versions are different, MCNP uses ENDFB VI 1.1 while COG uses ENDFB VIR7, (2) the code implementations are different, and (3) the models may differ slightly. We also limited the use of variance reduction methods when running the COG version of the problems.

  20. Benchmarking Multipacting Simulations in VORPAL

    SciTech Connect

    C. Nieter, C. Roark, P. Stoltz, K. Tian

    2009-05-01

    We will present the results of benchmarking simulations run to test the ability of VORPAL to model multipacting processes in Superconducting Radio Frequency structures. VORPAL is an electromagnetic (FDTD) particle-in-cell simulation code originally developed for applications in plasma and beam physics. The addition of conformal boundaries and algorithms for secondary electron emission allow VORPAL to be applied to multipacting processes. We start with simulations of multipacting between parallel plates where there are well understood theoretical predictions for the frequency bands where multipacting is expected to occur. We reproduce the predicted multipacting bands and demonstrate departures from the theoretical predictions when a more sophisticated model of secondary emission is used. Simulations of existing cavity structures developed at Jefferson National Laboratories will also be presented where we compare results from VORPAL to experimental data.

  1. Benchmarking ICRF simulations for ITER

    SciTech Connect

    R. V. Budny, L. Berry, R. Bilato, P. Bonoli, M. Brambilla, R.J. Dumont, A. Fukuyama, R. Harvey, E.F. Jaeger, E. Lerche, C.K. Phillips, V. Vdovin, J. Wright, and members of the ITPA-IOS

    2010-09-28

    Abstract Benchmarking of full-wave solvers for ICRF simulations is performed using plasma profiles and equilibria obtained from integrated self-consistent modeling predictions of four ITER plasmas. One is for a high performance baseline (5.3 T, 15 MA) DT H-mode plasma. The others are for half-field, half-current plasmas of interest for the pre-activation phase with bulk plasma ion species being either hydrogen or He4. The predicted profiles are used by seven groups to predict the ICRF electromagnetic fields and heating profiles. Approximate agreement is achieved for the predicted heating power partitions for the DT and He4 cases. Profiles of the heating powers and electromagnetic fields are compared.

  2. Energy in the Global Marketplace. Grades 9, 10, 11. Interdisciplinary Student/Teacher Materials in Energy, the Environment, and the Economy.

    ERIC Educational Resources Information Center

    National Science Teachers Association, Washington, DC.

    This instructional unit contains six classroom lessons in which 9th, 10th, or 11th grade social studies students examine the effects of competition among nations and world regions as demand for oil outstrips supply. The overall objective is to help students understand the concept that energy is a commodity to be bought and sold like any other…

  3. The Turn of the Century. Tenth Grade Lesson. Schools of California Online Resources for Education (SCORE): Connecting California's Classrooms to the World.

    ERIC Educational Resources Information Center

    Bartels, Dede

    In this 10th grade social studies and language arts interdisciplinary unit, students research and report on historical figures from the turn of the 20th century. Students are required to work in pairs to learn about famous and common individuals, including Andrew Carnegie, Samuel Gompers, Susan B. Anthony, Thomas Edison, Theodore Roosevelt, Booker…

  4. Adolescent Perceptions of Family Health Behavior: A Tenth Grade Educational Activity To Increase Family Awareness of a Community Cardiovascular Risk Reduction Program.

    ERIC Educational Resources Information Center

    Nader, Philip R.; And Others

    1982-01-01

    Two studies investigated 10th-grade students' health attitudes and practices related to cardiovascular risk. A survey determined students' perceptions of their families' health practices and needs and students' confidence in their ability to modify parents' behavior. The Heart Health Test Telecast (April 1980) was used to stimulate parent-child

  5. easyCBM Beginning Reading Measures: Grades K-1 Alternate Form Reliability and Criterion Validity with the SAT-10. Technical Report #1403

    ERIC Educational Resources Information Center

    Wray, Kraig; Lai, Cheng-Fei; Sez, Leilani; Alonzo, Julie; Tindal, Gerald

    2013-01-01

    We report the results of an alternate form reliability and criterion validity study of kindergarten and grade 1 (N = 84-199) reading measures from the easyCBM assessment system and Stanford Early School Achievement Test/Stanford Achievement Test, 10th edition (SESAT/SAT-10) across 5 time points. The alternate form reliabilities ranged from

  6. A Multiculturally-Responsive Tenth Grade English Curriculum Nurtures an Appreciation for Individual Differences by Immersing Students in the Lives of Others.

    ERIC Educational Resources Information Center

    Alexander, Jayne

    The goal of a practicum was that high school students in non-academic 10th grade English classes would develop an understanding of and appreciation for cultural differences. The English teacher challenged herself, her peers, her students, and the general school population (a regional high school in a growing community in the northeastern United

  7. NASA Software Engineering Benchmarking Study

    NASA Technical Reports Server (NTRS)

    Rarick, Heather L.; Godfrey, Sara H.; Kelly, John C.; Crumbley, Robert T.; Wifl, Joel M.

    2013-01-01

    To identify best practices for the improvement of software engineering on projects, NASA's Offices of Chief Engineer (OCE) and Safety and Mission Assurance (OSMA) formed a team led by Heather Rarick and Sally Godfrey to conduct this benchmarking study. The primary goals of the study are to identify best practices that: Improve the management and technical development of software intensive systems; Have a track record of successful deployment by aerospace industries, universities [including research and development (R&D) laboratories], and defense services, as well as NASA's own component Centers; and Identify candidate solutions for NASA's software issues. Beginning in the late fall of 2010, focus topics were chosen and interview questions were developed, based on the NASA top software challenges. Between February 2011 and November 2011, the Benchmark Team interviewed a total of 18 organizations, consisting of five NASA Centers, five industry organizations, four defense services organizations, and four university or university R and D laboratory organizations. A software assurance representative also participated in each of the interviews to focus on assurance and software safety best practices. Interviewees provided a wealth of information on each topic area that included: software policy, software acquisition, software assurance, testing, training, maintaining rigor in small projects, metrics, and use of the Capability Maturity Model Integration (CMMI) framework, as well as a number of special topics that came up in the discussions. NASA's software engineering practices compared favorably with the external organizations in most benchmark areas, but in every topic, there were ways in which NASA could improve its practices. Compared to defense services organizations and some of the industry organizations, one of NASA's notable weaknesses involved communication with contractors regarding its policies and requirements for acquired software. One of NASA's strengths was its software assurance practices, which seemed to rate well in comparison to the other organizational groups and also seemed to include a larger scope of activities. An unexpected benefit of the software benchmarking study was the identification of many opportunities for collaboration in areas including metrics, training, sharing of CMMI experiences and resources such as instructors and CMMI Lead Appraisers, and even sharing of assets such as documented processes. A further unexpected benefit of the study was the feedback on NASA practices that was received from some of the organizations interviewed. From that feedback, other potential areas where NASA could improve were highlighted, such as accuracy of software cost estimation and budgetary practices. The detailed report contains discussion of the practices noted in each of the topic areas, as well as a summary of observations and recommendations from each of the topic areas. The resulting 24 recommendations from the topic areas were then consolidated to eliminate duplication and culled into a set of 14 suggested actionable recommendations. This final set of actionable recommendations, listed below, are items that can be implemented to improve NASA's software engineering practices and to help address many of the items that were listed in the NASA top software engineering issues. 1. Develop and implement standard contract language for software procurements. 2. Advance accurate and trusted software cost estimates for both procured and in-house software and improve the capture of actual cost data to facilitate further improvements. 3. Establish a consistent set of objectives and expectations, specifically types of metrics at the Agency level, so key trends and models can be identified and used to continuously improve software processes and each software development effort. 4. Maintain the CMMI Maturity Level requirement for critical NASA projects and use CMMI to measure organizations developing software for NASA. 5.onsolidate, collect and, if needed, develop common processes principles and other assets across the Agency in order to provide more consistency in software development and acquisition practices and to reduce the overall cost of maintaining or increasing current NASA CMMI maturity levels. 6. Provide additional support for small projects that includes: (a) guidance for appropriate tailoring of requirements for small projects, (b) availability of suitable tools, including support tool set-up and training, and (c) training for small project personnel, assurance personnel and technical authorities on the acceptable options for tailoring requirements and performing assurance on small projects. 7. Develop software training classes for the more experienced software engineers using on-line training, videos, or small separate modules of training that can be accommodated as needed throughout a project. 8. Create guidelines to structure non-classroom training opportunities such as mentoring, peer reviews, lessons learned sessions, and on-the-job training. 9. Develop a set of predictive software defect data and a process for assessing software testing metric data against it. 10. Assess Agency-wide licenses for commonly used software tools. 11. Fill the knowledge gap in common software engineering practices for new hires and co-ops.12. Work through the Science, Technology, Engineering and Mathematics (STEM) program with universities in strengthening education in the use of common software engineering practices and standards. 13. Follow up this benchmark study with a deeper look into what both internal and external organizations perceive as the scope of software assurance, the value they expect to obtain from it, and the shortcomings they experience in the current practice. 14. Continue interactions with external software engineering environment through collaborations, knowledge sharing, and benchmarking.

  8. Beyond Benchmarking: Value-Adding Metrics

    ERIC Educational Resources Information Center

    Fitz-enz, Jac

    2007-01-01

    HR metrics has grown up a bit over the past two decades, moving away from simple benchmarking practices and toward a more inclusive approach to measuring institutional performance and progress. In this article, the acknowledged "father" of human capital performance benchmarking provides an overview of several aspects of today's HR metrics

  9. Benchmarking Learning and Teaching: Developing a Method

    ERIC Educational Resources Information Center

    Henderson-Smart, Cheryl; Winning, Tracey; Gerzina, Tania; King, Shalinie; Hyde, Sarah

    2006-01-01

    Purpose: To develop a method for benchmarking teaching and learning in response to an institutional need to validate a new program in Dentistry at the University of Sydney, Australia. Design/methodology/approach: After a collaborative partner, University of Adelaide, was identified, the areas of teaching and learning to be benchmarked, PBL

  10. Benchmarking Learning and Teaching: Developing a Method

    ERIC Educational Resources Information Center

    Henderson-Smart, Cheryl; Winning, Tracey; Gerzina, Tania; King, Shalinie; Hyde, Sarah

    2006-01-01

    Purpose: To develop a method for benchmarking teaching and learning in response to an institutional need to validate a new program in Dentistry at the University of Sydney, Australia. Design/methodology/approach: After a collaborative partner, University of Adelaide, was identified, the areas of teaching and learning to be benchmarked, PBL…

  11. A performance benchmark test for geodynamo simulations

    NASA Astrophysics Data System (ADS)

    Matsui, H.; Heien, E. M.

    2013-12-01

    In the last ten years, a number of numerical dynamo models have successfully represented basic characteristics of the geomagnetic field. As new models and numerical methods continue to be developed, it is important to update and extend benchmarks for testing these models. The first dynamo benchmark of Christensen et al. (2001) was applied to models based on spherical harmonic expansion methods. However, only a few groups have reported results of the dynamo benchmark using local methods (Harder and Hansen, 2005; Matsui and Okuda, 2005; Chan et al., 2007) because of the difficulty treating magnetic boundary conditions based on the local methods. On the other hand, spherical harmonics expansion methods perform poorly on massively parallel computers because global data communications are required for the spherical harmonics expansions to evaluate nonlinear terms. We perform benchmark tests to asses various numerical methods for the next generation of geodynamo simulations. The purpose of this benchmark test is to assess numerical geodynamo models on a massively parallel computational platform. To compare among many numerical methods as possible, we consider the model with the insulated magnetic boundary by Christensen et al. (2001) and with the pseudo vacuum magnetic boundary, because the pseudo vacuum boundaries are implemented easier by using the local method than the magnetic insulated boundaries. In the present study, we consider two kinds of benchmarks, so-called accuracy benchmark and performance benchmark. In the accuracy benchmark, we compare the dynamo models by using modest Ekman and Rayleigh numbers proposed by Christensen et. al. (2001). We investigate a required spatial resolution for each dynamo code to obtain less than 1% difference from the suggested solution of the benchmark test using the two magnetic boundary conditions. In the performance benchmark, we investigate computational performance under the same computational environment. We perform these dynamo models on XSEDE TACC Stampede, and investigate computational performance. To simplify the problem, we choose the same model and parameter regime as the accuracy benchmark test, but perform the simulations with much finer spatial resolutions to investigate computational capability under the closer condition to the Earth's outer core. We compare the results of the accuracy benchmark and performance benchmark tests by various codes and discuss characteristics of the simulation methods for geodynamo problems.

  12. A Seafloor Benchmark for 3-dimensional Geodesy

    NASA Astrophysics Data System (ADS)

    Chadwell, C. D.; Webb, S. C.; Nooner, S. L.

    2014-12-01

    We have developed an inexpensive, permanent seafloor benchmark to increase the longevity of seafloor geodetic measurements. The benchmark provides a physical tie to the sea floor lasting for decades (perhaps longer) on which geodetic sensors can be repeatedly placed and removed with millimeter resolution. Global coordinates estimated with seafloor geodetic techniques will remain attached to the benchmark allowing for the interchange of sensors as they fail or become obsolete, or for the sensors to be removed and used elsewhere, all the while maintaining a coherent series of positions referenced to the benchmark. The benchmark has been designed to free fall from the sea surface with transponders attached. The transponder can be recalled via an acoustic command sent from the surface to release from the benchmark and freely float to the sea surface for recovery. The duration of the sensor attachment to the benchmark will last from a few days to a few years depending on the specific needs of the experiment. The recovered sensors are then available to be reused at other locations, or again at the same site in the future. Three pins on the sensor frame mate precisely and unambiguously with three grooves on the benchmark. To reoccupy a benchmark a Remotely Operated Vehicle (ROV) uses its manipulator arm to place the sensor pins into the benchmark grooves. In June 2014 we deployed four benchmarks offshore central Oregon. We used the ROV Jason to successfully demonstrate the removal and replacement of packages onto the benchmark. We will show the benchmark design and its operational capabilities. Presently models of megathrust slip within the Cascadia Subduction Zone (CSZ) are mostly constrained by the sub-aerial GPS vectors from the Plate Boundary Observatory, a part of Earthscope. More long-lived seafloor geodetic measures are needed to better understand the earthquake and tsunami risk associated with a large rupture of the thrust fault within the Cascadia subduction zone. Using a ROV to place and remove sensors on the benchmarks will significantly reduce the number of sensors required by the community to monitor offshore strain in subduction zones.

  13. VENUS-2 MOX Core Benchmark: Results of ORNL Calculations Using HELIOS-1.4

    SciTech Connect

    Ellis, RJ

    2001-02-02

    The Task Force on Reactor-Based Plutonium Disposition, now an Expert Group, was set up through the Organization for Economic Cooperation and Development/Nuclear Energy Agency to facilitate technical assessments of burning weapons-grade plutonium mixed-oxide (MOX) fuel in U.S. pressurized-water reactors and Russian VVER nuclear reactors. More than ten countries participated to advance the work of the Task Force in a major initiative, which was a blind benchmark study to compare code benchmark calculations against experimental data for the VENUS-2 MOX core at SCK-CEN in Mol, Belgium. At the Oak Ridge National Laboratory, the HELIOS-1.4 code was used to perform a comprehensive study of pin-cell and core calculations for the VENUS-2 benchmark.

  14. Grading More Accurately

    ERIC Educational Resources Information Center

    Rom, Mark Carl

    2011-01-01

    Grades matter. College grading systems, however, are often ad hoc and prone to mistakes. This essay focuses on one factor that contributes to high-quality grading systems: grading accuracy (or "efficiency"). I proceed in several steps. First, I discuss the elements of "efficient" (i.e., accurate) grading. Next, I present analytical results…

  15. Consistency and Magnitude of Differences in Reading Curriculum-Based Measurement Slopes in Benchmark versus Strategic Monitoring

    ERIC Educational Resources Information Center

    Mercer, Sterett H.; Keller-Margulis, Milena A.

    2015-01-01

    Differences in oral reading curriculum-based measurement (R-CBM) slopes based on two commonly used progress monitoring practices in field-based data were compared in this study. Semester-specific R-CBM slopes were calculated for 150 Grade 1 and 2 students who completed benchmark (i.e., 3 R-CBM probes collected 3 times per year) and strategic

  16. ICSBEP Benchmarks For Nuclear Data Applications

    SciTech Connect

    Briggs, J. Blair

    2005-05-24

    The International Criticality Safety Benchmark Evaluation Project (ICSBEP) was initiated in 1992 by the United States Department of Energy. The ICSBEP became an official activity of the Organization for Economic Cooperation and Development (OECD) -- Nuclear Energy Agency (NEA) in 1995. Representatives from the United States, United Kingdom, France, Japan, the Russian Federation, Hungary, Republic of Korea, Slovenia, Serbia and Montenegro (formerly Yugoslavia), Kazakhstan, Spain, Israel, Brazil, Poland, and the Czech Republic are now participating. South Africa, India, China, and Germany are considering participation. The purpose of the ICSBEP is to identify, evaluate, verify, and formally document a comprehensive and internationally peer-reviewed set of criticality safety benchmark data. The work of the ICSBEP is published as an OECD handbook entitled ''International Handbook of Evaluated Criticality Safety Benchmark Experiments.'' The 2004 Edition of the Handbook contains benchmark specifications for 3331 critical or subcritical configurations that are intended for use in validation efforts and for testing basic nuclear data. New to the 2004 Edition of the Handbook is a draft criticality alarm / shielding type benchmark that should be finalized in 2005 along with two other similar benchmarks. The Handbook is being used extensively for nuclear data testing and is expected to be a valuable resource for code and data validation and improvement efforts for decades to come. Specific benchmarks that are useful for testing structural materials such as iron, chromium, nickel, and manganese; beryllium; lead; thorium; and 238U are highlighted.

  17. ICSBEP Benchmarks For Nuclear Data Applications

    NASA Astrophysics Data System (ADS)

    Briggs, J. Blair

    2005-05-01

    The International Criticality Safety Benchmark Evaluation Project (ICSBEP) was initiated in 1992 by the United States Department of Energy. The ICSBEP became an official activity of the Organization for Economic Cooperation and Development (OECD) — Nuclear Energy Agency (NEA) in 1995. Representatives from the United States, United Kingdom, France, Japan, the Russian Federation, Hungary, Republic of Korea, Slovenia, Serbia and Montenegro (formerly Yugoslavia), Kazakhstan, Spain, Israel, Brazil, Poland, and the Czech Republic are now participating. South Africa, India, China, and Germany are considering participation. The purpose of the ICSBEP is to identify, evaluate, verify, and formally document a comprehensive and internationally peer-reviewed set of criticality safety benchmark data. The work of the ICSBEP is published as an OECD handbook entitled "International Handbook of Evaluated Criticality Safety Benchmark Experiments." The 2004 Edition of the Handbook contains benchmark specifications for 3331 critical or subcritical configurations that are intended for use in validation efforts and for testing basic nuclear data. New to the 2004 Edition of the Handbook is a draft criticality alarm / shielding type benchmark that should be finalized in 2005 along with two other similar benchmarks. The Handbook is being used extensively for nuclear data testing and is expected to be a valuable resource for code and data validation and improvement efforts for decades to come. Specific benchmarks that are useful for testing structural materials such as iron, chromium, nickel, and manganese; beryllium; lead; thorium; and 238U are highlighted.

  18. Developing integrated benchmarks for DOE performance measurement

    SciTech Connect

    Barancik, J.I.; Kramer, C.F.; Thode, Jr. H.C.

    1992-09-30

    The objectives of this task were to describe and evaluate selected existing sources of information on occupational safety and health with emphasis on hazard and exposure assessment, abatement, training, reporting, and control identifying for exposure and outcome in preparation for developing DOE performance benchmarks. Existing resources and methodologies were assessed for their potential use as practical performance benchmarks. Strengths and limitations of current data resources were identified. Guidelines were outlined for developing new or improved performance factors, which then could become the basis for selecting performance benchmarks. Data bases for non-DOE comparison populations were identified so that DOE performance could be assessed relative to non-DOE occupational and industrial groups. Systems approaches were described which can be used to link hazards and exposure, event occurrence, and adverse outcome factors, as needed to generate valid, reliable, and predictive performance benchmarks. Data bases were identified which contain information relevant to one or more performance assessment categories . A list of 72 potential performance benchmarks was prepared to illustrate the kinds of information that can be produced through a benchmark development program. Current information resources which may be used to develop potential performance benchmarks are limited. There is need to develop an occupational safety and health information and data system in DOE, which is capable of incorporating demonstrated and documented performance benchmarks prior to, or concurrent with the development of hardware and software. A key to the success of this systems approach is rigorous development and demonstration of performance benchmark equivalents to users of such data before system hardware and software commitments are institutionalized.

  19. Benchmark problems in computational aeroacoustics

    NASA Technical Reports Server (NTRS)

    Porter-Locklear, Freda

    1994-01-01

    A recent directive at NASA Langley is aimed at numerically predicting principal noise sources. During my summer stay, I worked with high-order ENO code, developed by Dr. Harold Atkins, for solving the unsteady compressible Navier-Stokes equations, as it applies to computational aeroacoustics (CAA). A CAA workshop, composed of six categories of benchmark problems, has been organized to test various numerical properties of code. My task was to determine the robustness of Atkins' code for these test problems. In one category, we tested the nonlinear wave propagation of the code for the one-dimensional Euler equations, with initial pressure, density, and velocity conditions. Using freestream boundary conditions, our results were plausible. In another category, we solved the linearized two-dimensional Euler equations to test the effectiveness of radiation boundary conditions. Here we utilized MAPLE to compute eigenvalues and eigenvectors of the Jacobian given variable and flux vectors. We experienced a minor problem with inflow and outflow boundary conditions. Next, we solved the quasi one dimensional unsteady flow equations with an incoming acoustic wave of amplitude 10(exp -6). The small amplitude sound wave was incident on a convergent-divergent nozzle. After finding a steady-state solution and then marching forward, our solution indicated that after 30 periods the acoustic wave had dissipated (a period is time required for sound wave to traverse one end of nozzle to other end).

  20. Plans to update benchmarking tool.

    PubMed

    Stokoe, Mark

    2013-02-01

    The use of the current AssetMark system by hospital health facilities managers and engineers (in Australia) has decreased to a point of no activity occurring. A number of reasons have been cited, including cost, time to do, slow process, and level of information required. Based on current levels of activity, it would not be of any value to IHEA, or to its members, to continue with this form of AssetMark. For AssetMark to remain viable, it needs to be developed as a tool seen to be of value to healthcare facilities managers, and not just healthcare facility engineers. Benchmarking is still a very important requirement in the industry, and AssetMark can fulfil this need provided that it remains abreast of customer needs. The proposed future direction is to develop an online version of AssetMark with its current capabilities regarding capturing of data (12 Key Performance Indicators), reporting, and user interaction. The system would also provide end-users with access to live reporting features via a user-friendly web nterface linked through the IHEA web page. PMID:23513367

  1. ASIS healthcare security benchmarking study.

    PubMed

    2001-01-01

    Effective security has aligned itself into the everyday operations of a healthcare organization. This is evident in every regional market segment, regardless of size, location, and provider clinical expertise or organizational growth. This research addresses key security issues from an acute care provider to freestanding facilities, from rural hospitals and community hospitals to large urban teaching hospitals. Security issues and concerns are identified and addressed daily by senior and middle management. As provider campuses become larger and more diverse, the hospitals surveyed have identified critical changes and improvements that are proposed or pending. Mitigating liabilities and improving patient, visitor, and/or employee safety are consequential to the performance and viability of all healthcare providers. Healthcare organizations have identified the requirement to compete for patient volume and revenue. The facility that can deliver high-quality healthcare in a comfortable, safe, secure, and efficient atmosphere will have a significant competitive advantage over a facility where patient or visitor security and safety is deficient. Continuing changes in healthcare organizations' operating structure and healthcare geographic layout mean changes in leadership and direction. These changes have led to higher levels of corporate responsibility. As a result, each organization participating in this benchmark study has added value and will derive value for the overall benefit of the healthcare providers throughout the nation. This study provides a better understanding of how the fundamental security needs of security in healthcare organizations are being addressed and its solutions identified and implemented. PMID:11602980

  2. Metrics and Benchmarks for Visualization

    NASA Technical Reports Server (NTRS)

    Uselton, Samuel P.; Lasinski, T. A. (Technical Monitor)

    1995-01-01

    What is a "good" visualization? How can the quality of a visualization be measured? How can one tell whether one visualization is "better" than another? I claim that the true quality of a visualization can only be measured in the context of a particular purpose. The same image generated from the same data may be excellent for one purpose and abysmal for another. A good measure of visualization quality will correspond to the performance of users in accomplishing the intended purpose, so the "gold standard" is user testing. As a user of visualization software (or at least a consultant to such users) I don't expect visualization software to have been tested in this way for every possible use. In fact, scientific visualization (as distinct from more "production oriented" uses of visualization) will continually encounter new data, new questions and new purposes; user testing can never keep up. User need software they can trust, and advice on appropriate visualizations of particular purposes. Considering the following four processes, and their impact on visualization trustworthiness, reveals important work needed to create worthwhile metrics and benchmarks for visualization. These four processes are (1) complete system testing (user-in-loop), (2) software testing, (3) software design and (4) information dissemination. Additional information is contained in the original extended abstract.

  3. The Multimodal Brain Tumor Image Segmentation Benchmark (BRATS).

    PubMed

    Menze, Bjoern H; Jakab, Andras; Bauer, Stefan; Kalpathy-Cramer, Jayashree; Farahani, Keyvan; Kirby, Justin; Burren, Yuliya; Porz, Nicole; Slotboom, Johannes; Wiest, Roland; Lanczi, Levente; Gerstner, Elizabeth; Weber, Marc-Andr; Arbel, Tal; Avants, Brian B; Ayache, Nicholas; Buendia, Patricia; Collins, D Louis; Cordier, Nicolas; Corso, Jason J; Criminisi, Antonio; Das, Tilak; Delingette, Herv; Demiralp, a?atay; Durst, Christopher R; Dojat, Michel; Doyle, Senan; Festa, Joana; Forbes, Florence; Geremia, Ezequiel; Glocker, Ben; Golland, Polina; Guo, Xiaotao; Hamamci, Andac; Iftekharuddin, Khan M; Jena, Raj; John, Nigel M; Konukoglu, Ender; Lashkari, Danial; Mariz, Jos Antoni; Meier, Raphael; Pereira, Srgio; Precup, Doina; Price, Stephen J; Raviv, Tammy Riklin; Reza, Syed M S; Ryan, Michael; Sarikaya, Duygu; Schwartz, Lawrence; Shin, Hoo-Chang; Shotton, Jamie; Silva, Carlos A; Sousa, Nuno; Subbanna, Nagesh K; Szekely, Gabor; Taylor, Thomas J; Thomas, Owen M; Tustison, Nicholas J; Unal, Gozde; Vasseur, Flor; Wintermark, Max; Ye, Dong Hye; Zhao, Liang; Zhao, Binsheng; Zikic, Darko; Prastawa, Marcel; Reyes, Mauricio; Van Leemput, Koen

    2015-10-01

    In this paper we report the set-up and results of the Multimodal Brain Tumor Image Segmentation Benchmark (BRATS) organized in conjunction with the MICCAI 2012 and 2013 conferences. Twenty state-of-the-art tumor segmentation algorithms were applied to a set of 65 multi-contrast MR scans of low- and high-grade glioma patients-manually annotated by up to four raters-and to 65 comparable scans generated using tumor image simulation software. Quantitative evaluations revealed considerable disagreement between the human raters in segmenting various tumor sub-regions (Dice scores in the range 74%-85%), illustrating the difficulty of this task. We found that different algorithms worked best for different sub-regions (reaching performance comparable to human inter-rater variability), but that no single algorithm ranked in the top for all sub-regions simultaneously. Fusing several good algorithms using a hierarchical majority vote yielded segmentations that consistently ranked above all individual algorithms, indicating remaining opportunities for further methodological improvements. The BRATS image data and manual annotations continue to be publicly available through an online evaluation system as an ongoing benchmarking resource. PMID:25494501

  4. Updates to the integrated protein-protein interaction benchmarks: Docking benchmark version 5 and affinity benchmark version 2

    PubMed Central

    Vreven, Thom; Moal, Iain H.; Vangone, Anna; Pierce, Brian G.; Kastritis, Panagiotis L.; Torchala, Mieczyslaw; Chaleil, Raphael; Jiménez-García, Brian; Bates, Paul A.; Fernandez-Recio, Juan; Bonvin, Alexandre M.J.J.; Weng, Zhiping

    2015-01-01

    We present an updated and integrated version of our widely used protein-protein docking and binding affinity benchmarks. The benchmarks consist of non-redundant, high quality structures of protein-protein complexes along with the unbound structures of their components. Fifty-five new complexes were added to the docking benchmark, 35 of which have experimentally-measured binding affinities. These updated docking and affinity benchmarks now contain 230 and 179 entries, respectively. In particular, the number of antibody-antigen complexes has increased significantly, by 67% and 74% in the docking and affinity benchmarks, respectively. We tested previously developed docking and affinity prediction algorithms on the new cases. Considering only the top ten docking predictions per benchmark case, a prediction accuracy of 38% is achieved on all 55 cases, and up to 50% for the 32 rigid-body cases only. Predicted affinity scores are found to correlate with experimental binding energies up to r=0.52 overall, and r=0.72 for the rigid complexes. PMID:26231283

  5. Updates to the Integrated Protein-Protein Interaction Benchmarks: Docking Benchmark Version 5 and Affinity Benchmark Version 2.

    PubMed

    Vreven, Thom; Moal, Iain H; Vangone, Anna; Pierce, Brian G; Kastritis, Panagiotis L; Torchala, Mieczyslaw; Chaleil, Raphael; Jimnez-Garca, Brian; Bates, Paul A; Fernandez-Recio, Juan; Bonvin, Alexandre M J J; Weng, Zhiping

    2015-09-25

    We present an updated and integrated version of our widely used protein-protein docking and binding affinity benchmarks. The benchmarks consist of non-redundant, high-quality structures of protein-protein complexes along with the unbound structures of their components. Fifty-five new complexes were added to the docking benchmark, 35 of which have experimentally measured binding affinities. These updated docking and affinity benchmarks now contain 230 and 179 entries, respectively. In particular, the number of antibody-antigen complexes has increased significantly, by 67% and 74% in the docking and affinity benchmarks, respectively. We tested previously developed docking and affinity prediction algorithms on the new cases. Considering only the top 10 docking predictions per benchmark case, a prediction accuracy of 38% is achieved on all 55 cases and up to 50% for the 32 rigid-body cases only. Predicted affinity scores are found to correlate with experimental binding energies up to r=0.52 overall and r=0.72 for the rigid complexes. PMID:26231283

  6. Pollution prevention opportunity assessment benchmarking: Recommendations for Hanford

    SciTech Connect

    Engel, J.A.

    1994-05-01

    Pollution Prevention Opportunity Assessments (P2OAs) are an important first step in any pollution prevention program. While P2OAs have been and are being conducted at Hanford, there exists no standard guidance, training, tracking, or systematic approach to identifying and addressing the most important waste streams. The purpose of this paper then is to serve as a guide to the Pollution Prevention group at Westinghouse Hanford in developing and implementing P2OAs at Hanford. By searching the literature and benchmarks other sites and agencies, the best elements from those programs can be incorporated and pitfalls more easily avoided. This search began with the 1988 document that introduces P2OAs (then called Process Waste Assessments, PWAS) by the Environmental Protection Agency. This important document presented the basic framework of P20A features which appeared in almost all later programs. Major Department of Energy programs were also examined, with particular attention to the Defense Programs P20A method of a graded approach, as presented at the Kansas City Plant. The graded approach is a system of conducting P2OAs of varying levels of detail depending on the size and importance of the waste stream. Finally, private industry programs were examined briefly. While all the benchmarked programs had excellent features, it was determined that the size and mission of Hanford precluded lifting any one program for use. Thus, a series of recommendations were made, based on the literature review, in order to begin an extensive program of P2OAs at Hanford. These recommendations are in the areas of: facility Pollution Prevention teams, P20A scope and methodology, guidance documents, training for facilities (and management), technical and informational support, tracking and measuring success, and incentives.

  7. XWeB: The XML Warehouse Benchmark

    NASA Astrophysics Data System (ADS)

    Mahboubi, Hadj; Darmont, Jrme

    With the emergence of XML as a standard for representing business data, new decision support applications are being developed. These XML data warehouses aim at supporting On-Line Analytical Processing (OLAP) operations that manipulate irregular XML data. To ensure feasibility of these new tools, important performance issues must be addressed. Performance is customarily assessed with the help of benchmarks. However, decision support benchmarks do not currently support XML features. In this paper, we introduce the XML Warehouse Benchmark (XWeB), which aims at filling this gap. XWeB derives from the relational decision support benchmark TPC-H. It is mainly composed of a test data warehouse that is based on a unified reference model for XML warehouses and that features XML-specific structures, and its associate XQuery decision support workload. XWeB's usage is illustrated by experiments on several XML database management systems.

  8. DOE Commercial Building Benchmark Models: Preprint

    SciTech Connect

    Torcelini, P.; Deru, M.; Griffith, B.; Benne, K.; Halverson, M.; Winiarski, D.; Crawley, D. B.

    2008-07-01

    To provide a consistent baseline of comparison and save time conducting such simulations, the U.S. Department of Energy (DOE) has developed a set of standard benchmark building models. This paper will provide an executive summary overview of these benchmark buildings, and how they can save building analysts valuable time. Fully documented and implemented to use with the EnergyPlus energy simulation program, the benchmark models are publicly available and new versions will be created to maintain compatibility with new releases of EnergyPlus. The benchmark buildings will form the basis for research on specific building technologies, energy code development, appliance standards, and measurement of progress toward DOE energy goals. Having a common starting point allows us to better share and compare research results and move forward to make more energy efficient buildings.

  9. Toward Scalable Benchmarks for Mass Storage Systems

    NASA Technical Reports Server (NTRS)

    Miller, Ethan L.

    1996-01-01

    This paper presents guidelines for the design of a mass storage system benchmark suite, along with preliminary suggestions for programs to be included. The benchmarks will measure both peak and sustained performance of the system as well as predicting both short- and long-term behavior. These benchmarks should be both portable and scalable so they may be used on storage systems from tens of gigabytes to petabytes or more. By developing a standard set of benchmarks that reflect real user workload, we hope to encourage system designers and users to publish performance figures that can be compared with those of other systems. This will allow users to choose the system that best meets their needs and give designers a tool with which they can measure the performance effects of improvements to their systems.

  10. Benchmarking ENDF/B-VII.0

    NASA Astrophysics Data System (ADS)

    van der Marck, Steven C.

    2006-12-01

    The new major release VII.0 of the ENDF/B nuclear data library has been tested extensively using benchmark calculations. These were based upon MCNP-4C3 continuous-energy Monte Carlo neutronics simulations, together with nuclear data processed using the code NJOY. Three types of benchmarks were used, viz., criticality safety benchmarks, (fusion) shielding benchmarks, and reference systems for which the effective delayed neutron fraction is reported. For criticality safety, more than 700 benchmarks from the International Handbook of Criticality Safety Benchmark Experiments were used. Benchmarks from all categories were used, ranging from low-enriched uranium, compound fuel, thermal spectrum ones (LEU-COMP-THERM), to mixed uranium-plutonium, metallic fuel, fast spectrum ones (MIX-MET-FAST). For fusion shielding many benchmarks were based on IAEA specifications for the Oktavian experiments (for Al, Co, Cr, Cu, LiF, Mn, Mo, Si, Ti, W, Zr), Fusion Neutronics Source in Japan (for Be, C, N, O, Fe, Pb), and Pulsed Sphere experiments at Lawrence Livermore National Laboratory (for 6Li, 7Li, Be, C, N, O, Mg, Al, Ti, Fe, Pb, D 2O, H 2O, concrete, polyethylene and teflon). For testing delayed neutron data more than thirty measurements in widely varying systems were used. Among these were measurements in the Tank Critical Assembly (TCA in Japan) and IPEN/MB-01 (Brazil), both with a thermal spectrum, and two cores in Masurca (France) and three cores in the Fast Critical Assembly (FCA, Japan), all with fast spectra. In criticality safety, many benchmarks were chosen from the category with a thermal spectrum, low-enriched uranium, compound fuel (LEU-COMP-THERM), because this is typical of most current-day reactors, and because these benchmarks were previously underpredicted by as much as 0.5% by most nuclear data libraries (such as ENDF/B-VI.8, JEFF-3.0). The calculated results presented here show that this underprediction is no longer there for ENDF/B-VII.0. The average over 257 benchmarks deviates only 0.017% from the measured benchmark value. Moreover, no clear trends (with e.g. enrichment, lattice pitch, or spectrum) have been observed. Also for fast spectrum benchmarks, both for intermediately or highly enriched uranium and for plutonium, clear improvements are apparent from the calculations. The results for bare assemblies have improved, as well as those with a depleted or natural uranium reflector. On the other hand, the results for plutonium solutions (PU-SOL-THERM) are still high, on average (over 120 benchmarks) roughly 0.6%. Furthermore there still is a bias for a range of benchmarks based on cores in the Zero Power Reactor (ANL) with sizable amounts of tungsten in them. The results for the fusion shielding benchmarks have not changed significantly, compared to ENDF/B-VI.8, for most materials. The delayed neutron testing shows that the values for both thermal and fast spectrum cases are now well predicted, which is an improvement when compared with ENDF/B-VI.8.

  11. Public Relations in Accounting: A Benchmark Study.

    ERIC Educational Resources Information Center

    Pincus, J. David; Pincus, Karen V.

    1987-01-01

    Reports on a national study of one segment of the professional services market: the accounting profession. Benchmark data on CPA firms' attitudes toward and uses of public relations are presented and practical and theoretical/research issues are discussed. (JC)

  12. Numerical methods: Analytical benchmarking in transport theory

    SciTech Connect

    Ganapol, B.D. )

    1988-01-01

    Numerical methods applied to reactor technology have reached a high degree of maturity. Certainly one- and two-dimensional neutron transport calculations have become routine, with several programs available on personal computer and the most widely used programs adapted to workstation and minicomputer computational environments. With the introduction of massive parallelism and as experience with multitasking increases, even more improvement in the development of transport algorithms can be expected. Benchmarking an algorithm is usually not a very pleasant experience for the code developer. Proper algorithmic verification by benchmarking involves the following considerations: (1) conservation of particles, (2) confirmation of intuitive physical behavior, and (3) reproduction of analytical benchmark results. By using today's computational advantages, new basic numerical methods have been developed that allow a wider class of benchmark problems to be considered.

  13. Benchmarking of optical dimerizer systems.

    PubMed

    Pathak, Gopal P; Strickland, Devin; Vrana, Justin D; Tucker, Chandra L

    2014-11-21

    Optical dimerizers are a powerful new class of optogenetic tools that allow light-inducible control of protein-protein interactions. Such tools have been useful for regulating cellular pathways and processes with high spatiotemporal resolution in live cells, and a growing number of dimerizer systems are available. As these systems have been characterized by different groups using different methods, it has been difficult for users to compare their properties. Here, we set about to systematically benchmark the properties of four optical dimerizer systems, CRY2/CIB1, TULIPs, phyB/PIF3, and phyB/PIF6. Using a yeast transcriptional assay, we find significant differences in light sensitivity and fold-activation levels between the red light regulated systems but similar responses between the CRY2/CIB and TULIP systems. Further comparison of the ability of the CRY2/CIB1 and TULIP systems to regulate a yeast MAPK signaling pathway also showed similar responses, with slightly less background activity in the dark observed with CRY2/CIB. In the process of developing this work, we also generated an improved blue-light-regulated transcriptional system using CRY2/CIB in yeast. In addition, we demonstrate successful application of the CRY2/CIB dimerizers using a membrane-tethered CRY2, which may allow for better local control of protein interactions. Taken together, this work allows for a better understanding of the capacities of these different dimerization systems and demonstrates new uses of these dimerizers to control signaling and transcription in yeast. PMID:25350266

  14. Simple Benchmark Specifications for Space Radiation Protection

    NASA Technical Reports Server (NTRS)

    Singleterry, Robert C. Jr.; Aghara, Sukesh K.

    2013-01-01

    This report defines space radiation benchmark specifications. This specification starts with simple, monoenergetic, mono-directional particles on slabs and progresses to human models in spacecraft. This report specifies the models and sources needed to what the team performing the benchmark needs to produce in a report. Also included are brief descriptions of how OLTARIS, the NASA Langley website for space radiation analysis, performs its analysis.

  15. A framework for benchmarking land models

    SciTech Connect

    Luo, Yiqi; Randerson, J.; Abramowitz, G.; Bacour, C.; Blyth, E.; Carvalhais, N.; Ciais, Philippe; Dalmonech, D.; Fisher, J.B.; Fisher, R.; Friedlingstein, P.; Hibbard, Kathleen A.; Hoffman, F. M.; Huntzinger, Deborah; Jones, C.; Koven, C.; Lawrence, David M.; Li, D.J.; Mahecha, M.; Niu, S.L.; Norby, Richard J.; Piao, S.L.; Qi, X.; Peylin, P.; Prentice, I.C.; Riley, William; Reichstein, M.; Schwalm, C.; Wang, Y.; Xia, J. Y.; Zaehle, S.; Zhou, X. H.

    2012-10-09

    Land models, which have been developed by the modeling community in the past few decades to predict future states of ecosystems and climate, have to be critically evaluated for their performance skills of simulating ecosystem responses and feedback to climate change. Benchmarking is an emerging procedure to measure performance of models against a set of defined standards. This paper proposes a benchmarking framework for evaluation of land model performances and, meanwhile, highlights major challenges at this infant stage of benchmark analysis. The framework includes (1) targeted aspects of model performance to be evaluated, (2) a set of benchmarks as defined references to test model performance, (3) metrics to measure and compare performance skills among models so as to identify model strengths and deficiencies, and (4) model improvement. Land models are required to simulate exchange of water, energy, carbon and sometimes other trace gases between the atmosphere and land surface, and should be evaluated for their simulations of biophysical processes, biogeochemical cycles, and vegetation dynamics in response to climate change across broad temporal and spatial scales. Thus, one major challenge is to select and define a limited number of benchmarks to effectively evaluate land model performance. The second challenge is to develop metrics of measuring mismatches between models and benchmarks. The metrics may include (1) a priori thresholds of acceptable model performance and (2) a scoring system to combine datamodel mismatches for various processes at different temporal and spatial scales. The benchmark analyses should identify clues of weak model performance to guide future development, thus enabling improved predictions of future states of ecosystems and climate. The near-future research effort should be on development of a set of widely acceptable benchmarks that can be used to objectively, effectively, and reliably evaluate fundamental properties of land models to improve their prediction performance skills.

  16. A framework for benchmarking land models

    SciTech Connect

    Luo, Yiqi; Randerson, James T.; Hoffman, Forrest; Norby, Richard J

    2012-01-01

    Land models, which have been developed by the modeling community in the past few decades to predict future states of ecosystems and climate, have to be critically evaluated for their performance skills of simulating ecosystem responses and feedback to climate change. Benchmarking is an emerging procedure to measure performance of models against a set of defined standards. This paper proposes a benchmarking framework for evaluation of land model performances and, meanwhile, highlights major challenges at this infant stage of benchmark analysis. The framework includes (1) targeted aspects of model performance to be evaluated, (2) a set of benchmarks as defined references to test model performance, (3) metrics to measure and compare performance skills among models so as to identify model strengths and deficiencies, and (4) model improvement. Land models are required to simulate exchange of water, energy, carbon and sometimes other trace gases between the atmosphere and land surface, and should be evaluated for their simulations of biophysical processes, biogeochemical cycles, and vegetation dynamics in response to climate change across broad temporal and spatial scales. Thus, one major challenge is to select and define a limited number of benchmarks to effectively evaluate land model performance. The second challenge is to develop metrics of measuring mismatches between models and benchmarks. The metrics may include (1) a priori thresholds of acceptable model performance and (2) a scoring system to combine data model mismatches for various processes at different temporal and spatial scales. The benchmark analyses should identify clues of weak model performance to guide future development, thus enabling improved predictions of future states of ecosystems and climate. The near-future research effort should be on development of a set of widely acceptable benchmarks that can be used to objectively, effectively, and reliably evaluate fundamental properties of land models to improve their prediction performance skills.

  17. Implementation of NAS Parallel Benchmarks in Java

    NASA Technical Reports Server (NTRS)

    Frumkin, Michael; Schultz, Matthew; Jin, Hao-Qiang; Yan, Jerry

    2000-01-01

    A number of features make Java an attractive but a debatable choice for High Performance Computing (HPC). In order to gauge the applicability of Java to the Computational Fluid Dynamics (CFD) we have implemented NAS Parallel Benchmarks in Java. The performance and scalability of the benchmarks point out the areas where improvement in Java compiler technology and in Java thread implementation would move Java closer to Fortran in the competition for CFD applications.

  18. EDITORIAL: Selected papers from the 10th International Workshop on Micro and Nanotechnology for Power Generation and Energy Conversion Applications (PowerMEMS 2010) Selected papers from the 10th International Workshop on Micro and Nanotechnology for Power Generation and Energy Conversion Applications (PowerMEMS 2010)

    NASA Astrophysics Data System (ADS)

    Reynaerts, Dominiek; Vullers, Ruud

    2011-10-01

    This special section of Journal of Micromechanics and Microengineering features papers selected from the 10th International Workshop on Micro and Nanotechnology for Power Generation and Energy Conversion Applications (PowerMEMS 2010). The workshop was organized in Leuven, Belgium from 30 November to 3 December 2010 by Katholieke Universiteit Leuven and the imec/Holst Centre. This was a special PowerMEMS Workshop, for several reasons. First of all, we celebrated the 10th anniversary of the workshop: the first PowerMEMS meeting was organized in Sendai, Japan in 2000. None of the organizers or participants of this first meeting could have predicted the impact of the workshop over the next decade. The second reason was that, for the first time, the conference organization spanned two countries: Belgium and the Netherlands. Thanks to the advances in information technology, teams from Katholieke Universiteit Leuven (Belgium) and the imec/Holst Centre in Eindhoven (the Netherlands) have been able to work together seamlessly as one team. The objective of the PowerMEMS Workshop is to stimulate innovation in micro and nanotechnology for power generation and energy conversion applications. Its scope ranges from integrated microelectromechanical systems (MEMS) for power generation, dissipation, harvesting, and management, to novel nanostructures and materials for energy-related applications. True to the objective of the PowerMEMSWorkshop, the 2010 technical program covered a broad range of energy related research, ranging from the nanometer to the millimeter scale, discussed in 5 invited and 52 oral presentations, and 112 posters. This special section includes 14 papers covering vibration energy harvesters, thermal applications and micro power systems. Finally, we wish to express sincere appreciation to the members of the International Steering Committee, the Technical Program Committee and last but not least the Local Organizing Committee. This special issue was edited in collaboration with the staff of IOP Publishing. PowerMEMS 2010 contents Harvesting energy from airflow with a michromachined piezoelectric harvester inside a Helmholtz resonator S P Matova, R Elfrink, R J M Vullers and R van Schaijk Analysis and characterization of triangular electrode structures for electrostatic energy harvestingDaniel Hoffmann, Bernd Folkmer and Yiannos Manoli A smart and self-sufficient frequency tunable vibration energy harvesterC Eichhorn, R Tchagsim, N Wilhelm and P Woias Power output enhancement of a vibration-driven electret generator for wireless sensor applicationsTatsuakira Masaki, Kenji Sakurai, Toru Yokoyama, Masayo Ikuta, Hiroshi Sameshima, Masashi Doi, Tomonori Seki and Masatoshi Oba Harvesting traffic-induced vibrations for structural health monitoring of bridgesT V Galchev, J McCullagh, R L Peterson and K Najafi Dispenser-printed planar thick-film thermoelectric energy generatorsA Chen, D Madan, P K Wright and J W Evans Silicon nanowire arrays as thermoelectric material for a power microgeneratorD Dávila, A Tarancón, M Fernández-Regúlez, C Calaza, M Salleras, A San Paulo and L Fonseca A micro thermal switch with a stiffness-enhanced thermal isolation structureTakashiro Tsukamoto, Masayoshi Esashi and Shuji Tanaka A dielectric liquid contact thermal switch with electrowetting actuationA R McLanahan, C D Richards and R F Richards A self-regulating valve for single-phase liquid cooling of microelectronicsRadu Donose, Michaël De Volder, Jan Peirs and Dominiek Reynaerts A MEMS-enabled 3D zinc-air microbattery with improved discharge characteristics based on a multilayer metallic substructureA Armutlulu, Y Fang, S H Kim, C H Ji, S A Bidstrup Allen and M G Allen Design, fabrication and testing of an air-breathing micro direct methanol fuel cell with compound anode flow fieldLuwen Wang, Yufeng Zhang, Youran Zhao, Zijiang An, Zhiping Zhou and Xiaowei Liu A shadow-mask evaporated pyroMEMS igniterD A de Koninck, D Briand and N F de Rooij Aerodynamic journal bearing with a flexible, damped support operating at 7.2 million DNTobias Waumans, Jan Peirs, Farid Al-Bender and Dominiek Reynaerts Thermoelectric energy harvester on the heated human machineVladimir Leonov

  19. Machine characterization and benchmark performance prediction

    NASA Technical Reports Server (NTRS)

    Saavedra-Barrera, Rafael H.

    1988-01-01

    From runs of standard benchmarks or benchmark suites, it is not possible to characterize the machine nor to predict the run time of other benchmarks which have not been run. A new approach to benchmarking and machine characterization is reported. The creation and use of a machine analyzer is described, which measures the performance of a given machine on FORTRAN source language constructs. The machine analyzer yields a set of parameters which characterize the machine and spotlight its strong and weak points. Also described is a program analyzer, which analyzes FORTRAN programs and determines the frequency of execution of each of the same set of source language operations. It is then shown that by combining a machine characterization and a program characterization, we are able to predict with good accuracy the run time of a given benchmark on a given machine. Characterizations are provided for the Cray-X-MP/48, Cyber 205, IBM 3090/200, Amdahl 5840, Convex C-1, VAX 8600, VAX 11/785, VAX 11/780, SUN 3/50, and IBM RT-PC/125, and for the following benchmark programs or suites: Los Alamos (BMK8A1), Baskett, Linpack, Livermore Loops, Madelbrot Set, NAS Kernels, Shell Sort, Smith, Whetstone and Sieve of Erathostenes.

  20. Benchmarking infrastructure for mutation text mining

    PubMed Central

    2014-01-01

    Background Experimental research on the automatic extraction of information about mutations from texts is greatly hindered by the lack of consensus evaluation infrastructure for the testing and benchmarking of mutation text mining systems. Results We propose a community-oriented annotation and benchmarking infrastructure to support development, testing, benchmarking, and comparison of mutation text mining systems. The design is based on semantic standards, where RDF is used to represent annotations, an OWL ontology provides an extensible schema for the data and SPARQL is used to compute various performance metrics, so that in many cases no programming is needed to analyze results from a text mining system. While large benchmark corpora for biological entity and relation extraction are focused mostly on genes, proteins, diseases, and species, our benchmarking infrastructure fills the gap for mutation information. The core infrastructure comprises (1) an ontology for modelling annotations, (2) SPARQL queries for computing performance metrics, and (3) a sizeable collection of manually curated documents, that can support mutation grounding and mutation impact extraction experiments. Conclusion We have developed the principal infrastructure for the benchmarking of mutation text mining tasks. The use of RDF and OWL as the representation for corpora ensures extensibility. The infrastructure is suitable for out-of-the-box use in several important scenarios and is ready, in its current state, for initial community adoption. PMID:24568600

  1. Benchmarking for Cost Improvement. Final report

    SciTech Connect

    Not Available

    1993-09-01

    The US Department of Energy`s (DOE) Office of Environmental Restoration and Waste Management (EM) conducted the Benchmarking for Cost Improvement initiative with three objectives: Pilot test benchmarking as an EM cost improvement tool; identify areas for cost improvement and recommend actions to address these areas; provide a framework for future cost improvement. The benchmarking initiative featured the use of four principal methods (program classification, nationwide cost improvement survey, paired cost comparison and component benchmarking). Interested parties contributed during both the design and execution phases. The benchmarking initiative was conducted on an accelerated basis. Of necessity, it considered only a limited set of data that may not be fully representative of the diverse and complex conditions found at the many DOE installations. The initiative generated preliminary data about cost differences and it found a high degree of convergence on several issues. Based on this convergence, the report recommends cost improvement strategies and actions. This report describes the steps taken as part of the benchmarking initiative and discusses the findings and recommended actions for achieving cost improvement. The results and summary recommendations, reported below, are organized by the study objectives.

  2. Action-Oriented Benchmarking: Concepts and Tools

    SciTech Connect

    California Energy Commission; Mathew, Paul; Mills, Evan; Mathew, Paul; Piette, Mary Ann; Bourassa, Norman; Brook, Martha

    2008-02-13

    Most energy benchmarking tools provide static feedback on how one building compares to a larger set of loosely similar buildings, without providing information at the end-use level or on what can be done to reduce consumption, cost, or emissions. In this article--Part 1 of a two-part series--we describe an 'action-oriented benchmarking' approach, which extends whole-building energy benchmarking to include analysis of system and component energy use metrics and features. Action-oriented benchmarking thereby allows users to generate more meaningful metrics and to identify, screen and prioritize potential efficiency improvements. This opportunity assessment process can then be used to inform and optimize a full-scale audit or commissioning process. We introduce a new web-based action-oriented benchmarking system and associated software tool-EnergyIQ. The benchmarking methods, visualizations, and user interface design are informed by an end-user needs assessment survey and best-practice guidelines from ASHRAE.

  3. VENUS-2 Experimental Benchmark Analysis

    SciTech Connect

    Pavlovichev, A.M.

    2001-09-28

    The VENUS critical facility is a zero power reactor located at SCK-CEN, Mol, Belgium, which for the VENUS-2 experiment utilized a mixed-oxide core with near-weapons-grade plutonium. In addition to the VENUS-2 Core, additional computational variants based on each type of fuel cycle VENUS-2 core (3.3 wt. % UO{sub 2}, 4.0 wt. % UO{sub 2}, and 2.0/2.7 wt.% MOX) were also calculated. The VENUS-2 critical configuration and cell variants have been calculated with MCU-REA, which is a continuous energy Monte Carlo code system developed at Russian Research Center ''Kurchatov Institute'' and is used extensively in the Fissile Materials Disposition Program. The calculations resulted in a k{sub eff} of 0.99652 {+-} 0.00025 and relative pin powers within 2% for UO{sub 2} pins and 3% for MOX pins of the experimental values.

  4. 42 CFR 440.330 - Benchmark health benefits coverage.

    Code of Federal Regulations, 2010 CFR

    2010-10-01

    ... 42 Public Health 4 2010-10-01 2010-10-01 false Benchmark health benefits coverage. 440.330 Section 440.330 Public Health CENTERS FOR MEDICARE & MEDICAID SERVICES, DEPARTMENT OF HEALTH AND HUMAN... Benchmark-Equivalent Coverage § 440.330 Benchmark health benefits coverage. Benchmark coverage is...

  5. 42 CFR 440.330 - Benchmark health benefits coverage.

    Code of Federal Regulations, 2011 CFR

    2011-10-01

    ... 42 Public Health 4 2011-10-01 2011-10-01 false Benchmark health benefits coverage. 440.330 Section 440.330 Public Health CENTERS FOR MEDICARE & MEDICAID SERVICES, DEPARTMENT OF HEALTH AND HUMAN... Benchmark-Equivalent Coverage § 440.330 Benchmark health benefits coverage. Benchmark coverage is...

  6. 42 CFR 440.330 - Benchmark health benefits coverage.

    Code of Federal Regulations, 2012 CFR

    2012-10-01

    ... 42 Public Health 4 2012-10-01 2012-10-01 false Benchmark health benefits coverage. 440.330 Section 440.330 Public Health CENTERS FOR MEDICARE & MEDICAID SERVICES, DEPARTMENT OF HEALTH AND HUMAN... Benchmark-Equivalent Coverage § 440.330 Benchmark health benefits coverage. Benchmark coverage is...

  7. Grading. Research Brief

    ERIC Educational Resources Information Center

    Walker, Karen

    2006-01-01

    What do grades mean? Is there a commonly understood and acceptable definition for what a grade stands? The current grading system is approximately 100 years old. In the 1700s, students were given feedback but not grades. In 1780, Yale University began using a 4.0 system that is similar to the current system. In the 1830s, Harvard University

  8. Storage-Intensive Supercomputing Benchmark Study

    SciTech Connect

    Cohen, J; Dossa, D; Gokhale, M; Hysom, D; May, J; Pearce, R; Yoo, A

    2007-10-30

    Critical data science applications requiring frequent access to storage perform poorly on today's computing architectures. This project addresses efficient computation of data-intensive problems in national security and basic science by exploring, advancing, and applying a new form of computing called storage-intensive supercomputing (SISC). Our goal is to enable applications that simply cannot run on current systems, and, for a broad range of data-intensive problems, to deliver an order of magnitude improvement in price/performance over today's data-intensive architectures. This technical report documents much of the work done under LDRD 07-ERD-063 Storage Intensive Supercomputing during the period 05/07-09/07. The following chapters describe: (1) a new file I/O monitoring tool iotrace developed to capture the dynamic I/O profiles of Linux processes; (2) an out-of-core graph benchmark for level-set expansion of scale-free graphs; (3) an entity extraction benchmark consisting of a pipeline of eight components; and (4) an image resampling benchmark drawn from the SWarp program in the LSST data processing pipeline. The performance of the graph and entity extraction benchmarks was measured in three different scenarios: data sets residing on the NFS file server and accessed over the network; data sets stored on local disk; and data sets stored on the Fusion I/O parallel NAND Flash array. The image resampling benchmark compared performance of software-only to GPU-accelerated. In addition to the work reported here, an additional text processing application was developed that used an FPGA to accelerate n-gram profiling for language classification. The n-gram application will be presented at SC07 at the High Performance Reconfigurable Computing Technologies and Applications Workshop. The graph and entity extraction benchmarks were run on a Supermicro server housing the NAND Flash 40GB parallel disk array, the Fusion-io. The Fusion system specs are as follows: SuperMicro X7DBE Xeon Dual Socket Blackford Server Motherboard; 2 Intel Xeon Dual-Core 2.66 GHz processors; 1 GB DDR2 PC2-5300 RAM (2 x 512); 80GB Hard Drive (Seagate SATA II Barracuda). The Fusion board is presently capable of 4X in a PCIe slot. The image resampling benchmark was run on a dual Xeon workstation with NVIDIA graphics card (see Chapter 5 for full specification). An XtremeData Opteron+FPGA was used for the language classification application. We observed that these benchmarks are not uniformly I/O intensive. The only benchmark that showed greater that 50% of the time in I/O was the graph algorithm when it accessed data files over NFS. When local disk was used, the graph benchmark spent at most 40% of its time in I/O. The other benchmarks were CPU dominated. The image resampling benchmark and language classification showed order of magnitude speedup over software by using co-processor technology to offload the CPU-intensive kernels. Our experiments to date suggest that emerging hardware technologies offer significant benefit to boosting the performance of data-intensive algorithms. Using GPU and FPGA co-processors, we were able to improve performance by more than an order of magnitude on the benchmark algorithms, eliminating the processor bottleneck of CPU-bound tasks. Experiments with a prototype solid state nonvolative memory available today show 10X better throughput on random reads than disk, with a 2X speedup on a graph processing benchmark when compared to the use of local SATA disk.

  9. Test Nationally, Benchmark Locally: Using Local DIBELS Benchmarks to Predict Performance on the Pssa

    ERIC Educational Resources Information Center

    Ferchalk, Matthew R.

    2013-01-01

    The Dynamic Indicators of Basic Early Literacy Skills (DIBELS) benchmarks are frequently used to make important decision regarding student performance. More information, however, is needed to understand if the nationally-derived benchmarks created by the DIBELS system provide the most accurate criterion for evaluating reading proficiency. The

  10. Memorization techniques: Using mnemonics to learn fifth grade science terms

    NASA Astrophysics Data System (ADS)

    Garcia, Juan O.

    The purpose of this study was to determine whether mnemonic instruction could assist students in learning fifth-grade science terminology more effectively than traditional-study methods of recall currently in practice The task was to examine if fifth-grade students were able to learn a mnemonic and then use it to understand science vocabulary; subsequently, to determine if students were able to remember the science terms after a period of time. The problem is that in general, elementary school students are not being successful in science achievement at the fifth grade level. In view of this problem, if science performance is increased at the elementary level, then it is likely that students will be successful when tested at the 8th and 10th grade in science with the Texas Assessment of Knowledge and Skills (TAKS) in the future. Two research questions were posited: (1) Is there a difference in recall achievement when a mnemonic such as method of loci, pegword method, or keyword method is used in learning fifth-grade science vocabulary as compared to the traditional-study method? (2) If using a mnemonic in learning fifth-grade science vocabulary was effective on recall achievement, would this achievement be maintained over a span of time? The need for this study was to assist students in learning science terms and concepts for state accountability purposes. The first assumption was that memorization techniques are not commonly applied in fifth-grade science classes in elementary schools. A second assumption was that mnemonic devices could be used successfully in learning science terms and increase long term retention. The first limitation was that the study was conducted on one campus in one school district in South Texas which limited the generalization of the study. The second limitation was that it included random assigned intact groups as opposed to random student assignment to fifth-grade classroom groups.

  11. Action-Oriented Benchmarking: Using the CEUS Database to Benchmark Commercial Buildings in California

    SciTech Connect

    Mathew, Paul; Mills, Evan; Bourassa, Norman; Brook, Martha

    2008-02-01

    The 2006 Commercial End Use Survey (CEUS) database developed by the California Energy Commission is a far richer source of energy end-use data for non-residential buildings than has previously been available and opens the possibility of creating new and more powerful energy benchmarking processes and tools. In this article--Part 2 of a two-part series--we describe the methodology and selected results from an action-oriented benchmarking approach using the new CEUS database. This approach goes beyond whole-building energy benchmarking to more advanced end-use and component-level benchmarking that enables users to identify and prioritize specific energy efficiency opportunities - an improvement on benchmarking tools typically in use today.

  12. Toxicological benchmarks for wildlife: 1994 Revision

    SciTech Connect

    Opresko, D.M.; Sample, B.E.; Suter, G.W. II

    1994-09-01

    The process by which ecological risks of environmental contaminants are evaluated is two-tiered. The first tier is a screening assessment where concentrations of contaminants in the environment are compared to toxicological benchmarks which represent concentrations of chemicals in environmental media (water, sediment, soil, food, etc.) that are presumed to be nonhazardous to the surrounding biota. The second tier is a baseline ecological risk assessment where toxicological benchmarks are one of several lines of evidence used to support or refute the presence of ecological effects. The report presents toxicological benchmarks for assessment of effects of 76 chemicals on 8 representative mammalian wildlife species and 31 chemicals on 9 avian wildlife species. The chemicals are some of those that occur at United States Department of Energy waste sites; the wildlife species were chosen because they are widely distributed and provide a representative range of body sizes and diets. Further descriptions of the chosen wildlife species and chemicals are provided in the report. The benchmarks presented in this report represent values believed to be nonhazardous for the listed wildlife species. These benchmarks only consider contaminant exposure through oral ingestion of contaminated media; exposure through inhalation or direct dermal exposure are not considered in this report.

  13. Energy benchmarking of South Australian WWTPs.

    PubMed

    Krampe, J

    2013-01-01

    Optimising the energy consumption and energy generation of wastewater treatment plants (WWTPs) is a topic with increasing importance for water utilities in times of rising energy costs and pressures to reduce greenhouse gas (GHG) emissions. Assessing the energy efficiency and energy optimisation of a WWTP are difficult tasks as most plants vary greatly in size, process layout and other influencing factors. To overcome these limits it is necessary to compare energy efficiency with a statistically relevant base to identify shortfalls and optimisation potential. Such energy benchmarks have been successfully developed and used in central Europe over the last two decades. This paper demonstrates how the latest available energy benchmarks from Germany have been applied to 24 WWTPs in South Australia. It shows how energy benchmarking can be used to identify shortfalls in current performance, prioritise detailed energy assessments and help inform decisions on capital investment. PMID:23656950

  14. NAS Parallel Benchmarks, Multi-Zone Versions

    NASA Technical Reports Server (NTRS)

    vanderWijngaart, Rob F.; Haopiang, Jin

    2003-01-01

    We describe an extension of the NAS Parallel Benchmarks (NPB) suite that involves solving the application benchmarks LU, BT and SP on collections of loosely coupled discretization meshes. The solutions on the meshes are updated independently, but after each time step they exchange boundary value information. This strategy, which is common among structured-mesh production flow solver codes in use at NASA Ames and elsewhere, provides relatively easily exploitable coarse-grain parallelism between meshes. Since the individual application benchmarks also allow fine-grain parallelism themselves, this NPB extension, named NPB Multi-Zone (NPB-MZ), is a good candidate for testing hybrid and multi-level parallelization tools and strategies.

  15. Benchmarking criticality safety calculations with subcritical experiments

    SciTech Connect

    Mihalczo, J.T.

    1984-06-01

    Calculation of the neutron multiplication factor at delayed criticality may be necessary for benchmarking calculations but it may not be sufficient. The use of subcritical experiments to benchmark criticality safety calculations could result in substantial savings in fuel material costs for experiments. In some cases subcritical configurations could be used to benchmark calculations where sufficient fuel to achieve delayed criticality is not available. By performing a variety of measurements with subcritical configurations, much detailed information can be obtained which can be compared directly with calculations. This paper discusses several measurements that can be performed with subcritical assemblies and presents examples that include comparisons between calculation and experiment where possible. Where not, examples from critical experiments have been used but the measurement methods could also be used for subcritical experiments.

  16. Benchmark solutions for MHD solver development

    NASA Astrophysics Data System (ADS)

    Tao, Zhen; Ni, MingJiu

    2013-02-01

    A benchmark solution is of great importance in validating algorithms and codes for magnetohydrodynamic (MHD) flows. Hunt and Shercliff's solutions are usually employed as benchmarks for MHD flows in a duct with insulated walls or with thin conductive walls, in which wall effects on MHD are represented by the wall conductance ratio. With wall thickness resolved, it is stressed that the solution of Sloan and Smith's and the solution of Butler's can be used to check the error of the thin wall approximation condition used for Hunt's solutions. It is noted that Tao and Ni's solutions can be used as a benchmark for MHD flows in a duct with wall symmetrical or unsymmetrical, thick or thin. When the walls are symmetrical, Tao and Ni's solutions are reduced to Sloan and Smith's solution and Butler's solution, respectively.

  17. Benchmark field study of deep neutron penetration

    SciTech Connect

    Morgan, J.F.; Sale, K. ); Gold, R.; Roberts, J.H.; Preston, C.C. )

    1991-06-10

    A unique benchmark neutron field has been established at the Lawrence Livermore National Laboratory (LLNL) to study deep penetration neutron transport. At LLNL, a tandem accelerator is used to generate a monoenergetic neutron source that permits investigation of deep neutron penetration under conditions that are virtually ideal to model, namely the transport of mono-energetic neutrons through a single material in a simple geometry. General features of the Lawrence Tandem (LATAN) benchmark field are described with emphasis on neutron source characteristics and room return background. The single material chosen for the first benchmark, LATAN-1, is a steel representative of Light Water Reactor (LWR) Pressure Vessels (PV). Also included is a brief description of the Little Boy replica, a critical reactor assembly designed to mimic the radiation doses from the atomic bomb dropped on Hiroshima, and its us in neutron spectrometry. 18 refs.

  18. Benchmark field study of deep neutron penetration

    NASA Astrophysics Data System (ADS)

    Morgan, J. F.; Sale, K.; Gold, R.; Roberts, J. H.; Preston, C. C.

    1991-06-01

    A unique benchmark neutron field has been established at the Lawrence Livermore National Laboratory (LLNL) to study deep penetration neutron transport. At LLNL, a tandem accelerator is used to generate a monoenergetic neutron source that permits investigation of deep neutron penetration under conditions that are virtually ideal to model, namely the transport of mono-energetic neutrons through a single material in a simple geometry. General features of the Lawrence Tandem (LATAN) benchmark field are described with emphasis on neutron source characteristics and room return background. The single material chosen for the first benchmark, LATAN-1, is a steel representative of Light Water Reactor (LWR) Pressure Vessels (PV). Also included is a brief description of the Little Boy replica, a critical reactor assembly designed to mimic the radiation doses from the atomic bomb dropped on Hiroshima, and its us in neutron spectrometry.

  19. Algebra for All: California’s Eighth-Grade Algebra Initiative as Constrained Curricula

    PubMed Central

    Domina, Thurston; Penner, Andrew M.; Penner, Emily K.; Conley, Annemarie

    2015-01-01

    Background/Context Across the United States, secondary school curricula are intensifying as a growing proportion of students enroll in high-level academic math courses. In many districts, this intensification process occurs as early as eighth grade, where schools are effectively constraining their mathematics curricula by restricting course offerings and placing more students into Algebra I. This paper provides a quantitative single-case research study of policy-driven curricular intensification in one California school district. Research Questions (1a) What effect did 8th eighth grade curricular intensification have on mathematics course enrollment patterns in Towering Pines Unified schools? (2b) How did the distribution of prior achievement in Towering Pines math classrooms change as the district constrained the curriculum by universalizing 8th eighth grade Algebra? (3c) Did 8th eighth grade curricular intensification improve students’ mathematics achievement? Setting Towering Pines is an immigrant enclave in the inner-ring suburbs of a major metropolitan area. The district’s 10 middle schools together enroll approximately 4,000 eighth graders each year. The districts’ students are ethnically diverse and largely economically disadvantaged. The study draws upon administrative data describing 8th eighth graders in the district in the 2004–20-05 through 2007–20-08 school years. Intervention/Program/Practice During the study period, Towering Pines dramatically intensified middle school students’ math curricula: In the 2004–20-05 school year 32% of the district’s 8th eighth graders enrolled in Algebra or a higher- level mathematics course; by the 2007–20-08 school year that proportion had increased to 84%. Research Design We use an interrupted time-series design, comparing students’ 8th eighth grade math course enrollments, 10th grade math course enrollments, and 10th grade math test scores across the four cohorts, controlling for demographics and prior achievement. Findings/Results We find that students’ odds of taking higher level mathematics courses increased as this district implemented the state’s Algebra mandate. However, even as the district implemented a constrained curriculum strategy, mathematics achievement growth between 6th sixth and 10th grade slowed and the achievement advantages associated with 8th eighth grade Algebra declined. Conclusions/Recommendations Our analyses suggest that curricular intensification increased the inclusiveness and decreased the selectivity of the mathematics tracking regime in Towering Pines middle schools. However, the findings suggest that this constrained curriculum strategy may have may have unintended negative consequences for student achievement. PMID:26120219

  20. Benchmarks for the point kinetics equations

    SciTech Connect

    Ganapol, B.; Picca, P.; Previti, A.; Mostacci, D.

    2013-07-01

    A new numerical algorithm is presented for the solution to the point kinetics equations (PKEs), whose accurate solution has been sought for over 60 years. The method couples the simplest of finite difference methods, a backward Euler, with Richardsons extrapolation, also called an acceleration. From this coupling, a series of benchmarks have emerged. These include cases from the literature as well as several new ones. The novelty of this presentation lies in the breadth of reactivity insertions considered, covering both prescribed and feedback reactivities, and the extreme 8- to 9- digit accuracy achievable. The benchmarks presented are to provide guidance to those who wish to develop further numerical improvements. (authors)

  1. Benchmark testing of {sup 233}U evaluations

    SciTech Connect

    Wright, R.Q.; Leal, L.C.

    1997-07-01

    In this paper we investigate the adequacy of available {sup 233}U cross-section data (ENDF/B-VI and JENDL-3) for calculation of critical experiments. An ad hoc revised {sup 233}U evaluation is also tested and appears to give results which are improved relative to those obtained with either ENDF/B-VI or JENDL-3 cross sections. Calculations of k{sub eff} were performed for ten fast benchmarks and six thermal benchmarks using the three cross-section sets. Central reaction-rate-ratio calculations were also performed.

  2. Clinical benchmarking: implications for perinatal nursing.

    PubMed

    Fitzgerald, K

    1998-06-01

    Health care is a dynamic environment where expectations of quality must be balanced with appropriateness of treatment and cost of care. Managers often have inadequate information on which to base decisions, policy, and practice. Clinical benchmarking is a tool and a process of continuously comparing the practices and performances of one's operations against those of the best in the industry or the focused area of service and then using that information to enhance and improve performance and productivity. The article discusses the advantages and disadvantages of benchmarking as well as the factors influencing the need for such tools in health care and in perinatal nursing. PMID:9782874

  3. Los Alamos National Laboratory computer benchmarking 1982

    SciTech Connect

    Martin, J.L.

    1983-06-01

    Evaluating the performance of computing machinery is a continual effort of the Computer Research and Applications Group of the Los Alamos National Laboratory. This report summarizes the results of the group's benchmarking activities performed between October 1981 and September 1982, presenting compilation and execution times as well as megaflop rates for a set of benchmark codes. Tests were performed on the following computers: Cray Research, Inc. (CRI) Cray-1S; Control Data Corporation (CDC) 7600, 6600, Cyber 73, Cyber 825, Cyber 835, Cyber 855, and Cyber 205; Digital Equipment Corporation (DEC) VAX 11/780 and VAX 11/782; and Apollo Computer, Inc., Apollo.

  4. Toxicological benchmarks for wildlife: 1996 Revision

    SciTech Connect

    Sample, B.E.; Opresko, D.M.; Suter, G.W., II

    1996-06-01

    The purpose of this report is to present toxicological benchmarks for assessment of effects of certain chemicals on mammalian and avian wildlife species. Publication of this document meets a milestone for the Environmental Restoration (ER) Risk Assessment Program. This document provides the ER Program with toxicological benchmarks that may be used as comparative tools in screening assessments as well as lines of evidence to support or refute the presence of ecological effects in ecological risk assessments. The chemicals considered in this report are some that occur at US DOE waste sites, and the wildlife species evaluated herein were chosen because they represent a range of body sizes and diets.

  5. Three dimensional transport benchmark exercise using THREEDANT

    SciTech Connect

    Alcouffe, R.E.

    1990-01-01

    As part of the effort to assess the ability to perform three dimensional transport calculations to solve problems in reactor physics, we describe the THREEDANT code and its application to the set of three-dimensional benchmark problems proposed by Prof. T. Takeda. As part of this benchmarking activity, we display some key indicators as to computational performance and efficiency while displaying the sensitivity of the eigenvalue to Sn order and to spatial mesh size in each of the problems. In order to understand what is being displayed, we summarize the solution strategy incorporated in the code. 1 ref., 3 figs.

  6. Overview of TPC Benchmark E: The Next Generation of OLTP Benchmarks

    NASA Astrophysics Data System (ADS)

    Hogan, Trish

    Set to replace the aging TPC-C, the TPC Benchmark E is the next generation OLTP benchmark, which more accurately models client database usage. TPC-E addresses the shortcomings of TPC-C. It has a much more complex workload, requires the use of RAID-protected storage, generates much less I/O, and is much cheaper and easier to set up, run, and audit. After a period of overlap, it is expected that TPC-E will become the de facto OLTP benchmark.

  7. Highlights of 10th plasma chemistry meeting

    NASA Technical Reports Server (NTRS)

    Kitamura, K.; Hashimoto, H.; Hozumi, K.

    1981-01-01

    The chemical structure is given of a film formed by plasma polymerization from pyridine monomers. The film has a hydrophilic chemical structure, its molecular weight is 900, and the molecular system is C55H50N10O3. The electrical characteristics of a plasma polymerized film are described. The film has good insulating properties and was successfully applied as video disc coating. Etching resistance properties make it possible to use the film as a resist in etching. The characteristics of plasma polymer formed from monomers containing tetramethyltin are discussed. The polymer is in film form, displays good adhesiveness, is similar to UV film UV 35 in light absorption and is highly insulating.

  8. Part C Updates. 10th Edition

    ERIC Educational Resources Information Center

    Goode, Sue; Lazara, Alex; Danaher, Joan

    2008-01-01

    "Part C Updates" is a compilation of information on various aspects of the Early Intervention Program for Infants and Toddlers with Disabilities (Part C) of the Individuals with Disabilities Education Act (IDEA). This is the tenth volume in a series of compilations, which included two editions of Part H Updates, the former name of the program.…

  9. Win That Job! 10th Anniversary Edition.

    ERIC Educational Resources Information Center

    Stevens, Paul

    This book provides practical information on obtaining a job. Though it is published in Australia, 11 chapters introduce a universal range of job search methods, presenting: the importance of goals and self-knowledge; the resume; preparing job search correspondence; the interview; self-promotion; job search tips and unusual strategies; networking;

  10. 10th Anniversary P.S.

    ScienceCinema

    None

    2011-04-25

    John Adams parle de la préhistoire du P.S. avec présentation des dias. Le DG B.Gregory prend la parole. Les organisateurs présentent sous la direction du "Prof.Ocktette"(?) un sketch très humoristique (p.e.existence de Quark etc.....)

  11. NAS Parallel Benchmarks Results 3-95

    NASA Technical Reports Server (NTRS)

    Saini, Subhash; Bailey, David H.; Walter, Howard (Technical Monitor)

    1995-01-01

    The NAS Parallel Benchmarks (NPB) were developed in 1991 at NASA Ames Research Center to study the performance of parallel supercomputers. The eight benchmark problems are specified in a "pencil and paper" fashion, i.e., the complete details of the problem are given in a NAS technical document. Except for a few restrictions, benchmark implementors are free to select the language constructs and implementation techniques best suited for a particular system. In this paper, we present new NPB performance results for the following systems: (a) Parallel-Vector Processors: CRAY C90, CRAY T90 and Fujitsu VPP500; (b) Highly Parallel Processors: CRAY T3D, IBM SP2-WN (Wide Nodes), and IBM SP2-TN2 (Thin Nodes 2); and (c) Symmetric Multiprocessors: Convex Exemplar SPPIOOO, CRAY J90, DEC Alpha Server 8400 5/300, and SGI Power Challenge XL (75 MHz). We also present sustained performance per dollar for Class B LU, SP and BT benchmarks. We also mention future NAS plans for the NPB.

  12. Benchmarking Peer Production Mechanisms, Processes & Practices

    ERIC Educational Resources Information Center

    Fischer, Thomas; Kretschmer, Thomas

    2008-01-01

    This deliverable identifies key approaches for quality management in peer production by benchmarking peer production practices and processes in other areas. (Contains 29 footnotes, 13 figures and 2 tables.)[This report has been authored with contributions of: Kaisa Honkonen-Ratinen, Matti Auvinen, David Riley, Jose Pinzon, Thomas Fischer, Thomas

  13. What Is the Impact of Subject Benchmarking?

    ERIC Educational Resources Information Center

    Pidcock, Steve

    2006-01-01

    The introduction of subject benchmarking led to fears of increased external intervention in the activities of universities and a more restrictive view of institutional autonomy, accompanied by an undermining of the academic profession, particularly through the perceived threat of the introduction of a national curriculum for higher education. For…

  14. Cleanroom Energy Efficiency: Metrics and Benchmarks

    SciTech Connect

    International SEMATECH Manufacturing Initiative; Mathew, Paul A.; Tschudi, William; Sartor, Dale; Beasley, James

    2010-07-07

    Cleanrooms are among the most energy-intensive types of facilities. This is primarily due to the cleanliness requirements that result in high airflow rates and system static pressures, as well as process requirements that result in high cooling loads. Various studies have shown that there is a wide range of cleanroom energy efficiencies and that facility managers may not be aware of how energy efficient their cleanroom facility can be relative to other cleanroom facilities with the same cleanliness requirements. Metrics and benchmarks are an effective way to compare one facility to another and to track the performance of a given facility over time. This article presents the key metrics and benchmarks that facility managers can use to assess, track, and manage their cleanroom energy efficiency or to set energy efficiency targets for new construction. These include system-level metrics such as air change rates, air handling W/cfm, and filter pressure drops. Operational data are presented from over 20 different cleanrooms that were benchmarked with these metrics and that are part of the cleanroom benchmark dataset maintained by Lawrence Berkeley National Laboratory (LBNL). Overall production efficiency metrics for cleanrooms in 28 semiconductor manufacturing facilities in the United States and recorded in the Fabs21 database are also presented.

  15. Benchmark Results for Delayed Neutron Data

    SciTech Connect

    Marck, S.C. van der; Meulekamp, R. Klein; Hogenbirk, A.; Koning, A.J.

    2005-05-24

    We have calculated the effective delayed neutron fraction {beta}eff for 32 benchmark configurations for which measurements have been reported. We use these results to test the delayed neutron data of JEFF-3.0, ENDF/B-VI.8, and JENDL-3.3.

  16. Simple Benchmark for Complex Dose Finding Studies

    PubMed Central

    Cheung, Ying Kuen

    2014-01-01

    Summary While a general goal of early phase clinical studies is to identify an acceptable dose for further investigation, modern dose finding studies and designs are highly specific to individual clinical settings. In addition, as outcome-adaptive dose finding methods often involve complex algorithms, it is crucial to have diagnostic tools to evaluate the plausibility of a methods simulated performance and the adequacy of the algorithm. In this article, we propose a simple technique that provides an upper limit, or a benchmark, of accuracy for dose finding methods for a given design objective. The proposed benchmark is nonparametric optimal in the sense of OQuigley, Paoletti, and Maccario (2002), and is demonstrated by examples to be a practical accuracy upper bound for model-based dose finding methods. We illustrate the implementation of the technique in the context of phase I trials that consider multiple toxicities and phase I/II trials where dosing decisions are based on both toxicity and efficacy, and apply the benchmark to several clinical examples considered in the literature. By comparing the operating characteristics of a dose finding method to that of the benchmark, we can form quick initial assessments of whether the method is adequately calibrated and evaluate its sensitivity to the dose-outcome relationships. PMID:24571185

  17. Canadian Language Benchmarks 2000: Theoretical Framework.

    ERIC Educational Resources Information Center

    Pawlikowska-Smith, Grazyna

    This document provides indepth study and support of the "Canadian Language Benchmarks 2000" (CLB 2000). In order to make the CLB 2000 usable, the competencies and standards were considerably compressed and simplified, and much of the indepth discussion of language ability or proficiency was omitted, at publication. This document includes: (1)

  18. MHEC Survey Establishes Midwest Property Insurance Benchmarks.

    ERIC Educational Resources Information Center

    Midwestern Higher Education Commission Risk Management Institute Research Bulletin, 1994

    1994-01-01

    This publication presents the results of a survey of over 200 midwestern colleges and universities on their property insurance programs and establishes benchmarks to help these institutions evaluate their insurance programs. Findings included the following: (1) 51 percent of respondents currently purchase their property insurance as part of a…

  19. Closed benchmarks for network community structure characterization.

    PubMed

    Aldecoa, Rodrigo; Marn, Ignacio

    2012-02-01

    Characterizing the community structure of complex networks is a key challenge in many scientific fields. Very diverse algorithms and methods have been proposed to this end, many working reasonably well in specific situations. However, no consensus has emerged on which of these methods is the best to use in practice. In part, this is due to the fact that testing their performance requires the generation of a comprehensive, standard set of synthetic benchmarks, a goal not yet fully achieved. Here, we present a type of benchmark that we call "closed," in which an initial network of known community structure is progressively converted into a second network whose communities are also known. This approach differs from all previously published ones, in which networks evolve toward randomness. The use of this type of benchmark allows us to monitor the transformation of the community structure of a network. Moreover, we can predict the optimal behavior of the variation of information, a measure of the quality of the partitions obtained, at any moment of the process. This enables us in many cases to determine the best partition among those suggested by different algorithms. Also, since any network can be used as a starting point, extensive studies and comparisons can be performed using a heterogeneous set of structures, including random ones. These properties make our benchmarks a general standard for comparing community detection algorithms. PMID:22463281

  20. Benchmark data for sulcal pits extraction algorithms

    PubMed Central

    Auzias, G.; Brun, L.; Deruelle, C.; Coulon, O.

    2015-01-01

    This article contains data related to the research article Auzias et al. (2015) [1]. This data can be used as a benchmark for quantitative evaluation of sulcal pits extraction algorithm. In particular, it allows a quantitative comparison with our method, and the assessment of the consistency of the sulcal pits extraction across two well-matched populations. PMID:26958615

  1. Benchmarking 2009: Trends in Education Philanthropy

    ERIC Educational Resources Information Center

    Bearman, Jessica; Kilgore, Gin

    2009-01-01

    "Benchmarking 2009: Trends in Education Philanthropy" is Grantmakers for Education's (GFE) second annual study of grantmaking trends and priorities among members. As a national network dedicated to improving education outcomes through philanthropy, GFE members are mindful of their role in fostering greater knowledge in the field. They believe it's

  2. Benchmarks--Standards Comparisons. Math Competencies: EFF Benchmarks Comparison [and] Reading Competencies: EFF Benchmarks Comparison [and] Writing Competencies: EFF Benchmarks Comparison.

    ERIC Educational Resources Information Center

    Kent State Univ., OH. Ohio Literacy Resource Center.

    This document is intended to show the relationship between Ohio's Standards and Competencies, Equipped for the Future's (EFF's) Standards and Components of Performance, and Ohio's Revised Benchmarks. The document is divided into three parts, with Part 1 covering mathematics instruction, Part 2 covering reading instruction, and Part 3 covering

  3. 2010 Recruiting Benchmarks Survey. Research Brief

    ERIC Educational Resources Information Center

    National Association of Colleges and Employers (NJ1), 2010

    2010-01-01

    The National Association of Colleges and Employers conducted its annual survey of employer members from June 15, 2010 to August 15, 2010, to benchmark data relevant to college recruiting. From a base of 861 employers holding organizational membership, there were 268 responses for a response rate of 31 percent. Following are some of the major

  4. A MULTIMODEL APPROACH FOR CALCULATING BENCHMARK DOSE

    EPA Science Inventory


    A Multimodel Approach for Calculating Benchmark Dose
    Ramon I. Garcia and R. Woodrow Setzer

    In the assessment of dose response, a number of plausible dose- response models may give fits that are consistent with the data. If no dose response formulation had been speci...

  5. Benchmark problems for subsurface flow uncertainty quantification

    NASA Astrophysics Data System (ADS)

    Chang, Haibin; Liao, Qinzhuo; Zhang, Dongxiao

    2015-12-01

    In this work, we design a series of benchmark problems for subsurface flow uncertainty quantification. Three basic subsurface flow problems with increasing complexity are selected, which are steady state groundwater flow, groundwater contamination, and multi-phase flow. For the steady state groundwater flow, hydraulic conductivity is assumed to be uncertain, and the uncertain model parameter is assumed to be Gaussian random constant, Gaussian random field, and facies field, respectively. For the other two flow problems, the uncertain model parameter is assumed to be Gaussian random field and facies field, respectively. The statistical property of the uncertain model parameter is specified for each problem. The Monte Carlo (MC) method is used to obtain the benchmark results. The results include the first two statistical moments and the probability density function of the quantities of interest. To verify the MC results, we test the convergence of the results and the reliability of the sampling algorithm. For any existing and newly developed uncertainty quantification methods, which are not (fully) verified, the designed benchmark problems in this work can facilitate the verification process of those methods. For illustration, in this work, we provide a verification of the probabilistic collocation method using the benchmark results.

  6. Assessment Benchmarks for an Alternative Certification Program

    ERIC Educational Resources Information Center

    Denton, Jon; Davis, Trina; Smith, Ben; Beason, Lynn; Graham, Diane; Strader, Arlen

    2008-01-01

    The Accelerate Online/OPTIONS alternative certification program provides those possessing a degree an option leading to certification that can be completed in 18 months. The program consists of an online curriculum, an early field experience, and a year-long paid internship. To determine program effectiveness across five years, five benchmark

  7. Benchmarking 2008: Trends in Education Philanthropy

    ERIC Educational Resources Information Center

    Grantmakers for Education, 2008

    2008-01-01

    Through Grantmakers for Education's (GFE's) "Benchmarking 2008" research report, the researchers sought to make the act of sharing among GFE members as easy and worthwhile as possible. The researchers started with an online survey, which was completed by education grantmakers from more than 150 organizations. They analyzed their responses for

  8. Benchmark Problems for Spacecraft Formation Flying Missions

    NASA Technical Reports Server (NTRS)

    Carpenter, J. Russell; Leitner, Jesse A.; Burns, Richard D.; Folta, David C.

    2003-01-01

    To provide high-level focus to distributed space system flight dynamics and control research, several benchmark problems are suggested. These problems are not specific to any current or proposed mission, but instead are intended to capture high-level features that would be generic to many similar missions.

  9. VENUS-2 MOX Core Benchmark: Results of ORNL Calculations Using HELIOS-1.4 - Revised Report

    SciTech Connect

    Ellis, RJ

    2001-06-01

    The Task Force on Reactor-Based Plutonium Disposition (TFRPD) was formed by the Organization for Economic Cooperation and Development/Nuclear Energy Agency (OECD/NEA) to study reactor physics, fuel performance, and fuel cycle issues related to the disposition of weapons-grade (WG) plutonium as mixed-oxide (MOX) reactor fuel. To advance the goals of the TFRPD, 10 countries and 12 institutions participated in a major TFRPD activity: a blind benchmark study to compare code calculations to experimental data for the VENUS-2 MOX core at SCK-CEN in Mol, Belgium. At Oak Ridge National Laboratory, the HELIOS-1.4 code system was used to perform the comprehensive study of pin-cell and MOX core calculations for the VENUS-2 MOX core benchmark study.

  10. The Meaning of Grades.

    ERIC Educational Resources Information Center

    Teixeira, Serna E.

    1996-01-01

    Asserts that students see grades as an indicator of effort unconnected to the content of the course while teachers regard grades as a measure of achievement within a discipline. Discusses some of the current controversies and approaches concerning grades and how they relate to school reform. (MJP)

  11. General Graded Response Model.

    ERIC Educational Resources Information Center

    Samejima, Fumiko

    This paper describes the graded response model. The graded response model represents a family of mathematical models that deal with ordered polytomous categories, such as: (1) letter grading; (2) an attitude survey with "strongly disagree, disagree, agree, and strongly agree" choices; (3) partial credit given in accord with an individual's degree

  12. Bias in Grading

    ERIC Educational Resources Information Center

    Malouff, John

    2008-01-01

    Bias in grading can be conscious or unconscious. The author describes different types of bias, such as those based on student attractiveness or performance in prior courses, and a variety of methods of reducing bias, including keeping students anonymous during grading and using detailed criteria for subjective grading.

  13. Using Benchmarking To Influence Tuition and Fee Decisions.

    ERIC Educational Resources Information Center

    Hubbell, Loren W. Loomis; Massa, Robert J.; Lapovsky, Lucie

    2002-01-01

    Discusses the use of benchmarking in managing enrollment. Using a case study, illustrates how benchmarking can help administrators develop strategies for planning and implementing admissions and pricing practices. (EV)

  14. Results Oriented Benchmarking: The Evolution of Benchmarking at NASA from Competitive Comparisons to World Class Space Partnerships

    NASA Technical Reports Server (NTRS)

    Bell, Michael A.

    1999-01-01

    Informal benchmarking using personal or professional networks has taken place for many years at the Kennedy Space Center (KSC). The National Aeronautics and Space Administration (NASA) recognized early on, the need to formalize the benchmarking process for better utilization of resources and improved benchmarking performance. The need to compete in a faster, better, cheaper environment has been the catalyst for formalizing these efforts. A pioneering benchmarking consortium was chartered at KSC in January 1994. The consortium known as the Kennedy Benchmarking Clearinghouse (KBC), is a collaborative effort of NASA and all major KSC contractors. The charter of this consortium is to facilitate effective benchmarking, and leverage the resulting quality improvements across KSC. The KBC acts as a resource with experienced facilitators and a proven process. One of the initial actions of the KBC was to develop a holistic methodology for Center-wide benchmarking. This approach to Benchmarking integrates the best features of proven benchmarking models (i.e., Camp, Spendolini, Watson, and Balm). This cost-effective alternative to conventional Benchmarking approaches has provided a foundation for consistent benchmarking at KSC through the development of common terminology, tools, and techniques. Through these efforts a foundation and infrastructure has been built which allows short duration benchmarking studies yielding results gleaned from world class partners that can be readily implemented. The KBC has been recognized with the Silver Medal Award (in the applied research category) from the International Benchmarking Clearinghouse.

  15. Differences in physiological responses between short- vs. long-graded laboratory tests in road cyclists.

    PubMed

    Gonzlez-Haro, Carlos

    2015-04-01

    This study aimed to determine the effect of a short-graded with respect to a long-graded protocol laboratory test on the physiological responses of road cyclists. Twenty well-trained road cyclists performed a short-graded and long-graded laboratory tests within 1 week of each other in a randomized and crossover study design. Blood lactate concentration ([La-]b), heart rate (HR), oxygen consumption ((Equation is included in full-text article.)), and carbon dioxide production ((Equation is included in full-text article.)) were measured. Fat and carbohydrate oxidation rates (FAT(OxR) and CHO(OxR)) were estimated at the end of each stage during the short-graded and the long-graded (10th minute: T2.10) and in the middle of long-graded (fifth minute: T2.5) protocol. Lactate threshold (LT) and individual anaerobic threshold (IAT) were calculated. For maximal intensities, duration and maxFAT(OxR) were significantly higher in long-graded with respect to short-graded protocols. Peak power output (POPeak), HRPeak, [La-]bmax, (Equation is included in full-text article.), and maxCHO(OxR) were significantly higher in short-graded with respect to long-graded protocols. At submaximal intensities, short-graded protocol provoked higher demands on glycolytic metabolism than long-graded protocol; no differences were illustrated for HR or (Equation is included in full-text article.)between protocols. Crossover concept shifted to higher intensities in long-graded with respect to short-graded protocols due to the higher lipolytic response during the long-graded protocol. Both LT and IAT were reached at the same %(Equation is included in full-text article.), although significantly higher PO in short-graded with respect to long-graded protocols was reached. The long-graded proved to be more specific than the short-graded protocol to assess the physiological responses of road cyclists based on relative PO (Wkg(-1)). PMID:25330085

  16. Relationship between the TCAP and the Pearson Benchmark Assessment in Elementary Students' Reading and Math Performance in a Northeastern Tennessee School District

    ERIC Educational Resources Information Center

    Dugger-Roberts, Cherith A.

    2014-01-01

    The purpose of this quantitative study was to determine if there was a relationship between the TCAP test and Pearson Benchmark assessment in elementary students' reading and language arts and math performance in a northeastern Tennessee school district. This study involved 3rd, 4th, 5th, and 6th grade students. The study focused on the following…

  17. Relationship between the TCAP and the Pearson Benchmark Assessment in Elementary Students' Reading and Math Performance in a Northeastern Tennessee School District

    ERIC Educational Resources Information Center

    Dugger-Roberts, Cherith A.

    2014-01-01

    The purpose of this quantitative study was to determine if there was a relationship between the TCAP test and Pearson Benchmark assessment in elementary students' reading and language arts and math performance in a northeastern Tennessee school district. This study involved 3rd, 4th, 5th, and 6th grade students. The study focused on the following

  18. Benchmarking Text Understanding Systems to Human Performance: An Exploration.

    ERIC Educational Resources Information Center

    Butler, Frances A.; And Others

    This study, part of a larger effort to develop a methodology for evaluating intelligent computer systems (Artificial Intelligence Systems), explores the use of benchmarking as an evaluation technique. Benchmarking means comparing the performance of intelligent computer systems with human performance on the same task. Benchmarking in evaluation has

  19. Unstructured Adaptive (UA) NAS Parallel Benchmark. Version 1.0

    NASA Technical Reports Server (NTRS)

    Feng, Huiyu; VanderWijngaart, Rob; Biswas, Rupak; Mavriplis, Catherine

    2004-01-01

    We present a complete specification of a new benchmark for measuring the performance of modern computer systems when solving scientific problems featuring irregular, dynamic memory accesses. It complements the existing NAS Parallel Benchmark suite. The benchmark involves the solution of a stylized heat transfer problem in a cubic domain, discretized on an adaptively refined, unstructured mesh.

  20. Discovering and Implementing Best Practices to Strengthen SEAs: Collaborative Benchmarking

    ERIC Educational Resources Information Center

    Building State Capacity and Productivity Center, 2013

    2013-01-01

    This paper is written for state educational agency (SEA) leaders who are considering the benefits of collaborative benchmarking, and it addresses the following questions: (1) What does benchmarking of best practices entail?; (2) How does "collaborative benchmarking" enhance the process?; (3) How do SEAs control the process so that "their" needs…

  1. 42 CFR 422.258 - Calculation of benchmarks.

    Code of Federal Regulations, 2014 CFR

    2014-10-01

    ...) Weighting for the 2-year phase-in. (A) For 2012, the blended benchmark is the sum of one-half of the.... (ii) Weighting for the 4-year phase-in. The blended benchmark is the sum of the applicable amount at... specified amount for the area and year. (iii) Weighting for the 6-year phase-in. The blended benchmark...

  2. 45 CFR 156.100 - State selection of benchmark.

    Code of Federal Regulations, 2013 CFR

    2013-10-01

    ...-benchmark plan. If a State does not make a selection using the process defined in 156.100 of this section... 45 Public Welfare 1 2013-10-01 2013-10-01 false State selection of benchmark. 156.100 Section 156... Essential Health Benefits Package 156.100 State selection of benchmark. Each State may identify a...

  3. A Simple Alternative to Grading

    ERIC Educational Resources Information Center

    Potts, Glenda

    2010-01-01

    In this article, the author investigates whether an alternative grading system (contract grading) would yield the same final grades as traditional grading (letter grading), and whether or not it would be accepted by students. The author states that this study demonstrated that contract grading was widely, and for the most part, enthusiastically

  4. Experiences in Benchmarking of Autonomic Systems

    NASA Astrophysics Data System (ADS)

    Etchevers, Xavier; Coupaye, Thierry; Vachet, Guy

    Autonomic computing promises improvements of systems quality of service in terms of availability, reliability, performance, security, etc. However, little research and experimental results have so far demonstrated this assertion, nor provided proof of the return on investment stemming from the efforts that introducing autonomic features requires. Existing works in the area of benchmarking of autonomic systems can be characterized by their qualitative and fragmented approaches. Still a crucial need is to provide generic (i.e. independent from business, technology, architecture and implementation choices) autonomic computing benchmarking tools for evaluating and/or comparing autonomic systems from a technical and, ultimately, an economical point of view. This article introduces a methodology and a process for defining and evaluating factors, criteria and metrics in order to qualitatively and quantitatively assess autonomic features in computing systems. It also discusses associated experimental results on three different autonomic systems.

  5. Benchmark On Sensitivity Calculation (Phase III)

    SciTech Connect

    Ivanova, Tatiana; Laville, Cedric; Dyrda, James; Mennerdahl, Dennis; Golovko, Yury; Raskach, Kirill; Tsiboulia, Anatoly; Lee, Gil Soo; Woo, Sweng-Woong; Bidaud, Adrien; Patel, Amrit; Bledsoe, Keith C; Rearden, Bradley T; Gulliford, J.

    2012-01-01

    The sensitivities of the keff eigenvalue to neutron cross sections have become commonly used in similarity studies and as part of the validation algorithm for criticality safety assessments. To test calculations of the sensitivity coefficients, a benchmark study (Phase III) has been established by the OECD-NEA/WPNCS/EG UACSA (Expert Group on Uncertainty Analysis for Criticality Safety Assessment). This paper presents some sensitivity results generated by the benchmark participants using various computational tools based upon different computational methods: SCALE/TSUNAMI-3D and -1D, MONK, APOLLO2-MORET 5, DRAGON-SUSD3D and MMKKENO. The study demonstrates the performance of the tools. It also illustrates how model simplifications impact the sensitivity results and demonstrates the importance of 'implicit' (self-shielding) sensitivities. This work has been a useful step towards verification of the existing and developed sensitivity analysis methods.

  6. Toxicological benchmarks for wildlife. Environmental Restoration Program

    SciTech Connect

    Opresko, D.M.; Sample, B.E.; Suter, G.W.

    1993-09-01

    This report presents toxicological benchmarks for assessment of effects of 55 chemicals on six representative mammalian wildlife species (short-tailed shrew, white-footed mouse, cottontail ink, red fox, and whitetail deer) and eight avian wildlife species (American robin, woodcock, wild turkey, belted kingfisher, great blue heron, barred owl, Cooper`s hawk, and redtailed hawk) (scientific names are presented in Appendix C). These species were chosen because they are widely distributed and provide a representative range of body sizes and diets. The chemicals are some of those that occur at United States Department of Energy (DOE) waste sites. The benchmarks presented in this report are values believed to be nonhazardous for the listed wildlife species.

  7. TsunaFLASH Benchmark and Its Verifications

    NASA Astrophysics Data System (ADS)

    Pranowo, Widodo; Behrens, Joern

    2010-05-01

    In the end of year 2008 TsunAWI (Tsunami unstructured mesh finite element model developed at Alfred Wegener Institute) by Behrens et al. (2006 - 2008) [Behrens, 2008], had been launched as an operational model in the German - Indonesian Tsunami EarlyWarning System (GITEWS) framework. This model has been benchmarked and verified with 2004 Sumatra-Andaman mega tsunami event [Harig et al., 2008]. A new development uses adaptive mesh refinement to improve computational efficiency and accuracy, this approach is called TsunaFLASH [Pranowo et al., 2008]. After the initial development and verification phase with stabilization efforts, and study of refinement criteria, the code is now mature enough to be validated with data. This presentation will demonstrate results of TsunaFLASH for the experiments with diverse mesh refinement criteria, and benchmarks; in particular the problem set-1 of IWLRM, and field data of the Sumatra-Andaman 2004 event.

  8. A Simplified HTTR Diffusion Theory Benchmark

    SciTech Connect

    Rodolfo M. Ferrer; Abderrafi M. Ougouag; Farzad Rahnema

    2010-10-01

    The Georgia Institute of Technology (GA-Tech) recently developed a transport theory benchmark based closely on the geometry and the features of the HTTR reactor that is operational in Japan. Though simplified, the benchmark retains all the principal physical features of the reactor and thus provides a realistic and challenging test for the codes. The purpose of this paper is twofold. The first goal is an extension of the benchmark to diffusion theory applications by generating the additional data not provided in the GA-Tech prior work. The second goal is to use the benchmark on the HEXPEDITE code available to the INL. The HEXPEDITE code is a Green’s function-based neutron diffusion code in 3D hexagonal-z geometry. The results showed that the HEXPEDITE code accurately reproduces the effective multiplication factor of the reference HELIOS solution. A secondary, but no less important, conclusion is that in the testing against actual HTTR data of a full sequence of codes that would include HEXPEDITE, in the apportioning of inevitable discrepancies between experiment and models, the portion of error attributable to HEXPEDITE would be expected to be modest. If large discrepancies are observed, they would have to be explained by errors in the data fed into HEXPEDITE. Results based on a fully realistic model of the HTTR reactor are presented in a companion paper. The suite of codes used in that paper also includes HEXPEDITE. The results shown here should help that effort in the decision making process for refining the modeling steps in the full sequence of codes.

  9. SINBAD: Shielding integral benchmark archive and database

    SciTech Connect

    Hunter, H.T.; Ingersoll, D.T.; Roussin, R.W.

    1996-04-01

    SINBAD is a new electronic database developed to store a variety of radiation shielding benchmark data so that users can easily retrieve and incorporate the data into their calculations. SINBAD is an excellent data source for users who require the quality assurance necessary in developing cross-section libraries or radiation transport codes. The future needs of the scientific community are best served by the electronic database format of SINBAD and its user-friendly interface, combined with its data accuracy and integrity.

  10. Aeroelasticity Benchmark Assessment: Subsonic Fixed Wing Program

    NASA Technical Reports Server (NTRS)

    Florance, Jennifer P.; Chwalowski, Pawel; Wieseman, Carol D.

    2010-01-01

    The fundamental technical challenge in computational aeroelasticity is the accurate prediction of unsteady aerodynamic phenomena and the effect on the aeroelastic response of a vehicle. Currently, a benchmarking standard for use in validating the accuracy of computational aeroelasticity codes does not exist. Many aeroelastic data sets have been obtained in wind-tunnel and flight testing throughout the world; however, none have been globally presented or accepted as an ideal data set. There are numerous reasons for this. One reason is that often, such aeroelastic data sets focus on the aeroelastic phenomena alone (flutter, for example) and do not contain associated information such as unsteady pressures and time-correlated structural dynamic deflections. Other available data sets focus solely on the unsteady pressures and do not address the aeroelastic phenomena. Other discrepancies can include omission of relevant data, such as flutter frequency and / or the acquisition of only qualitative deflection data. In addition to these content deficiencies, all of the available data sets present both experimental and computational technical challenges. Experimental issues include facility influences, nonlinearities beyond those being modeled, and data processing. From the computational perspective, technical challenges include modeling geometric complexities, coupling between the flow and the structure, grid issues, and boundary conditions. The Aeroelasticity Benchmark Assessment task seeks to examine the existing potential experimental data sets and ultimately choose the one that is viewed as the most suitable for computational benchmarking. An initial computational evaluation of that configuration will then be performed using the Langley-developed computational fluid dynamics (CFD) software FUN3D1 as part of its code validation process. In addition to the benchmarking activity, this task also includes an examination of future research directions. Researchers within the Aeroelasticity Branch will examine other experimental efforts within the Subsonic Fixed Wing (SFW) program (such as testing of the NASA Common Research Model (CRM)) and other NASA programs and assess aeroelasticity issues and research topics.

  11. EXPERIMENTAL BENCHMARKING OF THE MAGNETIZED FRICTION FORCE.

    SciTech Connect

    FEDOTOV, A.V.; GALNANDER, B.; LITVINENKO, V.N.; LOFNES, T.; SIDORIN, A.O.; SMIRNOV, A.V.; ZIEMANN, V.

    2005-09-18

    High-energy electron cooling, presently considered as essential tool for several applications in high-energy and nuclear physics, requires accurate description of the friction force. A series of measurements were performed at CELSIUS with the goal to provide accurate data needed for the benchmarking of theories and simulations. Some results of accurate comparison of experimental data with the friction force formulas are presented.

  12. Benchmarking Global Food Safety Performances: The Era of Risk Intelligence.

    PubMed

    Valleé, Jean-Charles Le; Charlebois, Sylvain

    2015-10-01

    Food safety data segmentation and limitations hamper the world's ability to select, build up, monitor, and evaluate food safety performance. Currently, there is no metric that captures the entire food safety system, and performance data are not collected strategically on a global scale. Therefore, food safety benchmarking is essential not only to help monitor ongoing performance but also to inform continued food safety system design, adoption, and implementation toward more efficient and effective food safety preparedness, responsiveness, and accountability. This comparative study identifies and evaluates common elements among global food safety systems. It provides an overall world ranking of food safety performance for 17 Organisation for Economic Co-Operation and Development (OECD) countries, illustrated by 10 indicators organized across three food safety risk governance domains: risk assessment (chemical risks, microbial risks, and national reporting on food consumption), risk management (national food safety capacities, food recalls, food traceability, and radionuclides standards), and risk communication (allergenic risks, labeling, and public trust). Results show all countries have very high food safety standards, but Canada and Ireland, followed by France, earned excellent grades relative to their peers. However, any subsequent global ranking study should consider the development of survey instruments to gather adequate and comparable national evidence on food safety. PMID:26408141

  13. A PWR Thorium Pin Cell Burnup Benchmark

    SciTech Connect

    Weaver, Kevan Dean; Zhao, X.; Pilat, E. E; Hejzlar, P.

    2000-05-01

    As part of work to evaluate the potential benefits of using thorium in LWR fuel, a thorium fueled benchmark comparison was made in this study between state-of-the-art codes, MOCUP (MCNP4B + ORIGEN2), and CASMO-4 for burnup calculations. The MOCUP runs were done individually at MIT and INEEL, using the same model but with some differences in techniques and cross section libraries. Eigenvalue and isotope concentrations were compared on a PWR pin cell model up to high burnup. The eigenvalue comparison as a function of burnup is good: the maximum difference is within 2% and the average absolute difference less than 1%. The isotope concentration comparisons are better than a set of MOX fuel benchmarks and comparable to a set of uranium fuel benchmarks reported in the literature. The actinide and fission product data sources used in the MOCUP burnup calculations for a typical thorium fuel are documented. Reasons for code vs code differences are analyzed and discussed.

  14. Benchmarking numerical freeze/thaw models

    NASA Astrophysics Data System (ADS)

    Rhaak, Wolfram; Anbergen, Hauke; Molson, John; Grenier, Christophe; Sass, Ingo

    2015-04-01

    The modeling of freezing and thawing of water in porous media is of increasing interest, and for which very different application areas exist. For instance, the modeling of permafrost regression with respect to climate change issues is one area, while others include geotechnical applications in tunneling and for borehole heat exchangers which operate at temperatures below the freezing point. The modeling of these processes requires the solution of a coupled non-linear system of partial differential equations for flow and heat transport in space and time. Different code implementations have been developed in the past. Analytical solutions exist only for simple cases. Consequently, an interest has arisen in benchmarking different codes with analytical solutions, experiments and purely numerical results, similar to the long-standing DECOVALEX and the more recent "Geothermal Code Comparison" activities. The name for this freezing/ thawing benchmark consortium is INTERFROST. In addition to the well-known so-called Lunardini solution for a 1D case (case T1), two different 2D problems will be presented, one which represents melting of a frozen inclusion (case TH2) and another which represents the growth or thaw of permafrost around a talik (case TH3). These talik regions are important for controlling groundwater movement within a mainly frozen ground. First results of the different benchmark results will be shown and discussed.

  15. Introduction to the HPC Challenge Benchmark Suite

    SciTech Connect

    Luszczek, Piotr; Dongarra, Jack J.; Koester, David; Rabenseifner,Rolf; Lucas, Bob; Kepner, Jeremy; McCalpin, John; Bailey, David; Takahashi, Daisuke

    2005-04-25

    The HPC Challenge benchmark suite has been released by the DARPA HPCS program to help define the performance boundaries of future Petascale computing systems. HPC Challenge is a suite of tests that examine the performance of HPC architectures using kernels with memory access patterns more challenging than those of the High Performance Linpack (HPL) benchmark used in the Top500 list. Thus, the suite is designed to augment the Top500 list, providing benchmarks that bound the performance of many real applications as a function of memory access characteristics e.g., spatial and temporal locality, and providing a framework for including additional tests. In particular, the suite is composed of several well known computational kernels (STREAM, HPL, matrix multiply--DGEMM, parallel matrix transpose--PTRANS, FFT, RandomAccess, and bandwidth/latency tests--b{sub eff}) that attempt to span high and low spatial and temporal locality space. By design, the HPC Challenge tests are scalable with the size of data sets being a function of the largest HPL matrix for the tested system.

  16. Benchmarking and accounting for the (private) cloud

    NASA Astrophysics Data System (ADS)

    Belleman, J.; Schwickerath, U.

    2015-12-01

    During the past two years large parts of the CERN batch farm have been moved to virtual machines running on the CERN internal cloud. During this process a large fraction of the resources, which had previously been used as physical batch worker nodes, were converted into hypervisors. Due to the large spread of the per-core performance in the farm, caused by its heterogenous nature, it is necessary to have a good knowledge of the performance of the virtual machines. This information is used both for scheduling in the batch system and for accounting. While in the previous setup worker nodes were classified and benchmarked based on the purchase order number, for virtual batch worker nodes this is no longer possible; the information is now either hidden or hard to retrieve. Therefore we developed a new scheme to classify worker nodes according to their performance. The new scheme is flexible enough to be usable both for virtual and physical machines in the batch farm. With the new classification it is possible to have an estimation of the performance of worker nodes also in a very dynamic farm with worker nodes coming and going at a high rate, without the need to benchmark each new node again. An extension to public cloud resources is possible if all conditions under which the benchmark numbers have been obtained are fulfilled.

  17. Florida Course Descriptions: Exceptional Student Education, Grades 6-8, 1999 Revisions. Suggested Course Performance Objectives.

    ERIC Educational Resources Information Center

    Florida State Dept. of Education, Tallahassee. Bureau of Instructional Support and Community Services.

    This document presents the full revised course descriptions for Florida exceptional student education in grades 6-8. Courses incorporate the Florida standards for a special diploma. Requirements include a list of related benchmarks from the state standards for each level of functioning: independent, supported, and participatory. Introductory

  18. Why 12th Grade Must Be Redesigned Now--and How

    ERIC Educational Resources Information Center

    Vargas, Joel

    2015-01-01

    This first report in a new series by Jobs For the Future (JFF) provides the rationale for restructuring 12th grade and tying it more tightly to the first year of college through new high school and college partnerships. The paper proposes a new common benchmark of readiness that high schools and colleges can work together to meet to ensure

  19. Aiming High and Falling Short: California's Eighth-Grade Algebra-for-All Effort

    ERIC Educational Resources Information Center

    Domina, Thurston; McEachin, Andrew; Penner, Andrew; Penner, Emily

    2015-01-01

    The United States is in the midst of an effort to intensify middle school mathematics curricula by enrolling more 8th graders in Algebra. California is at the forefront of this effort, and in 2008, the state moved to make Algebra the accountability benchmark test for 8th-grade mathematics. This article takes advantage of this unevenly implemented…

  20. Aiming High and Falling Short: California's Eighth-Grade Algebra-for-All Effort

    ERIC Educational Resources Information Center

    Domina, Thurston; McEachin, Andrew; Penner, Andrew; Penner, Emily

    2015-01-01

    The United States is in the midst of an effort to intensify middle school mathematics curricula by enrolling more 8th graders in Algebra. California is at the forefront of this effort, and in 2008, the state moved to make Algebra the accountability benchmark test for 8th-grade mathematics. This article takes advantage of this unevenly implemented

  1. Student Performance on the 2013 Assessment Program in Primary Reading (Kindergarten to Grade 2). Memorandum

    ERIC Educational Resources Information Center

    Sanderson, Geoffrey T.

    2013-01-01

    This evaluation report provides details for school year 2013 on the percentage of primary students who met or exceeded grade-level reading benchmarks on the Montgomery County (Maryland) Public Schools' (MCPS) Assessment Program in Primary Reading (AP-PR). The percentage remained above 90% for kindergarten students and showed a small decrease for

  2. The Development of CBM Vocabulary Measures: Grade 2. Technical Report #1209

    ERIC Educational Resources Information Center

    Alonzo, Julie; Anderson, Daniel; Park, Bitnara Jasmine; Tindal, Gerald

    2012-01-01

    In this technical report, we describe the development and piloting of a series of vocabulary assessments intended for use with students in grades two through eight. These measures, available as part of easyCBM[TM], an online progress monitoring and benchmark/screening assessment system, were developed in 2010 and administered to approximately 1200…

  3. The Development of CBM Vocabulary Measures: Grade 4. Technical Report #1211

    ERIC Educational Resources Information Center

    Alonzo, Julie; Anderson, Daniel; Park, Bitnara Jasmine; Tindal, Gerald

    2012-01-01

    In this technical report, we describe the development and piloting of a series of vocabulary assessments intended for use with students in grades two through eight. These measures, available as part of easyCBM[TM], an online progress monitoring and benchmark/screening assessment system, were developed in 2010 and administered to approximately 1200…

  4. Historical benchmarks for medical therapy trials in surgery- and radiation-refractory meningioma: a RANO review

    PubMed Central

    Kaley, Thomas; Barani, Igor; Chamberlain, Marc; McDermott, Michael; Panageas, Katherine; Raizer, Jeffrey; Rogers, Leland; Schiff, David; Vogelbaum, Michael; Weber, Damien; Wen, Patrick

    2014-01-01

    Background The outcomes of patients with surgery- and radiation-refractory meningiomas treated with medical therapies are poorly defined. Published reports are limited by small patient numbers, selection bias, inclusion of mixed histologic grades and stages of illness, and World Health Organization (WHO) criteria changes. This analysis seeks to define outcome benchmarks for future clinical trial design. Methods A PubMed literature search was performed for all English language publications on medical therapy for meningioma. Reports were tabulated and analyzed for number of patients, histologic grade, prior therapy, overall survival, progression-free survival (PFS), and radiographic response. Results Forty-seven publications were identified and divided by histology and prior therapies, including only those that treated patients who were surgery and radiation refractory for further analysis. This included a variety of agents (hydroxyurea, temozolomide, irinotecan, interferon-?, mifepristone, octreotide analogues, megestrol acetate, bevacizumab, imatinib, erlotinib, and gefitinib) from retrospective, pilot, and phase II studies, exploratory arms of other studies, and a single phase III study. The only outcome extractable from all studies was the PFS 6-month rate, and a weighted average was calculated separately for WHO grade I meningioma and combined WHO grade II/III meningioma. For WHO I meningioma, the weighted average PFS-6 was 29% (95% confidence interval [CI]: 20.3%37.7%). For WHO II/III meningioma, the weighted average PFS-6 was 26% (95% CI: 19.3%32.7%). Conclusions This comprehensive review confirms the poor outcomes of medical therapy for surgery- and radiation-refractory meningioma. We recommend the above PFS-6 benchmarks for future trial design. PMID:24500419

  5. Guidebook for Using the Tool BEST Cement: Benchmarking and Energy Savings Tool for the Cement Industry

    SciTech Connect

    Galitsky, Christina; Price, Lynn; Zhou, Nan; Fuqiu , Zhou; Huawen, Xiong; Xuemin, Zeng; Lan, Wang

    2008-07-30

    The Benchmarking and Energy Savings Tool (BEST) Cement is a process-based tool based on commercially available efficiency technologies used anywhere in the world applicable to the cement industry. This version has been designed for use in China. No actual cement facility with every single efficiency measure included in the benchmark will likely exist; however, the benchmark sets a reasonable standard by which to compare for plants striving to be the best. The energy consumption of the benchmark facility differs due to differences in processing at a given cement facility. The tool accounts for most of these variables and allows the user to adapt the model to operational variables specific for his/her cement facility. Figure 1 shows the boundaries included in a plant modeled by BEST Cement. In order to model the benchmark, i.e., the most energy efficient cement facility, so that it represents a facility similar to the user's cement facility, the user is first required to input production variables in the input sheet (see Section 6 for more information on how to input variables). These variables allow the tool to estimate a benchmark facility that is similar to the user's cement plant, giving a better picture of the potential for that particular facility, rather than benchmarking against a generic one. The input variables required include the following: (1) the amount of raw materials used in tonnes per year (limestone, gypsum, clay minerals, iron ore, blast furnace slag, fly ash, slag from other industries, natural pozzolans, limestone powder (used post-clinker stage), municipal wastes and others); the amount of raw materials that are preblended (prehomogenized and proportioned) and crushed (in tonnes per year); (2) the amount of additives that are dried and ground (in tonnes per year); (3) the production of clinker (in tonnes per year) from each kiln by kiln type; (4) the amount of raw materials, coal and clinker that is ground by mill type (in tonnes per year); (5) the amount of production of cement by type and grade (in tonnes per year); (6) the electricity generated onsite; and, (7) the energy used by fuel type; and, the amount (in RMB per year) spent on energy. The tool offers the user the opportunity to do a quick assessment or a more detailed assessment--this choice will determine the level of detail of the energy input. The detailed assessment will require energy data for each stage of production while the quick assessment will require only total energy used at the entire facility (see Section 6 for more details on quick versus detailed assessments). The benchmarking tool provides two benchmarks--one for Chinese best practices and one for international best practices. Section 2 describes the differences between these two and how each benchmark was calculated. The tool also asks for a target input by the user for the user to set goals for the facility.

  6. Numisheet2005 Benchmark Analysis on Forming of an Automotive Deck Lid Inner Panel: Benchmark 1

    SciTech Connect

    Buranathiti, Thaweepat; Cao Jian

    2005-08-05

    Numerical simulations in sheet metal forming processes have been a very challenging topic in industry. There are many computer codes and modeling techniques existing today. However, there are many unknowns affecting the prediction accuracy. Systematic benchmark tests are needed to accelerate the future implementations and to provide as a reference. This report presents an international cooperative benchmark effort for an automotive deck lid inner panel. Predictions from simulations are analyzed and discussed against the corresponding experimental results. The correlations between accuracy of each parameter of interest are discussed in this report.

  7. Grain Grading and Handling.

    ERIC Educational Resources Information Center

    Rendleman, Matt; Legacy, James

    This publication provides an introduction to grain grading and handling for adult students in vocational and technical education programs. Organized in five chapters, the booklet provides a brief overview of the jobs performed at a grain elevator and of the techniques used to grade grain. The first chapter introduces the grain industry and…

  8. Are Grades Undermining Motivation?

    ERIC Educational Resources Information Center

    Berliner, David; Casanova, Ursula

    1988-01-01

    A study of fifth and sixth grade public school students in Israel suggests that grades and other norm-referenced methods are less effective than task-involving evaluation methods in enhancing student performance, interest, and motivation. This article consists of two commentaries on the findings, with suggestions for implementing task-involving

  9. Minimum Grading, Maximum Learning

    ERIC Educational Resources Information Center

    Carey, Theodore; Carifio, James

    2011-01-01

    Fair and effective schools should assign grades that align with clear and consistent evidence of student performance (Wormeli, 2006), but when a student's performance is inconsistent, traditional grading practices can prove inadequate. Understanding this, increasing numbers of schools have been experimenting with the practice of assigning minimum

  10. Beef grading by ultrasound

    NASA Technical Reports Server (NTRS)

    Gammell, P. M.

    1981-01-01

    Reflections in ultrasonic A-scan signatures of beef carcasses indicate USDA grade. Since reflections from within muscle are determined primarily by fat/muscle interface, richness of signals is direct indication of degree of marbling and quality. Method replaces subjective sight and feel tests by individual graders and is applicable to grade analysis of live cattle.

  11. Skill Schemes: Sixth Grade.

    ERIC Educational Resources Information Center

    Kempf, Jerry

    The sixth grade instructional unit, part of a field-tested grade school level career education series, is designed to assist learners in understanding how present experiences relate to past and future ones. Before the main body of the lessons is described, field testing results are reported and key items are presented: the concepts, the estimated

  12. Classroom: Efficient Grading

    ERIC Educational Resources Information Center

    Shaw, David D.; Pease, Leonard F., III.

    2014-01-01

    Grading can be accelerated to make time for more effective instruction. This article presents specific time management strategies selected to decrease administrative time required of faculty and teaching assistants, including a multiple answer multiple choice interface for exams, a three-tier grading system for open ended problem solving, and a…

  13. Grades out, Badges in

    ERIC Educational Resources Information Center

    Young, Jeffrey R.

    2012-01-01

    Grades are broken. Students grub for them, pick classes where good ones come easily, and otherwise hustle to win the highest scores for the least learning. As a result, college grades are inflated to the point of meaninglessness--especially to employers who want to know which diploma-holder is best qualified for their jobs. An alternative is to

  14. Third Grade Reading Policies

    ERIC Educational Resources Information Center

    Rose, Stephanie

    2012-01-01

    In 2012, 14 states passed legislation geared toward improving 3rd-grade literacy through identification, intervention, and/or retention initiatives. Today, a total of 32 states and the District of Columbia have policies in statute aimed at improving 3rd-grade reading proficiency. The majority of these states require early assessment and

  15. Classroom: Efficient Grading

    ERIC Educational Resources Information Center

    Shaw, David D.; Pease, Leonard F., III.

    2014-01-01

    Grading can be accelerated to make time for more effective instruction. This article presents specific time management strategies selected to decrease administrative time required of faculty and teaching assistants, including a multiple answer multiple choice interface for exams, a three-tier grading system for open ended problem solving, and a

  16. Growing beyond Grades

    ERIC Educational Resources Information Center

    Perchemlides, Natalia; Coutant, Carolyn

    2004-01-01

    Once students are asked to assess their own writing progress, they will begin to do their best for writing great prose instead of just great grades. Teachers will have to create a grade-free zone, allow students to set their own writing goals, provide a common language such as the Six Traits Model, and provide evaluation and instructional models

  17. Grading and Measuring.

    ERIC Educational Resources Information Center

    Diederich, Paul B.

    The low rate of agreement among readers of College Entrance Examination essays suggested the need to examine the qualities in student writing which caused wide variance in grading. To study this question, 300 homework papers by freshmen at three universities were graded by 60 distinguished readers in six fields. The following factors, by rank,

  18. The Grades Game

    ERIC Educational Resources Information Center

    Fleenor, Andy; Lamb, Sarah; Anton, Jennifer; Stinson, Todd; Donen, Tony

    2011-01-01

    It can be quite alarming (and eye-opening) to see exactly how many of the grades students receive are based on their behaviors rather than their learning. Students should be assessed on what they know and can use rather than on their behavior. The reality, unfortunately, is that the opposite is often the case. Grades for students who work hard are

  19. Grain Grading and Handling.

    ERIC Educational Resources Information Center

    Rendleman, Matt; Legacy, James

    This publication provides an introduction to grain grading and handling for adult students in vocational and technical education programs. Organized in five chapters, the booklet provides a brief overview of the jobs performed at a grain elevator and of the techniques used to grade grain. The first chapter introduces the grain industry and

  20. Benchmarking NSP Reactors with CORETRAN-01

    SciTech Connect

    Hines, Donald D.; Grow, Rodney L.; Agee, Lance J

    2004-10-15

    As part of an overall verification and validation effort, the Electric Power Research Institute's (EPRIs) CORETRAN-01 has been benchmarked against Northern States Power's Prairie Island and Monticello reactors through 12 cycles of operation. The two Prairie Island reactors are Westinghouse 2-loop units with 121 asymmetric 14 x 14 lattice assemblies utilizing up to 8 wt% gadolinium while Monticello is a General Electric 484 bundle boiling water reactor. All reactor cases were executed in full core utilizing 24 axial nodes per assembly in the fuel with 1 additional reflector node above, below, and around the perimeter of the core. Cross-section sets used in this benchmark effort were generated by EPRI's CPM-3 as well as Studsvik's CASMO-3 and CASMO-4 to allow for separation of the lattice calculation effect from the nodal simulation method. These cases exercised the depletion-shuffle-depletion sequence through four cycles for each unit using plant data to follow actual operations. Flux map calculations were performed for comparison to corresponding measurement statepoints. Additionally, start-up physics testing cases were used to predict cycle physics parameters for comparison to existing plant methods and measurements.These benchmark results agreed well with both current analysis methods and plant measurements, indicating that CORETRAN-01 may be appropriate for steady-state physics calculations of both the Prairie Island and Monticello reactors. However, only the Prairie Island results are discussed in this paper since Monticello results were of similar quality and agreement. No attempt was made in this work to investigate CORETRAN-01 kinetics capability by analyzing plant transients, but these steady-state results form a good foundation for moving in that direction.

  1. Towards Systematic Benchmarking of Climate Model Performance

    NASA Astrophysics Data System (ADS)

    Gleckler, P. J.

    2014-12-01

    The process by which climate models are evaluated has evolved substantially over the past decade, with the Coupled Model Intercomparison Project (CMIP) serving as a centralizing activity for coordinating model experimentation and enabling research. Scientists with a broad spectrum of expertise have contributed to the CMIP model evaluation process, resulting in many hundreds of publications that have served as a key resource for the IPCC process. For several reasons, efforts are now underway to further systematize some aspects of the model evaluation process. First, some model evaluation can now be considered routine and should not require "re-inventing the wheel" or a journal publication simply to update results with newer models. Second, the benefit of CMIP research to model development has not been optimal because the publication of results generally takes several years and is usually not reproducible for benchmarking newer model versions. And third, there are now hundreds of model versions and many thousands of simulations, but there is no community-based mechanism for routinely monitoring model performance changes. An important change in the design of CMIP6 can help address these limitations. CMIP6 will include a small set standardized experiments as an ongoing exercise (CMIP "DECK": ongoing Diagnostic, Evaluation and Characterization of Klima), so that modeling groups can submit them at any time and not be overly constrained by deadlines. In this presentation, efforts to establish routine benchmarking of existing and future CMIP simulations will be described. To date, some benchmarking tools have been made available to all CMIP modeling groups to enable them to readily compare with CMIP5 simulations during the model development process. A natural extension of this effort is to make results from all CMIP simulations widely available, including the results from newer models as soon as the simulations become available for research. Making the results from routine performance tests readily accessible will help advance a more transparent model evaluation process.

  2. A Uranium Bioremediation Reactive Transport Benchmark

    SciTech Connect

    Yabusaki, Steven B.; Sengor, Sevinc; Fang, Yilin

    2015-06-01

    A reactive transport benchmark problem set has been developed based on in situ uranium bio-immobilization experiments that have been performed at a former uranium mill tailings site in Rifle, Colorado, USA. Acetate-amended groundwater stimulates indigenous microorganisms to catalyze the reduction of U(VI) to a sparingly soluble U(IV) mineral. The interplay between the flow, acetate loading periods and rates, microbially-mediated and geochemical reactions leads to dynamic behavior in metal- and sulfate-reducing bacteria, pH, alkalinity, and reactive mineral surfaces. The benchmark is based on an 8.5 m long one-dimensional model domain with constant saturated flow and uniform porosity. The 159-day simulation introduces acetate and bromide through the upgradient boundary in 14-day and 85-day pulses separated by a 10 day interruption. Acetate loading is tripled during the second pulse, which is followed by a 50 day recovery period. Terminal electron accepting processes for goethite, phyllosilicate Fe(III), U(VI), and sulfate are modeled using Monod-type rate laws. Major ion geochemistry modeled includes mineral reactions, as well as aqueous and surface complexation reactions for UO2++, Fe++, and H+. In addition to the dynamics imparted by the transport of the acetate pulses, U(VI) behavior involves the interplay between bioreduction, which is dependent on acetate availability, and speciation-controlled surface complexation, which is dependent on pH, alkalinity and available surface complexation sites. The general difficulty of this benchmark is the large number of reactions (74), multiple rate law formulations, a multisite uranium surface complexation model, and the strong interdependency and sensitivity of the reaction processes. Results are presented for three simulators: HYDROGEOCHEM, PHT3D, and PHREEQC.

  3. Five Obstacles to Grading Reform

    ERIC Educational Resources Information Center

    Guskey, Thomas R.

    2011-01-01

    Educators seeking to reform grading must combat five long-held traditions that stand as formidable obstacles to change: (1) Grades should provide the basis for differentiating students; (2) grade distributions should resemble a bell-shaped curve; (3) grades should be based on students' standing among classmates; (4) poor grades prompt students to

  4. Paperless Grades and Faculty Development.

    ERIC Educational Resources Information Center

    Hardy, James C.; Jones, Dennis; Turner, Sandy

    2003-01-01

    Provides overview of process of switching from paper-based grade reporting to computer-based grading. Authors found that paperless grading decreased number of errors, made student access more immediate, and reduced costs incurred by purchasing and storing grade-scanning sheets. Authors also argue that direct entry grading encourages faculty to

  5. Five Obstacles to Grading Reform

    ERIC Educational Resources Information Center

    Guskey, Thomas R.

    2011-01-01

    Educators seeking to reform grading must combat five long-held traditions that stand as formidable obstacles to change: (1) Grades should provide the basis for differentiating students; (2) grade distributions should resemble a bell-shaped curve; (3) grades should be based on students' standing among classmates; (4) poor grades prompt students to…

  6. Eddy currents benchmark analysis with COMSOL

    NASA Astrophysics Data System (ADS)

    Etcheverry, Javier I.; Ziella, Daniel H.

    2014-02-01

    Several benchmark studies in eddy currents have been proposed by the WFNDEC. Also, there are many analytical exact and approximate solution of interest to quantitatively evaluate the capability of numerical codes to predict the magnitude of the induced currents effects. In this paper we present the results obtained using COMSOL for a representative set of cases of interest. We analyze the accuracy of the results, but also the elements we believe are relevant from the perspective of the NDE specialist (simplicity, computer requirements, things to take into account to get good results, problems faced when solving, failures, etc.).

  7. Benchmark simulations of ICRF antenna coupling

    SciTech Connect

    Louche, F.; Lamalle, P. U.; Messiaen, A. M.; Compernolle, B. van; Milanesio, D.; Maggiora, R.

    2007-09-28

    The paper reports on ongoing benchmark numerical simulations of antenna input impedance parameters in the ion cyclotron range of frequencies with different coupling codes: CST Microwave Studio, TOPICA and ANTITER 2. In particular we study the validity of the approximation of a magnetized plasma slab by a dielectric medium of suitably chosen permittivity. Different antenna models are considered: a single-strap antenna, a 4-strap antenna and the 24-strap ITER antenna array. Whilst the diagonal impedances are mostly in good agreement, some differences between the mutual terms predicted by Microwave Studio and TOPICA have yet to be resolved.

  8. Benchmarking boiler tube failures - Part 1

    SciTech Connect

    Patrick, J.; Oldani, R.; von Behren, D.

    2005-10-01

    Boiler tube failures continue to be the leading cause of downtime for steam power plants. That should not be a surprise; a typical steam generator has miles of tubes that operate at high temperatures and pressures. Are your experiences comparable to those of your peers? Could you learn something from tube-leak benchmarking data that could improve the operation of your plant? The Electric Utility Cost Group (EUCG) recently completed a boiler-tube failure study that is available only to its members. But Power magazine has been given exclusive access to some of the results, published in this article. 4 figs.

  9. A CPU benchmark for protein crystallographic refinement.

    PubMed

    Bourne, P E; Hendrickson, W A

    1990-01-01

    The CPU time required to complete a cycle of restrained least-squares refinement of a protein structure from X-ray crystallographic data using the FORTRAN codes PROTIN and PROLSQ are reported for 48 different processors, ranging from single-user workstations to supercomputers. Sequential, vector, VLIW, multiprocessor, and RISC hardware architectures are compared using both a small and a large protein structure. Representative compile times for each hardware type are also given, and the improvement in run-time when coding for a specific hardware architecture considered. The benchmarks involve scalar integer and vector floating point arithmetic and are representative of the calculations performed in many scientific disciplines. PMID:2225778

  10. NAS Parallel Benchmarks. 2.4

    NASA Technical Reports Server (NTRS)

    VanderWijngaart, Rob; Biegel, Bryan A. (Technical Monitor)

    2002-01-01

    We describe a new problem size, called Class D, for the NAS Parallel Benchmarks (NPB), whose MPI source code implementation is being released as NPB 2.4. A brief rationale is given for how the new class is derived. We also describe the modifications made to the MPI (Message Passing Interface) implementation to allow the new class to be run on systems with 32-bit integers, and with moderate amounts of memory. Finally, we give the verification values for the new problem size.

  11. Benchmarking Linac Codes for the HIPPI Project

    SciTech Connect

    Franchi, A.; Franchetti, G.; Groening, L.; Hofmann, I.; Orzhekhovskaya, A.; Yaramyschev, S.; Duperrier, R.; Uriot, D.; Gerigk, F.; Sauer, A.

    2005-06-08

    In the framework of the HIPPI project a 3D linac code comparison and benchmarking program have been initiated. Particular attention is devoted to the validation of the space charge solvers, comparing the calculated electric field of a common initial distribution with a semi-analytical solution. In order to study the effects of numerical noise on the single particle dynamics, also the calculated single particle tune is compared with an analytical prediction. Particle tracking is eventually compared using the lattice of UNILAC DTL section, in preparation of validation with experimental emittance measurements to be carried out in the next two years.

  12. Benchmarking East Tennessee`s economic capacity

    SciTech Connect

    1995-04-20

    This presentation is comprised of viewgraphs delineating major economic factors operating in 15 counties in East Tennessee. The purpose of the information presented is to provide a benchmark analysis of economic conditions for use in guiding economic growth in the region. The emphasis of the presentation is economic infrastructure, which is classified into six categories: human resources, technology, financial resources, physical infrastructure, quality of life, and tax and regulation. Data for analysis of key indicators in each of the categories are presented. Preliminary analyses, in the form of strengths and weaknesses and comparison to reference groups, are given.

  13. Positive predictive values of the International Classification of Diseases, 10th revision diagnoses of Gram-negative septicemia/sepsis and urosepsis for presence of Gram-negative bacteremia

    PubMed Central

    Sgaard, Kirstine Kobbere; Thomsen, Reimar Wernich; Schnheyder, Henrik Carl; Sgaard, Mette

    2015-01-01

    Background Health care databases are a valuable resource for infectious disease epidemiology if diagnoses are accurately coded. We examined the ability of diagnostic coding to accurately identify Gram-negative bacteremia. Methods We randomly selected 100 patients among 1,703 patients recorded in the Danish National Patient Register with a diagnosis of either septicemia/sepsis due to other Gram-negative organisms (International Classification of Diseases, 10th revision [ICD-10] code A41.5) or urosepsis (ICD-10 code A41.9B) who had been admitted at Aalborg University Hospital, Denmark between 1994 and 2012. We estimated the positive predictive value (PPV) of these diagnoses for presence of Gram-negative bacteremia, using microbiological results from blood cultures as standard reference. Complementary clinical information was obtained from the medical records. Results Of the 100 patients registered with Gram-negative septicemia/sepsis or urosepsis, 72 had blood culture confirmed Gram-negative bacteremia, four patients had monomicrobial Gram-positive bacteremia, 21 patients had a negative blood culture, and three had no blood culture taken. The overall PPV of a blood culture confirmed Gram-negative bacteremia diagnosis was 72% (95% confidence interval [CI]: 62%81%); for ICD-10 code A41.5 it was 86% (95% CI: 74%94%) and for ICD-10 code A41.9B it was 55% (95% CI: 39%70%). The highest PPV was achieved for diagnoses registered in the most recent calendar period (20092012) and for secondary discharge diagnoses. Conclusion Our findings indicated good agreement between ICD-10 code A41.5 septicemia/sepsis due to other Gram-negative organisms and Gram-negative bacteremia, whereas ICD-10 code A41.9B urosepsis was not suited for identification of Gram-negative bacteremia. PMID:25709502

  14. Benchmarking in healthcare: selecting and working with partners.

    PubMed

    Benson, H R

    1995-01-01

    The process of selecting a benchmarking partner begins with gathering information to establish industry standards, identifying potential partners and supplying data on the subject to be benchmarked. Suggested sources of information are business and trade publications; investment industry analysts; journalists; trade associations and professional organizations; government research reports; disclosure documents; current and former employees; and product and service providers. Potential partners should be approached only after careful preparation of a project plan that includes information about the benchmarking team's organization and purpose, description of the subject and a statement of benefits for the prospective partner. After obtaining a commitment from the benchmarking partner, relevant comparative data is gathered and analyzed, using some of the following methods: library research, questionnaires, telephone surveys, site visits and consultants. Because benchmarking often involves sharing information with competitors, a code of ethical conduct has been developed by the International Benchmarking Clearinghouse. PMID:10141100

  15. Development of a California commercial building benchmarking database

    SciTech Connect

    Kinney, Satkartar; Piette, Mary Ann

    2002-05-17

    Building energy benchmarking is a useful starting point for commercial building owners and operators to target energy savings opportunities. There are a number of tools and methods for benchmarking energy use. Benchmarking based on regional data can provides more relevant information for California buildings than national tools such as Energy Star. This paper discusses issues related to benchmarking commercial building energy use and the development of Cal-Arch, a building energy benchmarking database for California. Currently Cal-Arch uses existing survey data from California's Commercial End Use Survey (CEUS), a largely underutilized wealth of information collected by California's major utilities. Doe's Commercial Building Energy Consumption Survey (CBECS) is used by a similar tool, Arch, and by a number of other benchmarking tools. Future versions of Arch/Cal-Arch will utilize additional data sources including modeled data and individual buildings to expand the database.

  16. 1997-98 NACE Employer Benchmark Survey. Learning from Each Other's Experiences: Employers Benchmark Recruitment Activities.

    ERIC Educational Resources Information Center

    Allen, Claudia

    1998-01-01

    Reports results of the 1997-98 Employer Benchmark Survey conducted by the National Association of Colleges and Employers. Respondents reported that offering the soon-to-be graduate a position where they can utilize state-of-the-art technology is currently the most effective way to hire technical experts. Discusses using media to develop an image

  17. Nebraska Science Standards: Grades K-12

    ERIC Educational Resources Information Center

    Nebraska Department of Education, 2010

    2010-01-01

    This publication presents the Nebraska Science Standards for Grades K-12. The standards are presented according to the following grades: (1) Grades K-2; (2) Grades 3-5; (3) Grades 6-8; and (4) Grades 9-12.

  18. Functionally graded boron carbide

    SciTech Connect

    Petrovic, J.J.; McClellan, K.J.; Kise, C.D.; Hoover, R.C.; Scarborough, W.K.

    1998-12-31

    Lightweight body armor is important for the protection of US soldiers in the field. Here, fabrication techniques were developed for producing graded porosity B{sub 4}C, and for producing aluminum-B{sub 4}C and epoxy-B{sub 4}C functionally graded materials. The key fabrication aspect was obtaining the graded porosity B{sub 4}C. The feasibility of producing graded porosity B{sub 4}C using a grading of carbon densification aid produced from a gradient of furfuryl alcohol carbon precursor was demonstrated. This approach is quite promising, but it was not optimized in the present investigation. Graded porosity B{sub 4}C materials were produced by a layering approach using different size distributions of B{sub 4}C powders in the green state, and then densifying the layered assembly by hot pressing at 1,900 C. The hardness of uninfiltrated graded B{sub 4}C, aluminum infiltrated B{sub 4}C, and epoxy infiltrated B{sub 4}C was observed to be similar.

  19. Correlation of Parental Involvement on Students' Benchmarks among Urban Fourth-Grade Elementary Students

    ERIC Educational Resources Information Center

    Rutledge, Erika L.

    2013-01-01

    The No Child Left Behind Act of 2001 raised awareness of the educational quality gap between European Americans and minorities, especially in reading. This gap underscores the importance that students master basic reading skills in order to achieve higher levels of reading proficiency. The purpose of this quantitative quasi-experimental study was…

  20. Physical Properties of Normal Grade Biodiesel and Winter Grade Biodiesel

    PubMed Central

    Sadrolhosseini, Amir Reza; Moksin, Mohd Maarof; Nang, Harrison Lau Lik; Norozi, Monir; Yunus, W. Mahmood Mat; Zakaria, Azmi

    2011-01-01

    In this study, optical and thermal properties of normal grade and winter grade palm oil biodiesel were investigated. Surface Plasmon Resonance and Photopyroelectric technique were used to evaluate the samples. The dispersion curve and thermal diffusivity were obtained. Consequently, the variation of refractive index, as a function of wavelength in normal grade biodiesel is faster than winter grade palm oil biodiesel, and the thermal diffusivity of winter grade biodiesel is higher than the thermal diffusivity of normal grade biodiesel. This is attributed to the higher palmitic acid C16:0 content in normal grade than in winter grade palm oil biodiesel. PMID:21731429

  1. Physical properties of normal grade biodiesel and winter grade biodiesel.

    PubMed

    Sadrolhosseini, Amir Reza; Moksin, Mohd Maarof; Nang, Harrison Lau Lik; Norozi, Monir; Yunus, W Mahmood Mat; Zakaria, Azmi

    2011-01-01

    In this study, optical and thermal properties of normal grade and winter grade palm oil biodiesel were investigated. Surface Plasmon Resonance and Photopyroelectric technique were used to evaluate the samples. The dispersion curve and thermal diffusivity were obtained. Consequently, the variation of refractive index, as a function of wavelength in normal grade biodiesel is faster than winter grade palm oil biodiesel, and the thermal diffusivity of winter grade biodiesel is higher than the thermal diffusivity of normal grade biodiesel. This is attributed to the higher palmitic acid C(16:0) content in normal grade than in winter grade palm oil biodiesel. PMID:21731429

  2. The NAS Parallel Benchmarks 2.1 Results

    NASA Technical Reports Server (NTRS)

    Saphir, William; Woo, Alex; Yarrow, Maurice

    1996-01-01

    We present performance results for version 2.1 of the NAS Parallel Benchmarks (NPB) on the following architectures: IBM SP2/66 MHz; SGI Power Challenge Array/90 MHz; Cray Research T3D; and Intel Paragon. The NAS Parallel Benchmarks are a widely-recognized suite of benchmarks originally designed to compare the performance of highly parallel computers with that of traditional supercomputers.

  3. MARS code developments, benchmarking and applications

    SciTech Connect

    Mokhov, N.V.

    2000-03-24

    Recent developments of the MARS Monte Carlo code system for simulation of hadronic and electromagnetic cascades in shielding, accelerator and detector components in the energy range from a fraction of an electron volt up to 100 TeV are described. The physical model of hadron and lepton interactions with nuclei and atoms has undergone substantial improvements. These include a new nuclear cross section library, a model for soft prior production, a cascade-exciton model, a dual parton model, deuteron-nucleus and neutrino-nucleus interaction models, a detailed description of negative hadron and muon absorption, and a unified treatment of muon and charged hadron electro-magnetic interactions with matter. New algorithms have been implemented into the code and benchmarked against experimental data. A new Graphical-User Interface has been developed. The code capabilities to simulate cascades and generate a variety of results in complex systems have been enhanced. The MARS system includes links to the MCNP code for neutron and photon transport below 20 MeV, to the ANSYS code for thermal and stress analyses and to the STRUCT code for multi-turn particle tracking in large synchrotrons and collider rings. Results of recent benchmarking of the MARS code are presented. Examples of non-trivial code applications are given for the Fermilab Booster and Main Injector, for a 1.5 MW target station and a muon storage ring.

  4. Benchmarking Radiation Transport Codes for Space Missions

    NASA Astrophysics Data System (ADS)

    Tripathi, Ram; Wilson, John; Townsend, Larry; Gabriel, Tony; Pinsky, Larry; Slaba, Tony

    2008-04-01

    For long duration and/or deep space human missions, protection from severe space radiation exposure is a challenging design constraint and may be a potential limiting factor. The space radiation environment consists of galactic cosmic rays (GCR), solar particle events (SPE), trapped radiation, and includes ions of all the known elements over a very broad energy range. These ions penetrate spacecraft materials producing nuclear fragments and secondary particles that damage biological tissues, microelectronic devices, and materials. Accurate risk assessments critically depend on the accuracy of the input information about the interaction of ions with materials, electronics and tissues and the radiation transport codes. Due to complexity of the problem and paucity of huge amount of experimental data, it is prudent to benchmark leading radiation transport codes to build increasing confidence in exposure estimates. The deterministic code HZETRN2006, and the Monte Carlo Codes HETC-HEDS and FLUKA, are used for benchmarking efforts. The SPE Webber spectrum and 1977 GCR radiation environments has been taken to make radiation dose exposure studies on aluminum shield followed by water target.

  5. Computational Thermochemistry and Benchmarking of Reliable Methods

    SciTech Connect

    Feller, David F.; Dixon, David A.; Dunning, Thom H.; Dupuis, Michel; McClemore, Doug; Peterson, Kirk A.; Xantheas, Sotiris S.; Bernholdt, David E.; Windus, Theresa L.; Chalasinski, Grzegorz; Fosada, Rubicelia; Olguim, Jorge; Dobbs, Kerwin D.; Frurip, Donald; Stevens, Walter J.; Rondan, Nelson; Chase, Jared M.; Nichols, Jeffrey A.

    2006-06-20

    During the first and second years of the Computational Thermochemistry and Benchmarking of Reliable Methods project, we completed several studies using the parallel computing capabilities of the NWChem software and Molecular Science Computing Facility (MSCF), including large-scale density functional theory (DFT), second-order Moeller-Plesset (MP2) perturbation theory, and CCSD(T) calculations. During the third year, we continued to pursue the computational thermodynamic and benchmarking studies outlined in our proposal. With the issues affecting the robustness of the coupled cluster part of NWChem resolved, we pursued studies of the heats-of-formation of compounds containing 5 to 7 first- and/or second-row elements and approximately 10 to 14 hydrogens. The size of these systems, when combined with the large basis sets (cc-pVQZ and aug-cc-pVQZ) that are necessary for extrapolating to the complete basis set limit, creates a formidable computational challenge, for which NWChem on NWMPP1 is well suited.

  6. The Medical Library Association Benchmarking Network: results*

    PubMed Central

    Dudden, Rosalind Farnam; Corcoran, Kate; Kaplan, Janice; Magouirk, Jeff; Rand, Debra C.; Smith, Bernie Todd

    2006-01-01

    Objective: This article presents some limited results from the Medical Library Association (MLA) Benchmarking Network survey conducted in 2002. Other uses of the data are also presented. Methods: After several years of development and testing, a Web-based survey opened for data input in December 2001. Three hundred eighty-five MLA members entered data on the size of their institutions and the activities of their libraries. The data from 344 hospital libraries were edited and selected for reporting in aggregate tables and on an interactive site in the Members-Only area of MLANET. The data represent a 16% to 23% return rate and have a 95% confidence level. Results: Specific questions can be answered using the reports. The data can be used to review internal processes, perform outcomes benchmarking, retest a hypothesis, refute a previous survey findings, or develop library standards. The data can be used to compare to current surveys or look for trends by comparing the data to past surveys. Conclusions: The impact of this project on MLA will reach into areas of research and advocacy. The data will be useful in the everyday working of small health sciences libraries as well as provide concrete data on the current practices of health sciences libraries. PMID:16636703

  7. Experimentally Relevant Benchmarks for Gyrokinetic Codes

    NASA Astrophysics Data System (ADS)

    Bravenec, Ronald

    2010-11-01

    Although benchmarking of gyrokinetic codes has been performed in the past, e.g., The Numerical Tokamak, The Cyclone Project, The Plasma Microturbulence Project, and various informal activities, these efforts have typically employed simple plasma models. For example, the Cyclone ``base case'' assumed shifted-circle flux surfaces, no magnetic transport, adiabatic electrons, no collisions nor impurities, ?i << a (?i the ion gyroradius and a the minor radius), and no ExB flow shear. This work presents comparisons of linear frequencies and nonlinear fluxes from GYRO and GS2 with none of the above approximations except ?i << a and no ExB flow shear. The comparisons are performed at two radii of a DIII-D plasma, one in the confinement region (r/a = 0.5) and the other closer to the edge (r/a = 0.7). Many of the plasma parameters differ by a factor of two between these two locations. Good agreement between GYRO and GS2 is found when neglecting collisions. However, differences are found when including e-i collisions (Lorentz model). The sources of the discrepancy are unknown as of yet. Nevertheless, two collisionless benchmarks have been formulated with considerably different plasma parameters. Acknowledgements to J. Candy, E. Belli, and M. Barnes.

  8. Parallel Ada benchmarks for the SVMS

    NASA Technical Reports Server (NTRS)

    Collard, Philippe E.

    1990-01-01

    The use of parallel processing paradigm to design and develop faster and more reliable computers appear to clearly mark the future of information processing. NASA started the development of such an architecture: the Spaceborne VHSIC Multi-processor System (SVMS). Ada will be one of the languages used to program the SVMS. One of the unique characteristics of Ada is that it supports parallel processing at the language level through the tasking constructs. It is important for the SVMS project team to assess how efficiently the SVMS architecture will be implemented, as well as how efficiently Ada environment will be ported to the SVMS. AUTOCLASS II, a Bayesian classifier written in Common Lisp, was selected as one of the benchmarks for SVMS configurations. The purpose of the R and D effort was to provide the SVMS project team with the version of AUTOCLASS II, written in Ada, that would make use of Ada tasking constructs as much as possible so as to constitute a suitable benchmark. Additionally, a set of programs was developed that would measure Ada tasking efficiency on parallel architectures as well as determine the critical parameters influencing tasking efficiency. All this was designed to provide the SVMS project team with a set of suitable tools in the development of the SVMS architecture.

  9. Benchmarking database performance for genomic data.

    PubMed

    Khushi, Matloob

    2015-06-01

    Genomic regions represent features such as gene annotations, transcription factor binding sites and epigenetic modifications. Performing various genomic operations such as identifying overlapping/non-overlapping regions or nearest gene annotations are common research needs. The data can be saved in a database system for easy management, however, there is no comprehensive database built-in algorithm at present to identify overlapping regions. Therefore I have developed a novel region-mapping (RegMap) SQL-based algorithm to perform genomic operations and have benchmarked the performance of different databases. Benchmarking identified that PostgreSQL extracts overlapping regions much faster than MySQL. Insertion and data uploads in PostgreSQL were also better, although general searching capability of both databases was almost equivalent. In addition, using the algorithm pair-wise, overlaps of >1000 datasets of transcription factor binding sites and histone marks, collected from previous publications, were reported and it was found that HNF4G significantly co-locates with cohesin subunit STAG1 (SA1).Inc. PMID:25560631

  10. Recommendations for Benchmarking Preclinical Studies of Nanomedicines.

    PubMed

    Dawidczyk, Charlene M; Russell, Luisa M; Searson, Peter C

    2015-10-01

    Nanoparticle-based delivery systems provide new opportunities to overcome the limitations associated with traditional small-molecule drug therapy for cancer and to achieve both therapeutic and diagnostic functions in the same platform. Preclinical trials are generally designed to assess therapeutic potential and not to optimize the design of the delivery platform. Consequently, progress in developing design rules for cancer nanomedicines has been slow, hindering progress in the field. Despite the large number of preclinical trials, several factors restrict comparison and benchmarking of different platforms, including variability in experimental design, reporting of results, and the lack of quantitative data. To solve this problem, we review the variables involved in the design of preclinical trials and propose a protocol for benchmarking that we recommend be included in in vivo preclinical studies of drug-delivery platforms for cancer therapy. This strategy will contribute to building the scientific knowledge base that enables development of design rules and accelerates the translation of new technologies. PMID:26249177

  11. Direct data access protocols benchmarking on DPM

    NASA Astrophysics Data System (ADS)

    Furano, Fabrizio; Devresse, Adrien; Keeble, Oliver; Mancinelli, Valentina

    2015-12-01

    The Disk Pool Manager is an example of a multi-protocol, multi-VO system for data access on the Grid that went though a considerable technical evolution in the last years. Among other features, its architecture offers the opportunity of testing its different data access frontends under exactly the same conditions, including hardware and backend software. This characteristic inspired the idea of collecting monitoring information from various testbeds in order to benchmark the behaviour of the HTTP and Xrootd protocols for the use case of data analysis, batch or interactive. A source of information is the set of continuous tests that are run towards the worldwide endpoints belonging to the DPM Collaboration, which accumulated relevant statistics in its first year of activity. On top of that, the DPM releases are based on multiple levels of automated testing that include performance benchmarks of various kinds, executed regularly every day. At the same time, the recent releases of DPM can report monitoring information about any data access protocol to the same monitoring infrastructure that is used to monitor the Xrootd deployments. Our goal is to evaluate under which circumstances the HTTP-based protocols can be good enough for batch or interactive data access. In this contribution we show and discuss the results that our test systems have collected under the circumstances that include ROOT analyses using TTreeCache and stress tests on the metadata performance.

  12. EVA Health and Human Performance Benchmarking Study

    NASA Technical Reports Server (NTRS)

    Abercromby, A. F.; Norcross, J.; Jarvis, S. L.

    2016-01-01

    Multiple HRP Risks and Gaps require detailed characterization of human health and performance during exploration extravehicular activity (EVA) tasks; however, a rigorous and comprehensive methodology for characterizing and comparing the health and human performance implications of current and future EVA spacesuit designs does not exist. This study will identify and implement functional tasks and metrics, both objective and subjective, that are relevant to health and human performance, such as metabolic expenditure, suit fit, discomfort, suited postural stability, cognitive performance, and potentially biochemical responses for humans working inside different EVA suits doing functional tasks under the appropriate simulated reduced gravity environments. This study will provide health and human performance benchmark data for humans working in current EVA suits (EMU, Mark III, and Z2) as well as shirtsleeves using a standard set of tasks and metrics with quantified reliability. Results and methodologies developed during this test will provide benchmark data against which future EVA suits, and different suit configurations (eg, varied pressure, mass, CG) may be reliably compared in subsequent tests. Results will also inform fitness for duty standards as well as design requirements and operations concepts for future EVA suits and other exploration systems.

  13. REVISED STREAM CODE AND WASP5 BENCHMARK

    SciTech Connect

    Chen, K

    2005-05-01

    STREAM is an emergency response code that predicts downstream pollutant concentrations for releases from the SRS area to the Savannah River. The STREAM code uses an algebraic equation to approximate the solution of the one dimensional advective transport differential equation. This approach generates spurious oscillations in the concentration profile when modeling long duration releases. To improve the capability of the STREAM code to model long-term releases, its calculation module was replaced by the WASP5 code. WASP5 is a US EPA water quality analysis program that simulates one-dimensional pollutant transport through surface water. Test cases were performed to compare the revised version of STREAM with the existing version. For continuous releases, results predicted by the revised STREAM code agree with physical expectations. The WASP5 code was benchmarked with the US EPA 1990 and 1991 dye tracer studies, in which the transport of the dye was measured from its release at the New Savannah Bluff Lock and Dam downstream to Savannah. The peak concentrations predicted by the WASP5 agreed with the measurements within {+-}20.0%. The transport times of the dye concentration peak predicted by the WASP5 agreed with the measurements within {+-}3.6%. These benchmarking results demonstrate that STREAM should be capable of accurately modeling releases from SRS outfalls.

  14. Non-Markovianity in Randomized Benchmarking

    NASA Astrophysics Data System (ADS)

    Ball, Harrison; Stace, Tom M.; Biercuk, Michael J.

    2015-03-01

    Randomized benchmarking is routinely employed to recover information about the fidelity of a quantum operation by exploiting probabilistic twirling errors over an implementation of the Clifford group. Standard assumptions of Markovianity in the underlying noise environment, however, remain at odds with realistic, correlated noise encountered in real systems. We model single-qubit randomized benchmarking experiments as a sequence of ideal Clifford operations interleaved with stochastic dephasing errors, implemented as unitary rotations about ?z. Successive error rotations map to a sequence of random variables whose correlations introduce non-Markovian effects emulating realistic colored-noise environments. The Markovian limit is recovered by turning off all correlations, reducing each error to an independent Gaussian-distributed random variable. We examine the dependence of the statistical distribution of fidelity outcomes on these noise correlations, deriving analytic expressions for probability density functions and related statistics for relevant fidelity metrics. This enables us to characterize and bear out the distinction between the Markovian and non-Markovian cases, with implications for interpretation and handling of experimental data.

  15. Improving Mass Balance Modeling of Benchmark Glaciers

    NASA Astrophysics Data System (ADS)

    van Beusekom, A. E.; March, R. S.; O'Neel, S.

    2009-12-01

    The USGS monitors long-term glacier mass balance at three benchmark glaciers in different climate regimes. The coastal and continental glaciers are represented by Wolverine and Gulkana Glaciers in Alaska, respectively. Field measurements began in 1966 and continue. We have reanalyzed the published balance time series with more modern methods and recomputed reference surface and conventional balances. Addition of the most recent data shows a continuing trend of mass loss. We compare the updated balances to the previously accepted balances and discuss differences. Not all balance quantities can be determined from the field measurements. For surface processes, we model missing information with an improved degree-day model. Degree-day models predict ablation from the sum of daily mean temperatures and an empirical degree-day factor. We modernize the traditional degree-day model as well as derive new degree-day factors in an effort to closer match the balance time series and thus better predict the future state of the benchmark glaciers. For subsurface processes, we model the refreezing of meltwater for internal accumulation. We examine the sensitivity of the balance time series to the subsurface process of internal accumulation, with the goal of determining the best way to include internal accumulation into balance estimates.

  16. Simple mathematical law benchmarks human confrontations

    PubMed Central

    Johnson, Neil F.; Medina, Pablo; Zhao, Guannan; Messinger, Daniel S.; Horgan, John; Gill, Paul; Bohorquez, Juan Camilo; Mattson, Whitney; Gangi, Devon; Qi, Hong; Manrique, Pedro; Velasquez, Nicolas; Morgenstern, Ana; Restrepo, Elvira; Johnson, Nicholas; Spagat, Michael; Zarama, Roberto

    2013-01-01

    Many high-profile societal problems involve an individual or group repeatedly attacking another from child-parent disputes, sexual violence against women, civil unrest, violent conflicts and acts of terror, to current cyber-attacks on national infrastructure and ultrafast cyber-trades attacking stockholders. There is an urgent need to quantify the likely severity and timing of such future acts, shed light on likely perpetrators, and identify intervention strategies. Here we present a combined analysis of multiple datasets across all these domains which account for >100,000 events, and show that a simple mathematical law can benchmark them all. We derive this benchmark and interpret it, using a minimal mechanistic model grounded by state-of-the-art fieldwork. Our findings provide quantitative predictions concerning future attacks; a tool to help detect common perpetrators and abnormal behaviors; insight into the trajectory of a lone wolf'; identification of a critical threshold for spreading a message or idea among perpetrators; an intervention strategy to erode the most lethal clusters; and more broadly, a quantitative starting point for cross-disciplinary theorizing about human aggression at the individual and group level, in both real and online worlds. PMID:24322528

  17. Simple mathematical law benchmarks human confrontations

    NASA Astrophysics Data System (ADS)

    Johnson, Neil F.; Medina, Pablo; Zhao, Guannan; Messinger, Daniel S.; Horgan, John; Gill, Paul; Bohorquez, Juan Camilo; Mattson, Whitney; Gangi, Devon; Qi, Hong; Manrique, Pedro; Velasquez, Nicolas; Morgenstern, Ana; Restrepo, Elvira; Johnson, Nicholas; Spagat, Michael; Zarama, Roberto

    2013-12-01

    Many high-profile societal problems involve an individual or group repeatedly attacking another - from child-parent disputes, sexual violence against women, civil unrest, violent conflicts and acts of terror, to current cyber-attacks on national infrastructure and ultrafast cyber-trades attacking stockholders. There is an urgent need to quantify the likely severity and timing of such future acts, shed light on likely perpetrators, and identify intervention strategies. Here we present a combined analysis of multiple datasets across all these domains which account for >100,000 events, and show that a simple mathematical law can benchmark them all. We derive this benchmark and interpret it, using a minimal mechanistic model grounded by state-of-the-art fieldwork. Our findings provide quantitative predictions concerning future attacks; a tool to help detect common perpetrators and abnormal behaviors; insight into the trajectory of a `lone wolf' identification of a critical threshold for spreading a message or idea among perpetrators; an intervention strategy to erode the most lethal clusters; and more broadly, a quantitative starting point for cross-disciplinary theorizing about human aggression at the individual and group level, in both real and online worlds.

  18. Performance Evaluation of Supercomputers using HPCC and IMB Benchmarks

    NASA Technical Reports Server (NTRS)

    Saini, Subhash; Ciotti, Robert; Gunney, Brian T. N.; Spelce, Thomas E.; Koniges, Alice; Dossa, Don; Adamidis, Panagiotis; Rabenseifner, Rolf; Tiyyagura, Sunil R.; Mueller, Matthias; Fatoohi, Rod

    2006-01-01

    The HPC Challenge (HPCC) benchmark suite and the Intel MPI Benchmark (IMB) are used to compare and evaluate the combined performance of processor, memory subsystem and interconnect fabric of five leading supercomputers - SGI Altix BX2, Cray XI, Cray Opteron Cluster, Dell Xeon cluster, and NEC SX-8. These five systems use five different networks (SGI NUMALINK4, Cray network, Myrinet, InfiniBand, and NEC IXS). The complete set of HPCC benchmarks are run on each of these systems. Additionally, we present Intel MPI Benchmarks (IMB) results to study the performance of 11 MPI communication functions on these systems.

  19. Hospital Energy Benchmarking Guidance - Version 1.0

    SciTech Connect

    Singer, Brett C.

    2009-09-08

    This document describes an energy benchmarking framework for hospitals. The document is organized as follows. The introduction provides a brief primer on benchmarking and its application to hospitals. The next two sections discuss special considerations including the identification of normalizing factors. The presentation of metrics is preceded by a description of the overall framework and the rationale for the grouping of metrics. Following the presentation of metrics, a high-level protocol is provided. The next section presents draft benchmarks for some metrics; benchmarks are not available for many metrics owing to a lack of data. This document ends with a list of research needs for further development.

  20. Research on PSD-based spatial benchmark system

    NASA Astrophysics Data System (ADS)

    Xu, Weilong; Gu, Jinliang; Luo, Honge; Xia, Yan

    2014-09-01

    This paper proposes a kind of ballistic range spatial benchmark control system, which is based on the Position Sensitive Detector (PSD) and multidimensional stepper motor console. The method how to control a stepper motor automatic implement space benchmark according to the position of light spot on the PSD is also given. Besides, the camera stations in ballistic range have been calibrated by the spatial benchmark devices. Experimental results show that PSD-based spatial benchmark system can effectively eliminate errors caused by the artificial operation between different stations, and its spatial precision is superior to 0.1mm.

  1. Rethinking Science and Technology Education To Meet the Demands of Future Generations in a Changing World. International Organization for Science and Technology Education (IOSTE) Symposium Proceedings (10th, Foz do Iguacu, Parana, Brazil, July 28-August 2, 2002). Volumes I [and] II.

    ERIC Educational Resources Information Center

    Bizzo, Nelio, Ed.; Kawasaki, Clarice Sumi, Ed.; Ferracioli, Laercio, Ed.; Leyser da Rosa, Vivian, Ed.

    This document is the proceedings of the 10th annual meeting of the International Organization for Science and Technology Education (IOSTE). Papers include: (1) "Liberal Education, Information Assessment and Argumentation in Science-LIA" (Andreas Quale, Anders Isnes, Terje Kristensen, and Ketil Mathiassen); (2) "Placing the History and the…

  2. PIALA 2000: Libraries and Archives--Where Information and Language Literacy Begin [and] Engaged Readers and Writers in Multicultural Island Communities. Selected Papers from the 10th Pacific Islands Association of Libraries and Archives Conference Joint with the 13th Annual Regional Language Arts Conference (Tumon, Guam, November 9-11, 2000)

    ERIC Educational Resources Information Center

    Cohen, Arlene, Ed.; Quan, Clarisa G., Ed.

    2004-01-01

    This proceedings combines presentations from the jointly held 10th Annual PIALA 2000 Conference and the 13th Annual Regional Language Arts Conference. The volume begins with the welcoming remarks of Mary L. Silk, Christine Ku Scott-Smith, Antonio R. Umpingco, Delia Munoz Rosal, Lawrence Kasperbauer, Rosie Tainatongo, Richard S. Tom, Mary L.…

  3. How Consistent Are Course Grades? An Examination of Differential Grading

    ERIC Educational Resources Information Center

    Rauschenberg, Samuel

    2014-01-01

    Differential grading occurs when students in courses with the same content and curriculum receive inconsistent grades across teachers, schools, or districts. It may be due to many factors, including differences in teacher grading standards, district grading policies, student behavior, teacher stereotypes, teacher quality, and curriculum adherence.

  4. Measuring grade inflation: a clinical grade discrepancy score.

    PubMed

    Paskausky, Anna L; Simonelli, M Colleen

    2014-08-01

    Grade inflation presents pedagogical and safety concerns for nursing educators and is defined as a "greater percentage of excellent scores than student performances warrant" (Speer etal., 2000, p. 112). This descriptive correlational study evaluated the relationship of licensure exam-style final written exams and faculty assigned clinical grades from undergraduate students (N=281) for evidence of grade inflation at a private undergraduate nursing program in the Northeast of the United States and developed a new measurement of grade inflation, the clinical grade discrepancy score. This measurement can be used in programs where clinical competency is graded on a numeric scale. Evidence suggested grade inflation was present and the clinical grade discrepancy score was an indicator of the severity of grade inflation. The correlation between licensure-style final written exams and faculty assigned clinical grades was moderate to low at 0.357. The clinical grade discrepancy scores were 98% positive indicating likely grade inflation. Some 70% of clinical grade discrepancy scores indicated a difference of student licensure-style final written exams and faculty assigned clinical grades of at least one full letter grade (10 points out of 100). Use of this new measure as a tool in exploring the prevalence of grade inflation and implications for patient safety are discussed. PMID:24602828

  5. NASA Indexing Benchmarks: Evaluating Text Search Engines

    NASA Technical Reports Server (NTRS)

    Esler, Sandra L.; Nelson, Michael L.

    1997-01-01

    The current proliferation of on-line information resources underscores the requirement for the ability to index collections of information and search and retrieve them in a convenient manner. This study develops criteria for analytically comparing the index and search engines and presents results for a number of freely available search engines. A product of this research is a toolkit capable of automatically indexing, searching, and extracting performance statistics from each of the focused search engines. This toolkit is highly configurable and has the ability to run these benchmark tests against other engines as well. Results demonstrate that the tested search engines can be grouped into two levels. Level one engines are efficient on small to medium sized data collections, but show weaknesses when used for collections 100MB or larger. Level two search engines are recommended for data collections up to and beyond 100MB.

  6. DRG benchmarking study establishes national coding norms.

    PubMed

    Vaul, J H

    1998-05-01

    With the increase in fraud and abuse investigations, healthcare financial managers should examine their organization's medical record coding procedures. The Federal government and third-party payers are looking specifically for improper billing of outpatient services, unbundling of procedures to increase payment, assigning higher-paying DRG codes for inpatient claims, and other abuses. A recent benchmarking study of Medicare Provider Analysis and Review (MEDPAR) data has established national norms for hospital coding and case mix based on DRGs and has revealed the majority of atypical coding cases fall into six DRG pairs. Organizations with a greater percentage of atypical cases--those more likely to be scrutinized by Federal investigators--will want to conduct suitable review and be sure appropriate documentation exists to justify the coding. PMID:10179440

  7. Benchmarking and testing the "Sea Level Equation

    NASA Astrophysics Data System (ADS)

    Spada, G.; Barletta, V. R.; Klemann, V.; van der Wal, W.; James, T. S.; Simon, K.; Riva, R. E. M.; Martinec, Z.; Gasperini, P.; Lund, B.; Wolf, D.; Vermeersen, L. L. A.; King, M. A.

    2012-04-01

    The study of the process of Glacial Isostatic Adjustment (GIA) and of the consequent sea level variations is gaining an increasingly important role within the geophysical community. Understanding the response of the Earth to the waxing and waning ice sheets is crucial in various contexts, ranging from the interpretation of modern satellite geodetic measurements to the projections of future sea level trends in response to climate change. All the processes accompanying GIA can be described solving the so-called Sea Level Equation (SLE), an integral equation that accounts for the interactions between the ice sheets, the solid Earth, and the oceans. Modern approaches to the SLE are based on various techniques that range from purely analytical formulations to fully numerical methods. Despite various teams independently investigating GIA, we do not have a suitably large set of agreed numerical results through which the methods may be validated. Following the example of the mantle convection community and our recent successful Benchmark for Post Glacial Rebound codes (Spada et al., 2011, doi: 10.1111/j.1365-246X.2011.04952.x), here we present the results of a benchmark study of independently developed codes designed to solve the SLE. This study has taken place within a collaboration facilitated through the European Cooperation in Science and Technology (COST) Action ES0701. The tests involve predictions of past and current sea level variations, and 3D deformations of the Earth surface. In spite of the signi?cant differences in the numerical methods employed, the test computations performed so far show a satisfactory agreement between the results provided by the participants. The differences found, which can be often attributed to the different numerical algorithms employed within the community, help to constrain the intrinsic errors in model predictions. These are of fundamental importance for a correct interpretation of the geodetic variations observed today, and particularly for the evaluation of climate-driven sea level variations.

  8. Benchmarking Competitiveness: Is America's Technological Hegemony Waning?

    NASA Astrophysics Data System (ADS)

    Lubell, Michael S.

    2006-03-01

    For more than half a century, by almost every standard, the United States has been the world's leader in scientific discovery, innovation and technological competitiveness. To a large degree, that dominant position stemmed from the circumstances our nation inherited at the conclusion of the World War Two: we were, in effect, the only major nation left standing that did not have to repair serious war damage. And we found ourselves with an extraordinary science and technology base that we had developed for military purposes. We had the laboratories -- industrial, academic and government -- as well as the scientific and engineering personnel -- many of them immigrants who had escaped from war-time Europe. What remained was to convert the wartime machinery into peacetime uses. We adopted private and public policies that accomplished the transition remarkably well, and we have prospered ever since. Our higher education system, our protection of intellectual property rights, our venture capital system, our entrepreneurial culture and our willingness to commit government funds for the support of science and engineering have been key components to our success. But recent competitiveness benchmarks suggest that our dominance is waning rapidly, in part because other nations have begun to emulate our successful model, in part because globalization has ``flattened'' the world and in part because we have been reluctant to pursue the public policies that are necessary to ensure our leadership. We will examine these benchmarks and explore the policy changes that are needed to keep our nation's science and technology enterprise vibrant and our economic growth on an upward trajectory.

  9. Performance Evaluation and Benchmarking of Intelligent Systems

    SciTech Connect

    Madhavan, Raj; Messina, Elena; Tunstel, Edward

    2009-09-01

    To design and develop capable, dependable, and affordable intelligent systems, their performance must be measurable. Scientific methodologies for standardization and benchmarking are crucial for quantitatively evaluating the performance of emerging robotic and intelligent systems technologies. There is currently no accepted standard for quantitatively measuring the performance of these systems against user-defined requirements; and furthermore, there is no consensus on what objective evaluation procedures need to be followed to understand the performance of these systems. The lack of reproducible and repeatable test methods has precluded researchers working towards a common goal from exchanging and communicating results, inter-comparing system performance, and leveraging previous work that could otherwise avoid duplication and expedite technology transfer. Currently, this lack of cohesion in the community hinders progress in many domains, such as manufacturing, service, healthcare, and security. By providing the research community with access to standardized tools, reference data sets, and open source libraries of solutions, researchers and consumers will be able to evaluate the cost and benefits associated with intelligent systems and associated technologies. In this vein, the edited book volume addresses performance evaluation and metrics for intelligent systems, in general, while emphasizing the need and solutions for standardized methods. To the knowledge of the editors, there is not a single book on the market that is solely dedicated to the subject of performance evaluation and benchmarking of intelligent systems. Even books that address this topic do so only marginally or are out of date. The research work presented in this volume fills this void by drawing from the experiences and insights of experts gained both through theoretical development and practical implementation of intelligent systems in a variety of diverse application domains. The book presents a detailed and coherent picture of state-of-the-art, recent developments, and further research areas in intelligent systems.

  10. Ground truth and benchmarks for performance evaluation

    NASA Astrophysics Data System (ADS)

    Takeuchi, Ayako; Shneier, Michael; Hong, Tsai Hong; Chang, Tommy; Scrapper, Christopher; Cheok, Geraldine S.

    2003-09-01

    Progress in algorithm development and transfer of results to practical applications such as military robotics requires the setup of standard tasks, of standard qualitative and quantitative measurements for performance evaluation and validation. Although the evaluation and validation of algorithms have been discussed for over a decade, the research community still faces a lack of well-defined and standardized methodology. The range of fundamental problems include a lack of quantifiable measures of performance, a lack of data from state-of-the-art sensors in calibrated real-world environments, and a lack of facilities for conducting realistic experiments. In this research, we propose three methods for creating ground truth databases and benchmarks using multiple sensors. The databases and benchmarks will provide researchers with high quality data from suites of sensors operating in complex environments representing real problems of great relevance to the development of autonomous driving systems. At NIST, we have prototyped a High Mobility Multi-purpose Wheeled Vehicle (HMMWV) system with a suite of sensors including a Riegl ladar, GDRS ladar, stereo CCD, several color cameras, Global Position System (GPS), Inertial Navigation System (INS), pan/tilt encoders, and odometry . All sensors are calibrated with respect to each other in space and time. This allows a database of features and terrain elevation to be built. Ground truth for each sensor can then be extracted from the database. The main goal of this research is to provide ground truth databases for researchers and engineers to evaluate algorithms for effectiveness, efficiency, reliability, and robustness, thus advancing the development of algorithms.

  11. Benchmarking management practices in Australian public healthcare.

    PubMed

    Agarwal, Renu; Green, Roy; Agarwal, Neeru; Randhawa, Krithika

    2016-03-21

    Purpose - The purpose of this paper is to investigate the quality of management practices of public hospitals in the Australian healthcare system, specifically those in the state-managed health systems of Queensland and New South Wales (NSW). Further, the authors assess the management practices of Queensland and NSW public hospitals jointly and globally benchmark against those in the health systems of seven other countries, namely, USA, UK, Sweden, France, Germany, Italy and Canada. Design/methodology/approach - In this study, the authors adapt the unique and globally deployed Bloom et al. (2009) survey instrument that uses a "double blind, double scored" methodology and an interview-based scoring grid to measure and internationally benchmark the management practices in Queensland and NSW public hospitals based on 21 management dimensions across four broad areas of management - operations, performance monitoring, targets and people management. Findings - The findings reveal the areas of strength and potential areas of improvement in the Queensland and NSW Health hospital management practices when compared with public hospitals in seven countries, namely, USA, UK, Sweden, France, Germany, Italy and Canada. Together, Queensland and NSW Health hospitals perform best in operations management followed by performance monitoring. While target management presents scope for improvement, people management is the sphere where these Australian hospitals lag the most. Practical implications - This paper is of interest to both hospital administrators and health care policy-makers aiming to lift management quality at the hospital level as well as at the institutional level, as a vehicle to consistently deliver sustainable high-quality health services. Originality/value - This study provides the first internationally comparable robust measure of management capability in Australian public hospitals, where hospitals are run independently by the state-run healthcare systems. Additionally, this research study contributes to the empirical evidence base on the quality of management practices in the Australian public healthcare systems of Queensland and NSW. PMID:26964848

  12. A Benchmark Study on Casting Residual Stress

    SciTech Connect

    Johnson, Eric M.; Watkins, Thomas R; Schmidlin, Joshua E; Dutler, S. A.

    2012-01-01

    Stringent regulatory requirements, such as Tier IV norms, have pushed the cast iron for automotive applications to its limit. The castings need to be designed with closer tolerances by incorporating hitherto unknowns, such as residual stresses arising due to thermal gradients, phase and microstructural changes during solidification phenomenon. Residual stresses were earlier neglected in the casting designs by incorporating large factors of safety. Experimental measurement of residual stress in a casting through neutron or X-ray diffraction, sectioning or hole drilling, magnetic, electric or photoelastic measurements is very difficult and time consuming exercise. A detailed multi-physics model, incorporating thermo-mechanical and phase transformation phenomenon, provides an attractive alternative to assess the residual stresses generated during casting. However, before relying on the simulation methodology, it is important to rigorously validate the prediction capability by comparing it to experimental measurements. In the present work, a benchmark study was undertaken for casting residual stress measurements through neutron diffraction, which was subsequently used to validate the accuracy of simulation prediction. The stress lattice specimen geometry was designed such that subsequent castings would generate adequate residual stresses during solidification and cooling, without any cracks. The residual stresses in the cast specimen were measured using neutron diffraction. Considering the difficulty in accessing the neutron diffraction facility, these measurements can be considered as benchmark for casting simulation validations. Simulations were performed using the identical specimen geometry and casting conditions for predictions of residual stresses. The simulation predictions were found to agree well with the experimentally measured residual stresses. The experimentally validated model can be subsequently used to predict residual stresses in different cast components. This enables incorporation of the residual stresses at the design phase along with external loads for accurate predictions of fatigue and fracture performance of the cast components.

  13. Yoga May Mitigate Decreases in High School Grades

    PubMed Central

    Butzer, Bethany; van Over, Max; Noggle Taylor, Jessica J.; Khalsa, Sat Bir S.

    2015-01-01

    This study involves an exploratory examination of the effects of a 12-week school-based yoga intervention on changes in grade point average (GPA) in 9th and 10th grade students. Participants included 95 high school students who had registered for physical education (PE) in spring 2010. PE class sections were group randomized to receive either a yoga intervention or a PE-as-usual control condition. The yoga intervention took place during the entire third quarter and half of the fourth quarter of the school year, and quarterly GPA was collected via school records at the end of the school year. Results revealed a significant interaction between group and quarter suggesting that GPA differed between the yoga and control groups over time. Post hoc tests revealed that while both groups exhibited a general decline in GPA over the school year, the control group exhibited a significantly greater decline in GPA from quarter 1 to quarter 3 than the yoga group. Both groups showed equivalent declines in GPA in quarter 4 after the yoga intervention had ended. The results suggest that yoga may have a protective effect on academic performance by preventing declines in GPA; however these preventive effects may not persist once yoga practice is discontinued. PMID:26347787

  14. Benchmarks for Psychotherapy Efficacy in Adult Major Depression

    ERIC Educational Resources Information Center

    Minami, Takuya; Wampold, Bruce E.; Serlin, Ronald C.; Kircher, John C.; Brown, George S.

    2007-01-01

    This study estimates pretreatment-posttreatment effect size benchmarks for the treatment of major depression in adults that may be useful in evaluating psychotherapy effectiveness in clinical practice. Treatment efficacy benchmarks for major depression were derived for 3 different types of outcome measures: the Hamilton Rating Scale for Depression

  15. 7 CFR 1709.5 - Determination of energy cost benchmarks.

    Code of Federal Regulations, 2011 CFR

    2011-01-01

    ... 7 Agriculture 11 2011-01-01 2011-01-01 false Determination of energy cost benchmarks. 1709.5... SERVICE, DEPARTMENT OF AGRICULTURE ASSISTANCE TO HIGH ENERGY COST COMMUNITIES General Requirements § 1709.5 Determination of energy cost benchmarks. (a) The Administrator shall establish, using the...

  16. Enhancing knowledge of rangeland ecological processes with benchmark ecological sites

    Technology Transfer Automated Retrieval System (TEKTRAN)

    A benchmark ecological site is one that has the greatest potential to yield data and information about ecological functions, processes, and the effects of management or climate changes on a broad area or critical ecological zone. A benchmark ecological site represents other similar sites in a major ...

  17. Constructing Benchmarks for Monitoring Purposes: Evidence from South Africa

    ERIC Educational Resources Information Center

    Scherman, Vanessa; Howie, Sarah J.; Bosker, Roel J.

    2011-01-01

    In information-rich environments, schools are often presented with a myriad of data from which decisions need to be made. The use of the information on a classroom level may be facilitated if performance could be described in terms of levels of proficiency or benchmarks. The aim of this article is to explore benchmarks using data from a monitoring

  18. Teaching Benchmark Strategy for Fifth-Graders in Taiwan

    ERIC Educational Resources Information Center

    Yang, Der-Ching; Lai, M. L.

    2013-01-01

    The key purpose of this study was how we taught the use of benchmark strategy when comparing fraction for fifth-graders in Taiwan. 26 fifth graders from a public elementary in south Taiwan were selected to join this study. Results of this case study showed that students had a much progress on the use of benchmark strategy when comparing fraction…

  19. Benchmark Evaluation of HTR-PROTEUS Pebble Bed Experimental Program

    DOE PAGESBeta

    Bess, John D.; Montierth, Leland; Köberl, Oliver; Snoj, Luka

    2014-10-09

    Benchmark models were developed to evaluate 11 critical core configurations of the HTR-PROTEUS pebble bed experimental program. Various additional reactor physics measurements were performed as part of this program; currently only a total of 37 absorber rod worth measurements have been evaluated as acceptable benchmark experiments for Cores 4, 9, and 10. Dominant uncertainties in the experimental keff for all core configurations come from uncertainties in the ²³⁵U enrichment of the fuel, impurities in the moderator pebbles, and the density and impurity content of the radial reflector. Calculations of keff with MCNP5 and ENDF/B-VII.0 neutron nuclear data are greatermore » than the benchmark values but within 1% and also within the 3σ uncertainty, except for Core 4, which is the only randomly packed pebble configuration. Repeated calculations of keff with MCNP6.1 and ENDF/B-VII.1 are lower than the benchmark values and within 1% (~3σ) except for Cores 5 and 9, which calculate lower than the benchmark eigenvalues within 4σ. The primary difference between the two nuclear data libraries is the adjustment of the absorption cross section of graphite. Simulations of the absorber rod worth measurements are within 3σ of the benchmark experiment values. The complete benchmark evaluation details are available in the 2014 edition of the International Handbook of Evaluated Reactor Physics Benchmark Experiments.« less

  20. What Are the ACT College Readiness Benchmarks? Information Brief

    ERIC Educational Resources Information Center

    ACT, Inc., 2013

    2013-01-01

    The ACT College Readiness Benchmarks are the minimum ACT college readiness assessment scores required for students to have a high probability of success in credit-bearing college courses--English Composition, social sciences courses, College Algebra, or Biology. This report identifies the College Readiness Benchmarks on the ACT Compass scale