Kannry, Joseph; Mukani, Sonia; Myers, Kristin
2006-01-01
The experience of Mount Sinai Hospital is representative of the challenges and problems facing large academic medical centers in selecting an ambulatory EMR. The facility successfully revived a stalled process in a challenging financial climate, using a framework of science and rigorous investigation. The process incorporated several innovations: 1) There was a thorough review of medical informatics literature to develop a mission statement, determine practical objectives and guide the demonstration process; 2) The process involved rigorous investigation of vendor statements, industry statements and other institution's views of vendors; 3) The initiative focused on user-centric selection, and the survey instrument was scientifically and specifically designed to assess user feedback; 4) There was scientific analysis of validated findings and survey results at all steering meetings; 5) The process included an assessment of vendors' ability to support research by identifying funded and published research; 6) Selection involved meticulous total cost of ownership analysis to assess and compare real costs of implementing a vendor solution; and finally, 7) There were iterative meetings with stakeholders, executives and users to understand needs, address concerns and communicate the vision.
Hughes, Brianna H; Greenberg, Neil J; Yang, Tom C; Skonberg, Denise I
2015-01-01
High-pressure processing (HPP) is used to increase meat safety and shelf-life, with conflicting quality effects depending on rigor status during HPP. In the seafood industry, HPP is used to shuck and pasteurize oysters, but its use on abalones has only been minimally evaluated and the effect of rigor status during HPP on abalone quality has not been reported. Farm-raised abalones (Haliotis rufescens) were divided into 12 HPP treatments and 1 unprocessed control treatment. Treatments were processed pre-rigor or post-rigor at 2 pressures (100 and 300 MPa) and 3 processing times (1, 3, and 5 min). The control was analyzed post-rigor. Uniform plugs were cut from adductor and foot meat for texture profile analysis, shear force, and color analysis. Subsamples were used for scanning electron microscopy of muscle ultrastructure. Texture profile analysis revealed that post-rigor processed abalone was significantly (P < 0.05) less firm and chewy than pre-rigor processed irrespective of muscle type, processing time, or pressure. L values increased with pressure to 68.9 at 300 MPa for pre-rigor processed foot, 73.8 for post-rigor processed foot, 90.9 for pre-rigor processed adductor, and 89.0 for post-rigor processed adductor. Scanning electron microscopy images showed fraying of collagen fibers in processed adductor, but did not show pressure-induced compaction of the foot myofibrils. Post-rigor processed abalone meat was more tender than pre-rigor processed meat, and post-rigor processed foot meat was lighter in color than pre-rigor processed foot meat, suggesting that waiting for rigor to resolve prior to processing abalones may improve consumer perceptions of quality and market value. © 2014 Institute of Food Technologists®
ERIC Educational Resources Information Center
Regional Resource Center Program, 2014
2014-01-01
One component of the recently required State Systemic Improvement Plan (SSIP) for State Departments of Education calls for the selection and implementation of evidence-based practices (EBPs). This report provides six steps to guide the process of selecting evidence based practices (EBP): (1) Begin with the End in Mind--Determine Targeted Outcomes;…
ERIC Educational Resources Information Center
Jung, Jae Yup
2013-01-01
This study developed and tested a new model of the cognitive processes associated with occupational/career indecision for gifted adolescents. A survey instrument with rigorous psychometric properties, developed from a number of existing instruments, was administered to a sample of 687 adolescents attending three academically selective high schools…
Instrument Selection for Randomized Controlled Trials Why This and Not That?
Records, Kathie; Keller, Colleen; Ainsworth, Barbara; Permana, Paska
2011-01-01
A fundamental linchpin for obtaining rigorous findings in quantitative research involves the selection of survey instruments. Psychometric recommendations are available for the processes for scale development and testing and guidance for selection of established scales. These processes are necessary to address the validity link between the phenomena under investigation, the empirical measures and, ultimately, the theoretical ties between these and the world views of the participants. Detailed information is most often provided about study design and protocols, but far less frequently is a detailed theoretical explanation provided for why specific instruments are chosen. Guidance to inform choices is often difficult to find when scales are needed for specific cultural, ethnic, or racial groups. This paper details the rationale underlying instrument selection for measurement of the major processes (intervention, mediator and moderator variables, outcome variables) in an ongoing study of postpartum Latinas, Madres para la Salud [Mothers for Health]. The rationale underpinning our choices includes a discussion of alternatives, when appropriate. These exemplars may provide direction for other intervention researchers who are working with specific cultural, racial, or ethnic groups or for other investigators who are seeking to select the ‘best’ instrument. Thoughtful consideration of measurement and articulation of the rationale underlying our choices facilitates the maintenance of rigor within the study design and improves our ability to assess study outcomes. PMID:21986392
A conceptual framework for invasion in microbial communities.
Kinnunen, Marta; Dechesne, Arnaud; Proctor, Caitlin; Hammes, Frederik; Johnson, David; Quintela-Baluja, Marcos; Graham, David; Daffonchio, Daniele; Fodelianakis, Stilianos; Hahn, Nicole; Boon, Nico; Smets, Barth F
2016-12-01
There is a growing interest in controlling-promoting or avoiding-the invasion of microbial communities by new community members. Resource availability and community structure have been reported as determinants of invasion success. However, most invasion studies do not adhere to a coherent and consistent terminology nor always include rigorous interpretations of the processes behind invasion. Therefore, we suggest that a consistent set of definitions and a rigorous conceptual framework are needed. We define invasion in a microbial community as the establishment of an alien microbial type in a resident community and argue how simple criteria to define aliens, residents, and alien establishment can be applied for a wide variety of communities. In addition, we suggest an adoption of the community ecology framework advanced by Vellend (2010) to clarify potential determinants of invasion. This framework identifies four fundamental processes that control community dynamics: dispersal, selection, drift and diversification. While selection has received ample attention in microbial community invasion research, the three other processes are often overlooked. Here, we elaborate on the relevance of all four processes and conclude that invasion experiments should be designed to elucidate the role of dispersal, drift and diversification, in order to obtain a complete picture of invasion as a community process.
A conceptual framework for invasion in microbial communities
Kinnunen, Marta; Dechesne, Arnaud; Proctor, Caitlin; Hammes, Frederik; Johnson, David; Quintela-Baluja, Marcos; Graham, David; Daffonchio, Daniele; Fodelianakis, Stilianos; Hahn, Nicole; Boon, Nico; Smets, Barth F
2016-01-01
There is a growing interest in controlling—promoting or avoiding—the invasion of microbial communities by new community members. Resource availability and community structure have been reported as determinants of invasion success. However, most invasion studies do not adhere to a coherent and consistent terminology nor always include rigorous interpretations of the processes behind invasion. Therefore, we suggest that a consistent set of definitions and a rigorous conceptual framework are needed. We define invasion in a microbial community as the establishment of an alien microbial type in a resident community and argue how simple criteria to define aliens, residents, and alien establishment can be applied for a wide variety of communities. In addition, we suggest an adoption of the community ecology framework advanced by Vellend (2010) to clarify potential determinants of invasion. This framework identifies four fundamental processes that control community dynamics: dispersal, selection, drift and diversification. While selection has received ample attention in microbial community invasion research, the three other processes are often overlooked. Here, we elaborate on the relevance of all four processes and conclude that invasion experiments should be designed to elucidate the role of dispersal, drift and diversification, in order to obtain a complete picture of invasion as a community process. PMID:27137125
Selection of species and sampling areas: The importance of inference
Paul Stephen Corn
2009-01-01
Inductive inference, the process of drawing general conclusions from specific observations, is fundamental to the scientific method. Platt (1964) termed conclusions obtained through rigorous application of the scientific method as "strong inference" and noted the following basic steps: generating alternative hypotheses; devising experiments, the...
Stages in the research process.
Gelling, Leslie
2015-03-04
Research should be conducted in a systematic manner, allowing the researcher to progress from a general idea or clinical problem to scientifically rigorous research findings that enable new developments to improve clinical practice. Using a research process helps guide this process. This article is the first in a 26-part series on nursing research. It examines the process that is common to all research, and provides insights into ten different stages of this process: developing the research question, searching and evaluating the literature, selecting the research approach, selecting research methods, gaining access to the research site and data, pilot study, sampling and recruitment, data collection, data analysis, and dissemination of results and implementation of findings.
Effective Practices in High Performing Districts Serving Students in Special Education
ERIC Educational Resources Information Center
Huberman, Mette; Navo, Matt; Parrish, Tom
2012-01-01
Through a rigorous selection process based on special education performance over four years, this study identified eight unified districts in California that showed unusually strong academic performance for their special education population compared to similar districts in the state. Researchers conducted interviews with these districts' special…
Interface Pattern Selection in Directional Solidification
NASA Technical Reports Server (NTRS)
Trivedi, Rohit; Tewari, Surendra N.
2001-01-01
The central focus of this research is to establish key scientific concepts that govern the selection of cellular and dendritic patterns during the directional solidification of alloys. Ground-based studies have established that the conditions under which cellular and dendritic microstructures form are precisely where convection effects are dominant in bulk samples. Thus, experimental data can not be obtained terrestrially under pure diffusive regime. Furthermore, reliable theoretical models are not yet possible which can quantitatively incorporate fluid flow in the pattern selection criterion. Consequently, microgravity experiments on cellular and dendritic growth are designed to obtain benchmark data under diffusive growth conditions that can be quantitatively analyzed and compared with the rigorous theoretical model to establish the fundamental principles that govern the selection of specific microstructure and its length scales. In the cellular structure, different cells in an array are strongly coupled so that the cellular pattern evolution is controlled by complex interactions between thermal diffusion, solute diffusion and interface effects. These interactions give infinity of solutions, and the system selects only a narrow band of solutions. The aim of this investigation is to obtain benchmark data and develop a rigorous theoretical model that will allow us to quantitatively establish the physics of this selection process.
ERIC Educational Resources Information Center
Duhon-Haynes, Gwendolyn; And Others
This paper examines alternative certification programs in terms of entrance requirements, supervision and mentoring, and post-certification professional support. A good alternative program uses rigorous screening processes to ensure the selection of qualified teacher interns; provides high-quality preservice training in methodology, classroom…
Birkeland, S; Akse, L
2010-01-01
Improved slaughtering procedures in the salmon industry have caused a delayed onset of rigor mortis and, thus, a potential for pre-rigor secondary processing. The aim of this study was to investigate the effect of rigor status at time of processing on quality traits color, texture, sensory, microbiological, in injection salted, and cold-smoked Atlantic salmon (Salmo salar). Injection of pre-rigor fillets caused a significant (P<0.001) contraction (-7.9%± 0.9%) on the caudal-cranial axis. No significant differences in instrumental color (a*, b*, C*, or h*), texture (hardness), or sensory traits (aroma, color, taste, and texture) were observed between pre- or post-rigor processed fillets; however, post-rigor (1477 ± 38 g) fillets had a significant (P>0.05) higher fracturability than pre-rigor fillets (1369 ± 71 g). Pre-rigor fillets were significantly (P<0.01) lighter, L*, (39.7 ± 1.0) than post-rigor fillets (37.8 ± 0.8) and had significantly lower (P<0.05) aerobic plate count (APC), 1.4 ± 0.4 log CFU/g against 2.6 ± 0.6 log CFU/g, and psychrotrophic count (PC), 2.1 ± 0.2 log CFU/g against 3.0 ± 0.5 log CFU/g, than post-rigor processed fillets. This study showed that similar quality characteristics can be obtained in cold-smoked products processed either pre- or post-rigor when using suitable injection salting protocols and smoking techniques. © 2010 Institute of Food Technologists®
Dunn-Walters, Deborah K.; Belelovsky, Alex; Edelman, Hanna; Banerjee, Monica; Mehr, Ramit
2002-01-01
We have developed a rigorous graph-theoretical algorithm for quantifying the shape properties of mutational lineage trees. We show that information about the dynamics of hypermutation and antigen-driven clonal selection during the humoral immune response is contained in the shape of mutational lineage trees deduced from the responding clones. Age and tissue related differences in the selection process can be studied using this method. Thus, tree shape analysis can be used as a means of elucidating humoral immune response dynamics in various situations. PMID:15144020
Single-case synthesis tools I: Comparing tools to evaluate SCD quality and rigor.
Zimmerman, Kathleen N; Ledford, Jennifer R; Severini, Katherine E; Pustejovsky, James E; Barton, Erin E; Lloyd, Blair P
2018-03-03
Tools for evaluating the quality and rigor of single case research designs (SCD) are often used when conducting SCD syntheses. Preferred components include evaluations of design features related to the internal validity of SCD to obtain quality and/or rigor ratings. Three tools for evaluating the quality and rigor of SCD (Council for Exceptional Children, What Works Clearinghouse, and Single-Case Analysis and Design Framework) were compared to determine if conclusions regarding the effectiveness of antecedent sensory-based interventions for young children changed based on choice of quality evaluation tool. Evaluation of SCD quality differed across tools, suggesting selection of quality evaluation tools impacts evaluation findings. Suggestions for selecting an appropriate quality and rigor assessment tool are provided and across-tool conclusions are drawn regarding the quality and rigor of studies. Finally, authors provide guidance for using quality evaluations in conjunction with outcome analyses when conducting syntheses of interventions evaluated in the context of SCD. Copyright © 2018 Elsevier Ltd. All rights reserved.
ERIC Educational Resources Information Center
Marsh, Herbert W.; Hau, Kit-Tai; Artelt, Cordula; Baumert, Jurgen; Peschar, Jules L.
2006-01-01
Through a rigorous process of selecting educational psychology's most useful affective constructs, the Organisation for Economic Co-operation and Development (OECD) constructed the Students' Approaches to Learning (SAL) instrument, which requires only 10 min to measure 14 factors that assess self-regulated learning strategies, self-beliefs,…
De Brún, Aoife; Flynn, Darren; Ternent, Laura; Price, Christopher I; Rodgers, Helen; Ford, Gary A; Rudd, Matthew; Lancsar, Emily; Simpson, Stephen; Teah, John; Thomson, Richard G
2018-06-22
A discrete choice experiment (DCE) is a method used to elicit participants' preferences and the relative importance of different attributes and levels within a decision-making process. DCEs have become popular in healthcare; however, approaches to identify the attributes/levels influencing a decision of interest and to selection methods for their inclusion in a DCE are under-reported. Our objectives were: to explore the development process used to select/present attributes/levels from the identified range that may be influential; to describe a systematic and rigorous development process for design of a DCE in the context of thrombolytic therapy for acute stroke; and, to discuss the advantages of our five-stage approach to enhance current guidance for developing DCEs. A five-stage DCE development process was undertaken. Methods employed included literature review, qualitative analysis of interview and ethnographic data, expert panel discussions, a quantitative structured prioritisation (ranking) exercise and pilot testing of the DCE using a 'think aloud' approach. The five-stage process reported helped to reduce the list of 22 initial patient-related factors to a final set of nine variable factors and six fixed factors for inclusion in a testable DCE using a vignette model of presentation. In order for the data and conclusions generated by DCEs to be deemed valid, it is crucial that the methods of design and development are documented and reported. This paper has detailed a rigorous and systematic approach to DCE development which may be useful to researchers seeking to establish methods for reducing and prioritising attributes for inclusion in future DCEs.
Preserving pre-rigor meat functionality for beef patty production.
Claus, J R; Sørheim, O
2006-06-01
Three methods were examined for preserving pre-rigor meat functionality in beef patties. Hot-boned semimembranosus muscles were processed as follows: (1) pre-rigor ground, salted, patties immediately cooked; (2) pre-rigor ground, salted and stored overnight; (3) pre-rigor injected with brine; and (4) post-rigor ground and salted. Raw patties contained 60% lean beef, 19.7% beef fat trim, 1.7% NaCl, 3.6% starch, and 15% water. Pre-rigor processing occurred at 3-3.5h postmortem. Patties made from pre-rigor ground meat had higher pH values; greater protein solubility; firmer, more cohesive, and chewier texture; and substantially lower cooking losses than the other treatments. Addition of salt was sufficient to reduce the rate and extent of glycolysis. Brine injection of intact pre-rigor muscles resulted in some preservation of the functional properties but not as pronounced as with salt addition to pre-rigor ground meat.
ERIC Educational Resources Information Center
Cosner, Shelby; Tozer, Steve; Smylie, Mark
2012-01-01
This article describes the process of replacing a modest Master's level school leader preparation program with an innovative Ed.D. program at the University of Illinois at Chicago (UIC). The new doctoral program is intensive, highly selective, intellectually rigorous, and field-based. The authors provide side-by-side comparisons of the difference…
ERIC Educational Resources Information Center
Robbins, William Shane
2011-01-01
"We know that education can be an arduous process. Countries use different approaches based on societal acceptances, but effective education always requires enormous efforts. Whether success is achieved, depends on the development of a rigorous and progressive curriculum, while at the same time providing all students the opportunity to…
ERIC Educational Resources Information Center
Diouf, Boucar; Rioux, Pierre
1999-01-01
Presents the rigor mortis process in brook charr (Salvelinus fontinalis) as a tool for better understanding skeletal muscle metabolism. Describes an activity that demonstrates how rigor mortis is related to the post-mortem decrease of muscular glycogen and ATP, how glycogen degradation produces lactic acid that lowers muscle pH, and how…
Rohwer, Anke; Schoonees, Anel; Young, Taryn
2014-11-02
This paper describes the process, our experience and the lessons learnt in doing document reviews of health science curricula. Since we could not find relevant literature to guide us on how to approach these reviews, we feel that sharing our experience would benefit researchers embarking on similar projects. We followed a rigorous, transparent, pre-specified approach that included the preparation of a protocol, a pre-piloted data extraction form and coding schedule. Data were extracted, analysed and synthesised. Quality checks were included at all stages of the process. The main lessons we learnt related to time and project management, continuous quality assurance, selecting the software that meets the needs of the project, involving experts as needed and disseminating the findings to relevant stakeholders. A complete curriculum evaluation comprises, apart from a document review, interviews with students and lecturers to assess the learnt and taught curricula respectively. Rigorous methods must be used to ensure an objective assessment.
Forecasting volatility with neural regression: a contribution to model adequacy.
Refenes, A N; Holt, W T
2001-01-01
Neural nets' usefulness for forecasting is limited by problems of overfitting and the lack of rigorous procedures for model identification, selection and adequacy testing. This paper describes a methodology for neural model misspecification testing. We introduce a generalization of the Durbin-Watson statistic for neural regression and discuss the general issues of misspecification testing using residual analysis. We derive a generalized influence matrix for neural estimators which enables us to evaluate the distribution of the statistic. We deploy Monte Carlo simulation to compare the power of the test for neural and linear regressors. While residual testing is not a sufficient condition for model adequacy, it is nevertheless a necessary condition to demonstrate that the model is a good approximation to the data generating process, particularly as neural-network estimation procedures are susceptible to partial convergence. The work is also an important step toward developing rigorous procedures for neural model identification, selection and adequacy testing which have started to appear in the literature. We demonstrate its applicability in the nontrivial problem of forecasting implied volatility innovations using high-frequency stock index options. Each step of the model building process is validated using statistical tests to verify variable significance and model adequacy with the results confirming the presence of nonlinear relationships in implied volatility innovations.
Kim, Minseok; Eleftheriades, George V
2016-10-15
We propose a highly efficient (nearly lossless and impedance-matched) all-dielectric optical tensor impedance metasurface that mimics chiral effects at optical wavelengths. By cascading an array of rotated crossed silicon nanoblocks, we realize chiral optical tensor impedance metasurfaces that operate as circular polarization selective surfaces. Their efficiencies are maximized through a nonlinear numerical optimization process in which the tensor impedance metasurfaces are modeled via multi-conductor transmission line theory. From rigorous full-wave simulations that include all material losses, we show field transmission efficiencies of 94% for right- and left-handed circular polarization selective surfaces at 800 nm.
Forty Years of Psychological and Psychiatric Selection of NASA Astronauts
NASA Technical Reports Server (NTRS)
Holland, Albert W.
2000-01-01
The purpose of this presentation is to chronicle the history and development of the psychological selection process for NASA astronauts. For over 40 years, astronaut applicants have undergone rigorous medical testing to qualify for candidacy. Psychological selection has an equally long history, dating back to 1958, when psychological requirements were established for astronauts during the Mercury program. However, for many years, psychological selection consisted of psychiatric screening for psychopathology. As we approach the day in which the first ISS crew will live and work in space for months at a time, it becomes clear that both the psychological criteria and the selection system to detect said criteria have changed. This presentation discusses the events that led to the current, dual-phase selection system that is used to select individuals into the astronaut corps. Future directions for psychological selection will also be addressed.
ISC-EHB: Reconstruction of a robust earthquake dataset
NASA Astrophysics Data System (ADS)
Weston, J.; Engdahl, E. R.; Harris, J.; Di Giacomo, D.; Storchak, D. A.
2018-04-01
The EHB Bulletin of hypocentres and associated travel-time residuals was originally developed with procedures described by Engdahl, Van der Hilst and Buland (1998) and currently ends in 2008. It is a widely used seismological dataset, which is now expanded and reconstructed, partly by exploiting updated procedures at the International Seismological Centre (ISC), to produce the ISC-EHB. The reconstruction begins in the modern period (2000-2013) to which new and more rigorous procedures for event selection, data preparation, processing, and relocation are applied. The selection criteria minimise the location bias produced by unmodelled 3D Earth structure, resulting in events that are relatively well located in any given region. Depths of the selected events are significantly improved by a more comprehensive review of near station and secondary phase travel-time residuals based on ISC data, especially for the depth phases pP, pwP and sP, as well as by a rigorous review of the event depths in subduction zone cross sections. The resulting cross sections and associated maps are shown to provide details of seismicity in subduction zones in much greater detail than previously achievable. The new ISC-EHB dataset will be especially useful for global seismicity studies and high-frequency regional and global tomographic inversions.
Process qualification and testing of LENS deposited AY1E0125 D-bottle brackets.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Atwood, Clinton J.; Smugeresky, John E.; Jew, Michael
2006-11-01
The LENS Qualification team had the goal of performing a process qualification for the Laser Engineered Net Shaping{trademark}(LENS{reg_sign}) process. Process Qualification requires that a part be selected for process demonstration. The AY1E0125 D-Bottle Bracket from the W80-3 was selected for this work. The repeatability of the LENS process was baselined to determine process parameters. Six D-Bottle brackets were deposited using LENS, machined to final dimensions, and tested in comparison to conventionally processed brackets. The tests, taken from ES1E0003, included a mass analysis and structural dynamic testing including free-free and assembly-level modal tests, and Haversine shock tests. The LENS brackets performedmore » with very similar characteristics to the conventionally processed brackets. Based on the results of the testing, it was concluded that the performance of the brackets made them eligible for parallel path testing in subsystem level tests. The testing results and process rigor qualified the LENS process as detailed in EER200638525A.« less
Decision support systems in water and wastewater treatment process selection and design: a review.
Hamouda, M A; Anderson, W B; Huck, P M
2009-01-01
The continuously changing drivers of the water treatment industry, embodied by rigorous environmental and health regulations and the challenge of emerging contaminants, necessitates the development of decision support systems for the selection of appropriate treatment trains. This paper explores a systematic approach to developing decision support systems, which includes the analysis of the treatment problem(s), knowledge acquisition and representation, and the identification and evaluation of criteria controlling the selection of optimal treatment systems. The objective of this article is to review approaches and methods used in decision support systems developed to aid in the selection, sequencing of unit processes and design of drinking water, domestic wastewater, and industrial wastewater treatment systems. Not surprisingly, technical considerations were found to dominate the logic of the developed systems. Most of the existing decision-support tools employ heuristic knowledge. It has been determined that there is a need to develop integrated decision support systems that are generic, usable and consider a system analysis approach.
Application of Crossover Design for Conducting Rigorous Extension Evaluations
ERIC Educational Resources Information Center
Jayaratne, K. S. U.; Bird, Carolyn L.; McClelland, Jacquelyn W.
2013-01-01
With the increasing demand for accountability of Extension programming, Extension professionals need to apply rigorous evaluation designs. Randomized designs are useful to eliminate selection biases of program participants and to improve the accuracy of evaluation. However, randomized control designs are not practical to apply in Extension program…
NASA Astrophysics Data System (ADS)
Carrico, James D.; Tyler, Tom; Leang, Kam K.
2017-10-01
Smart polymeric and gel actuators change shape or size in response to stimuli like electricity, heat, or light. These smart polymeric- and gel-based actuators are compliant and well suited for development of soft mechatronic and robotic devices. This paper provides a thorough review of select smart polymeric and gel actuator materials where an automated and freeform fabrication process, like 3D printing, is exploited to create custom shaped monolithic devices. In particular, the advantages and limitations, examples of applications, manufacturing and fabrication techniques, and methods for actuator control are discussed. Finally, a rigorous comparison and analysis of some of the advantages and limitations, as well as manufacturing processes, for these materials, are presented.
The genealogy of samples in models with selection.
Neuhauser, C; Krone, S M
1997-02-01
We introduce the genealogy of a random sample of genes taken from a large haploid population that evolves according to random reproduction with selection and mutation. Without selection, the genealogy is described by Kingman's well-known coalescent process. In the selective case, the genealogy of the sample is embedded in a graph with a coalescing and branching structure. We describe this graph, called the ancestral selection graph, and point out differences and similarities with Kingman's coalescent. We present simulations for a two-allele model with symmetric mutation in which one of the alleles has a selective advantage over the other. We find that when the allele frequencies in the population are already in equilibrium, then the genealogy does not differ much from the neutral case. This is supported by rigorous results. Furthermore, we describe the ancestral selection graph for other selective models with finitely many selection classes, such as the K-allele models, infinitely-many-alleles models. DNA sequence models, and infinitely-many-sites models, and briefly discuss the diploid case.
The Genealogy of Samples in Models with Selection
Neuhauser, C.; Krone, S. M.
1997-01-01
We introduce the genealogy of a random sample of genes taken from a large haploid population that evolves according to random reproduction with selection and mutation. Without selection, the genealogy is described by Kingman's well-known coalescent process. In the selective case, the genealogy of the sample is embedded in a graph with a coalescing and branching structure. We describe this graph, called the ancestral selection graph, and point out differences and similarities with Kingman's coalescent. We present simulations for a two-allele model with symmetric mutation in which one of the alleles has a selective advantage over the other. We find that when the allele frequencies in the population are already in equilibrium, then the genealogy does not differ much from the neutral case. This is supported by rigorous results. Furthermore, we describe the ancestral selection graph for other selective models with finitely many selection classes, such as the K-allele models, infinitely-many-alleles models, DNA sequence models, and infinitely-many-sites models, and briefly discuss the diploid case. PMID:9071604
Monitoring muscle optical scattering properties during rigor mortis
NASA Astrophysics Data System (ADS)
Xia, J.; Ranasinghesagara, J.; Ku, C. W.; Yao, G.
2007-09-01
Sarcomere is the fundamental functional unit in skeletal muscle for force generation. In addition, sarcomere structure is also an important factor that affects the eating quality of muscle food, the meat. The sarcomere structure is altered significantly during rigor mortis, which is the critical stage involved in transforming muscle to meat. In this paper, we investigated optical scattering changes during the rigor process in Sternomandibularis muscles. The measured optical scattering parameters were analyzed along with the simultaneously measured passive tension, pH value, and histology analysis. We found that the temporal changes of optical scattering, passive tension, pH value and fiber microstructures were closely correlated during the rigor process. These results suggested that sarcomere structure changes during rigor mortis can be monitored and characterized by optical scattering, which may find practical applications in predicting meat quality.
The future of fMRI in cognitive neuroscience.
Poldrack, Russell A
2012-08-15
Over the last 20 years, fMRI has revolutionized cognitive neuroscience. Here I outline a vision for what the next 20 years of fMRI in cognitive neuroscience might look like. Some developments that I hope for include increased methodological rigor, an increasing focus on connectivity and pattern analysis as opposed to "blobology", a greater focus on selective inference powered by open databases, and increased use of ontologies and computational models to describe underlying processes. Copyright © 2011 Elsevier Inc. All rights reserved.
ISDRS 2016 special issue foreword
NASA Astrophysics Data System (ADS)
Iliadis, Agis A.; Akturk, Akin; Tompkins, Randy P.
2017-10-01
This 8th ISDRS Special Issue of Solid-State Electronics contains manuscripts presented at the 2016 International Semiconductor Device Research Symposium (ISDRS 2016) that was held at the Hyatt Regency Hotel, in Bethesda, Maryland, on December 7-9, 2016. A total of 32 manuscripts were submitted, which were sent out for review by both the ISDRS Guest Editors, Agis Iliadis, Akin Akturk, Randy Tompkins, and the Solid-State Electronics Editor, Alex Zaslavsky. This ISDRS Special Issue of Solid-State Electronics contains the manuscripts selected through this rigorous review process.
The updating of clinical practice guidelines: insights from an international survey
2011-01-01
Background Clinical practice guidelines (CPGs) have become increasingly popular, and the methodology to develop guidelines has evolved enormously. However, little attention has been given to the updating process, in contrast to the appraisal of the available literature. We conducted an international survey to identify current practices in CPG updating and explored the need to standardize and improve the methods. Methods We developed a questionnaire (28 items) based on a review of the existing literature about guideline updating and expert comments. We carried out the survey between March and July 2009, and it was sent by email to 106 institutions: 69 members of the Guidelines International Network who declared that they developed CPGs; 30 institutions included in the U.S. National Guideline Clearinghouse database that published more than 20 CPGs; and 7 institutions selected by an expert committee. Results Forty-four institutions answered the questionnaire (42% response rate). In the final analysis, 39 completed questionnaires were included. Thirty-six institutions (92%) reported that they update their guidelines. Thirty-one institutions (86%) have a formal procedure for updating their guidelines, and 19 (53%) have a formal procedure for deciding when a guideline becomes out of date. Institutions describe the process as moderately rigorous (36%) or acknowledge that it could certainly be more rigorous (36%). Twenty-two institutions (61%) alert guideline users on their website when a guideline is older than three to five years or when there is a risk of being outdated. Twenty-five institutions (64%) support the concept of "living guidelines," which are continuously monitored and updated. Eighteen institutions (46%) have plans to design a protocol to improve their guideline-updating process, and 21 (54%) are willing to share resources with other organizations. Conclusions Our study is the first to describe the process of updating CPGs among prominent guideline institutions across the world, providing a comprehensive picture of guideline updating. There is an urgent need to develop rigorous international standards for this process and to minimize duplication of effort internationally. PMID:21914177
Another View: In Defense of Vigor over Rigor in Classroom Demonstrations
ERIC Educational Resources Information Center
Dunn, Dana S.
2008-01-01
Scholarship of teaching and learning (SoTL) demands greater empirical rigor on the part of authors and the editorial process than ever before. Although admirable and important, I worry that this increasing rigor will limit opportunities and outlets for a form of pedagogical vigor--the publication of simple, experiential, but empirically…
Computer modeling of lung cancer diagnosis-to-treatment process
Ju, Feng; Lee, Hyo Kyung; Osarogiagbon, Raymond U.; Yu, Xinhua; Faris, Nick
2015-01-01
We introduce an example of a rigorous, quantitative method for quality improvement in lung cancer care-delivery. Computer process modeling methods are introduced for lung cancer diagnosis, staging and treatment selection process. Two types of process modeling techniques, discrete event simulation (DES) and analytical models, are briefly reviewed. Recent developments in DES are outlined and the necessary data and procedures to develop a DES model for lung cancer diagnosis, leading up to surgical treatment process are summarized. The analytical models include both Markov chain model and closed formulas. The Markov chain models with its application in healthcare are introduced and the approach to derive a lung cancer diagnosis process model is presented. Similarly, the procedure to derive closed formulas evaluating the diagnosis process performance is outlined. Finally, the pros and cons of these methods are discussed. PMID:26380181
Quantifying male attractiveness.
McNamara, John M; Houston, Alasdair I; Marques Dos Santos, Miguel; Kokko, Hanna; Brooks, Rob
2003-01-01
Genetic models of sexual selection are concerned with a dynamic process in which female preference and male trait values coevolve. We present a rigorous method for characterizing evolutionary endpoints of this process in phenotypic terms. In our phenotypic characterization the mate-choice strategy of female population members determines how attractive females should find each male, and a population is evolutionarily stable if population members are actually behaving in this way. This provides a justification of phenotypic explanations of sexual selection and the insights into sexual selection that they provide. Furthermore, the phenotypic approach also has enormous advantages over a genetic approach when computing evolutionarily stable mate-choice strategies, especially when strategies are allowed to be complex time-dependent preference rules. For simplicity and clarity our analysis deals with haploid mate-choice genetics and a male trait that is inherited phenotypically, for example by vertical cultural transmission. The method is, however, easily extendible to other cases. An example illustrates that the sexy son phenomenon can occur when there is phenotypic inheritance of the male trait. PMID:14561306
Genetic and environmental effects on the muscle structure response post-mortem.
Thompson, J M; Perry, D; Daly, B; Gardner, G E; Johnston, D J; Pethick, D W
2006-09-01
This paper reviewed the mechanisms by which glycolytic rate and pre-rigor stretching of muscle impact on meat quality. If muscle is free to shorten during the rigor process extremes in glycolytic rate can impact negatively on meat quality by inducing either cold or rigor shortening. Factors that contribute to variation in glycolytic rate include the glycogen concentration at slaughter and fibre type of the muscle. Glycolysis is highly sensitive to temperature, which is an important factor in heavy grain fed carcasses. An alternative solution to controlling glycolysis is to stretch the muscle pre-rigor so that it cannot shorten, thus providing an insurance against extremes in processing conditions. Results are presented which show a large reduction in variance (both additive and phenotypic) in tenderness caused by pre-rigor stretching. Whilst this did not impact on the heritability of shear force, it did reduce genotype differences. The implications of these results on the magnitude of genotype effects on tenderness is discussed.
Liu, Jianfeng; Laird, Carl Damon
2017-09-22
Optimal design of a gas detection systems is challenging because of the numerous sources of uncertainty, including weather and environmental conditions, leak location and characteristics, and process conditions. Rigorous CFD simulations of dispersion scenarios combined with stochastic programming techniques have been successfully applied to the problem of optimal gas detector placement; however, rigorous treatment of sensor failure and nonuniform unavailability has received less attention. To improve reliability of the design, this paper proposes a problem formulation that explicitly considers nonuniform unavailabilities and all backup detection levels. The resulting sensor placement problem is a large-scale mixed-integer nonlinear programming (MINLP) problem thatmore » requires a tailored solution approach for efficient solution. We have developed a multitree method which depends on iteratively solving a sequence of upper-bounding master problems and lower-bounding subproblems. The tailored global solution strategy is tested on a real data problem and the encouraging numerical results indicate that our solution framework is promising in solving sensor placement problems. This study was selected for the special issue in JLPPI from the 2016 International Symposium of the MKO Process Safety Center.« less
DOE Office of Scientific and Technical Information (OSTI.GOV)
Liu, Jianfeng; Laird, Carl Damon
Optimal design of a gas detection systems is challenging because of the numerous sources of uncertainty, including weather and environmental conditions, leak location and characteristics, and process conditions. Rigorous CFD simulations of dispersion scenarios combined with stochastic programming techniques have been successfully applied to the problem of optimal gas detector placement; however, rigorous treatment of sensor failure and nonuniform unavailability has received less attention. To improve reliability of the design, this paper proposes a problem formulation that explicitly considers nonuniform unavailabilities and all backup detection levels. The resulting sensor placement problem is a large-scale mixed-integer nonlinear programming (MINLP) problem thatmore » requires a tailored solution approach for efficient solution. We have developed a multitree method which depends on iteratively solving a sequence of upper-bounding master problems and lower-bounding subproblems. The tailored global solution strategy is tested on a real data problem and the encouraging numerical results indicate that our solution framework is promising in solving sensor placement problems. This study was selected for the special issue in JLPPI from the 2016 International Symposium of the MKO Process Safety Center.« less
Editorial: Special Issue on Experimental Vibration Analysis
NASA Astrophysics Data System (ADS)
Serra, Roger
2018-04-01
The vibratory analyses are particularly present today in the various fields of industry, from aeronautics to manufacturing, from machining and maintenance to civil engineering, to mention a few areas, which have made this special issue a true need. The International Journal of Mechanics & Industry compiles a Special Issue on Experimental Vibration Analysis. More than thirty manuscripts were received by the international scientific committee on the 6th congress AVE2016 and only eight papers have been selected after completing a careful and rigorous peer-review process for the Special Issue, which are briefly summarized below.
EFL/ESL Textbook Selection in Korea and East Asia - Relevant Issues and Literature Review
NASA Astrophysics Data System (ADS)
Meurant, Robert C.
EFL/ESL departments periodically face the problem of textbook selection. Cogent issues are that non-native speakers will use L2 English mainly to communicate with other non-native English speakers, so an American accent is becoming less important. L2 English will mainly be used in computer-mediated communication, hence the importance of L2 Digital Literacy. The convergence of Information Communication Technologies is radically impacting Second Language Acquisition, which is integrating web-hosted Assessment and Learning Management Systems. EFL/ESL textbooks need to be compatible with blended learning, prepare students for a globalized world, and foster autonomous learning. I summarize five papers on EFL/ESL textbook evaluation and selection, and include relevant material for adaptation. Textbooks are major sources of contact with the target language, so selection is an important decision. Educators need to be systematic and objective in their approach, adopting a selection process that is open, transparent, accountable, participatory, informed and rigorous.
Harris, Sarah Parker; Gould, Robert; Fujiura, Glenn
2015-01-01
There is increasing theoretical consideration about the use of systematic and scoping reviews of evidence in informing disability and rehabilitation research and practice. Indicative of this trend, this journal published a piece by Rumrill, Fitzgerald and Merchant in 2010 explaining the utility and process for conducting reviews of intervention-based research. There is still need to consider how to apply such rigor when conducting more exploratory reviews of heterogeneous research. This article explores the challenges, benefits, and procedures for conducting rigorous exploratory scoping reviews of diverse evidence. The article expands upon Rumrill, Fitzgerald and Merchant's framework and considers its application to more heterogeneous evidence on the impact of social policy. A worked example of a scoping review of the Americans with Disabilities Act is provided with a procedural framework for conducting scoping reviews on the effects of a social policy. The need for more nuanced techniques for enhancing rigor became apparent during the review process. There are multiple methodological steps that can enhance the utility of exploratory scoping reviews. The potential of systematic consideration during the exploratory review process is shown as a viable method to enhance the rigor in reviewing diverse bodies of evidence.
Ruiz-Meana, M; Garcia-Dorado, D; Juliá, M; Inserte, J; Siegmund, B; Ladilov, Y; Piper, M; Tritto, F P; González, M A; Soler-Soler, J
2000-01-01
The objective of this study was to investigate the effect of Na+-H+ exchange (NHE) and HCO3--Na+ symport inhibition on the development of rigor contracture. Freshly isolated adult rat cardiomyocytes were subjected to 60 min metabolic inhibition (MI) and 5 min re-energization (Rx). The effects of perfusion of HCO3- or HCO3--free buffer with or without the NHE inhibitor HOE642 (7 microM) were investigated during MI and Rx. In HCO3--free conditions, HOE642 reduced the percentage of cells developing rigor during MI from 79 +/- 1% to 40 +/- 4% (P < 0.001) without modifying the time at which rigor appeared. This resulted in a 30% reduction of hypercontracture during Rx (P < 0.01). The presence of HCO3- abolished the protective effect of HOE642 against rigor. Cells that had developed rigor underwent hypercontracture during Rx independently of treatment allocation. Ratiofluorescence measurement demonstrated that the rise in cytosolic Ca2+ (fura-2) occurred only after the onset of rigor, and was not influenced by HOE642. NHE inhibition did not modify Na+ rise (SBFI) during MI, but exaggerated the initial fall of intracellular pH (BCEFC). In conclusion, HOE642 has a protective effect against rigor during energy deprivation, but only when HCO3--dependent transporters are inhibited. This effect is independent of changes in cytosolic Na+ or Ca2+ concentrations.
Cypress, Brigitte S
Issues are still raised even now in the 21st century by the persistent concern with achieving rigor in qualitative research. There is also a continuing debate about the analogous terms reliability and validity in naturalistic inquiries as opposed to quantitative investigations. This article presents the concept of rigor in qualitative research using a phenomenological study as an exemplar to further illustrate the process. Elaborating on epistemological and theoretical conceptualizations by Lincoln and Guba, strategies congruent with qualitative perspective for ensuring validity to establish the credibility of the study are described. A synthesis of the historical development of validity criteria evident in the literature during the years is explored. Recommendations are made for use of the term rigor instead of trustworthiness and the reconceptualization and renewed use of the concept of reliability and validity in qualitative research, that strategies for ensuring rigor must be built into the qualitative research process rather than evaluated only after the inquiry, and that qualitative researchers and students alike must be proactive and take responsibility in ensuring the rigor of a research study. The insights garnered here will move novice researchers and doctoral students to a better conceptual grasp of the complexity of reliability and validity and its ramifications for qualitative inquiry.
Cumplido-Hernández, Gustavo; Campos-Arciniega, María Faustina; Chávez-López, Arturo
2007-01-01
Medical specialty training courses have peculiar characteristics that probably influence the learning process of the residents. These training courses take place in large hospitals; the residents are subjected to a rigorous selection process, and at the same time they are affiliated employees of the institution. They work long shifts and are immersed in complex academic and occupational relationships. This study aims to ascertain the significance that these future specialists give to the environment where the training course takes place in relation with their learning process. We used the social anthropology narrative analysis method. A theoretical social perspective was used to emphasize on the context to explain the reality in which the residents live. Discipline, workload, conflictive relationships and strength of family ties were the most significant elements.
Navy electroplating pollution control technology assessment manual
NASA Astrophysics Data System (ADS)
Cushnie, G. C., Jr.
1984-02-01
The report provides information on more than 27 separate technologies encompassing conventional treatment, alternate treatment, material recovery techniques and processes and new plating bath formulations. In addition, the incorporation of a section on in-plant process changes enhances the usefulness of the product in that it highlights noncapital-intensive changes to current practices and/or processes that may have significant bearing on reducing overall chemical and water usage costs as well as consequent wastewater treatment needs and disposal costs. This document was prepared as a joint Air Force-Navy effort. It is intended to serve as a guide for technical personnel making decisions on an appropriate means of meeting effluent limits. The selection of any of the described technologies should be done only after a rigorous identification of site requirements has been performed.
Putrefactive rigor: apparent rigor mortis due to gas distension.
Gill, James R; Landi, Kristen
2011-09-01
Artifacts due to decomposition may cause confusion for the initial death investigator, leading to an incorrect suspicion of foul play. Putrefaction is a microorganism-driven process that results in foul odor, skin discoloration, purge, and bloating. Various decompositional gases including methane, hydrogen sulfide, carbon dioxide, and hydrogen will cause the body to bloat. We describe 3 instances of putrefactive gas distension (bloating) that produced the appearance of inappropriate rigor, so-called putrefactive rigor. These gases may distend the body to an extent that the extremities extend and lose contact with their underlying support surface. The medicolegal investigator must recognize that this is not true rigor mortis and the body was not necessarily moved after death for this gravity-defying position to occur.
Collisional damping rates for plasma waves
NASA Astrophysics Data System (ADS)
Tigik, S. F.; Ziebell, L. F.; Yoon, P. H.
2016-06-01
The distinction between the plasma dynamics dominated by collisional transport versus collective processes has never been rigorously addressed until recently. A recent paper [P. H. Yoon et al., Phys. Rev. E 93, 033203 (2016)] formulates for the first time, a unified kinetic theory in which collective processes and collisional dynamics are systematically incorporated from first principles. One of the outcomes of such a formalism is the rigorous derivation of collisional damping rates for Langmuir and ion-acoustic waves, which can be contrasted to the heuristic customary approach. However, the results are given only in formal mathematical expressions. The present brief communication numerically evaluates the rigorous collisional damping rates by considering the case of plasma particles with Maxwellian velocity distribution function so as to assess the consequence of the rigorous formalism in a quantitative manner. Comparison with the heuristic ("Spitzer") formula shows that the accurate damping rates are much lower in magnitude than the conventional expression, which implies that the traditional approach over-estimates the importance of attenuation of plasma waves by collisional relaxation process. Such a finding may have a wide applicability ranging from laboratory to space and astrophysical plasmas.
Striation Patterns of Ox Muscle in Rigor Mortis
Locker, Ronald H.
1959-01-01
Ox muscle in rigor mortis offers a selection of myofibrils fixed at varying degrees of contraction from sarcomere lengths of 3.7 to 0.7 µ. A study of this material by phase contrast and electron microscopy has revealed four distinct successive patterns of contraction, including besides the familiar relaxed and contracture patterns, two intermediate types (2.4 to 1.9 µ, 1.8 to 1.5 µ) not previously well described. PMID:14417790
Selecting participants when testing new drugs: the implications of age and gender discrimination.
Ferguson, Pamela R
2002-01-01
Pharmaceutical products are rigorously tested for safety and efficacy prior to being licensed for use. During this testing process the archetypal research subject is a young male; women and older people are less frequently invited to participate. This is especially true at the early stages, but can also occur in the later phases of drug testing. This paper considers the reasons for the relative under-representation of these groups, and the legal implications of failing to include as research subjects the very types of people who will ultimately consume these drugs.
The case for character displacement in plants
Beans, Carolyn M
2014-01-01
The evidence for character displacement as a widespread response to competition is now building. This progress is largely the result of the establishment of rigorous criteria for demonstrating character displacement in the animal literature. There are, however, relatively few well-supported examples of character displacement in plants. This review explores the potential for character displacement in plants by addressing the following questions: (1) Why aren't examples of character displacement in plants more common? (2) What are the requirements for character displacement to occur and how do plant populations meet those requirements? (3) What are the criteria for testing the pattern and process of character displacement and what methods can and have been used to address these criteria in the plant literature? (4) What are some additional approaches for studying character displacement in plants? While more research is needed, the few plant systems in which character displacement hypotheses have been rigorously tested suggest that character displacement may play a role in shaping plant communities. Plants are especially amenable to character displacement studies because of the experimental ease with which they can be used in common gardens, selection analyses, and breeding designs. A deeper investigation of character displacement in plants is critical for a more complete understanding of the ecological and evolutionary processes that permit the coexistence of plant species. PMID:24683467
Sukumaran, Anuraj T; Holtcamp, Alexander J; Campbell, Yan L; Burnett, Derris; Schilling, Mark W; Dinh, Thu T N
2018-06-07
The objective of this study was to determine the effects of deboning time (pre- and post-rigor), processing steps (grinding - GB; salting - SB; batter formulation - BB), and storage time on the quality of raw beef mixtures and vacuum-packaged cooked sausage, produced using a commercial formulation with 0.25% phosphate. The pH was greater in pre-rigor GB and SB than in post-rigor GB and SB (P < .001). However, deboning time had no effect on metmyoglobin reducing activity, cooking loss, and color of raw beef mixtures. Protein solubility of pre-rigor beef mixtures (124.26 mg/kg) was greater than that of post-rigor beef (113.93 mg/kg; P = .071). TBARS were increased in BB but decreased during vacuum storage of cooked sausage (P ≤ .018). Except for chewiness and saltiness being 52.9 N-mm and 0.3 points greater in post-rigor sausage (P = .040 and 0.054, respectively), texture profile analysis and trained panelists detected no difference in texture between pre- and post-rigor sausage. Published by Elsevier Ltd.
NASA Technical Reports Server (NTRS)
Rey, P. A.; Gourinard, Y.; Cambou, F. (Principal Investigator); Guyader, J. C.; Gouaux, P.; Letoan, T.; Monchant, M.; Donville, B.; Loubet, D.
1973-01-01
The author has identified the following significant results. Significant results of the ARNICA program (February - December 1973) were: (1) The quantitative processing of ERTS-1 data was developed along two lines: the study of geological structures and lineaments of Spanish Catalonia, and the phytogeographical study of the forest region of the Landes of Gascony (France). In both cases it is shown that the ERTS-1 imagery can be used in establishing zonings of equal quantitative interpretation value. (2) In keeping with the operational transfer program proposed in previous reports between exploration of the imagery and charting of the object, a precise data processing method was developed, concerning more particularly the selection of digital equidensity samples computer display and rigorous referencing.
Adaptive convex combination approach for the identification of improper quaternion processes.
Ujang, Bukhari Che; Jahanchahi, Cyrus; Took, Clive Cheong; Mandic, Danilo P
2014-01-01
Data-adaptive optimal modeling and identification of real-world vector sensor data is provided by combining the fractional tap-length (FT) approach with model order selection in the quaternion domain. To account rigorously for the generality of such processes, both second-order circular (proper) and noncircular (improper), the proposed approach in this paper combines the FT length optimization with both the strictly linear quaternion least mean square (QLMS) and widely linear QLMS (WL-QLMS). A collaborative approach based on QLMS and WL-QLMS is shown to both identify the type of processes (proper or improper) and to track their optimal parameters in real time. Analysis shows that monitoring the evolution of the convex mixing parameter within the collaborative approach allows us to track the improperness in real time. Further insight into the properties of those algorithms is provided by establishing a relationship between the steady-state error and optimal model order. The approach is supported by simulations on model order selection and identification of both strictly linear and widely linear quaternion-valued systems, such as those routinely used in renewable energy (wind) and human-centered computing (biomechanics).
ERIC Educational Resources Information Center
Nehm, Ross H.; Schonfeld, Irvin Sam
2008-01-01
Growing recognition of the central importance of fostering an in-depth understanding of natural selection has, surprisingly, failed to stimulate work on the development and rigorous evaluation of instruments that measure knowledge of it. We used three different methodological tools, the Conceptual Inventory of Natural Selection (CINS), a modified…
Newton's second law and the multiplication of distributions
NASA Astrophysics Data System (ADS)
Sarrico, C. O. R.; Paiva, A.
2018-01-01
Newton's second law is applied to study the motion of a particle subjected to a time dependent impulsive force containing a Dirac delta distribution. Within this setting, we prove that this problem can be rigorously solved neither by limit processes nor by using the theory of distributions (limited to the classical Schwartz products). However, using a distributional multiplication, not defined by a limit process, a rigorous solution emerges.
A selection criterion for patterns in reaction–diffusion systems
2014-01-01
Background Alan Turing’s work in Morphogenesis has received wide attention during the past 60 years. The central idea behind his theory is that two chemically interacting diffusible substances are able to generate stable spatial patterns, provided certain conditions are met. Ever since, extensive work on several kinds of pattern-generating reaction diffusion systems has been done. Nevertheless, prediction of specific patterns is far from being straightforward, and a great deal of interest in deciphering how to generate specific patterns under controlled conditions prevails. Results Techniques allowing one to predict what kind of spatial structure will emerge from reaction–diffusion systems remain unknown. In response to this need, we consider a generalized reaction diffusion system on a planar domain and provide an analytic criterion to determine whether spots or stripes will be formed. Our criterion is motivated by the existence of an associated energy function that allows bringing in the intuition provided by phase transitions phenomena. Conclusions Our criterion is proved rigorously in some situations, generalizing well-known results for the scalar equation where the pattern selection process can be understood in terms of a potential. In more complex settings it is investigated numerically. Our work constitutes a first step towards rigorous pattern prediction in arbitrary geometries/conditions. Advances in this direction are highly applicable to the efficient design of Biotechnology and Developmental Biology experiments, as well as in simplifying the analysis of morphogenetic models. PMID:24476200
Processing capacity under perceptual and cognitive load: a closer look at load theory.
Fitousi, Daniel; Wenger, Michael J
2011-06-01
Variations in perceptual and cognitive demands (load) play a major role in determining the efficiency of selective attention. According to load theory (Lavie, Hirst, Fockert, & Viding, 2004) these factors (a) improve or hamper selectivity by altering the way resources (e.g., processing capacity) are allocated, and (b) tap resources rather than data limitations (Norman & Bobrow, 1975). Here we provide an extensive and rigorous set of tests of these assumptions. Predictions regarding changes in processing capacity are tested using the hazard function of the response time (RT) distribution (Townsend & Ashby, 1978; Wenger & Gibson, 2004). The assumption that load taps resource rather than data limitations is examined using measures of sensitivity and bias drawn from signal detection theory (Swets, 1964). All analyses were performed at two levels: the individual and the aggregate. Hypotheses regarding changes in processing capacity were confirmed at the level of the aggregate. Hypotheses regarding resource and data limitations were not completely supported at either level of analysis. And in all of the analyses, we observed substantial individual differences. In sum, the results suggest a need to expand the theoretical vocabulary of load theory, rather than a need to discard it.
Stochastic dynamics of adaptive trait and neutral marker driven by eco-evolutionary feedbacks.
Billiard, Sylvain; Ferrière, Régis; Méléard, Sylvie; Tran, Viet Chi
2015-11-01
How the neutral diversity is affected by selection and adaptation is investigated in an eco-evolutionary framework. In our model, we study a finite population in continuous time, where each individual is characterized by a trait under selection and a completely linked neutral marker. Population dynamics are driven by births and deaths, mutations at birth, and competition between individuals. Trait values influence ecological processes (demographic events, competition), and competition generates selection on trait variation, thus closing the eco-evolutionary feedback loop. The demographic effects of the trait are also expected to influence the generation and maintenance of neutral variation. We consider a large population limit with rare mutation, under the assumption that the neutral marker mutates faster than the trait under selection. We prove the convergence of the stochastic individual-based process to a new measure-valued diffusive process with jumps that we call Substitution Fleming-Viot Process (SFVP). When restricted to the trait space this process is the Trait Substitution Sequence first introduced by Metz et al. (1996). During the invasion of a favorable mutation, a genetical bottleneck occurs and the marker associated with this favorable mutant is hitchhiked. By rigorously analysing the hitchhiking effect and how the neutral diversity is restored afterwards, we obtain the condition for a time-scale separation; under this condition, we show that the marker distribution is approximated by a Fleming-Viot distribution between two trait substitutions. We discuss the implications of the SFVP for our understanding of the dynamics of neutral variation under eco-evolutionary feedbacks and illustrate the main phenomena with simulations. Our results highlight the joint importance of mutations, ecological parameters, and trait values in the restoration of neutral diversity after a selective sweep.
Hassenforder, Emeline; Ducrot, Raphaëlle; Ferrand, Nils; Barreteau, Olivier; Anne Daniell, Katherine; Pittock, Jamie
2016-09-15
Participatory approaches are now increasingly recognized and used as an essential element of policies and programs, especially in regards to natural resource management (NRM). Most practitioners, decision-makers and researchers having adopted participatory approaches also acknowledge the need to monitor and evaluate such approaches in order to audit their effectiveness, support decision-making or improve learning. Many manuals and frameworks exist on how to carry out monitoring and evaluation (M&E) for participatory processes. However, few provide guidelines on the selection and implementation of M&E methods, an aspect which is also often obscure in published studies, at the expense of the transparency, reliability and validity of the study. In this paper, we argue that the selection and implementation of M&E methods are particularly strategic when monitoring and evaluating a participatory process. We demonstrate that evaluators of participatory processes have to tackle a quadruple challenge when selecting and implementing methods: using mixed-methods, both qualitative and quantitative; assessing the participatory process, its outcomes, and its context; taking into account both the theory and participants' views; and being both rigorous and adaptive. The M&E of a participatory planning process in the Rwenzori Region, Uganda, is used as an example to show how these challenges unfold on the ground and how they can be tackled. Based on this example, we conclude by providing tools and strategies that can be used by evaluators to ensure that they make utile, feasible, coherent, transparent and adaptive methodological choices when monitoring and evaluating participatory processes for NRM. Copyright © 2016 Elsevier Ltd. All rights reserved.
A Randomized Study of How Physicians Interpret Research Funding Disclosures
Kesselheim, Aaron S.; Robertson, Christopher T.; Myers, Jessica A.; Rose, Susannah L.; Gillet, Victoria; Ross, Kathryn M.; Glynn, Robert J.; Joffe, Steven; Avorn, Jerry
2012-01-01
BACKGROUND The effects of clinical-trial funding on the interpretation of trial results are poorly understood. We examined how such support affects physicians’ reactions to trials with a high, medium, or low level of methodologic rigor. METHODS We presented 503 board-certified internists with abstracts that we designed describing clinical trials of three hypothetical drugs. The trials had high, medium, or low methodologic rigor, and each report included one of three support disclosures: funding from a pharmaceutical company, NIH funding, or none. For both factors studied (rigor and funding), one of the three possible variations was randomly selected for inclusion in the abstracts. Follow-up questions assessed the physicians’ impressions of the trials’ rigor, their confidence in the results, and their willingness to prescribe the drugs. RESULTS The 269 respondents (53.5% response rate) perceived the level of study rigor accurately. Physicians reported that they would be less willing to prescribe drugs tested in low-rigor trials than those tested in medium-rigor trials (odds ratio, 0.64; 95% confidence interval [CI], 0.46 to 0.89; P = 0.008) and would be more willing to prescribe drugs tested in high-rigor trials than those tested in medium-rigor trials (odds ratio, 3.07; 95% CI, 2.18 to 4.32; P<0.001). Disclosure of industry funding, as compared with no disclosure of funding, led physicians to downgrade the rigor of a trial (odds ratio, 0.63; 95% CI, 0.46 to 0.87; P = 0.006), their confidence in the results (odds ratio, 0.71; 95% CI, 0.51 to 0.98; P = 0.04), and their willingness to prescribe the hypothetical drugs (odds ratio, 0.68; 95% CI, 0.49 to 0.94; P = 0.02). Physicians were half as willing to prescribe drugs studied in industry-funded trials as they were to prescribe drugs studied in NIH-funded trials (odds ratio, 0.52; 95% CI, 0.37 to 0.71; P<0.001). These effects were consistent across all levels of methodologic rigor. CONCLUSIONS Physicians discriminate among trials of varying degrees of rigor, but industry sponsorship negatively influences their perception of methodologic quality and reduces their willingness to believe and act on trial findings, independently of the trial’s quality. These effects may influence the translation of clinical research into practice. PMID:22992075
Pattern-oriented modelling: a ‘multi-scope’ for predictive systems ecology
Grimm, Volker; Railsback, Steven F.
2012-01-01
Modern ecology recognizes that modelling systems across scales and at multiple levels—especially to link population and ecosystem dynamics to individual adaptive behaviour—is essential for making the science predictive. ‘Pattern-oriented modelling’ (POM) is a strategy for doing just this. POM is the multi-criteria design, selection and calibration of models of complex systems. POM starts with identifying a set of patterns observed at multiple scales and levels that characterize a system with respect to the particular problem being modelled; a model from which the patterns emerge should contain the right mechanisms to address the problem. These patterns are then used to (i) determine what scales, entities, variables and processes the model needs, (ii) test and select submodels to represent key low-level processes such as adaptive behaviour, and (iii) find useful parameter values during calibration. Patterns are already often used in these ways, but a mini-review of applications of POM confirms that making the selection and use of patterns more explicit and rigorous can facilitate the development of models with the right level of complexity to understand ecological systems and predict their response to novel conditions. PMID:22144392
Akse, L; Birkeland, S; Tobiassen, T; Joensen, S; Larsen, R
2008-10-01
Processing of fish is generally conducted postrigor, but prerigor processing is associated with some potential advantages. The aim of this study was to study how 5 processing regimes of cold-smoked cod and salmon conducted at different stages of rigor influenced yield, fillet shrinkage, and gaping. Farmed cod and salmon was filleted, salted by brine injection of 25% NaCl, and smoked for 2 h at different stages of rigor. Filleting and salting prerigor resulted in increased fillet shrinkage and less increase in weight during brine injection, which in turn was correlated to the salt content of the fillet. These effects were more pronounced in cod fillets when compared to salmon. Early processing reduced fillet gaping and fillets were evaluated as having a firmer texture. In a follow-up trial with cod, shrinkage and weight gain during injection was studied as an effect of processing time postmortem. No changes in weight gain were observed for fillets salted the first 24 h postmortem; however, by delaying the processing 12 h postmortem, the high and rapid shrinking of cod fillets during brine injection was halved.
An analysis of phenotypic selection in natural stands of northern red oak (Quercus rubra L.)
Jeffery W. Stringer; David B. Wagner; Scott E. Schlarbaum; Daniel B. Houston
1995-01-01
Comparison of growth and stem quality parameters of 19-year-old progeny from superior and comparison trees indicates that rigorous phenotypic selection of trees in natural stands may not be an efficient method of parent tree selection for Quercus rubra L. Total tree height, dbh, number of branches in the butt log, fork height, and number of mainstem...
Jenke, Dennis
2015-01-01
An accelerating trend in the pharmaceutical industry is the use of plastic components in systems used to produce an active pharmaceutical ingredient or a finished drug product. If the active pharmaceutical ingredient, the finished drug product, or any solution used to generate them (for example, a process stream such as media, buffers, eluents, and the like) is contacted by a plastic component at any time during the production process, substances leached from the component may accumulate in the active pharmaceutical ingredient or finished drug product, affecting its safety and/or efficacy. In this article the author develops and justifies a semi-quantitative risk evaluation matrix that is used to determine the amount and rigor of component testing necessary and appropriate to establish that the component is chemically suitable for its intended use. By considering key properties of the component, the contact medium, the contact conditions, and the active pharmaceutical ingredient's or finished drug product's clinical conditions of use, use of the risk evaluation matrix produces a risk score whose magnitude reflects the accumulated risk that the component will interact with the contact solution to such an extent that component-related extractables will accumulate in the active pharmaceutical ingredient or finished drug product as leachables at levels sufficiently high to adversely affect user safety. The magnitude of the risk score establishes the amount and rigor of the testing that is required to select and qualify the component, and such testing is broadly grouped into three categories: baseline assessment, general testing, and full testing (extractables profiling). Production suites used to generate pharmaceuticals can include plastic components. It is possible that substances in the components could leach into manufacturing solutions and accumulate in the pharmaceutical product. In this article the author develops and justifies a semi-quantitative risk evaluation matrix that can be used to determine the amount and rigor of component testing that may be necessary and appropriate to establish that the component is suitable for its intended use. Use of the risk evaluation matrix allows a user of a component to determine the type and amount of testing that should be performed to establish the patient safety risk associated with using that component in order to manufacture an active pharmaceutical ingredient or a finished drug product. © PDA, Inc. 2015.
Rotter, Thomas; Kinsman, Leigh; James, Erica; Machotta, Andreas; Steyerberg, Ewout W
2012-06-18
The purpose of this article is to report on the quality of the existing evidence base regarding the effectiveness of clinical pathway (CPW) research in the hospital setting. The analysis is based on a recently published Cochrane review of the effectiveness of CPWs. An integral component of the review process was a rigorous appraisal of the methodological quality of published CPW evaluations. This allowed the identification of strengths and limitations of the evidence base for CPW effectiveness. We followed the validated Cochrane Effective Practice and Organisation of Care Group (EPOC) criteria for randomized and non-randomized clinical pathway evaluations. In addition, we tested the hypotheses that simple pre-post studies tend to overestimate CPW effects reported. Out of the 260 primary studies meeting CPW content criteria, only 27 studies met the EPOC study design criteria, with the majority of CPW studies (more than 70 %) excluded from the review on the basis that they were simple pre-post evaluations, mostly comparing two or more annual patient cohorts. Methodologically poor study designs are often used to evaluate CPWs and this compromises the quality of the existing evidence base. Cochrane EPOC methodological criteria, including the selection of rigorous study designs along with detailed descriptions of CPW development and implementation processes, are recommended for quantitative evaluations to improve the evidence base for the use of CPWs in hospitals.
Benson, Neil; van der Graaf, Piet H; Peletier, Lambertus A
2017-11-15
A key element of the drug discovery process is target selection. Although the topic is subject to much discussion and experimental effort, there are no defined quantitative rules around optimal selection. Often 'rules of thumb', that have not been subject to rigorous exploration, are used. In this paper we explore the 'rule of thumb' notion that the molecule that initiates a pathway signal is the optimal target. Given the multi-factorial and complex nature of this question, we have simplified an example pathway to its logical minimum of two steps and used a mathematical model of this to explore the different options in the context of typical small and large molecule drugs. In this paper, we report the conclusions of our analysis and describe the analysis tool and methods used. These provide a platform to enable a more extensive enquiry into this important topic. Copyright © 2017 Elsevier B.V. All rights reserved.
34 CFR 645.6 - What definitions apply to the Upward Bound Program?
Code of Federal Regulations, 2012 CFR
2012-07-01
..., chemistry, and physics. (iv) Three years of social studies. (v) One year of a language other than English... by the individual's State. Rigorous secondary school program of study means a program of study that... recognized as a rigorous secondary school program of study by the Secretary through the process described in...
34 CFR 645.6 - What definitions apply to the Upward Bound Program?
Code of Federal Regulations, 2014 CFR
2014-07-01
..., chemistry, and physics. (iv) Three years of social studies. (v) One year of a language other than English... by the individual's State. Rigorous secondary school program of study means a program of study that... recognized as a rigorous secondary school program of study by the Secretary through the process described in...
34 CFR 645.6 - What definitions apply to the Upward Bound Program?
Code of Federal Regulations, 2013 CFR
2013-07-01
..., chemistry, and physics. (iv) Three years of social studies. (v) One year of a language other than English... by the individual's State. Rigorous secondary school program of study means a program of study that... recognized as a rigorous secondary school program of study by the Secretary through the process described in...
34 CFR 645.6 - What definitions apply to the Upward Bound Program?
Code of Federal Regulations, 2011 CFR
2011-07-01
..., chemistry, and physics. (iv) Three years of social studies. (v) One year of a language other than English... by the individual's State. Rigorous secondary school program of study means a program of study that... recognized as a rigorous secondary school program of study by the Secretary through the process described in...
Quasi-experimental study designs series-paper 6: risk of bias assessment.
Waddington, Hugh; Aloe, Ariel M; Becker, Betsy Jane; Djimeu, Eric W; Hombrados, Jorge Garcia; Tugwell, Peter; Wells, George; Reeves, Barney
2017-09-01
Rigorous and transparent bias assessment is a core component of high-quality systematic reviews. We assess modifications to existing risk of bias approaches to incorporate rigorous quasi-experimental approaches with selection on unobservables. These are nonrandomized studies using design-based approaches to control for unobservable sources of confounding such as difference studies, instrumental variables, interrupted time series, natural experiments, and regression-discontinuity designs. We review existing risk of bias tools. Drawing on these tools, we present domains of bias and suggest directions for evaluation questions. The review suggests that existing risk of bias tools provide, to different degrees, incomplete transparent criteria to assess the validity of these designs. The paper then presents an approach to evaluating the internal validity of quasi-experiments with selection on unobservables. We conclude that tools for nonrandomized studies of interventions need to be further developed to incorporate evaluation questions for quasi-experiments with selection on unobservables. Copyright © 2017 Elsevier Inc. All rights reserved.
Hughes, Brianna H; Perkins, L Brian; Yang, Tom C; Skonberg, Denise I
2016-03-01
High pressure processing (HPP) of post-rigor abalone at 300MPa for 10min extended the refrigerated shelf-life to four times that of unprocessed controls. Shucked abalone meats were processed at 100 or 300MPa for 5 or 10min, and stored at 2°C for 35days. Treatments were analyzed for aerobic plate count (APC), total volatile base nitrogen (TVBN), K-value, biogenic amines, color, and texture. APC did not exceed 10(6) and TVBN levels remained below 35mg/100g for 35days for the 300MPa treatments. No biogenic amines were detected in the 300MPa treatments, but putrescine and cadaverine were detected in the control and 100MPa treatments. Color and texture were not affected by HPP or storage time. These results indicate that post-rigor processing at 300MPa for 10min can significantly increase refrigerated shelf-life of abalone without affecting chemical or physical quality characteristics important to consumers. Copyright © 2015 Elsevier Ltd. All rights reserved.
Five Guidelines for Selecting Hydrological Signatures
NASA Astrophysics Data System (ADS)
McMillan, H. K.; Westerberg, I.; Branger, F.
2017-12-01
Hydrological signatures are index values derived from observed or modeled series of hydrological data such as rainfall, flow or soil moisture. They are designed to extract relevant information about hydrological behavior, such as to identify dominant processes, and to determine the strength, speed and spatiotemporal variability of the rainfall-runoff response. Hydrological signatures play an important role in model evaluation. They allow us to test whether particular model structures or parameter sets accurately reproduce the runoff generation processes within the watershed of interest. Most modeling studies use a selection of different signatures to capture different aspects of the catchment response, for example evaluating overall flow distribution as well as high and low flow extremes and flow timing. Such studies often choose their own set of signatures, or may borrow subsets of signatures used in multiple other works. The link between signature values and hydrological processes is not always straightforward, leading to uncertainty and variability in hydrologists' signature choices. In this presentation, we aim to encourage a more rigorous approach to hydrological signature selection, which considers the ability of signatures to represent hydrological behavior and underlying processes for the catchment and application in question. To this end, we propose a set of guidelines for selecting hydrological signatures. We describe five criteria that any hydrological signature should conform to: Identifiability, Robustness, Consistency, Representativeness, and Discriminatory Power. We describe an example of the design process for a signature, assessing possible signature designs against the guidelines above. Due to their ubiquity, we chose a signature related to the Flow Duration Curve, selecting the FDC mid-section slope as a proposed signature to quantify catchment overall behavior and flashiness. We demonstrate how assessment against each guideline could be used to compare or choose between alternative signature definitions. We believe that reaching a consensus on selection criteria for hydrological signatures will assist modelers to choose between competing signatures, facilitate comparison between hydrological studies, and help hydrologists to fully evaluate their models.
Efficiency versus speed in quantum heat engines: Rigorous constraint from Lieb-Robinson bound
NASA Astrophysics Data System (ADS)
Shiraishi, Naoto; Tajima, Hiroyasu
2017-08-01
A long-standing open problem whether a heat engine with finite power achieves the Carnot efficiency is investgated. We rigorously prove a general trade-off inequality on thermodynamic efficiency and time interval of a cyclic process with quantum heat engines. In a first step, employing the Lieb-Robinson bound we establish an inequality on the change in a local observable caused by an operation far from support of the local observable. This inequality provides a rigorous characterization of the following intuitive picture that most of the energy emitted from the engine to the cold bath remains near the engine when the cyclic process is finished. Using this description, we prove an upper bound on efficiency with the aid of quantum information geometry. Our result generally excludes the possibility of a process with finite speed at the Carnot efficiency in quantum heat engines. In particular, the obtained constraint covers engines evolving with non-Markovian dynamics, which almost all previous studies on this topic fail to address.
Efficiency versus speed in quantum heat engines: Rigorous constraint from Lieb-Robinson bound.
Shiraishi, Naoto; Tajima, Hiroyasu
2017-08-01
A long-standing open problem whether a heat engine with finite power achieves the Carnot efficiency is investgated. We rigorously prove a general trade-off inequality on thermodynamic efficiency and time interval of a cyclic process with quantum heat engines. In a first step, employing the Lieb-Robinson bound we establish an inequality on the change in a local observable caused by an operation far from support of the local observable. This inequality provides a rigorous characterization of the following intuitive picture that most of the energy emitted from the engine to the cold bath remains near the engine when the cyclic process is finished. Using this description, we prove an upper bound on efficiency with the aid of quantum information geometry. Our result generally excludes the possibility of a process with finite speed at the Carnot efficiency in quantum heat engines. In particular, the obtained constraint covers engines evolving with non-Markovian dynamics, which almost all previous studies on this topic fail to address.
Re-establishment of rigor mortis: evidence for a considerably longer post-mortem time span.
Crostack, Chiara; Sehner, Susanne; Raupach, Tobias; Anders, Sven
2017-07-01
Re-establishment of rigor mortis following mechanical loosening is used as part of the complex method for the forensic estimation of the time since death in human bodies and has formerly been reported to occur up to 8-12 h post-mortem (hpm). We recently described our observation of the phenomenon in up to 19 hpm in cases with in-hospital death. Due to the case selection (preceding illness, immobilisation), transfer of these results to forensic cases might be limited. We therefore examined 67 out-of-hospital cases of sudden death with known time points of death. Re-establishment of rigor mortis was positive in 52.2% of cases and was observed up to 20 hpm. In contrast to the current doctrine that a recurrence of rigor mortis is always of a lesser degree than its first manifestation in a given patient, muscular rigidity at re-establishment equalled or even exceeded the degree observed before dissolving in 21 joints. Furthermore, this is the first study to describe that the phenomenon appears to be independent of body or ambient temperature.
Tailoring Systems Engineering Projects for Small Satellite Missions
NASA Technical Reports Server (NTRS)
Horan, Stephen; Belvin, Keith
2013-01-01
NASA maintains excellence in its spaceflight systems by utilizing rigorous engineering processes based on over 50 years of experience. The NASA systems engineering process for flight projects described in NPR 7120.5E was initially developed for major flight projects. The design and development of low-cost small satellite systems does not entail the financial and risk consequences traditionally associated with spaceflight projects. Consequently, an approach is offered to tailoring of the processes such that the small satellite missions will benefit from the engineering rigor without overly burdensome overhead. In this paper we will outline the approaches to tailoring the standard processes for these small missions and describe how it will be applied in a proposed small satellite mission.
Kobayashi, Masahiko; Takemori, Shigeru; Yamaguchi, Maki
2004-02-10
Based on the molecular mechanism of rigor mortis, we have proposed that stiffness (elastic modulus evaluated with tension response against minute length perturbations) can be a suitable index of post-mortem rigidity in skeletal muscle. To trace the developmental process of rigor mortis, we measured stiffness and tension in both red and white rat skeletal muscle kept in liquid paraffin at 37 and 25 degrees C. White muscle (in which type IIB fibres predominate) developed stiffness and tension significantly more slowly than red muscle, except for soleus red muscle at 25 degrees C, which showed disproportionately slow rigor development. In each of the examined muscles, stiffness and tension developed more slowly at 25 degrees C than at 37 degrees C. In each specimen, tension always reached its maximum level earlier than stiffness, and then decreased more rapidly and markedly than stiffness. These phenomena may account for the sequential progress of rigor mortis in human cadavers.
Ghosh, Monisankar; Saha, Suchandrima; Dutta, Samir Kumar
2016-02-07
Herein, we synthesize and elucidate the potential of a novel 'dual hit' molecule, LDCA, to constitutively block lactate dehydrogenase isoform-A (LDH-A) to selectively subvert apoptosis and rigorously attenuate breast tumor progression in a mouse model, comprehensively delineating the therapeutic prospectus of LDCA in the field of cancer metabolics.
The difficulties of systematic reviews.
Westgate, Martin J; Lindenmayer, David B
2017-10-01
The need for robust evidence to support conservation actions has driven the adoption of systematic approaches to research synthesis in ecology. However, applying systematic review to complex or open questions remains challenging, and this task is becoming more difficult as the quantity of scientific literature increases. We drew on the science of linguistics for guidance as to why the process of identifying and sorting information during systematic review remains so labor intensive, and to provide potential solutions. Several linguistic properties of peer-reviewed corpora-including nonrandom selection of review topics, small-world properties of semantic networks, and spatiotemporal variation in word meaning-greatly increase the effort needed to complete the systematic review process. Conversely, the resolution of these semantic complexities is a common motivation for narrative reviews, but this process is rarely enacted with the rigor applied during linguistic analysis. Therefore, linguistics provides a unifying framework for understanding some key challenges of systematic review and highlights 2 useful directions for future research. First, in cases where semantic complexity generates barriers to synthesis, ecologists should consider drawing on existing methods-such as natural language processing or the construction of research thesauri and ontologies-that provide tools for mapping and resolving that complexity. These tools could help individual researchers classify research material in a more robust manner and provide valuable guidance for future researchers on that topic. Second, a linguistic perspective highlights that scientific writing is a rich resource worthy of detailed study, an observation that can sometimes be lost during the search for data during systematic review or meta-analysis. For example, mapping semantic networks can reveal redundancy and complementarity among scientific concepts, leading to new insights and research questions. Consequently, wider adoption of linguistic approaches may facilitate improved rigor and richness in research synthesis. © 2017 Society for Conservation Biology.
NASA Technical Reports Server (NTRS)
Xu, Xidong; Ulrey, Mike L.; Brown, John A.; Mast, James; Lapis, Mary B.
2013-01-01
NextGen is a complex socio-technical system and, in many ways, it is expected to be more complex than the current system. It is vital to assess the safety impact of the NextGen elements (technologies, systems, and procedures) in a rigorous and systematic way and to ensure that they do not compromise safety. In this study, the NextGen elements in the form of Operational Improvements (OIs), Enablers, Research Activities, Development Activities, and Policy Issues were identified. The overall hazard situation in NextGen was outlined; a high-level hazard analysis was conducted with respect to multiple elements in a representative NextGen OI known as OI-0349 (Automation Support for Separation Management); and the hazards resulting from the highly dynamic complexity involved in an OI-0349 scenario were illustrated. A selected but representative set of the existing safety methods, tools, processes, and regulations was then reviewed and analyzed regarding whether they are sufficient to assess safety in the elements of that OI and ensure that safety will not be compromised and whether they might incur intolerably high costs.
Accurate force field for molybdenum by machine learning large materials data
NASA Astrophysics Data System (ADS)
Chen, Chi; Deng, Zhi; Tran, Richard; Tang, Hanmei; Chu, Iek-Heng; Ong, Shyue Ping
2017-09-01
In this work, we present a highly accurate spectral neighbor analysis potential (SNAP) model for molybdenum (Mo) developed through the rigorous application of machine learning techniques on large materials data sets. Despite Mo's importance as a structural metal, existing force fields for Mo based on the embedded atom and modified embedded atom methods do not provide satisfactory accuracy on many properties. We will show that by fitting to the energies, forces, and stress tensors of a large density functional theory (DFT)-computed dataset on a diverse set of Mo structures, a Mo SNAP model can be developed that achieves close to DFT accuracy in the prediction of a broad range of properties, including elastic constants, melting point, phonon spectra, surface energies, grain boundary energies, etc. We will outline a systematic model development process, which includes a rigorous approach to structural selection based on principal component analysis, as well as a differential evolution algorithm for optimizing the hyperparameters in the model fitting so that both the model error and the property prediction error can be simultaneously lowered. We expect that this newly developed Mo SNAP model will find broad applications in large and long-time scale simulations.
Inferring the nature of anthropogenic threats from long-term abundance records.
Shoemaker, Kevin T; Akçakaya, H Resit
2015-02-01
Diagnosing the processes that threaten species persistence is critical for recovery planning and risk forecasting. Dominant threats are typically inferred by experts on the basis of a patchwork of informal methods. Transparent, quantitative diagnostic tools would contribute much-needed consistency, objectivity, and rigor to the process of diagnosing anthropogenic threats. Long-term census records, available for an increasingly large and diverse set of taxa, may exhibit characteristic signatures of specific threatening processes and thereby provide information for threat diagnosis. We developed a flexible Bayesian framework for diagnosing threats on the basis of long-term census records and diverse ancillary sources of information. We tested this framework with simulated data from artificial populations subjected to varying degrees of exploitation and habitat loss and several real-world abundance time series for which threatening processes are relatively well understood: bluefin tuna (Thunnus maccoyii) and Atlantic cod (Gadus morhua) (exploitation) and Red Grouse (Lagopus lagopus scotica) and Eurasian Skylark (Alauda arvensis) (habitat loss). Our method correctly identified the process driving population decline for over 90% of time series simulated under moderate to severe threat scenarios. Successful identification of threats approached 100% for severe exploitation and habitat loss scenarios. Our method identified threats less successfully when threatening processes were weak and when populations were simultaneously affected by multiple threats. Our method selected the presumed true threat model for all real-world case studies, although results were somewhat ambiguous in the case of the Eurasian Skylark. In the latter case, incorporation of an ancillary source of information (records of land-use change) increased the weight assigned to the presumed true model from 70% to 92%, illustrating the value of the proposed framework in bringing diverse sources of information into a common rigorous framework. Ultimately, our framework may greatly assist conservation organizations in documenting threatening processes and planning species recovery. © 2014 Society for Conservation Biology.
A practical guide for nurses in diluent selection for subcutaneous infusion using a syringe driver.
McLeod, Fiona; Flowers, Charne
2006-12-01
Appropriate diluent selection in continuous subcutaneous infusion optimises symptom management and client well-being. The responsibility of diluent selection is commonly one of the attending nurse. This paper was developed with the intention of providing nurses with practical instruction for diluent selection when preparing medications for administration subcutaneously using a syringe driver. A literature review was undertaken of published journal databases and published guidelines sites. Recommendations regarding diluent choice were reviewed in two iterations by an expert panel of palliative care nurse clinicians. The principles for diluent selection are presented. They are based primarily on expert opinion level of evidence given a lack of primary research evidence in the area of diluent selection. There is a pressing need for manufacturers' guidance on diluent selection and independent research to establish the impact of diluents on drug and drug combinations when using syringe drivers. Until such time that this evidence is available to guide practice, clinicians need to be trained to inspect solutions and assess the effectiveness of the medication in controlling symptoms. The capacity of this paper to provide practical instruction has been limited by the lack of rigorous evidence available, and indeed, the process of developing this guide identified perhaps more questions than answers available at the present time.
Multilocus patterns of polymorphism and selection across the X chromosome of Caenorhabditis remanei.
Cutter, Asher D
2008-03-01
Natural selection and neutral processes such as demography, mutation, and gene conversion all contribute to patterns of polymorphism within genomes. Identifying the relative importance of these varied components in evolution provides the principal challenge for population genetics. To address this issue in the nematode Caenorhabditis remanei, I sampled nucleotide polymorphism at 40 loci across the X chromosome. The site-frequency spectrum for these loci provides no evidence for population size change, and one locus presents a candidate for linkage to a target of balancing selection. Selection for codon usage bias leads to the non-neutrality of synonymous sites, and despite its weak magnitude of effect (N(e)s approximately 0.1), is responsible for profound patterns of diversity and divergence in the C. remanei genome. Although gene conversion is evident for many loci, biased gene conversion is not identified as a significant evolutionary process in this sample. No consistent association is observed between synonymous-site diversity and linkage-disequilibrium-based estimators of the population recombination parameter, despite theoretical predictions about background selection or widespread genetic hitchhiking, but genetic map-based estimates of recombination are needed to rigorously test for a diversity-recombination relationship. Coalescent simulations also illustrate how a spurious correlation between diversity and linkage-disequilibrium-based estimators of recombination can occur, due in part to the presence of unbiased gene conversion. These results illustrate the influence that subtle natural selection can exert on polymorphism and divergence, in the form of codon usage bias, and demonstrate the potential of C. remanei for detecting natural selection from genomic scans of polymorphism.
Ong, MS.
2015-01-01
Summary Objective To summarize significant contributions to the research on human factors and organizational issues in medical informatics. Methods An extensive search using PubMed/Medline and Web of Science® was conducted to identify the scientific contributions, published in 2014, to human factors and organizational issues in medical informatics, with a focus on health information technology (HIT) usability. The selection process comprised three steps: (i) 15 candidate best papers were selected by the two section editors, (ii) external reviewers from a pool of international experts reviewed each candidate best paper, and (iii) the final selection of three best papers was made by the editorial board of the IMIA Yearbook. Results Noteworthy papers published in 2014 describe an efficient, easy to implement, and useful process for detecting and mitigating human factors and ergonomics (HFE) issues of HIT. They contribute to promote the HFE approach with interventions based on rigorous and well-conducted methods when designing and implementing HIT. Conclusion The application of HFE in the design and implementation of HIT remains limited, and the impact of incorporating HFE principles on patient safety is understudied. Future works should be conducted to advance this field of research, so that the safety and quality of patient care are not compromised by the increasing adoption of HIT. PMID:26293852
A method to select human-system interfaces for nuclear power plants
Hugo, Jacques Victor; Gertman, David Ira
2015-10-19
The new generation of nuclear power plants (NPPs) will likely make use of state-of-the-art technologies in many areas of the plant. The analysis, design, and selection of advanced human–system interfaces (HSIs) constitute an important part of power plant engineering. Designers need to consider the new capabilities afforded by these technologies in the context of current regulations and new operational concepts, which is why they need a more rigorous method by which to plan the introduction of advanced HSIs in NPP work areas. Much of current human factors research stops at the user interface and fails to provide a definitive processmore » for integration of end user devices with instrumentation and control (I&C) and operational concepts. The current lack of a clear definition of HSI technology, including the process for integration, makes characterization and implementation of new and advanced HSIs difficult. This paper describes how new design concepts in the nuclear industry can be analyzed and how HSI technologies associated with new industrial processes might be considered. Furthermore, it also describes a basis for an understanding of human as well as technology characteristics that could be incorporated into a prioritization scheme for technology selection and deployment plans.« less
The MINERVA Software Development Process
NASA Technical Reports Server (NTRS)
Narkawicz, Anthony; Munoz, Cesar A.; Dutle, Aaron M.
2017-01-01
This paper presents a software development process for safety-critical software components of cyber-physical systems. The process is called MINERVA, which stands for Mirrored Implementation Numerically Evaluated against Rigorously Verified Algorithms. The process relies on formal methods for rigorously validating code against its requirements. The software development process uses: (1) a formal specification language for describing the algorithms and their functional requirements, (2) an interactive theorem prover for formally verifying the correctness of the algorithms, (3) test cases that stress the code, and (4) numerical evaluation on these test cases of both the algorithm specifications and their implementations in code. The MINERVA process is illustrated in this paper with an application to geo-containment algorithms for unmanned aircraft systems. These algorithms ensure that the position of an aircraft never leaves a predetermined polygon region and provide recovery maneuvers when the region is inadvertently exited.
Krompecher, Thomas; Gilles, André; Brandt-Casadevall, Conception; Mangin, Patrice
2008-04-07
Objective measurements were carried out to study the possible re-establishment of rigor mortis on rats after "breaking" (mechanical solution). Our experiments showed that: *Cadaveric rigidity can re-establish after breaking. *A significant rigidity can reappear if the breaking occurs before the process is complete. *Rigidity will be considerably weaker after the breaking. *The time course of the intensity does not change in comparison to the controls: --the re-establishment begins immediately after the breaking; --maximal values are reached at the same time as in the controls; --the course of the resolution is the same as in the controls.
Conceptualizing Rigor and Its Implications for Education in the Era of the Common Core
ERIC Educational Resources Information Center
Paige, David D.; Smith, Grant S.; Sizemore, John M.
2015-01-01
The adoption of Common Core State Standards in the USA by 46 states and the District of Columbia has provided several new foci for K-12 instruction, not the least of which is the reading and understanding of complex text, a higher order thinking process. Closely associated with this is the notion of rigor, the focus of the present study. As…
Rigorous mathematical modelling for a Fast Corrector Power Supply in TPS
NASA Astrophysics Data System (ADS)
Liu, K.-B.; Liu, C.-Y.; Chien, Y.-C.; Wang, B.-S.; Wong, Y. S.
2017-04-01
To enhance the stability of beam orbit, a Fast Orbit Feedback System (FOFB) eliminating undesired disturbances was installed and tested in the 3rd generation synchrotron light source of Taiwan Photon Source (TPS) of National Synchrotron Radiation Research Center (NSRRC). The effectiveness of the FOFB greatly depends on the output performance of Fast Corrector Power Supply (FCPS); therefore, the design and implementation of an accurate FCPS is essential. A rigorous mathematical modelling is very useful to shorten design time and improve design performance of a FCPS. A rigorous mathematical modelling derived by the state-space averaging method for a FCPS in the FOFB of TPS composed of a full-bridge topology is therefore proposed in this paper. The MATLAB/SIMULINK software is used to construct the proposed mathematical modelling and to conduct the simulations of the FCPS. Simulations for the effects of the different resolutions of ADC on the output accuracy of the FCPS are investigated. A FCPS prototype is realized to demonstrate the effectiveness of the proposed rigorous mathematical modelling for the FCPS. Simulation and experimental results show that the proposed mathematical modelling is helpful for selecting the appropriate components to meet the accuracy requirements of a FCPS.
Flower, Andrew; Lewith, George T; Little, Paul
2007-11-01
For most complementary and alternative medicine interventions, the absence of a high-quality evidence base to define good practice presents a serious problem for clinicians, educators, and researchers. The Delphi process may offer a pragmatic way to establish good practice guidelines until more rigorous forms of assessment can be undertaken. To use a modified Delphi to develop good practice guidelines for a feasibility study exploring the role of Chinese herbal medicine (CHM) in the treatment of endometriosis. To compare the outcomes from Delphi with data derived from a systematic review of the Chinese language database. An expert group was convened for a three-round Delphi that initially produced key statements relating to the CHM diagnosis and treatment of endometriosis (round 1) and then anonymously rated these on a 1-7 Likert scale (rounds 2 and 3). Statements with a median score of 5 and above were regarded as demonstrating positive group consensus. The differential diagnoses within Chinese Medicine and rating of the clinical value of individual herbs were then contrasted with comparable data from a review of Chinese language reports in the Chinese Biomedical Retrieval System (1978-2002), and China Academy of Traditional Chinese Medicine (1985-2002) databases and the Chinese TCM and magazine literature (1984-2004) databases. Consensus (good practice) guidelines for the CHM treatment of endometriosis relating to common diagnostic patterns, herb selection, dosage, and patient management were produced. The Delphi guidelines demonstrated a high degree of congruence with the information from the Chinese language databases. In the absence of rigorous evidence, Delphi offers a way to synthesize expert knowledge relating to diagnosis, patient management, and herbal selection in the treatment of endometriosis. The limitations of the expert group and the inability of Delphi to capture the subtle nuances of individualized clinical decision-making limit the usefulness of this approach.
Peer Review of EPA's Draft BMDS Document: Exponential ...
BMDS is one of the Agency's premier tools for estimating risk assessments, therefore the validity and reliability of its statistical models are of paramount importance. This page provides links to peer review of the BMDS applications and its models as they were developed and eventually released documenting the rigorous review process taken to provide the best science tools available for statistical modeling. This page provides links to peer review of the BMDS applications and its models as they were developed and eventually released documenting the rigorous review process taken to provide the best science tools available for statistical modeling.
A Rigorous Treatment of Energy Extraction from a Rotating Black Hole
NASA Astrophysics Data System (ADS)
Finster, F.; Kamran, N.; Smoller, J.; Yau, S.-T.
2009-05-01
The Cauchy problem is considered for the scalar wave equation in the Kerr geometry. We prove that by choosing a suitable wave packet as initial data, one can extract energy from the black hole, thereby putting supperradiance, the wave analogue of the Penrose process, into a rigorous mathematical framework. We quantify the maximal energy gain. We also compute the infinitesimal change of mass and angular momentum of the black hole, in agreement with Christodoulou’s result for the Penrose process. The main mathematical tool is our previously derived integral representation of the wave propagator.
Abiiro, Gilbert Abotisem; Leppert, Gerald; Mbera, Grace Bongololo; Robyn, Paul J; De Allegri, Manuela
2014-05-22
Discrete choice experiments (DCEs) are attribute-driven experimental techniques used to elicit stakeholders' preferences to support the design and implementation of policy interventions. The validity of a DCE, therefore, depends on the appropriate specification of the attributes and their levels. There have been recent calls for greater rigor in implementing and reporting on the processes of developing attributes and attribute-levels for discrete choice experiments (DCEs). This paper responds to such calls by carefully reporting a systematic process of developing micro health insurance attributes and attribute-levels for the design of a DCE in rural Malawi. Conceptual attributes and attribute-levels were initially derived from a literature review which informed the design of qualitative data collection tools to identify context specific attributes and attribute-levels. Qualitative data was collected in August-September 2012 from 12 focus group discussions with community residents and 8 in-depth interviews with health workers. All participants were selected according to stratified purposive sampling. The material was tape-recorded, fully transcribed, and coded by three researchers to identify context-specific attributes and attribute-levels. Expert opinion was used to scale down the attributes and levels. A pilot study confirmed the appropriateness of the selected attributes and levels for a DCE. First, a consensus, emerging from an individual level analysis of the qualitative transcripts, identified 10 candidate attributes. Levels were assigned to all attributes based on data from transcripts and knowledge of the Malawian context, derived from literature. Second, through further discussions with experts, four attributes were discarded based on multiple criteria. The 6 remaining attributes were: premium level, unit of enrollment, management structure, health service benefit package, transportation coverage and copayment levels. A final step of revision and piloting confirmed that the retained attributes satisfied the credibility criteria of DCE attributes. This detailed description makes our attribute development process transparent, and provides the reader with a basis to assess the rigor of this stage of constructing the DCE. This paper contributes empirical evidence to the limited methodological literature on attributes and levels development for DCE, thereby providing further empirical guidance on the matter, specifically within rural communities of low- and middle-income countries.
Investigating the utility of a GPA institutional adjustment index.
Didier, Thomas; Kreiter, Clarence D; Buri, Russell; Solow, Catherine
2006-05-01
Grading standards vary widely across undergraduate institutions. If, during the medical school admissions process, GPA is considered without reference to the institution attended, it will disadvantage applicants from undergraduate institutions employing rigorous grading standards. A regression-based GPA institutional equating method using historical MCAT and GPA information is described. Classes selected from eight applicant pools demonstrate the impact of the GPA adjustment. The validity of the adjustment is examined by comparing adjusted and unadjusted GPAs' correlation with USMLE and medical college grades. The adjusted GPA demonstrated significantly improved congruence with MCAT estimates of applicant preparedness. The adjustment changed selection decisions for 21% of those admitted. The adjusted GPA enhanced prediction of USMLE and medical school grades only for students from institutions which required large adjustments. Unlike other indices, the adjustment described uses the same metric as GPA and is based only on an institution's history of preparing medical school applicants. The institutional adjustment is consequential in selection, significantly enhances congruence with a standardized measure of academic preparedness and may enhance the validity of the GPA.
Selecting at-risk populations for sexually transmitted disease/HIV intervention studies.
Wu, Zunyou; Rotheram-Borus, Mary Jane; Detels, Roger; Li, Li; Guan, Jihui; Liang, Guojun; Yap, Lorraine
2007-12-01
This paper describes one option to select populations for randomized, controlled trials (RCT). We used a popular opinion leader intervention in Fuzhou, China, to: (1) identify population selection criteria; (2) systematically examine the suitability of potential target populations and settings; (3) briefly evaluate risk and stability in the population; and (4) evaluate regional and organizational support among administrators and government officials. After comparing migrant villagers, truck drivers, factory workers, construction workers, and market employees in five regions of China, market employees in Fuzhou were identified as the optimal target population. Markets were the optimal sites for several reasons: (1) the population demonstrated a sufficient base rate of sexually transmitted diseases; (2) the population was stable over time; (3) a sufficient number of sites of manageable sizes were available; (4) stable networks existed; (5) local gatekeepers/stakeholders supported the intervention; (6) there was organizational capacity in the local health department to mount the intervention; (7) the demographic profile was similar across potential sites; and (8) the sites were sufficiently distanced to minimize contamination. Evaluating intervention efficacy in an RCT requires a time-consuming and rigorous process that systematically and routinely documents selection criteria, evaluates multiple populations, sites, and organizations for their appropriateness.
Raymond, Nancy C; Wyman, Jean F; Dighe, Satlaj; Harwood, Eileen M; Hang, Mikow
2018-06-01
Process evaluation is an important tool in quality improvement efforts. This article illustrates how a systematic and continuous evaluation process can be used to improve the quality of faculty career development programs by using the University of Minnesota's Building Interdisciplinary Research Careers in Women's Health (BIRCWH) K12 program as an exemplar. Data from a rigorous process evaluation incorporating quantitative and qualitative measurements were analyzed and reviewed by the BIRCWH program leadership on a regular basis. Examples are provided of how this evaluation model and processes were used to improve many aspects of the program, thereby improving scholar, mentor, and advisory committee members' satisfaction and scholar outcomes. A rigorous evaluation plan can increase the effectiveness and impact of a research career development plan.
Osorio-Cuellar, Gisel Viviana; Pacichana-Quinayáz, Sara Gabriela; Bonilla-Escobar, Francisco Javier; Fandiño-Losada, Andrés; Gutiérrez-Martinez, Maria Isabel
2017-09-01
Given the context and the number of armed conflict victims in the Colombian Pacific coast and their difficulties to access psycho-social care, Narrative Community-based Group Therapy appears as a viable mental health intervention. The objective of this study is to describe the process of implementation and results of the intervention in Afro-Colombian victims of violence, in the municipalities of Buenaventura and Quibdó. More specifically, we will be looking at the perspectives of workers and supervisors, through evaluative case studies and individual in-depth interviews. The therapy allows us to identify support and coping systems through coexistence, communication and interaction. It requires an adaptation process to the diversity of knowledge and expressions of victims of Colombian violence, greater empathy from care providers and rigor in their profiles selection, facilities ensuring security and confidentiality, and links with other educational, employment and recreational organizations. It is important to include these results while improving current and future intervention processes.
NASA Astrophysics Data System (ADS)
Mustac, M.; Kim, S.; Tkalcic, H.; Rhie, J.; Chen, Y.; Ford, S. R.; Sebastian, N.
2015-12-01
Conventional approaches to inverse problems suffer from non-linearity and non-uniqueness in estimations of seismic structures and source properties. Estimated results and associated uncertainties are often biased by applied regularizations and additional constraints, which are commonly introduced to solve such problems. Bayesian methods, however, provide statistically meaningful estimations of models and their uncertainties constrained by data information. In addition, hierarchical and trans-dimensional (trans-D) techniques are inherently implemented in the Bayesian framework to account for involved error statistics and model parameterizations, and, in turn, allow more rigorous estimations of the same. Here, we apply Bayesian methods throughout the entire inference process to estimate seismic structures and source properties in Northeast Asia including east China, the Korean peninsula, and the Japanese islands. Ambient noise analysis is first performed to obtain a base three-dimensional (3-D) heterogeneity model using continuous broadband waveforms from more than 300 stations. As for the tomography of surface wave group and phase velocities in the 5-70 s band, we adopt a hierarchical and trans-D Bayesian inversion method using Voronoi partition. The 3-D heterogeneity model is further improved by joint inversions of teleseismic receiver functions and dispersion data using a newly developed high-efficiency Bayesian technique. The obtained model is subsequently used to prepare 3-D structural Green's functions for the source characterization. A hierarchical Bayesian method for point source inversion using regional complete waveform data is applied to selected events from the region. The seismic structure and source characteristics with rigorously estimated uncertainties from the novel Bayesian methods provide enhanced monitoring and discrimination of seismic events in northeast Asia.
Testing for Mutagens Using Fruit Flies.
ERIC Educational Resources Information Center
Liebl, Eric C.
1998-01-01
Describes a laboratory employed in undergraduate teaching that uses fruit flies to test student-selected compounds for their ability to cause mutations. Requires no prior experience with fruit flies, incorporates a student design component, and employs both rigorous controls and statistical analyses. (DDR)
Development of an evidence-based review with recommendations using an online iterative process.
Rudmik, Luke; Smith, Timothy L
2011-01-01
The practice of modern medicine is governed by evidence-based principles. Due to the plethora of medical literature, clinicians often rely on systematic reviews and clinical guidelines to summarize the evidence and provide best practices. Implementation of an evidence-based clinical approach can minimize variation in health care delivery and optimize the quality of patient care. This article reports a method for developing an "Evidence-based Review with Recommendations" using an online iterative process. The manuscript describes the following steps involved in this process: Clinical topic selection, Evidence-hased review assignment, Literature review and initial manuscript preparation, Iterative review process with author selection, and Manuscript finalization. The goal of this article is to improve efficiency and increase the production of evidence-based reviews while maintaining the high quality and transparency associated with the rigorous methodology utilized for clinical guideline development. With the rise of evidence-based medicine, most medical and surgical specialties have an abundance of clinical topics which would benefit from a formal evidence-based review. Although clinical guideline development is an important methodology, the associated challenges limit development to only the absolute highest priority clinical topics. As outlined in this article, the online iterative approach to the development of an Evidence-based Review with Recommendations may improve productivity without compromising the quality associated with formal guideline development methodology. Copyright © 2011 American Rhinologic Society-American Academy of Otolaryngic Allergy, LLC.
Mathematical Rigor vs. Conceptual Change: Some Early Results
NASA Astrophysics Data System (ADS)
Alexander, W. R.
2003-05-01
Results from two different pedagogical approaches to teaching introductory astronomy at the college level will be presented. The first of these approaches is a descriptive, conceptually based approach that emphasizes conceptual change. This descriptive class is typically an elective for non-science majors. The other approach is a mathematically rigorous treatment that emphasizes problem solving and is designed to prepare students for further study in astronomy. The mathematically rigorous class is typically taken by science majors. It also fulfills an elective science requirement for these science majors. The Astronomy Diagnostic Test version 2 (ADT 2.0) was used as an assessment instrument since the validity and reliability have been investigated by previous researchers. The ADT 2.0 was administered as both a pre-test and post-test to both groups. Initial results show no significant difference between the two groups in the post-test. However, there is a slightly greater improvement for the descriptive class between the pre and post testing compared to the mathematically rigorous course. There was great care to account for variables. These variables included: selection of text, class format as well as instructor differences. Results indicate that the mathematically rigorous model, doesn't improve conceptual understanding any better than the conceptual change model. Additional results indicate that there is a similar gender bias in favor of males that has been measured by previous investigators. This research has been funded by the College of Science and Mathematics at James Madison University.
NASA Astrophysics Data System (ADS)
Röpke, G.
2018-01-01
One of the fundamental problems in physics that are not yet rigorously solved is the statistical mechanics of nonequilibrium processes. An important contribution to describing irreversible behavior starting from reversible Hamiltonian dynamics was given by D. N. Zubarev, who invented the method of the nonequilibrium statistical operator. We discuss this approach, in particular, the extended von Neumann equation, and as an example consider the electrical conductivity of a system of charged particles. We consider the selection of the set of relevant observables. We show the relation between kinetic theory and linear response theory. Using thermodynamic Green's functions, we present a systematic treatment of correlation functions, but the convergence needs investigation. We compare different expressions for the conductivity and list open questions.
Accelerated recovery of Atlantic salmon (Salmo salar) from effects of crowding by swimming.
Veiseth, Eva; Fjaera, Svein Olav; Bjerkeng, Bjørn; Skjervold, Per Olav
2006-07-01
The effects of post-crowding swimming velocity (0, 0.35, and 0.70 m/s) and recovery time (1.5, 6, and 12 h) on physiological recovery and processing quality parameters of adult Atlantic salmon (Salmo salar) were determined. Atlantic salmon crowded to a density similar to that of a commercial slaughter process (>200 kg/m(3), 40 min) were transferred to a swimming chamber for recovery treatment. Osmolality and concentrations of cortisol, glucose and lactate in blood plasma were used as physiological stress indicators, whereas image analyses of extent and duration of rigor contraction, and fillet gaping were used as measures of processing quality. Crowded salmon had a 5.8-fold higher plasma cortisol concentration than control salmon (P<0.05). The elevated plasma cortisol concentration was reduced by increasing the swimming velocity, and had returned to control levels after 6 h recovery at high water velocity. Similar effects of swimming velocity were observed for plasma osmolality and lactate concentration. A lower plasma glucose concentration was present in crowded than in control fish (P<0.05), although a typical post-stress elevation in plasma glucose was observed after the recovery treatments. Lower muscle pH was found in crowded compared with control salmon (P<0.05), but muscle pH returned to control levels after 6 h recovery at intermediate and high swimming velocities and after 12 h in the low velocity group. Crowding caused an early onset of rigor mortis contraction. However, subjecting crowded salmon to active swimming for 6 h before slaughter delayed the onset of rigor mortis contraction from 2.5 to 7.5 h post mortem. The extent of rigor mortis contraction was also affected by crowding and post-stress swimming activity (P<0.05), and the largest degree of contraction was found in crowded salmon. In conclusion, active swimming accelerated the return of plasma cortisol, hydromineral balance, and the energy metabolism of adult Atlantic salmon to pre-stress levels. Moreover, an active swimming period delayed the onset of rigor mortis contraction, which has a positive technological implication for the salmon processing industry.
Ice-sheet response to oceanic forcing.
Joughin, Ian; Alley, Richard B; Holland, David M
2012-11-30
The ice sheets of Greenland and Antarctica are losing ice at accelerating rates, much of which is a response to oceanic forcing, especially of the floating ice shelves. Recent observations establish a clear correspondence between the increased delivery of oceanic heat to the ice-sheet margin and increased ice loss. In Antarctica, most of these processes are reasonably well understood but have not been rigorously quantified. In Greenland, an understanding of the processes by which warmer ocean temperatures drive the observed retreat remains elusive. Experiments designed to identify the relevant processes are confounded by the logistical difficulties of instrumenting ice-choked fjords with actively calving glaciers. For both ice sheets, multiple challenges remain before the fully coupled ice-ocean-atmosphere models needed for rigorous sea-level projection are available.
Kallio, Hanna; Pietilä, Anna-Maija; Johnson, Martin; Kangasniemi, Mari
2016-12-01
To produce a framework for the development of a qualitative semi-structured interview guide. Rigorous data collection procedures fundamentally influence the results of studies. The semi-structured interview is a common data collection method, but methodological research on the development of a semi-structured interview guide is sparse. Systematic methodological review. We searched PubMed, CINAHL, Scopus and Web of Science for methodological papers on semi-structured interview guides from October 2004-September 2014. Having examined 2,703 titles and abstracts and 21 full texts, we finally selected 10 papers. We analysed the data using the qualitative content analysis method. Our analysis resulted in new synthesized knowledge on the development of a semi-structured interview guide, including five phases: (1) identifying the prerequisites for using semi-structured interviews; (2) retrieving and using previous knowledge; (3) formulating the preliminary semi-structured interview guide; (4) pilot testing the guide; and (5) presenting the complete semi-structured interview guide. Rigorous development of a qualitative semi-structured interview guide contributes to the objectivity and trustworthiness of studies and makes the results more plausible. Researchers should consider using this five-step process to develop a semi-structured interview guide and justify the decisions made during it. © 2016 John Wiley & Sons Ltd.
Medicine, methodology, and values: trade-offs in clinical science and practice.
Ho, Vincent K Y
2011-01-01
The current guidelines of evidence-based medicine (EBM) presuppose that clinical research and clinical practice should advance from rigorous scientific tests as they generate reliable, value-free knowledge. Under this presupposition, hypotheses postulated by doctors and patients in the process of their decision making are preferably tested in randomized clinical trials (RCTs), and in systematic reviews and meta-analyses summarizing outcomes from multiple RCTs. Since testing under this scheme is predominantly focused on the criteria of generality and precision achieved through methodological rigor, at the cost of the criterion of realism, translating test results to clinical practice is often problematic. Choices concerning which methodological criteria should have priority are inevitable, however, as clinical trials, and scientific research in general, cannot meet all relevant criteria at the same time. Since these choices may be informed by considerations external to science, we must acknowledge that science cannot be value-free in a strict sense, and this invites a more prominent role for value-laden considerations in evaluating clinical research. The urgency for this becomes even more apparent when we consider the important yet implicit role of scientific theories in EBM, which may also be subjected to methodological evaluation and for which selectiveness in methodological focus is likewise inevitable.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Hugo, Jacques Victor; Gertman, David Ira
The new generation of nuclear power plants (NPPs) will likely make use of state-of-the-art technologies in many areas of the plant. The analysis, design, and selection of advanced human–system interfaces (HSIs) constitute an important part of power plant engineering. Designers need to consider the new capabilities afforded by these technologies in the context of current regulations and new operational concepts, which is why they need a more rigorous method by which to plan the introduction of advanced HSIs in NPP work areas. Much of current human factors research stops at the user interface and fails to provide a definitive processmore » for integration of end user devices with instrumentation and control (I&C) and operational concepts. The current lack of a clear definition of HSI technology, including the process for integration, makes characterization and implementation of new and advanced HSIs difficult. This paper describes how new design concepts in the nuclear industry can be analyzed and how HSI technologies associated with new industrial processes might be considered. Furthermore, it also describes a basis for an understanding of human as well as technology characteristics that could be incorporated into a prioritization scheme for technology selection and deployment plans.« less
Resistance of Grape Rootstocks to Plant-parasitic Nematodes.
Ferris, H; Zheng, L; Walker, M A
2012-12-01
Candidate grape rootstocks were selected through a rigorous screening program initiated with important sources of resistance to Meloidogyne pathotypes and to Xiphinema index in Muscadinia rotundifolia and Vitis species native to North America. Based on their rooting capability and horticultural characteristics, 200 candidates were selected from 5,000 progeny of multiple crosses between commercial grape rootstocks and wild grape species that exhibited resistance to nematodes. After a 15-year screening process, 13 selections emerged with either almost complete or complete combined resistance to M. incognita Race 3, M. incognita pathotype Harmony C, M. arenaria pathotype Harmony A, and X. index, important nematode pests of grapevines. Durability of this broad resistance was tested by challenging the selections with the target nematodes in combination and with the target nematodes in combinations with species not included in the screening process. Durability of resistance of the candidate rootstocks was also tested by exposure to the nematode communities of infested field soils from different locations. Breadth of resistance was determined on the basis of their host status to non-target nematodes, including Mesocriconema xenoplax, Pratylenchus vulnus, Tylenchulus semipenetrans and Paratylenchus hamatus. After a total of 204 separate trials, the rootstocks were released to the grape industry as UCD GRN1, UCD GRN2, UCD GRN3, UCD GRN4, and UCD GRN5. We provide a compilation of current knowledge of the host status of these five newly released rootstocks and of 27 other rootstock cultivars to plant-parasitic nematodes.
Resistance of Grape Rootstocks to Plant-parasitic Nematodes
Ferris, H.; Zheng, L.; Walker, M. A.
2012-01-01
Candidate grape rootstocks were selected through a rigorous screening program initiated with important sources of resistance to Meloidogyne pathotypes and to Xiphinema index in Muscadinia rotundifolia and Vitis species native to North America. Based on their rooting capability and horticultural characteristics, 200 candidates were selected from 5,000 progeny of multiple crosses between commercial grape rootstocks and wild grape species that exhibited resistance to nematodes. After a 15-year screening process, 13 selections emerged with either almost complete or complete combined resistance to M. incognita Race 3, M. incognita pathotype Harmony C, M. arenaria pathotype Harmony A, and X. index, important nematode pests of grapevines. Durability of this broad resistance was tested by challenging the selections with the target nematodes in combination and with the target nematodes in combinations with species not included in the screening process. Durability of resistance of the candidate rootstocks was also tested by exposure to the nematode communities of infested field soils from different locations. Breadth of resistance was determined on the basis of their host status to non-target nematodes, including Mesocriconema xenoplax, Pratylenchus vulnus, Tylenchulus semipenetrans and Paratylenchus hamatus. After a total of 204 separate trials, the rootstocks were released to the grape industry as UCD GRN1, UCD GRN2, UCD GRN3, UCD GRN4, and UCD GRN5. We provide a compilation of current knowledge of the host status of these five newly released rootstocks and of 27 other rootstock cultivars to plant-parasitic nematodes. PMID:23482972
Autism and Pervasive Developmental Disorders
ERIC Educational Resources Information Center
Volkmar, Fred R.; Lord, Catherine; Bailey, Anthony; Schultz, Robert T.; Klin, Ami
2004-01-01
The quantity and quality of research into autism and related conditions have increased dramatically in recent years. Consequently we selectively review key accomplishments and highlight directions for future research. More consistent approaches to diagnosis and more rigorous assessment methods have significantly advanced research, although the…
Injection-salting of pre rigor fillets of Atlantic salmon (Salmo salar).
Birkeland, Sveinung; Akse, Leif; Joensen, Sjurdur; Tobiassen, Torbjørn; Skåra, Torstein
2007-01-01
The effects of temperature (-1, 4, and 10 degrees C), brine concentration (12% and 25% NaCl), injection volumes, and needle densities were investigated on fillet weight gain (%), salt content (%), fillet contraction (%), and muscle gaping in pre rigor brine-injected fillets of Atlantic salmon (Salmo salar). Increased brine concentration (12% to 25%) significantly increased the initial (< 5 min after injection) and final contraction (24 h after injection) of pre rigor fillets. Increased brine concentration significantly reduced weight gain and increased salt content but had no significant effect on muscle gaping. The temperatures tested did not significantly affect weight gain, fillet contraction, or gaping score. Significant regressions (P < 0.01) between the injection volume and weight gain (range: 2.5% to 15.5%) and salt content (range: 1.7% to 6.5%) were observed for injections of pre rigor fillets. Double injections significantly increased the weight gain and salt content compared to single injections. Initial fillet contraction measured 30 min after brine injection increased significantly (P < 0.01) with increasing brine injection volume but no significant difference in the fillet contraction was observed 12 h after brine injection (range: 7.9% to 8.9%). Brine-injected post rigor control fillets obtained higher weight gain, higher salt content, more muscle gaping, and significantly lower fillet contraction compared to the pre rigor injected fillets. Injection-salting is an applicable technology as a means to obtain satisfactory salt contents and homogenously distribute the salt into the muscle of pre rigor fillets of Atlantic salmon before further processing steps such as drying and smoking.
Social selection and peer influence in an online social network.
Lewis, Kevin; Gonzalez, Marco; Kaufman, Jason
2012-01-03
Disentangling the effects of selection and influence is one of social science's greatest unsolved puzzles: Do people befriend others who are similar to them, or do they become more similar to their friends over time? Recent advances in stochastic actor-based modeling, combined with self-reported data on a popular online social network site, allow us to address this question with a greater degree of precision than has heretofore been possible. Using data on the Facebook activity of a cohort of college students over 4 years, we find that students who share certain tastes in music and in movies, but not in books, are significantly likely to befriend one another. Meanwhile, we find little evidence for the diffusion of tastes among Facebook friends-except for tastes in classical/jazz music. These findings shed light on the mechanisms responsible for observed network homogeneity; provide a statistically rigorous assessment of the coevolution of cultural tastes and social relationships; and suggest important qualifications to our understanding of both homophily and contagion as generic social processes.
Yang, Yuting; Gourinath, S; Kovács, Mihály; Nyitray, László; Reutzel, Robbie; Himmel, Daniel M; O'Neall-Hennessey, Elizabeth; Reshetnikova, Ludmilla; Szent-Györgyi, Andrew G; Brown, Jerry H; Cohen, Carolyn
2007-05-01
Unlike processive cellular motors such as myosin V, whose structure has recently been determined in a "rigor-like" conformation, myosin II from contracting muscle filaments necessarily spends most of its time detached from actin. By using squid and sea scallop sources, however, we have now obtained similar rigor-like atomic structures for muscle myosin heads (S1). The significance of the hallmark closed actin-binding cleft in these crystal structures is supported here by actin/S1-binding studies. These structures reveal how different duty ratios, and hence cellular functions, of the myosin isoforms may be accounted for, in part, on the basis of detailed differences in interdomain contacts. Moreover, the rigor-like position of switch II turns out to be unique for myosin V. The overall arrangements of subdomains in the motor are relatively conserved in each of the known contractile states, and we explore qualitatively the energetics of these states.
NASA Technical Reports Server (NTRS)
Tanveer, S.; Foster, M. R.
2002-01-01
We report progress in three areas of investigation related to dendritic crystal growth. Those items include: 1. Selection of tip features dendritic crystal growth; 2) Investigation of nonlinear evolution for two-sided model; and 3) Rigorous mathematical justification.
Smartphone apps for the self-management of low back pain: A systematic review.
Machado, Gustavo C; Pinheiro, Marina B; Lee, Hopin; Ahmed, Osman H; Hendrick, Paul; Williams, Chris; Kamper, Steven J
2016-12-01
Guidelines for low back pain (LBP) often recommend the use of self-management such as unsupervised exercise, booklets, and online education. Another potentially useful way for patients to self-manage LBP is by using smartphone applications (apps). However, to date, there has been no rigorous evaluation of LBP apps and no guidance for consumers on how to select high-quality, evidence-based apps. This chapter reviews smartphone apps for the self-management of LBP and evaluates their content quality and whether they recommend evidence-based interventions. This chapter shows that generally app developers are selecting interventions that are endorsed by guidelines, although their quality is low. There are many apps available for the self-management of LBP, but their effectiveness in improving patient outcomes has not been rigorously assessed. App developers need to work closely with healthcare professionals, researchers, and patients to ensure app content is accurate, evidence based, and engaging. Copyright © 2017 Elsevier Ltd. All rights reserved.
Ridenour, Ty A; Pineo, Thomas Z; Maldonado Molina, Mildred M; Hassmiller Lich, Kristen
2013-06-01
Psychosocial prevention research lacks evidence from intensive within-person lines of research to understand idiographic processes related to development and response to intervention. Such data could be used to fill gaps in the literature and expand the study design options for prevention researchers, including lower-cost yet rigorous studies (e.g., for program evaluations), pilot studies, designs to test programs for low prevalence outcomes, selective/indicated/adaptive intervention research, and understanding of differential response to programs. This study compared three competing analytic strategies designed for this type of research: autoregressive moving average, mixed model trajectory analysis, and P-technique. Illustrative time series data were from a pilot study of an intervention for nursing home residents with diabetes (N = 4) designed to improve control of blood glucose. A within-person, intermittent baseline design was used. Intervention effects were detected using each strategy for the aggregated sample and for individual patients. The P-technique model most closely replicated observed glucose levels. ARIMA and P-technique models were most similar in terms of estimated intervention effects and modeled glucose levels. However, ARIMA and P-technique also were more sensitive to missing data, outliers and number of observations. Statistical testing suggested that results generalize both to other persons as well as to idiographic, longitudinal processes. This study demonstrated the potential contributions of idiographic research in prevention science as well as the need for simulation studies to delineate the research circumstances when each analytic approach is optimal for deriving the correct parameter estimates.
Pineo, Thomas Z.; Maldonado Molina, Mildred M.; Lich, Kristen Hassmiller
2013-01-01
Psychosocial prevention research lacks evidence from intensive within-person lines of research to understand idiographic processes related to development and response to intervention. Such data could be used to fill gaps in the literature and expand the study design options for prevention researchers, including lower-cost yet rigorous studies (e.g., for program evaluations), pilot studies, designs to test programs for low prevalence outcomes, selective/indicated/ adaptive intervention research, and understanding of differential response to programs. This study compared three competing analytic strategies designed for this type of research: autoregressive moving average, mixed model trajectory analysis, and P-technique. Illustrative time series data were from a pilot study of an intervention for nursing home residents with diabetes (N=4) designed to improve control of blood glucose. A within-person, intermittent baseline design was used. Intervention effects were detected using each strategy for the aggregated sample and for individual patients. The P-technique model most closely replicated observed glucose levels. ARIMA and P-technique models were most similar in terms of estimated intervention effects and modeled glucose levels. However, ARIMA and P-technique also were more sensitive to missing data, outliers and number of observations. Statistical testing suggested that results generalize both to other persons as well as to idiographic, longitudinal processes. This study demonstrated the potential contributions of idiographic research in prevention science as well as the need for simulation studies to delineate the research circumstances when each analytic approach is optimal for deriving the correct parameter estimates. PMID:23299558
Zeni, Mary Beth
2012-03-01
The purpose of this study was to evaluate if paediatric asthma educational intervention studies included in the Cochrane Collaboration database incorporated concepts of health literacy. Inclusion criteria were established to identify review categories in the Cochrane Collaboration database specific to paediatric asthma educational interventions. Articles that met the inclusion criteria were selected from the Cochrane Collaboration database in 2010. The health literacy definition from Healthy People 2010 was used to develop a 4-point a priori rating scale to determine the extent a study reported aspects of health literacy in the development of an educational intervention for parents and/or children. Five Cochrane review categories met the inclusion criteria; 75 studies were rated for health literacy content regarding educational interventions with families and children living with asthma. A priori criteria were used for the rating process. While 52 (69%) studies had no information pertaining to health literacy, 23 (31%) reported an aspect of health literacy. Although all studies maintained the rigorous standards of randomized clinical trials, a model of health literacy was not reported regarding the design and implementation of interventions. While a more comprehensive health literacy model for the development of educational interventions with families and children may have been available after the reviewed studies were conducted, general literacy levels still could have been addressed. The findings indicate a need to incorporate health literacy in the design of client-centred educational interventions and in the selection criteria of relevant Cochrane reviews. Inclusion assures that health literacy is as important as randomization and statistical analyses in the research design of educational interventions and may even assure participation of people with literacy challenges. © 2012 The Author. International Journal of Evidence-Based Healthcare © 2012 The Joanna Briggs Institute.
Araújo, Luciano V; Malkowski, Simon; Braghetto, Kelly R; Passos-Bueno, Maria R; Zatz, Mayana; Pu, Calton; Ferreira, João E
2011-12-22
Recent medical and biological technology advances have stimulated the development of new testing systems that have been providing huge, varied amounts of molecular and clinical data. Growing data volumes pose significant challenges for information processing systems in research centers. Additionally, the routines of genomics laboratory are typically characterized by high parallelism in testing and constant procedure changes. This paper describes a formal approach to address this challenge through the implementation of a genetic testing management system applied to human genome laboratory. We introduced the Human Genome Research Center Information System (CEGH) in Brazil, a system that is able to support constant changes in human genome testing and can provide patients updated results based on the most recent and validated genetic knowledge. Our approach uses a common repository for process planning to ensure reusability, specification, instantiation, monitoring, and execution of processes, which are defined using a relational database and rigorous control flow specifications based on process algebra (ACP). The main difference between our approach and related works is that we were able to join two important aspects: 1) process scalability achieved through relational database implementation, and 2) correctness of processes using process algebra. Furthermore, the software allows end users to define genetic testing without requiring any knowledge about business process notation or process algebra. This paper presents the CEGH information system that is a Laboratory Information Management System (LIMS) based on a formal framework to support genetic testing management for Mendelian disorder studies. We have proved the feasibility and showed usability benefits of a rigorous approach that is able to specify, validate, and perform genetic testing using easy end user interfaces.
2011-01-01
Background Recent medical and biological technology advances have stimulated the development of new testing systems that have been providing huge, varied amounts of molecular and clinical data. Growing data volumes pose significant challenges for information processing systems in research centers. Additionally, the routines of genomics laboratory are typically characterized by high parallelism in testing and constant procedure changes. Results This paper describes a formal approach to address this challenge through the implementation of a genetic testing management system applied to human genome laboratory. We introduced the Human Genome Research Center Information System (CEGH) in Brazil, a system that is able to support constant changes in human genome testing and can provide patients updated results based on the most recent and validated genetic knowledge. Our approach uses a common repository for process planning to ensure reusability, specification, instantiation, monitoring, and execution of processes, which are defined using a relational database and rigorous control flow specifications based on process algebra (ACP). The main difference between our approach and related works is that we were able to join two important aspects: 1) process scalability achieved through relational database implementation, and 2) correctness of processes using process algebra. Furthermore, the software allows end users to define genetic testing without requiring any knowledge about business process notation or process algebra. Conclusions This paper presents the CEGH information system that is a Laboratory Information Management System (LIMS) based on a formal framework to support genetic testing management for Mendelian disorder studies. We have proved the feasibility and showed usability benefits of a rigorous approach that is able to specify, validate, and perform genetic testing using easy end user interfaces. PMID:22369688
Near Identifiability of Dynamical Systems
NASA Technical Reports Server (NTRS)
Hadaegh, F. Y.; Bekey, G. A.
1987-01-01
Concepts regarding approximate mathematical models treated rigorously. Paper presents new results in analysis of structural identifiability, equivalence, and near equivalence between mathematical models and physical processes they represent. Helps establish rigorous mathematical basis for concepts related to structural identifiability and equivalence revealing fundamental requirements, tacit assumptions, and sources of error. "Structural identifiability," as used by workers in this field, loosely translates as meaning ability to specify unique mathematical model and set of model parameters that accurately predict behavior of corresponding physical system.
Cavitt, L C; Sams, A R
2003-07-01
Studies were conducted to develop a non-destructive method for monitoring the rate of rigor mortis development in poultry and to evaluate the effectiveness of electrical stimulation (ES). In the first study, 36 male broilers in each of two trials were processed at 7 wk of age. After being bled, half of the birds received electrical stimulation (400 to 450 V, 400 to 450 mA, for seven pulses of 2 s on and 1 s off), and the other half were designated as controls. At 0.25 and 1.5 h postmortem (PM), carcasses were evaluated for the angles of the shoulder, elbow, and wing tip and the distance between the elbows. Breast fillets were harvested at 1.5 h PM (after chilling) from all carcasses. Fillet samples were excised and frozen for later measurement of pH and R-value, and the remainder of each fillet was held on ice until 24 h postmortem. Shear value and pH means were significantly lower, but R-value means were higher (P < 0.05) for the ES fillets compared to the controls, suggesting acceleration of rigor mortis by ES. The physical dimensions of the shoulder and elbow changed (P < 0.05) during rigor mortis development and with ES. These results indicate that physical measurements of the wings maybe useful as a nondestructive indicator of rigor development and for monitoring the effectiveness of ES. In the second study, 60 male broilers in each of two trials were processed at 7 wk of age. At 0.25, 1.5, 3.0, and 6.0 h PM, carcasses were evaluated for the distance between the elbows. At each time point, breast fillets were harvested from each carcass. Fillet samples were excised and frozen for later measurement of pH and sacromere length, whereas the remainder of each fillet was held on ice until 24 h PM. Shear value and pH means (P < 0.05) decreased, whereas sarcomere length means (P < 0.05) increased over time, indicating rigor mortis development. Elbow distance decreased (P < 0.05) with rigor development and was correlated (P < 0.01) with shear value (r = 0.2581), sarcomere length (r = -0.3079), and pH (r = 0.6303). These results suggest that elbow distance could be used in conjunction with other detection methods for optically automating measurement of rigor mortis development in broiler carcasses.
Rigor + Results = Impact: Measuring Impact with Integrity (Invited)
NASA Astrophysics Data System (ADS)
Davis, H. B.; Scalice, D.
2013-12-01
Are you struggling to measure and explain the impact of your EPO efforts? The NASA Astrobiology Institute (NAI) is using an evaluation process to determine the impact of its 15 EPO projects with over 200 activities. What is the current impact? How can it be improved in the future? We have developed a process that preserves autonomy at the project implementation level while still painting a picture of the entire portfolio. The impact evaluation process looks at an education/public outreach activity through its entire project cycle. Working with an external evaluator, education leads: 1) rate the quality/health of an activity in each stage of its cycle, and 2) determine the impact based on the results of the evaluation and the rigor of the methods used. The process has created a way to systematically codify a project's health and its impact, while offering support for improving both impact and how it is measured.
McCaig, Chris; Begon, Mike; Norman, Rachel; Shankland, Carron
2011-03-01
Changing scale, for example, the ability to move seamlessly from an individual-based model to a population-based model, is an important problem in many fields. In this paper, we introduce process algebra as a novel solution to this problem in the context of models of infectious disease spread. Process algebra allows us to describe a system in terms of the stochastic behaviour of individuals, and is a technique from computer science. We review the use of process algebra in biological systems, and the variety of quantitative and qualitative analysis techniques available. The analysis illustrated here solves the changing scale problem: from the individual behaviour we can rigorously derive equations to describe the mean behaviour of the system at the level of the population. The biological problem investigated is the transmission of infection, and how this relates to individual interactions.
NASA Astrophysics Data System (ADS)
Katsoulakis, Markos A.; Vlachos, Dionisios G.
2003-11-01
We derive a hierarchy of successively coarse-grained stochastic processes and associated coarse-grained Monte Carlo (CGMC) algorithms directly from the microscopic processes as approximations in larger length scales for the case of diffusion of interacting particles on a lattice. This hierarchy of models spans length scales between microscopic and mesoscopic, satisfies a detailed balance, and gives self-consistent fluctuation mechanisms whose noise is asymptotically identical to the microscopic MC. Rigorous, detailed asymptotics justify and clarify these connections. Gradient continuous time microscopic MC and CGMC simulations are compared under far from equilibrium conditions to illustrate the validity of our theory and delineate the errors obtained by rigorous asymptotics. Information theory estimates are employed for the first time to provide rigorous error estimates between the solutions of microscopic MC and CGMC, describing the loss of information during the coarse-graining process. Simulations under periodic boundary conditions are used to verify the information theory error estimates. It is shown that coarse-graining in space leads also to coarse-graining in time by q2, where q is the level of coarse-graining, and overcomes in part the hydrodynamic slowdown. Operation counting and CGMC simulations demonstrate significant CPU savings in continuous time MC simulations that vary from q3 for short potentials to q4 for long potentials. Finally, connections of the new coarse-grained stochastic processes to stochastic mesoscopic and Cahn-Hilliard-Cook models are made.
NASA Astrophysics Data System (ADS)
Cucinotta, Francis A.; Hamada, Nobuyuki; Little, Mark P.
2016-08-01
Previous analysis has shown that astronauts have a significantly lower standardized mortality ratio for circulatory disease mortality compared to the U.S. population, which is consistent with the rigorous selection process and healthy lifestyles of astronauts, and modest space radiation exposures from past space missions. However, a recent report by Delp et al. estimated the proportional mortality ratio for ages of 55-64 y of Apollo lunar mission astronauts to claim a high risk of cardiovascular disease due to space radiation compared to the U.S. population or to non-flight astronauts. In this Commentary we discuss important deficiencies in the methods and assumptions on radiation exposures used by Delp et al. that we judge cast serious doubt on their conclusions.
Towards a Credibility Assessment of Models and Simulations
NASA Technical Reports Server (NTRS)
Blattnig, Steve R.; Green, Lawrence L.; Luckring, James M.; Morrison, Joseph H.; Tripathi, Ram K.; Zang, Thomas A.
2008-01-01
A scale is presented to evaluate the rigor of modeling and simulation (M&S) practices for the purpose of supporting a credibility assessment of the M&S results. The scale distinguishes required and achieved levels of rigor for a set of M&S elements that contribute to credibility including both technical and process measures. The work has its origins in an interest within NASA to include a Credibility Assessment Scale in development of a NASA standard for models and simulations.
Evidence on the Returns to Secondary Vocational Education
ERIC Educational Resources Information Center
Meer, Jonathan
2007-01-01
Vocational education in high schools has frequently been stigmatized as an anachronistic, dead-end path for students. We use data from the National Education Longitudinal Survey of 1988 to examine claims that students on a vocational track would benefit from a more academically rigorous education. Clearly, selection bias confounds attempts to…
A Practical Guide to Regression Discontinuity
ERIC Educational Resources Information Center
Jacob, Robin; Zhu, Pei; Somers, Marie-Andrée; Bloom, Howard
2012-01-01
Regression discontinuity (RD) analysis is a rigorous nonexperimental approach that can be used to estimate program impacts in situations in which candidates are selected for treatment based on whether their value for a numeric rating exceeds a designated threshold or cut-point. Over the last two decades, the regression discontinuity approach has…
Federal Register 2010, 2011, 2012, 2013, 2014
2013-02-12
... that can demonstrate impact through other methodological approaches such as a quasi-experimental design... definition of ``relevant outcome.'' Lastly, quasi-experimental designs are already included in the definition... paragraph (b) of this definition, provided they are rigorous and comparable across schools. (b) For non...
Federal Register 2010, 2011, 2012, 2013, 2014
2010-06-23
... academic achievement in large public high schools. These smaller units include freshman academies, multi... academic supports to help students succeed in rigorous academic courses; (4) Deliver comprehensive guidance and academic advising to students and their parents that includes assistance in selecting courses and...
Theoretical rationale for music selection in oncology intervention research: an integrative review.
Burns, Debra S
2012-01-01
Music-based interventions have helped patients with cancer improve their quality of life, decrease treatment related distress, and manage pain. However, quantitative findings from music intervention studies are inconsistent. The purpose of this review was to explore the theoretical underpinnings for the selection of the music stimuli used to influence targeted outcomes. It was hypothesized that disparate findings were due in part to the atheoretical nature of music selection and the resulting diversity in music stimuli between and within studies. A systematic research synthesis including a comprehensive database and reference list search resulted in 22 studies. Included studies were compiled into two tables cataloging intervention theory, intervention content, and outcomes. A majority of studies did not provide a rationale or intervention theory for the delivery of music or choice of outcomes. Recorded music was the most common delivery method, but the specific music was rarely included within the report. Only two studies that included a theoretical framework reported null results on at least some of the outcomes. Null results are partially explained by an incomplete or mismatch in intervention theory and music selection and delivery. While the inclusion of an intervention theory does not guarantee positive results, including a theoretical rationale for the use of music, particular therapeutic processes or mechanisms, and the specifics of how music is selected and delivered increases scientific rigor and the probability of clinical translation.
Inferring fitness landscapes and selection on phenotypic states from single-cell genealogical data
Kussell, Edo
2017-01-01
Recent advances in single-cell time-lapse microscopy have revealed non-genetic heterogeneity and temporal fluctuations of cellular phenotypes. While different phenotypic traits such as abundance of growth-related proteins in single cells may have differential effects on the reproductive success of cells, rigorous experimental quantification of this process has remained elusive due to the complexity of single cell physiology within the context of a proliferating population. We introduce and apply a practical empirical method to quantify the fitness landscapes of arbitrary phenotypic traits, using genealogical data in the form of population lineage trees which can include phenotypic data of various kinds. Our inference methodology for fitness landscapes determines how reproductivity is correlated to cellular phenotypes, and provides a natural generalization of bulk growth rate measures for single-cell histories. Using this technique, we quantify the strength of selection acting on different cellular phenotypic traits within populations, which allows us to determine whether a change in population growth is caused by individual cells’ response, selection within a population, or by a mixture of these two processes. By applying these methods to single-cell time-lapse data of growing bacterial populations that express a resistance-conferring protein under antibiotic stress, we show how the distributions, fitness landscapes, and selection strength of single-cell phenotypes are affected by the drug. Our work provides a unified and practical framework for quantitative measurements of fitness landscapes and selection strength for any statistical quantities definable on lineages, and thus elucidates the adaptive significance of phenotypic states in time series data. The method is applicable in diverse fields, from single cell biology to stem cell differentiation and viral evolution. PMID:28267748
An ArcGIS decision support tool for artificial reefs site selection (ArcGIS ARSS)
NASA Astrophysics Data System (ADS)
Stylianou, Stavros; Zodiatis, George
2017-04-01
Although the use and benefits of artificial reefs, both socio-economic and environmental, have been recognized with research and national development programmes worldwide their development is rarely subjected to a rigorous site selection process and the majority of the projects use the traditional (non-GIS) approach, based on trial and error mode. Recent studies have shown that the use of Geographic Information Systems, unlike to traditional methods, for the identification of suitable areas for artificial reefs siting seems to offer a number of distinct advantages minimizing possible errors, time and cost. A decision support tool (DSS) has been developed based on the existing knowledge, the multi-criteria decision analysis techniques and the GIS approach used in previous studies in order to help the stakeholders to identify the optimal locations for artificial reefs deployment on the basis of the physical, biological, oceanographic and socio-economic features of the sites. The tool provides to the users the ability to produce a final report with the results and suitability maps. The ArcGIS ARSS support tool runs within the existing ArcMap 10.2.x environment and for the development the VB .NET high level programming language has been used along with ArcObjects 10.2.x. Two local-scale case studies were conducted in order to test the application of the tool focusing on artificial reef siting. The results obtained from the case studies have shown that the tool can be successfully integrated within the site selection process in order to select objectively the optimal site for artificial reefs deployment.
Orford, Jim
2008-06-01
To identify possible reasons for the disappointingly negative results of methodologically rigorous controlled trials of psychological treatments in the addictions field. A selective overview of the literature on addictive behaviour change. Eight failings of existing research are described: failing to account for the outcome equivalence paradox; neglecting relationships in favour of techniques; failing to integrate treatment research and research on unaided change; imposing an inappropriate time-scale on the change process; failing to take a systems or social network view; ignoring therapists' tacit theories; not including the patient's view; and displaying an ignorance of modern developments in the philosophy of science. Treatment research has been asking the wrong questions in the wrong way. Three necessary shifts in ways of conducting research are proposed: (i) the field should stop studying named techniques and focus instead on change processes; (ii) change processes should be studied within the broader, longer-acting systems of which treatment is part; and (iii) science in the field should be brought up to date by acknowledging a variety of sources of useful knowledge.
O’Suilleabhain, Padraig E.; Sanghera, Manjit; Patel, Neepa; Khemani, Pravin; Lacritz, Laura H.; Chitnis, Shilpa; Whitworth, Louis A.; Dewey, Richard B.
2016-01-01
Objective To develop a process to improve patient outcomes from deep brain stimulation (DBS) surgery for Parkinson disease (PD), essential tremor (ET), and dystonia. Methods We employed standard quality improvement methodology using the Plan-Do-Study-Act process to improve patient selection, surgical DBS lead implantation, postoperative programming, and ongoing assessment of patient outcomes. Results The result of this quality improvement process was the development of a neuromodulation network. The key aspect of this program is rigorous patient assessment of both motor and non-motor outcomes tracked longitudinally using a REDCap database. We describe how this information is used to identify problems and to initiate Plan-Do-Study-Act cycles to address them. Preliminary outcomes data is presented for the cohort of PD and ET patients who have received surgery since the creation of the neuromodulation network. Conclusions Careful outcomes tracking is essential to ensure quality in a complex therapeutic endeavor like DBS surgery for movement disorders. The REDCap database system is well suited to store outcomes data for the purpose of ongoing quality assurance monitoring. PMID:27711133
Dewey, Richard B; O'Suilleabhain, Padraig E; Sanghera, Manjit; Patel, Neepa; Khemani, Pravin; Lacritz, Laura H; Chitnis, Shilpa; Whitworth, Louis A; Dewey, Richard B
2016-01-01
To develop a process to improve patient outcomes from deep brain stimulation (DBS) surgery for Parkinson disease (PD), essential tremor (ET), and dystonia. We employed standard quality improvement methodology using the Plan-Do-Study-Act process to improve patient selection, surgical DBS lead implantation, postoperative programming, and ongoing assessment of patient outcomes. The result of this quality improvement process was the development of a neuromodulation network. The key aspect of this program is rigorous patient assessment of both motor and non-motor outcomes tracked longitudinally using a REDCap database. We describe how this information is used to identify problems and to initiate Plan-Do-Study-Act cycles to address them. Preliminary outcomes data is presented for the cohort of PD and ET patients who have received surgery since the creation of the neuromodulation network. Careful outcomes tracking is essential to ensure quality in a complex therapeutic endeavor like DBS surgery for movement disorders. The REDCap database system is well suited to store outcomes data for the purpose of ongoing quality assurance monitoring.
Miltner, M; Makaruk, A; Krischan, J; Harasek, M
2012-01-01
In the present work chemical-oxidative scrubbing as a novel method for the desulphurisation of raw biogas is presented with a special focus on the process potentials and economics. The selective absorption of hydrogen sulphide from gas streams containing high amounts of carbon dioxide using caustic solutions is not trivial but has been treated in literature. However, the application of this method to biogas desulphurisation has not been established so far. Based on rigorous experimental work, an industrial-scale pilot plant has been designed, erected and commissioned at a biogas plant with biogas upgrading and gas grid injection in Austria. Data collected from the 12-month monitored operation has been used to elaborate performance as well as economic parameters for the novel desulphurisation method. The proposed technology offers significant operational advantages regarding the degree of automation and the flexibility towards fluctuations in process boundary conditions. Furthermore, the economic assessment revealed the high competitiveness of the chemical-oxidative scrubbing process compared with other desulphurisation technologies with the named advantageous operational behaviour.
Stakeholder-focused evaluation of an online course for health care providers.
Dunet, Diane O; Reyes, Michele
2006-01-01
Different people who have a stake or interest in a training course (stakeholders) may have markedly different definitions of what constitutes "training success" and how they will use evaluation results. Stakeholders at multiple levels within and outside of the organization guided the development of an evaluation plan for a Web-based training course on hemochromatosis. Stakeholder interests and values were reflected in the type, level, and rigor of evaluation methods selected. Our mixed-method evaluation design emphasized small sample sizes and repeated measures. Limited resources for evaluation were leveraged by focusing on the data needs of key stakeholders, understanding how they wanted to use evaluation results, and collecting data needed for stakeholder decision making. Regular feedback to key stakeholders provided opportunities for updating the course evaluation plan to meet emerging needs for new or different information. Early and repeated involvement of stakeholders in the evaluation process also helped build support for the final product. Involving patient advocacy groups, managers, and representative course participants improved the course and enhanced product dissemination. For training courses, evaluation planning is an opportunity to tailor methods and data collection to meet the information needs of particular stakeholders. Rigorous evaluation research of every training course may be infeasible or unwarranted; however, course evaluations can be improved by good planning. A stakeholder-focused approach can build a picture of the results and impact of training while fostering the practical use of evaluation data.
Tsyshevsky, Roman V; Sharia, Onise; Kuklja, Maija M
2016-02-19
This review presents a concept, which assumes that thermal decomposition processes play a major role in defining the sensitivity of organic energetic materials to detonation initiation. As a science and engineering community we are still far away from having a comprehensive molecular detonation initiation theory in a widely agreed upon form. However, recent advances in experimental and theoretical methods allow for a constructive and rigorous approach to design and test the theory or at least some of its fundamental building blocks. In this review, we analyzed a set of select experimental and theoretical articles, which were augmented by our own first principles modeling and simulations, to reveal new trends in energetic materials and to refine known existing correlations between their structures, properties, and functions. Our consideration is intentionally limited to the processes of thermally stimulated chemical reactions at the earliest stage of decomposition of molecules and materials containing defects.
Tsyshevsky, Roman; Sharia, Onise; Kuklja, Maija
2016-02-19
Our review presents a concept, which assumes that thermal decomposition processes play a major role in defining the sensitivity of organic energetic materials to detonation initiation. As a science and engineering community we are still far away from having a comprehensive molecular detonation initiation theory in a widely agreed upon form. However, recent advances in experimental and theoretical methods allow for a constructive and rigorous approach to design and test the theory or at least some of its fundamental building blocks. In this review, we analyzed a set of select experimental and theoretical articles, which were augmented by our ownmore » first principles modeling and simulations, to reveal new trends in energetic materials and to refine known existing correlations between their structures, properties, and functions. Lastly, our consideration is intentionally limited to the processes of thermally stimulated chemical reactions at the earliest stage of decomposition of molecules and materials containing defects.« less
Statistical Model Selection for TID Hardness Assurance
NASA Technical Reports Server (NTRS)
Ladbury, R.; Gorelick, J. L.; McClure, S.
2010-01-01
Radiation Hardness Assurance (RHA) methodologies against Total Ionizing Dose (TID) degradation impose rigorous statistical treatments for data from a part's Radiation Lot Acceptance Test (RLAT) and/or its historical performance. However, no similar methods exist for using "similarity" data - that is, data for similar parts fabricated in the same process as the part under qualification. This is despite the greater difficulty and potential risk in interpreting of similarity data. In this work, we develop methods to disentangle part-to-part, lot-to-lot and part-type-to-part-type variation. The methods we develop apply not just for qualification decisions, but also for quality control and detection of process changes and other "out-of-family" behavior. We begin by discussing the data used in ·the study and the challenges of developing a statistic providing a meaningful measure of degradation across multiple part types, each with its own performance specifications. We then develop analysis techniques and apply them to the different data sets.
The MIXED framework: A novel approach to evaluating mixed-methods rigor.
Eckhardt, Ann L; DeVon, Holli A
2017-10-01
Evaluation of rigor in mixed-methods (MM) research is a persistent challenge due to the combination of inconsistent philosophical paradigms, the use of multiple research methods which require different skill sets, and the need to combine research at different points in the research process. Researchers have proposed a variety of ways to thoroughly evaluate MM research, but each method fails to provide a framework that is useful for the consumer of research. In contrast, the MIXED framework is meant to bridge the gap between an academic exercise and practical assessment of a published work. The MIXED framework (methods, inference, expertise, evaluation, and design) borrows from previously published frameworks to create a useful tool for the evaluation of a published study. The MIXED framework uses an experimental eight-item scale that allows for comprehensive integrated assessment of MM rigor in published manuscripts. Mixed methods are becoming increasingly prevalent in nursing and healthcare research requiring researchers and consumers to address issues unique to MM such as evaluation of rigor. © 2017 John Wiley & Sons Ltd.
2014-01-01
Background Discrete choice experiments (DCEs) are attribute-driven experimental techniques used to elicit stakeholders’ preferences to support the design and implementation of policy interventions. The validity of a DCE, therefore, depends on the appropriate specification of the attributes and their levels. There have been recent calls for greater rigor in implementing and reporting on the processes of developing attributes and attribute-levels for discrete choice experiments (DCEs). This paper responds to such calls by carefully reporting a systematic process of developing micro health insurance attributes and attribute-levels for the design of a DCE in rural Malawi. Methods Conceptual attributes and attribute-levels were initially derived from a literature review which informed the design of qualitative data collection tools to identify context specific attributes and attribute-levels. Qualitative data was collected in August-September 2012 from 12 focus group discussions with community residents and 8 in-depth interviews with health workers. All participants were selected according to stratified purposive sampling. The material was tape-recorded, fully transcribed, and coded by three researchers to identify context-specific attributes and attribute-levels. Expert opinion was used to scale down the attributes and levels. A pilot study confirmed the appropriateness of the selected attributes and levels for a DCE. Results First, a consensus, emerging from an individual level analysis of the qualitative transcripts, identified 10 candidate attributes. Levels were assigned to all attributes based on data from transcripts and knowledge of the Malawian context, derived from literature. Second, through further discussions with experts, four attributes were discarded based on multiple criteria. The 6 remaining attributes were: premium level, unit of enrollment, management structure, health service benefit package, transportation coverage and copayment levels. A final step of revision and piloting confirmed that the retained attributes satisfied the credibility criteria of DCE attributes. Conclusion This detailed description makes our attribute development process transparent, and provides the reader with a basis to assess the rigor of this stage of constructing the DCE. This paper contributes empirical evidence to the limited methodological literature on attributes and levels development for DCE, thereby providing further empirical guidance on the matter, specifically within rural communities of low- and middle-income countries. PMID:24884920
On Large Time Behavior and Selection Principle for a Diffusive Carr-Penrose Model
NASA Astrophysics Data System (ADS)
Conlon, Joseph G.; Dabkowski, Michael; Wu, Jingchen
2016-04-01
This paper is concerned with the study of a diffusive perturbation of the linear LSW model introduced by Carr and Penrose. A main subject of interest is to understand how the presence of diffusion acts as a selection principle, which singles out a particular self-similar solution of the linear LSW model as determining the large time behavior of the diffusive model. A selection principle is rigorously proven for a model which is a semiclassical approximation to the diffusive model. Upper bounds on the rate of coarsening are also obtained for the full diffusive model.
Selection theory of free dendritic growth in a potential flow.
von Kurnatowski, Martin; Grillenbeck, Thomas; Kassner, Klaus
2013-04-01
The Kruskal-Segur approach to selection theory in diffusion-limited or Laplacian growth is extended via combination with the Zauderer decomposition scheme. This way nonlinear bulk equations become tractable. To demonstrate the method, we apply it to two-dimensional crystal growth in a potential flow. We omit the simplifying approximations used in a preliminary calculation for the same system [Fischaleck, Kassner, Europhys. Lett. 81, 54004 (2008)], thus exhibiting the capability of the method to extend mathematical rigor to more complex problems than hitherto accessible.
Why Open-Ended Survey Questions Are Unlikely to Support Rigorous Qualitative Insights.
LaDonna, Kori A; Taylor, Taryn; Lingard, Lorelei
2018-03-01
Health professions education researchers are increasingly relying on a combination of quantitative and qualitative research methods to explore complex questions in the field. This important and necessary development, however, creates new methodological challenges that can affect both the rigor of the research process and the quality of the findings. One example is "qualitatively" analyzing free-text responses to survey or assessment instrument questions. In this Invited Commentary, the authors explain why analysis of such responses rarely meets the bar for rigorous qualitative research. While the authors do not discount the potential for free-text responses to enhance quantitative findings or to inspire new research questions, they caution that these responses rarely produce data rich enough to generate robust, stand-alone insights. The authors consider exemplars from health professions education research and propose strategies for treating free-text responses appropriately.
Rigorous ILT optimization for advanced patterning and design-process co-optimization
NASA Astrophysics Data System (ADS)
Selinidis, Kosta; Kuechler, Bernd; Cai, Howard; Braam, Kyle; Hoppe, Wolfgang; Domnenko, Vitaly; Poonawala, Amyn; Xiao, Guangming
2018-03-01
Despite the large difficulties involved in extending 193i multiple patterning and the slow ramp of EUV lithography to full manufacturing readiness, the pace of development for new technology node variations has been accelerating. Multiple new variations of new and existing technology nodes have been introduced for a range of device applications; each variation with at least a few new process integration methods, layout constructs and/or design rules. This had led to a strong increase in the demand for predictive technology tools which can be used to quickly guide important patterning and design co-optimization decisions. In this paper, we introduce a novel hybrid predictive patterning method combining two patterning technologies which have each individually been widely used for process tuning, mask correction and process-design cooptimization. These technologies are rigorous lithography simulation and inverse lithography technology (ILT). Rigorous lithography simulation has been extensively used for process development/tuning, lithography tool user setup, photoresist hot-spot detection, photoresist-etch interaction analysis, lithography-TCAD interactions/sensitivities, source optimization and basic lithography design rule exploration. ILT has been extensively used in a range of lithographic areas including logic hot-spot fixing, memory layout correction, dense memory cell optimization, assist feature (AF) optimization, source optimization, complex patterning design rules and design-technology co-optimization (DTCO). The combined optimization capability of these two technologies will therefore have a wide range of useful applications. We investigate the benefits of the new functionality for a few of these advanced applications including correction for photoresist top loss and resist scumming hotspots.
Decontamination and disposal of PCB wastes.
Johnston, L E
1985-01-01
Decontamination and disposal processes for PCB wastes are reviewed. Processes are classed as incineration, chemical reaction or decontamination. Incineration technologies are not limited to the rigorous high temperature but include those where innovations in use of oxident, heat transfer and residue recycle are made. Chemical processes include the sodium processes, radiant energy processes and low temperature oxidations. Typical processing rates and associated costs are provided where possible. PMID:3928363
Mehl, Steffen W.; Hill, Mary C.
2013-01-01
This report documents the addition of ghost node Local Grid Refinement (LGR2) to MODFLOW-2005, the U.S. Geological Survey modular, transient, three-dimensional, finite-difference groundwater flow model. LGR2 provides the capability to simulate groundwater flow using multiple block-shaped higher-resolution local grids (a child model) within a coarser-grid parent model. LGR2 accomplishes this by iteratively coupling separate MODFLOW-2005 models such that heads and fluxes are balanced across the grid-refinement interface boundary. LGR2 can be used in two-and three-dimensional, steady-state and transient simulations and for simulations of confined and unconfined groundwater systems. Traditional one-way coupled telescopic mesh refinement methods can have large, often undetected, inconsistencies in heads and fluxes across the interface between two model grids. The iteratively coupled ghost-node method of LGR2 provides a more rigorous coupling in which the solution accuracy is controlled by convergence criteria defined by the user. In realistic problems, this can result in substantially more accurate solutions and require an increase in computer processing time. The rigorous coupling enables sensitivity analysis, parameter estimation, and uncertainty analysis that reflects conditions in both model grids. This report describes the method used by LGR2, evaluates accuracy and performance for two-and three-dimensional test cases, provides input instructions, and lists selected input and output files for an example problem. It also presents the Boundary Flow and Head (BFH2) Package, which allows the child and parent models to be simulated independently using the boundary conditions obtained through the iterative process of LGR2.
Mehl, Steffen W.; Hill, Mary C.
2006-01-01
This report documents the addition of shared node Local Grid Refinement (LGR) to MODFLOW-2005, the U.S. Geological Survey modular, transient, three-dimensional, finite-difference ground-water flow model. LGR provides the capability to simulate ground-water flow using one block-shaped higher-resolution local grid (a child model) within a coarser-grid parent model. LGR accomplishes this by iteratively coupling two separate MODFLOW-2005 models such that heads and fluxes are balanced across the shared interfacing boundary. LGR can be used in two-and three-dimensional, steady-state and transient simulations and for simulations of confined and unconfined ground-water systems. Traditional one-way coupled telescopic mesh refinement (TMR) methods can have large, often undetected, inconsistencies in heads and fluxes across the interface between two model grids. The iteratively coupled shared-node method of LGR provides a more rigorous coupling in which the solution accuracy is controlled by convergence criteria defined by the user. In realistic problems, this can result in substantially more accurate solutions and require an increase in computer processing time. The rigorous coupling enables sensitivity analysis, parameter estimation, and uncertainty analysis that reflects conditions in both model grids. This report describes the method used by LGR, evaluates LGR accuracy and performance for two- and three-dimensional test cases, provides input instructions, and lists selected input and output files for an example problem. It also presents the Boundary Flow and Head (BFH) Package, which allows the child and parent models to be simulated independently using the boundary conditions obtained through the iterative process of LGR.
Incentives, Selection, and Teacher Performance: Evidence from IMPACT
ERIC Educational Resources Information Center
Dee, Thomas S.; Wyckoff, James
2015-01-01
Teachers in the United States are compensated largely on the basis of fixed schedules that reward experience and credentials. However, there is a growing interest in whether performance-based incentives based on rigorous teacher evaluations can improve teacher retention and performance. The evidence available to date has been mixed at best. This…
76 FR 74076 - Notice of Random Assignment Study To Evaluate the YouthBuild Program; Final Notice
Federal Register 2010, 2011, 2012, 2013, 2014
2011-11-30
... Evaluate the YouthBuild Program; Final Notice AGENCY: Employment and Training Administration (ETA), Labor... rigorous, nationally-representative estimates of the net impacts of the YouthBuild program. The Department... study. In the sites randomly selected to participate in this evaluation, all applicants for YouthBuild...
Standardized Test Results: KEEP and Control Students. 1975-1976, Technical Report #69.
ERIC Educational Resources Information Center
Antill, Ellen; Speidel, Gisela E.
This report presents the results of various standardized measures administered to Kamehameha Early Education Program (KEEP) students and control students in the school year 1975-1976. In contrast to previous comparisons, KEEP employed more rigorous procedures for the selection of the control students and for the conditions of test administration.…
Neurocognitive Functioning in AD/HD, Predominantly Inattentive and Combined Subtypes
ERIC Educational Resources Information Center
Solanto, Mary V.; Gilbert, Sharone N.; Raj, Anu; Zhu, John; Pope-Boyd, Sa'brina; Stepak, Brenda; Vail, Lucia; Newcorn, Jeffrey H.
2007-01-01
The Predominantly Inattentive (PI) and Combined (CB) subtypes of AD/HD differ in cognitive tempo, age of onset, gender ratio, and comorbidity, yet a differentiating endophenotype has not been identified. The aim of this study was to test rigorously diagnosed PI, CB, and typical children on measures selected for their potential to reveal…
Mechanical properties of frog skeletal muscles in iodoacetic acid rigor.
Mulvany, M J
1975-01-01
1. Methods have been developed for describing the length: tension characteristics of frog skeletal muscles which go into rigor at 4 degrees C following iodoacetic acid poisoning either in the presence of Ca2+ (Ca-rigor) or its absence (Ca-free-rigor). 2. Such rigor muscles showed less resistance to slow stretch (slow rigor resistance) that to fast stretch (fast rigor resistance). The slow and fast rigor resistances of Ca-free-rigor muscles were much lower than those of Ca-rigor muscles. 3. The slow rigor resistance of Ca-rigor muscles was proportional to the amount of overlap between the contractile filaments present when the muscles were put into rigor. 4. Withdrawing Ca2+ from Ca-rigor muscles (induced-Ca-free rigor) reduced their slow and fast rigor resistances. Readdition of Ca2+ (but not Mg2+, Mn2+ or Sr2+) reversed the effect. 5. The slow and fast rigor resistances of Ca-rigor muscles (but not of Ca-free-rigor muscles) decreased with time. 6.The sarcomere structure of Ca-rigor and induced-Ca-free rigor muscles stretched by 0.2lo was destroyed in proportion to the amount of stretch, but the lengths of the remaining intact sarcomeres were essentially unchanged. This suggests that there had been a successive yielding of the weakeast sarcomeres. 7. The difference between the slow and fast rigor resistance and the effect of calcium on these resistances are discussed in relation to possible variations in the strength of crossbridges between the thick and thin filaments. Images Plate 1 Plate 2 PMID:1082023
Ensuring Effective Curriculum Approval Processes: A Guide for Local Senates
ERIC Educational Resources Information Center
Academic Senate for California Community Colleges, 2016
2016-01-01
Curriculum is the heart of the mission of every college. College curriculum approval processes have been established to ensure that rigorous, high quality curriculum is offered that meets the needs of students. While some concerns may exist regarding the effectiveness and efficiency of local curriculum processes, all participants in the process…
Human-rated Safety Certification of a High Voltage Robonaut Lithium-ion Battery
NASA Technical Reports Server (NTRS)
Jeevarajan, Judith; Yayathi, S.; Johnson, M.; Waligora, T.; Verdeyen, W.
2013-01-01
NASA's rigorous certification process is being followed for the R2 high voltage battery program for use of R2 on International Space Station (ISS). Rigorous development testing at appropriate levels to credible off-nominal conditions and review of test data led to design improvements for safety at the virtual cell, cartridge and battery levels. Tests were carried out at all levels to confirm that both hardware and software controls work. Stringent flight acceptance testing of the flight battery will be completed before launch for mission use on ISS.
Perspective: Optical measurement of feature dimensions and shapes by scatterometry
NASA Astrophysics Data System (ADS)
Diebold, Alain C.; Antonelli, Andy; Keller, Nick
2018-05-01
The use of optical scattering to measure feature shape and dimensions, scatterometry, is now routine during semiconductor manufacturing. Scatterometry iteratively improves an optical model structure using simulations that are compared to experimental data from an ellipsometer. These simulations are done using the rigorous coupled wave analysis for solving Maxwell's equations. In this article, we describe the Mueller matrix spectroscopic ellipsometry based scatterometry. Next, the rigorous coupled wave analysis for Maxwell's equations is presented. Following this, several example measurements are described as they apply to specific process steps in the fabrication of gate-all-around (GAA) transistor structures. First, simulations of measurement sensitivity for the inner spacer etch back step of horizontal GAA transistor processing are described. Next, the simulated metrology sensitivity for sacrificial (dummy) amorphous silicon etch back step of vertical GAA transistor processing is discussed. Finally, we present the application of plasmonically active test structures for improving the sensitivity of the measurement of metal linewidths.
Hardy, Micael; Zielonka, Jacek; Karoui, Hakim; Sikora, Adam; Michalski, Radosław; Podsiadły, Radosław; Lopez, Marcos; Vasquez-Vivar, Jeannette; Kalyanaraman, Balaraman; Ouari, Olivier
2018-05-20
Since the discovery of the superoxide dismutase enzyme, the generation and fate of short-lived oxidizing, nitrosating, nitrating, and halogenating species in biological systems has been of great interest. Despite the significance of reactive oxygen species (ROS) and reactive nitrogen species (RNS) in numerous diseases and intracellular signaling, the rigorous detection of ROS and RNS has remained a challenge. Recent Advances: Chemical characterization of the reactions of selected ROS and RNS with electron paramagnetic resonance (EPR) spin traps and fluorescent probes led to the establishment of species-specific products, which can be used for specific detection of several forms of ROS and RNS in cell-free systems and in cultured cells in vitro and in animals in vivo. Profiling oxidation products from the ROS and RNS probes provides a rigorous method for detection of those species in biological systems. Formation and detection of species-specific products from the probes enables accurate characterization of the oxidative environment in cells. Measurement of the total signal (fluorescence, chemiluminescence, etc.) intensity does not allow for identification of the ROS/RNS formed. It is critical to identify the products formed by using chromatographic or other rigorous techniques. Product analyses should be accompanied by monitoring of the intracellular probe level, another factor controlling the yield of the product(s) formed. More work is required to characterize the chemical reactivity of the ROS/RNS probes, and to develop new probes/detection approaches enabling real-time, selective monitoring of the specific products formed from the probes. Antioxid. Redox Signal. 28, 1416-1432.
Observational Research Rigor Alone Does Not Justify Causal Inference
Ejima, Keisuke; Li, Peng; Smith, Daniel L.; Nagy, Tim R.; Kadish, Inga; van Groen, Thomas; Dawson, John A.; Yang, Yongbin; Patki, Amit; Allison, David B.
2016-01-01
Background Differing opinions exist on whether associations obtained in observational studies can be reliable indicators of a causal effect if the observational study is sufficiently well controlled and executed. Materials and methods To test this, we conducted two animal observational studies that were rigorously controlled and executed beyond what is achieved in studies of humans. In study 1, we randomized 332 genetically identical C57BL/6J mice into three diet groups with differing food energy allotments and recorded individual self-selected daily energy intake and lifespan. In study 2, 60 male mice (CD1) were paired and divided into two groups for a 2-week feeding regimen. We evaluated the association between weight gain and food consumption. Within each pair, one animal was randomly assigned to an S group in which the animals had free access to food. The second paired animal (R group) was provided exactly the same diet that their S partner ate the day before. Results In study 1, across all three groups, we found a significant negative effect of energy intake on lifespan. However, we found a positive association between food intake and lifespan among the ad libitum feeding group: 29.99 (95% CI: 8.2 to 51.7) days per daily kcal. In study 2, we found a significant (P=0.003) group (randomized vs self-selected)-by-food consumption interaction effect on weight gain. Conclusions At least in nutrition research, associations derived from observational studies may not be reliable indicators of causal effects, even with the most rigorous study designs achievable. PMID:27711975
Bringing a transgenic crop to market: where compositional analysis fits.
Privalle, Laura S; Gillikin, Nancy; Wandelt, Christine
2013-09-04
In the process of developing a biotechnology product, thousands of genes and transformation events are evaluated to select the event that will be commercialized. The ideal event is identified on the basis of multiple characteristics including trait efficacy, the molecular characteristics of the insert, and agronomic performance. Once selected, the commercial event is subjected to a rigorous safety evaluation taking a multipronged approach including examination of the safety of the gene and gene product - the protein, plant performance, impact of cultivating the crop on the environment, agronomic performance, and equivalence of the crop/food to conventional crops/food - by compositional analysis. The compositional analysis is composed of a comparison of the nutrient and antinutrient composition of the crop containing the event, its parental line (variety), and other conventional lines (varieties). Different geographies have different requirements for the compositional analysis studies. Parameters that vary include the number of years (seasons) and locations (environments) to be evaluated, the appropriate comparator(s), analytes to be evaluated, and statistical analysis. Specific examples of compositional analysis results will be presented.
Social selection and peer influence in an online social network
Lewis, Kevin; Gonzalez, Marco; Kaufman, Jason
2012-01-01
Disentangling the effects of selection and influence is one of social science's greatest unsolved puzzles: Do people befriend others who are similar to them, or do they become more similar to their friends over time? Recent advances in stochastic actor-based modeling, combined with self-reported data on a popular online social network site, allow us to address this question with a greater degree of precision than has heretofore been possible. Using data on the Facebook activity of a cohort of college students over 4 years, we find that students who share certain tastes in music and in movies, but not in books, are significantly likely to befriend one another. Meanwhile, we find little evidence for the diffusion of tastes among Facebook friends—except for tastes in classical/jazz music. These findings shed light on the mechanisms responsible for observed network homogeneity; provide a statistically rigorous assessment of the coevolution of cultural tastes and social relationships; and suggest important qualifications to our understanding of both homophily and contagion as generic social processes. PMID:22184242
NASA Astrophysics Data System (ADS)
Hong, JaeSub; van den Berg, Maureen; Schlegel, Eric M.; Grindlay, Jonathan E.; Koenig, Xavier; Laycock, Silas; Zhao, Ping
2005-12-01
We describe the X-ray analysis procedure of the ongoing Chandra Multiwavelength Plane (ChaMPlane) Survey and report the initial results from the analysis of 15 selected anti-Galactic center observations (90deg
Analyzing Single-Molecule Time Series via Nonparametric Bayesian Inference
Hines, Keegan E.; Bankston, John R.; Aldrich, Richard W.
2015-01-01
The ability to measure the properties of proteins at the single-molecule level offers an unparalleled glimpse into biological systems at the molecular scale. The interpretation of single-molecule time series has often been rooted in statistical mechanics and the theory of Markov processes. While existing analysis methods have been useful, they are not without significant limitations including problems of model selection and parameter nonidentifiability. To address these challenges, we introduce the use of nonparametric Bayesian inference for the analysis of single-molecule time series. These methods provide a flexible way to extract structure from data instead of assuming models beforehand. We demonstrate these methods with applications to several diverse settings in single-molecule biophysics. This approach provides a well-constrained and rigorously grounded method for determining the number of biophysical states underlying single-molecule data. PMID:25650922
Cucinotta, Francis A; Hamada, Nobuyuki; Little, Mark P
2016-08-01
Previous analysis has shown that astronauts have a significantly lower standardized mortality ratio for circulatory disease mortality compared to the U.S. population, which is consistent with the rigorous selection process and healthy lifestyles of astronauts, and modest space radiation exposures from past space missions. However, a recent report by Delp et al. estimated the proportional mortality ratio for ages of 55-64 y of Apollo lunar mission astronauts to claim a high risk of cardiovascular disease due to space radiation compared to the U.S. population or to non-flight astronauts. In this Commentary we discuss important deficiencies in the methods and assumptions on radiation exposures used by Delp et al. that we judge cast serious doubt on their conclusions. Copyright © 2016 The Committee on Space Research (COSPAR). All rights reserved.
Rigor mortis development in turkey breast muscle and the effect of electrical stunning.
Alvarado, C Z; Sams, A R
2000-11-01
Rigor mortis development in turkey breast muscle and the effect of electrical stunning on this process are not well characterized. Some electrical stunning procedures have been known to inhibit postmortem (PM) biochemical reactions, thereby delaying the onset of rigor mortis in broilers. Therefore, this study was designed to characterize rigor mortis development in stunned and unstunned turkeys. A total of 154 turkey toms in two trials were conventionally processed at 20 to 22 wk of age. Turkeys were either stunned with a pulsed direct current (500 Hz, 50% duty cycle) at 35 mA (40 V) in a saline bath for 12 seconds or left unstunned as controls. At 15 min and 1, 2, 4, 8, 12, and 24 h PM, pectoralis samples were collected to determine pH, R-value, L* value, sarcomere length, and shear value. In Trial 1, the samples obtained for pH, R-value, and sarcomere length were divided into surface and interior samples. There were no significant differences between the surface and interior samples among any parameters measured. Muscle pH significantly decreased over time in stunned and unstunned birds through 2 h PM. The R-values increased to 8 h PM in unstunned birds and 24 h PM in stunned birds. The L* values increased over time, with no significant differences after 1 h PM for the controls and 2 h PM for the stunned birds. Sarcomere length increased through 2 h PM in the controls and 12 h PM in the stunned fillets. Cooked meat shear values decreased through the 1 h PM deboning time in the control fillets and 2 h PM in the stunned fillets. These results suggest that stunning delayed the development of rigor mortis through 2 h PM, but had no significant effect on the measured parameters at later time points, and that deboning turkey breasts at 2 h PM or later will not significantly impair meat tenderness.
Quantitative phosphoproteomic analysis of caprine muscle with high and low meat quality.
Liu, Manshun; Wei, Yanchao; Li, Xin; Quek, Siew Young; Zhao, Jing; Zhong, Huazhen; Zhang, Dequan; Liu, Yongfeng
2018-07-01
During the conversion of muscle to meat, protein phosphorylation can regulate various biological processes that have important effects on meat quality. To investigate the phosphorylation pattern of protein on rigor mortis, goat longissimus thoracis and external intercostals were classified into two groups (high quality and low quality), and meat quality was evaluated according to meat quality attributes (Warner-Bratzler shear force, Color, pH and drip loss). A quantitative mass spectrometry-based phosphoproteomic study was conducted to analyze the caprine muscle at 12h postmortem applying the TiO 2 -SIMAC-HILIC (TiSH) phosphopeptide enrichment strategy. A total of 2125 phosphopeptides were identified from 750 phosphoproteins. Among them, 96 proteins had differed in phosphorylation levels. The majority of these proteins are involved in glucose metabolism and muscle contraction. The differential phosphorylation level of proteins (PFK, MYL2 and HSP27) in two groups may be the crucial factors of regulating muscle rigor mortis. This study provides a comprehensive view for the phosphorylation status of caprine muscle at rigor mortis, it also gives a better understanding of the regulation of protein phosphorylation on various biological processes that affect the final meat quality attributes. Copyright © 2018. Published by Elsevier Ltd.
Validation of a 30-year-old process for the manufacture of L-asparaginase from Erwinia chrysanthemi.
Gervais, David; Allison, Nigel; Jennings, Alan; Jones, Shane; Marks, Trevor
2013-04-01
A 30-year-old manufacturing process for the biologic product L-asparaginase from the plant pathogen Erwinia chrysanthemi was rigorously qualified and validated, with a high level of agreement between validation data and the 6-year process database. L-Asparaginase exists in its native state as a tetrameric protein and is used as a chemotherapeutic agent in the treatment regimen for Acute Lymphoblastic Leukaemia (ALL). The manufacturing process involves fermentation of the production organism, extraction and purification of the L-asparaginase to make drug substance (DS), and finally formulation and lyophilisation to generate drug product (DP). The extensive manufacturing experience with the product was used to establish ranges for all process parameters and product quality attributes. The product and in-process intermediates were rigorously characterised, and new assays, such as size-exclusion and reversed-phase UPLC, were developed, validated, and used to analyse several pre-validation batches. Finally, three prospective process validation batches were manufactured and product quality data generated using both the existing and the new analytical methods. These data demonstrated the process to be robust, highly reproducible and consistent, and the validation was successful, contributing to the granting of an FDA product license in November, 2011.
Sikes, Anita L; Mawson, Raymond; Stark, Janet; Warner, Robyn
2014-11-01
The delivery of a consistent quality product to the consumer is vitally important for the food industry. The aim of this study was to investigate the potential for using high frequency ultrasound applied to pre- and post-rigor beef muscle on the metabolism and subsequent quality. High frequency ultrasound (600kHz at 48kPa and 65kPa acoustic pressure) applied to post-rigor beef striploin steaks resulted in no significant effect on the texture (peak force value) of cooked steaks as measured by a Tenderometer. There was no added benefit of ultrasound treatment above that of the normal ageing process after ageing of the steaks for 7days at 4°C. Ultrasound treatment of post-rigor beef steaks resulted in a darkening of fresh steaks but after ageing for 7days at 4°C, the ultrasound-treated steaks were similar in colour to that of the aged, untreated steaks. High frequency ultrasound (2MHz at 48kPa acoustic pressure) applied to pre-rigor beef neck muscle had no effect on the pH, but the calculated exhaustion factor suggested that there was some effect on metabolism and actin-myosin interaction. However, the resultant texture of cooked, ultrasound-treated muscle was lower in tenderness compared to the control sample. After ageing for 3weeks at 0°C, the ultrasound-treated samples had the same peak force value as the control. High frequency ultrasound had no significant effect on the colour parameters of pre-rigor beef neck muscle. This proof-of-concept study showed no effect of ultrasound on quality but did indicate that the application of high frequency ultrasound to pre-rigor beef muscle shows potential for modifying ATP turnover and further investigation is warranted. Crown Copyright © 2014. Published by Elsevier B.V. All rights reserved.
Pullini, Daniele; Repetto, Piermario; Bernard, Stefano; Doskolovich, Leonid; Perlo, Pietro
2005-08-20
The use of metal 2D subwavelength structures (SWSs) is a promising solution for all those applications where a selective emission from a thermal source is desirable, e.g., photovoltaic and blackbody emission. The investigation of the SWS's photonic bandgap properties is challenging, especially for the infrared and visible spectra, where the fabrication difficulties have always represented an obstacle. In this paper, the anodization of aluminum films as a self-assembly method for the SWS fabrication is proposed. A rigorous calculation of 2D SWSs of gold having high absorptivity in the visible and low absorptivity in the NIR, their fabrication by DC-sputtering deposition through anodic porous alumina templates, and their optical and topographic characterization are presented.
The Analytic Hierarchy Process and Participatory Decisionmaking
Daniel L. Schmoldt; Daniel L. Peterson; Robert L. Smith
1995-01-01
Managing natural resource lands requires social, as well as biophysical, considerations. Unfortunately, it is extremely difficult to accurately assess and quantify changing social preferences, and to aggregate conflicting opinions held by diverse social groups. The Analytic Hierarchy Process (AHP) provides a systematic, explicit, rigorous, and robust mechanism for...
Accessing Social Capital through the Academic Mentoring Process
ERIC Educational Resources Information Center
Smith, Buffy
2007-01-01
This article explores how mentors and mentees create and maintain social capital during the mentoring process. I employ a sociological conceptual framework and rigorous qualitative analytical techniques to examine how students of color and first-generation college students access social capital through mentoring relationships. The findings…
ERIC Educational Resources Information Center
Booth, Sara
2013-01-01
Benchmarking has traditionally been viewed as a way to compare data only; however, its utilisation as a more investigative, research-informed process to add rigor to decision-making processes at the institutional level is gaining momentum in the higher education sector. Indeed, with recent changes in the Australian quality environment from the…
The Optimal Size for Discussion Groups. Exchange Bibliography No. 378.
ERIC Educational Resources Information Center
Petty, Robert M.
Many variables relate to the successful functioning of groups, but one that is fundamental is the size of the group. Part 1 of this bibliography includes a selection of studies from small-group research in experimental social psychology. Part 2 of this report represents an attempt at a rigorous review of the feelings of clinicians and counselors…
Mental Disorders among Gifted and Nongifted Youth: A Selected Review of the Epidemiologic Literature
ERIC Educational Resources Information Center
Martin, Laurie T.; Burns, Rachel M.; Schonlau, Matthias
2010-01-01
Given the ongoing debate over whether giftedness is associated with mental health disorders, there is a great need to highlight and compare results from the most methodologically rigorous studies. Surprisingly, the vast majority of literature reviews and background sections of research articles include studies that do not directly compare gifted…
Transmission and Prevention of Mood Disorders among Children of Affectively Ill Parents: A Review
ERIC Educational Resources Information Center
Beardslee, William R.; Gladstone, Tracy R. G.; O'Connor, Erin E.
2011-01-01
Objective: To provide a conceptual review of the literature on children of depressed parents over the past 12 years. Method: This selective review focused on published studies that delineate the diagnosis of depression in parents, have large samples, describe children 6 to 17 years old, and are methodologically rigorous. The review emphasized…
Incentives, Selection, and Teacher Performance: Evidence from IMPACT. NBER Working Paper No. 19529
ERIC Educational Resources Information Center
Dee, Thomas; Wyckoff, James
2013-01-01
Teachers in the United States are compensated largely on the basis of fixed schedules that reward experience and credentials. However, there is a growing interest in whether performance-based incentives based on rigorous teacher evaluations can improve teacher retention and performance. The evidence available to date has been mixed at best. This…
Assurance of lubricant supply in wet-lubricated space bearings
NASA Technical Reports Server (NTRS)
Glassow, F. A.
1976-01-01
Conventional lubrication techniques appear to be satisfactory, but rigorous proof of meeting a ten-year life requirement is lacking. One approach provides additional lubricant only when commanded from ground control, while the other passively augments lubrication at all times. Each technique has specific advantages, and selection should be related to the application to obtain optimum performance.
ERIC Educational Resources Information Center
Vineberg, Robert; Joyner, John N.
Instructional System Development (ISD) methodologies and practices were examined in the Army, Navy, Marine Corps, and Air Force, each of which prescribes the ISD system involving rigorous derivation of training requirements from job requirements, selection of instructional strategies to maximize training efficiency, and revision of instruction…
Predicting SAT Performance from Advanced Course Content and Timing of Matriculation
ERIC Educational Resources Information Center
Patterson, Jonathan Sparks
2012-01-01
As record numbers of students are applying to selective colleges and universities, students are attempting to set themselves apart from their peers by taking rigorous advanced courses in high school. The race for improving a student's academic record has resulted in more and more students taking these courses earlier and earlier in their high…
Quantifying falsifiability of scientific theories
NASA Astrophysics Data System (ADS)
Nemenman, Ilya
I argue that the notion of falsifiability, a key concept in defining a valid scientific theory, can be quantified using Bayesian Model Selection, which is a standard tool in modern statistics. This relates falsifiability to the quantitative version of the statistical Occam's razor, and allows transforming some long-running arguments about validity of scientific theories from philosophical discussions to rigorous mathematical calculations.
Guidance for updating clinical practice guidelines: a systematic review of methodological handbooks.
Vernooij, Robin W M; Sanabria, Andrea Juliana; Solà, Ivan; Alonso-Coello, Pablo; Martínez García, Laura
2014-01-02
Updating clinical practice guidelines (CPGs) is a crucial process for maintaining the validity of recommendations. Methodological handbooks should provide guidance on both developing and updating CPGs. However, little is known about the updating guidance provided by these handbooks. We conducted a systematic review to identify and describe the updating guidance provided by CPG methodological handbooks and included handbooks that provide updating guidance for CPGs. We searched in the Guidelines International Network library, US National Guidelines Clearinghouse and MEDLINE (PubMed) from 1966 to September 2013. Two authors independently selected the handbooks and extracted the data. We used descriptive statistics to analyze the extracted data and conducted a narrative synthesis. We included 35 handbooks. Most handbooks (97.1%) focus mainly on developing CPGs, including variable degrees of information about updating. Guidance on identifying new evidence and the methodology of assessing the need for an update is described in 11 (31.4%) and eight handbooks (22.8%), respectively. The period of time between two updates is described in 25 handbooks (71.4%), two to three years being the most frequent (40.0%). The majority of handbooks do not provide guidance for the literature search, evidence selection, assessment, synthesis, and external review of the updating process. Guidance for updating CPGs is poorly described in methodological handbooks. This guidance should be more rigorous and explicit. This could lead to a more optimal updating process, and, ultimately to valid trustworthy guidelines.
NASA Astrophysics Data System (ADS)
Mailfert, Julien; Van de Kerkhove, Jeroen; De Bisschop, Peter; De Meyer, Kristin
2014-03-01
A Metal1-layer (M1) patterning study is conducted on 20nm node (N20) for random-logic applications. We quantified the printability performance on our test vehicle for N20, corresponding to Poly/M1 pitches of 90/64nm, and with a selected minimum M1 gap size of 70nm. The Metal1 layer is patterned with 193nm immersion lithography (193i) using Negative Tone Developer (NTD) resist, and a double-patterning Litho-Etch-Litho-Etch (LELE) process. Our study is based on Logic test blocks that we OPCed with a combination of calibrated models for litho and for etch. We report the Overlapping Process Window (OPW), based on a selection of test structures measured after-etch. We find that most of the OPW limiting structures are EOL (End-of-Line) configurations. Further analysis of these individual OPW limiters will reveal that they belong to different types, such as Resist 3D (R3D) and Mask 3D (M3D) sensitive structures, limiters related to OPC (Optical Proximity Corrections) options such as assist placement, or the choice of CD metrics and tolerances for calculation of the process windows itself. To guide this investigation, we will consider a `reference OPC' case to be compared with other solutions. In addition, rigorous simulations and OPC verifications will complete the after-etch measurements to help us to validate our experimental findings.
The Markov process admits a consistent steady-state thermodynamic formalism
NASA Astrophysics Data System (ADS)
Peng, Liangrong; Zhu, Yi; Hong, Liu
2018-01-01
The search for a unified formulation for describing various non-equilibrium processes is a central task of modern non-equilibrium thermodynamics. In this paper, a novel steady-state thermodynamic formalism was established for general Markov processes described by the Chapman-Kolmogorov equation. Furthermore, corresponding formalisms of steady-state thermodynamics for the master equation and Fokker-Planck equation could be rigorously derived in mathematics. To be concrete, we proved that (1) in the limit of continuous time, the steady-state thermodynamic formalism for the Chapman-Kolmogorov equation fully agrees with that for the master equation; (2) a similar one-to-one correspondence could be established rigorously between the master equation and Fokker-Planck equation in the limit of large system size; (3) when a Markov process is restrained to one-step jump, the steady-state thermodynamic formalism for the Fokker-Planck equation with discrete state variables also goes to that for master equations, as the discretization step gets smaller and smaller. Our analysis indicated that general Markov processes admit a unified and self-consistent non-equilibrium steady-state thermodynamic formalism, regardless of underlying detailed models.
Primer-Free Aptamer Selection Using A Random DNA Library
Pan, Weihua; Xin, Ping; Patrick, Susan; Dean, Stacey; Keating, Christine; Clawson, Gary
2010-01-01
Aptamers are highly structured oligonucleotides (DNA or RNA) that can bind to targets with affinities comparable to antibodies 1. They are identified through an in vitro selection process called Systematic Evolution of Ligands by EXponential enrichment (SELEX) to recognize a wide variety of targets, from small molecules to proteins and other macromolecules 2-4. Aptamers have properties that are well suited for in vivo diagnostic and/or therapeutic applications: Besides good specificity and affinity, they are easily synthesized, survive more rigorous processing conditions, they are poorly immunogenic, and their relatively small size can result in facile penetration of tissues. Aptamers that are identified through the standard SELEX process usually comprise ~80 nucleotides (nt), since they are typically selected from nucleic acid libraries with ~40 nt long randomized regions plus fixed primer sites of ~20 nt on each side. The fixed primer sequences thus can comprise nearly ~50% of the library sequences, and therefore may positively or negatively compromise identification of aptamers in the selection process 3, although bioinformatics approaches suggest that the fixed sequences do not contribute significantly to aptamer structure after selection 5. To address these potential problems, primer sequences have been blocked by complementary oligonucleotides or switched to different sequences midway during the rounds of SELEX 6, or they have been trimmed to 6-9 nt 7, 8. Wen and Gray 9 designed a primer-free genomic SELEX method, in which the primer sequences were completely removed from the library before selection and were then regenerated to allow amplification of the selected genomic fragments. However, to employ the technique, a unique genomic library has to be constructed, which possesses limited diversity, and regeneration after rounds of selection relies on a linear reamplification step. Alternatively, efforts to circumvent problems caused by fixed primer sequences using high efficiency partitioning are met with problems regarding PCR amplification 10. We have developed a primer-free (PF) selection method that significantly simplifies SELEX procedures and effectively eliminates primer-interference problems 11, 12. The protocols work in a straightforward manner. The central random region of the library is purified without extraneous flanking sequences and is bound to a suitable target (for example to a purified protein or complex mixtures such as cell lines). Then the bound sequences are obtained, reunited with flanking sequences, and re-amplified to generate selected sub-libraries. As an example, here we selected aptamers to S100B, a protein marker for melanoma. Binding assays showed Kd s in the 10-7 - 10-8 M range after a few rounds of selection, and we demonstrate that the aptamers function effectively in a sandwich binding format. PMID:20689511
STAR Graduate and GRO Undergraduate Fellowship Recipient List
EPA's STAR graduate fellowship program supports masters and doctoral candidates in environmental studies. Each year, students in the United States compete for STAR fellowships through a rigorous review process.
Software service history report
DOT National Transportation Integrated Search
2002-01-01
The safe and reliable operation of software within civil aviation systems and equipment has historically been assured through the application of rigorous design assurance applied during the software development process. Increasingly, manufacturers ar...
Quality and rigor of the concept mapping methodology: a pooled study analysis.
Rosas, Scott R; Kane, Mary
2012-05-01
The use of concept mapping in research and evaluation has expanded dramatically over the past 20 years. Researchers in academic, organizational, and community-based settings have applied concept mapping successfully without the benefit of systematic analyses across studies to identify the features of a methodologically sound study. Quantitative characteristics and estimates of quality and rigor that may guide for future studies are lacking. To address this gap, we conducted a pooled analysis of 69 concept mapping studies to describe characteristics across study phases, generate specific indicators of validity and reliability, and examine the relationship between select study characteristics and quality indicators. Individual study characteristics and estimates were pooled and quantitatively summarized, describing the distribution, variation and parameters for each. In addition, variation in the concept mapping data collection in relation to characteristics and estimates was examined. Overall, results suggest concept mapping yields strong internal representational validity and very strong sorting and rating reliability estimates. Validity and reliability were consistently high despite variation in participation and task completion percentages across data collection modes. The implications of these findings as a practical reference to assess the quality and rigor for future concept mapping studies are discussed. Copyright © 2011 Elsevier Ltd. All rights reserved.
Mathematics interventions for children and adolescents with Down syndrome: a research synthesis.
Lemons, C J; Powell, S R; King, S A; Davidson, K A
2015-08-01
Many children and adolescents with Down syndrome fail to achieve proficiency in mathematics. Researchers have suggested that tailoring interventions based on the behavioural phenotype may enhance efficacy. The research questions that guided this review were (1) what types of mathematics interventions have been empirically evaluated with children and adolescents with Down syndrome?; (2) do the studies demonstrate sufficient methodological rigor?; (3) is there evidence of efficacy for the evaluated mathematics interventions?; and (4) to what extent have researchers considered aspects of the behavioural phenotype in selecting, designing and/or implementing mathematics interventions for children and adolescents with Down syndrome? Nine studies published between 1989 and 2012 were identified for inclusion. Interventions predominantly focused on early mathematics skills and reported positive outcomes. However, no study met criteria for methodological rigor. Further, no authors explicitly considered the behavioural phenotype. Additional research using rigorous experimental designs is needed to evaluate the efficacy of mathematics interventions for children and adolescents with Down syndrome. Suggestions for considering the behavioural phenotype in future research are provided. © 2015 MENCAP and International Association of the Scientific Study of Intellectual and Developmental Disabilities and John Wiley & Sons Ltd.
Snodgrass, Melinda R; Chung, Moon Y; Meadan, Hedda; Halle, James W
2018-03-01
Single-case research (SCR) has been a valuable methodology in special education research. Montrose Wolf (1978), an early pioneer in single-case methodology, coined the term "social validity" to refer to the social importance of the goals selected, the acceptability of procedures employed, and the effectiveness of the outcomes produced in applied investigations. Since 1978, many contributors to SCR have included social validity as a feature of their articles and several authors have examined the prevalence and role of social validity in SCR. We systematically reviewed all SCR published in six highly-ranked special education journals from 2005 to 2016 to establish the prevalence of social validity assessments and to evaluate their scientific rigor. We found relatively low, but stable prevalence with only 28 publications addressing all three factors of the social validity construct (i.e., goals, procedures, outcomes). We conducted an in-depth analysis of the scientific rigor of these 28 publications. Social validity remains an understudied construct in SCR, and the scientific rigor of social validity assessments is often lacking. Implications and future directions are discussed. Copyright © 2018 Elsevier Ltd. All rights reserved.
Collet-Brose, Justine
2016-01-01
The aim of this study was, at the assay development stage and thus with an appropriate degree of rigor, to select the most appropriate technology platform and sample pretreatment procedure for a clinical ADA assay. Thus, ELISA, MSD, Gyrolab, and AlphaLISA immunoassay platforms were evaluated in association with target depletion and acid dissociation sample pretreatment steps. An acid dissociation step successfully improved the drug tolerance for all 4 technology platforms and the required drug tolerance was achieved with the Gyrolab and MSD platforms. The target tolerance was shown to be better for the ELISA format, where an acid dissociation treatment step alone was sufficient to achieve the desired target tolerance. However, inclusion of a target depletion step in conjunction with the acid treatment raised the target tolerance to the desired level for all of the technologies. A higher sensitivity was observed for the MSD and Gyrolab assays and the ELISA, MSD, and Gyrolab all displayed acceptable interdonor variability. This study highlights the usefulness of evaluating the performance of different assay platforms at an early stage in the assay development process to aid in the selection of the best fit-for-purpose technology platform and sample pretreatment steps. PMID:27243038
International Data on Radiological Sources
DOE Office of Scientific and Technical Information (OSTI.GOV)
Martha Finck; Margaret Goldberg
2010-07-01
ABSTRACT The mission of radiological dispersal device (RDD) nuclear forensics is to identify the provenance of nuclear and radiological materials used in RDDs and to aid law enforcement in tracking nuclear materials and routes. The application of databases to radiological forensics is to match RDD source material to a source model in the database, provide guidance regarding a possible second device, and aid the FBI by providing a short list of manufacturers and distributors, and ultimately to the last legal owner of the source. The Argonne/Idaho National Laboratory RDD attribution database is a powerful technical tool in radiological forensics. Themore » database (1267 unique vendors) includes all sealed sources and a device registered in the U.S., is complemented by data from the IAEA Catalogue, and is supported by rigorous in-lab characterization of selected sealed sources regarding physical form, radiochemical composition, and age-dating profiles. Close working relationships with global partners in the commercial sealed sources industry provide invaluable technical information and expertise in the development of signature profiles. These profiles are critical to the down-selection of potential candidates in either pre- or post- event RDD attribution. The down-selection process includes a match between an interdicted (or detonated) source and a model in the database linked to one or more manufacturers and distributors.« less
Wild, Diane; Furtado, Tamzin; Angalakuditi, Mallik
2012-01-01
Background The Child Behavior Checklist (CBCL) is a caregiver rating scale for assessing the behavioral profile of children. It was developed in the US, and has been extensively translated and used in a large number of studies internationally. Objective The objective of this study was to translate the CBCL into six languages using a rigorous translation methodology, placing particular emphasis on cultural adaptation and ensuring that the measure has content validity with carers of children with epilepsy. Methods A rigorous translation and cultural adaptation methodology was used. This is a process which includes two forward translations, reconciliation, two back-translations, and cognitive debriefing interviews with five carers of children with epilepsy in each country. In addition, a series of open-ended questions were asked of the carers in order to provide evidence of content validity. Results A number of cultural adaptations were made during the translation process. This included adaptations to the examples of sports and hobbies. An addition of “milk delivery” was made to the job examples in the Malayalam translation. In addition, two sexual problem items were removed from the Hebrew translation for Israel. Conclusion An additional six translations of the CBCL are now available for use in multinational studies. These translations have evidence of content validity for use with parents of children with epilepsy and have been appropriately culturally adapted so that they are acceptable for use in the target countries. The study highlights the importance of a rigorous translation process and the process of cultural adaptation. PMID:22715318
Documenting the Engineering Design Process
ERIC Educational Resources Information Center
Hollers, Brent
2017-01-01
Documentation of ideas and the engineering design process is a critical, daily component of a professional engineer's job. While patent protection is often cited as the primary rationale for documentation, it can also benefit the engineer, the team, company, and stakeholders through creating a more rigorously designed and purposeful solution.…
Response mechanisms of conifers to air pollutants
DOE Office of Scientific and Technical Information (OSTI.GOV)
Matyssek, R.; Reich, P.; Oren, R.
1995-07-01
Conifers are known to respond to SO{sub 2}, O{sub 3}, NO{sub x} and acid deposition. Of these pollutants, O{sub 3} is likely the most widespread and phytotoxic compound, and therefore of great interest to individuals concerned with forest resources Direct biological responses have a toxicological effects on metabolism which can then scale to effects on tree growth and forest ecology, including processes of competition and succession. Air pollution can cause reductions in photosynthesis and stomatal conductance, which are the physiological parameters most rigorously studied for conifers. Some effects air pollutants can have on plants are influenced by the presence ofmore » co-occurring environmental stresses. For example, drought usually reduces vulnerability of plants to air pollution. In addition, air pollution sensitivity may differ among species and with plant/leaf age. Plants may make short-term physiological adjustments to compensate for air pollution or may evolve resistance to air pollution through the processes of selection. Models are necessary to understand how physiological processes, growth processes, and ecological processes are affected by air pollutants. The process of defining the ecological risk that air pollutants pose for coniferous forests requires approaches that exploit existing databases, environmental monitoring of air pollutants and forest resources, experiments with well-defined air pollution treatments and environmental control/monitoring, modeling, predicting air pollution-caused changes in productivity and ecological processes over time and space, and integration of social values.« less
Imaging 2D optical diffuse reflectance in skeletal muscle
NASA Astrophysics Data System (ADS)
Ranasinghesagara, Janaka; Yao, Gang
2007-04-01
We discovered a unique pattern of optical reflectance from fresh prerigor skeletal muscles, which can not be described using existing theories. A numerical fitting function was developed to quantify the equiintensity contours of acquired reflectance images. Using this model, we studied the changes of reflectance profile during stretching and rigor process. We found that the prominent anisotropic features diminished after rigor completion. These results suggested that muscle sarcomere structures played important roles in modulating light propagation in whole muscle. When incorporating the sarcomere diffraction in a Monte Carlo model, we showed that the resulting reflectance profiles quantitatively resembled the experimental observation.
NASA Astrophysics Data System (ADS)
Cuntz, Matthias; Mai, Juliane; Zink, Matthias; Thober, Stephan; Kumar, Rohini; Schäfer, David; Schrön, Martin; Craven, John; Rakovec, Oldrich; Spieler, Diana; Prykhodko, Vladyslav; Dalmasso, Giovanni; Musuuza, Jude; Langenberg, Ben; Attinger, Sabine; Samaniego, Luis
2015-08-01
Environmental models tend to require increasing computational time and resources as physical process descriptions are improved or new descriptions are incorporated. Many-query applications such as sensitivity analysis or model calibration usually require a large number of model evaluations leading to high computational demand. This often limits the feasibility of rigorous analyses. Here we present a fully automated sequential screening method that selects only informative parameters for a given model output. The method requires a number of model evaluations that is approximately 10 times the number of model parameters. It was tested using the mesoscale hydrologic model mHM in three hydrologically unique European river catchments. It identified around 20 informative parameters out of 52, with different informative parameters in each catchment. The screening method was evaluated with subsequent analyses using all 52 as well as only the informative parameters. Subsequent Sobol's global sensitivity analysis led to almost identical results yet required 40% fewer model evaluations after screening. mHM was calibrated with all and with only informative parameters in the three catchments. Model performances for daily discharge were equally high in both cases with Nash-Sutcliffe efficiencies above 0.82. Calibration using only the informative parameters needed just one third of the number of model evaluations. The universality of the sequential screening method was demonstrated using several general test functions from the literature. We therefore recommend the use of the computationally inexpensive sequential screening method prior to rigorous analyses on complex environmental models.
NASA Astrophysics Data System (ADS)
Mai, Juliane; Cuntz, Matthias; Zink, Matthias; Thober, Stephan; Kumar, Rohini; Schäfer, David; Schrön, Martin; Craven, John; Rakovec, Oldrich; Spieler, Diana; Prykhodko, Vladyslav; Dalmasso, Giovanni; Musuuza, Jude; Langenberg, Ben; Attinger, Sabine; Samaniego, Luis
2016-04-01
Environmental models tend to require increasing computational time and resources as physical process descriptions are improved or new descriptions are incorporated. Many-query applications such as sensitivity analysis or model calibration usually require a large number of model evaluations leading to high computational demand. This often limits the feasibility of rigorous analyses. Here we present a fully automated sequential screening method that selects only informative parameters for a given model output. The method requires a number of model evaluations that is approximately 10 times the number of model parameters. It was tested using the mesoscale hydrologic model mHM in three hydrologically unique European river catchments. It identified around 20 informative parameters out of 52, with different informative parameters in each catchment. The screening method was evaluated with subsequent analyses using all 52 as well as only the informative parameters. Subsequent Sobol's global sensitivity analysis led to almost identical results yet required 40% fewer model evaluations after screening. mHM was calibrated with all and with only informative parameters in the three catchments. Model performances for daily discharge were equally high in both cases with Nash-Sutcliffe efficiencies above 0.82. Calibration using only the informative parameters needed just one third of the number of model evaluations. The universality of the sequential screening method was demonstrated using several general test functions from the literature. We therefore recommend the use of the computationally inexpensive sequential screening method prior to rigorous analyses on complex environmental models.
Chang, Zhiwei; Halle, Bertil
2013-10-14
In complex biological or colloidal samples, magnetic relaxation dispersion (MRD) experiments using the field-cycling technique can characterize molecular motions on time scales ranging from nanoseconds to microseconds, provided that a rigorous theory of nuclear spin relaxation is available. In gels, cross-linked proteins, and biological tissues, where an immobilized macromolecular component coexists with a mobile solvent phase, nuclear spins residing in solvent (or cosolvent) species relax predominantly via exchange-mediated orientational randomization (EMOR) of anisotropic nuclear (electric quadrupole or magnetic dipole) couplings. The physical or chemical exchange processes that dominate the MRD typically occur on a time scale of microseconds or longer, where the conventional perturbation theory of spin relaxation breaks down. There is thus a need for a more general relaxation theory. Such a theory, based on the stochastic Liouville equation (SLE) for the EMOR mechanism, is available for a single quadrupolar spin I = 1. Here, we present the corresponding theory for a dipole-coupled spin-1/2 pair. To our knowledge, this is the first treatment of dipolar MRD outside the motional-narrowing regime. Based on an analytical solution of the spatial part of the SLE, we show how the integral longitudinal relaxation rate can be computed efficiently. Both like and unlike spins, with selective or non-selective excitation, are treated. For the experimentally important dilute regime, where only a small fraction of the spin pairs are immobilized, we obtain simple analytical expressions for the auto-relaxation and cross-relaxation rates which generalize the well-known Solomon equations. These generalized results will be useful in biophysical studies, e.g., of intermittent protein dynamics. In addition, they represent a first step towards a rigorous theory of water (1)H relaxation in biological tissues, which is a prerequisite for unravelling the molecular basis of soft-tissue contrast in clinical magnetic resonance imaging.
Becan, Jennifer E; Bartkowski, John P; Knight, Danica K; Wiley, Tisha R A; DiClemente, Ralph; Ducharme, Lori; Welsh, Wayne N; Bowser, Diana; McCollister, Kathryn; Hiller, Matthew; Spaulding, Anne C; Flynn, Patrick M; Swartzendruber, Andrea; Dickson, Megan F; Fisher, Jacqueline Horan; Aarons, Gregory A
2018-04-13
This paper describes the means by which a United States National Institute on Drug Abuse (NIDA)-funded cooperative, Juvenile Justice-Translational Research on Interventions for Adolescents in the Legal System (JJ-TRIALS), utilized an established implementation science framework in conducting a multi-site, multi-research center implementation intervention initiative. The initiative aimed to bolster the ability of juvenile justice agencies to address unmet client needs related to substance use while enhancing inter-organizational relationships between juvenile justice and local behavioral health partners. The EPIS (Exploration, Preparation, Implementation, Sustainment) framework was selected and utilized as the guiding model from inception through project completion; including the mapping of implementation strategies to EPIS stages, articulation of research questions, and selection, content, and timing of measurement protocols. Among other key developments, the project led to a reconceptualization of its governing implementation science framework into cyclical form as the EPIS Wheel. The EPIS Wheel is more consistent with rapid-cycle testing principles and permits researchers to track both progressive and recursive movement through EPIS. Moreover, because this randomized controlled trial was predicated on a bundled strategy method, JJ-TRIALS was designed to rigorously test progress through the EPIS stages as promoted by facilitation of data-driven decision making principles. The project extended EPIS by (1) elucidating the role and nature of recursive activity in promoting change (yielding the circular EPIS Wheel), (2) by expanding the applicability of the EPIS framework beyond a single evidence-based practice (EBP) to address varying process improvement efforts (representing varying EBPs), and (3) by disentangling outcome measures of progression through EPIS stages from the a priori established study timeline. The utilization of EPIS in JJ-TRIALS provides a model for practical and applied use of implementation frameworks in real-world settings that span outer service system and inner organizational contexts in improving care for vulnerable populations. NCT02672150 . Retrospectively registered on 22 January 2016.
Lucchini Raies, Camila; Márquez Doren, Francisca; Rivera Martínez, María Soledad
2017-01-01
Breastfeeding is the most beneficial feeding practice for infants. However, it is not always the first choice for mothers and their encouragement and support from health professionals is variable. To understand the experience of mothers who had difficulties with their breastfeeding process. A phenomenological study was conducted in a University Health center. Twelve breastfeeding women were included. Data collection technique was in depth interviews, taped recorded with participants consent. Phenomenological analysis of data followed Streubert´s method. The rigor of the study was guarded by criteria for qualitative research and the research process. Ethical aspects were sheltered through the informed consent process, confidentiality and methodological rigor. The experience of living difficulties in the breastfeeding process is revealed in five comprehensive categories: recognizing the difficulties with breastfeeding; emotional impact when unable to breastfeed; motivation to overcome the difficulty and ask for help; support for breastfeeding recovery; and transition process from stress and anxiety to peace, gratification and empowerment. The understanding of this experience is qualitative evidence that contributes to a comprehensive understanding of the situation of each mother and child, allowing to improve support care interventions in health.
Alborz, Alison; McNally, Rosalind
2004-12-01
To develop methods to facilitate the 'systematic' review of evidence from a range of methodologies on diffuse or 'soft' topics, as exemplified by 'access to health care'. Twenty-eight bibliographic databases, research registers, organizational websites or library catalogues. Reference lists from identified studies. Contact with experts and service users. Current awareness and contents alerting services in the area of learning disabilities. Inclusion criteria were English language literature from 1980 onwards, relating to people with learning disabilities of any age and all study designs. The main criteria for assessment was relevance to Guillifords' model of access to health care which was adapted to the circumstances of people with learning disabilities. Selected studies were evaluated for scientific rigour then data was extracted and the results synthesized. Quality assessment was by an initial set of 'generic' quality indicators. This enabled further evidence selection before evaluation of findings according to specific criteria for qualitative, quantitative or mixed-method studies. Eighty-two studies were fully evaluated. Five studies were rated 'highly rigorous', 22 'rigorous', 46 'less rigorous' and nine 'poor' papers were retained as the sole evidence covering aspects of the guiding model. The majority of studies were quantitative but used only descriptive statistics. Most evidence lacked methodological detail, which often lowered final quality ratings. The application of a consistent structure to quality evaluation can facilitate data appraisal, extraction and synthesis across a range of methodologies in diffuse or 'soft' topics. Synthesis can be facilitated further by using software, such as the microsoft 'access' database, for managing information.
Designing cost effective water demand management programs in Australia.
White, S B; Fane, S A
2002-01-01
This paper describes recent experience with integrated resource planning (IRP) and the application of least cost planning (LCP) for the evaluation of demand management strategies in urban water. Two Australian case studies, Sydney and Northern New South Wales (NSW) are used in illustration. LCP can determine the most cost effective means of providing water services or alternatively the cheapest forms of water conservation. LCP contrasts to a traditional approach of evaluation which looks only at means of increasing supply. Detailed investigation of water usage, known as end-use analysis, is required for LCP. End-use analysis allows both rigorous demand forecasting, and the development and evaluation of conservation strategies. Strategies include education campaigns, increasing water use efficiency and promoting wastewater reuse or rainwater tanks. The optimal mix of conservation strategies and conventional capacity expansion is identified based on levelised unit cost. IRP uses LCP in the iterative process, evaluating and assessing options, investing in selected options, measuring the results, and then re-evaluating options. Key to this process is the design of cost effective demand management programs. IRP however includes a range of parameters beyond least economic cost in the planning process and program designs, including uncertainty, benefit partitioning and implementation considerations.
Development of rigor mortis is not affected by muscle volume.
Kobayashi, M; Ikegaya, H; Takase, I; Hatanaka, K; Sakurada, K; Iwase, H
2001-04-01
There is a hypothesis suggesting that rigor mortis progresses more rapidly in small muscles than in large muscles. We measured rigor mortis as tension determined isometrically in rat musculus erector spinae that had been cut into muscle bundles of various volumes. The muscle volume did not influence either the progress or the resolution of rigor mortis, which contradicts the hypothesis. Differences in pre-rigor load on the muscles influenced the onset and resolution of rigor mortis in a few pairs of samples, but did not influence the time taken for rigor mortis to reach its full extent after death. Moreover, the progress of rigor mortis in this muscle was biphasic; this may reflect the early rigor of red muscle fibres and the late rigor of white muscle fibres.
Shaver, Aaron C; Greig, Bruce W; Mosse, Claudio A; Seegmiller, Adam C
2015-05-01
Optimizing a clinical flow cytometry panel can be a subjective process dependent on experience. We develop a quantitative method to make this process more rigorous and apply it to B lymphoblastic leukemia/lymphoma (B-ALL) minimal residual disease (MRD) testing. We retrospectively analyzed our existing three-tube, seven-color B-ALL MRD panel and used our novel method to develop an optimized one-tube, eight-color panel, which was tested prospectively. The optimized one-tube, eight-color panel resulted in greater efficiency of time and resources with no loss in diagnostic power. Constructing a flow cytometry panel using a rigorous, objective, quantitative method permits optimization and avoids problems of interdependence and redundancy in a large, multiantigen panel. Copyright© by the American Society for Clinical Pathology.
ERIC Educational Resources Information Center
MOAKLEY, FRANCIS X.
EFFECTS OF PERIODIC VARIATIONS IN AN INSTRUCTIONAL FILM'S NORMAL LOUDNESS LEVEL FOR RELEVANT AND IRRELEVANT FILM SEQUENCES WERE MEASURED BY A MULTIPLE CHOICE TEST. RIGOROUS PILOT STUDIES, RANDOM GROUPING OF SEVENTH GRADERS FOR TREATMENTS, AND RATINGS OF RELEVANT AND IRRELEVANT PORTIONS OF THE FILM BY AN UNSPECIFIED NUMBER OF JUDGES PRECEDED THE…
ERIC Educational Resources Information Center
Corcoran, Sean Patrick; Baker-Smith, E. Christine
2018-01-01
New York City's public specialized high schools have a long history of offering a rigorous, college preparatory education to the city's most academically talented students. Though immensely popular and highly selective, their policy of admitting students using a single entrance exam has raised questions about diversity and equity in access. In…
What Does Research Tell Us about Trends in Dissertations on PBL?
ERIC Educational Resources Information Center
Erdogan, Tolga
2017-01-01
The aim of this study is to investigate the research trends in dissertations on PBL from 2002 to 2015 in Turkey. For this purpose, the master's and doctorate dissertations in the National Thesis Database of Council of Higher Education (CoHE) were selected for rigorous content analysis. The analysis was utilized to classify the type of study, the…
Evaluation of powder metallurgy superalloy disk materials
NASA Technical Reports Server (NTRS)
Evans, D. J.
1975-01-01
A program was conducted to develop nickel-base superalloy disk material using prealloyed powder metallurgy techniques. The program included fabrication of test specimens and subscale turbine disks from four different prealloyed powders (NASA-TRW-VIA, AF2-1DA, Mar-M-432 and MERL 80). Based on evaluation of these specimens and disks, two alloys (AF2-1DA and Mar-M-432) were selected for scale-up evaluation. Using fabricating experience gained in the subscale turbine disk effort, test specimens and full scale turbine disks were formed from the selected alloys. These specimens and disks were then subjected to a rigorous test program to evaluate their physical properties and determine their suitability for use in advanced performance turbine engines. A major objective of the program was to develop processes which would yield alloy properties that would be repeatable in producing jet engine disks from the same powder metallurgy alloys. The feasibility of manufacturing full scale gas turbine engine disks by thermomechanical processing of pre-alloyed metal powders was demonstrated. AF2-1DA was shown to possess tensile and creep-rupture properties in excess of those of Astroloy, one of the highest temperature capability disk alloys now in production. It was determined that metallographic evaluation after post-HIP elevated temperature exposure should be used to verify the effectiveness of consolidation of hot isostatically pressed billets.
Clinical Trials Targeting Aging and Age-Related Multimorbidity
Crimmins, Eileen M; Grossardt, Brandon R; Crandall, Jill P; Gelfond, Jonathan A L; Harris, Tamara B; Kritchevsky, Stephen B; Manson, JoAnn E; Robinson, Jennifer G; Rocca, Walter A; Temprosa, Marinella; Thomas, Fridtjof; Wallace, Robert; Barzilai, Nir
2017-01-01
Abstract Background There is growing interest in identifying interventions that may increase health span by targeting biological processes underlying aging. The design of efficient and rigorous clinical trials to assess these interventions requires careful consideration of eligibility criteria, outcomes, sample size, and monitoring plans. Methods Experienced geriatrics researchers and clinical trialists collaborated to provide advice on clinical trial design. Results Outcomes based on the accumulation and incidence of age-related chronic diseases are attractive for clinical trials targeting aging. Accumulation and incidence rates of multimorbidity outcomes were developed by selecting at-risk subsets of individuals from three large cohort studies of older individuals. These provide representative benchmark data for decisions on eligibility, duration, and assessment protocols. Monitoring rules should be sensitive to targeting aging-related, rather than disease-specific, outcomes. Conclusions Clinical trials targeting aging are feasible, but require careful design consideration and monitoring rules. PMID:28364543
Recent advances in applying decision science to managing national forests
Marcot, Bruce G.; Thompson, Matthew P.; Runge, Michael C.; Thompson, Frank R.; McNulty, Steven; Cleaves, David; Tomosy, Monica; Fisher, Larry A.; Andrew, Bliss
2012-01-01
Management of federal public forests to meet sustainability goals and multiple use regulations is an immense challenge. To succeed, we suggest use of formal decision science procedures and tools in the context of structured decision making (SDM). SDM entails four stages: problem structuring (framing the problem and defining objectives and evaluation criteria), problem analysis (defining alternatives, evaluating likely consequences, identifying key uncertainties, and analyzing tradeoffs), decision point (identifying the preferred alternative), and implementation and monitoring the preferred alternative with adaptive management feedbacks. We list a wide array of models, techniques, and tools available for each stage, and provide three case studies of their selected use in National Forest land management and project plans. Successful use of SDM involves participation by decision-makers, analysts, scientists, and stakeholders. We suggest specific areas for training and instituting SDM to foster transparency, rigor, clarity, and inclusiveness in formal decision processes regarding management of national forests.
Gómez, Aina G; Ondiviela, Bárbara; Puente, Araceli; Juanes, José A
2015-05-15
This work presents a standard and unified procedure for assessment of environmental risks at the contaminant source level in port aquatic systems. Using this method, port managers and local authorities will be able to hierarchically classify environmental hazards and proceed with the most suitable management actions. This procedure combines rigorously selected parameters and indicators to estimate the environmental risk of each contaminant source based on its probability, consequences and vulnerability. The spatio-temporal variability of multiple stressors (agents) and receptors (endpoints) is taken into account to provide accurate estimations for application of precisely defined measures. The developed methodology is tested on a wide range of different scenarios via application in six European ports. The validation process confirms its usefulness, versatility and adaptability as a management tool for port water quality in Europe and worldwide. Copyright © 2015 Elsevier Ltd. All rights reserved.
RF Behavior of Cylindrical Cavity Based 240 GHz, 1 MW Gyrotron for Future Tokamak System
NASA Astrophysics Data System (ADS)
Kumar, Nitin; Singh, Udaybir; Bera, Anirban; Sinha, A. K.
2017-11-01
In this paper, we present the RF behavior of conventional cylindrical interaction cavity for 240 GHz, 1 MW gyrotron for futuristic plasma fusion reactors. Very high-order TE mode is searched for this gyrotron to minimize the Ohmic wall loading at the interaction cavity. The mode selection process is carried out rigorously to analyze the mode competition and design feasibility. The cold cavity analysis and beam-wave interaction computation are carried out to finalize the cavity design. The detail parametric analyses for interaction cavity are performed in terms of mode stability, interaction efficiency and frequency. In addition, the design of triode type magnetron injection gun is also discussed. The electron beam parameters such as velocity ratio and velocity spread are optimized as per the requirement at interaction cavity. The design studies presented here confirm the realization of CW, 1 MW power at 240 GHz frequency at TE46,17 mode.
Image-algebraic design of multispectral target recognition algorithms
NASA Astrophysics Data System (ADS)
Schmalz, Mark S.; Ritter, Gerhard X.
1994-06-01
In this paper, we discuss methods for multispectral ATR (Automated Target Recognition) of small targets that are sensed under suboptimal conditions, such as haze, smoke, and low light levels. In particular, we discuss our ongoing development of algorithms and software that effect intelligent object recognition by selecting ATR filter parameters according to ambient conditions. Our algorithms are expressed in terms of IA (image algebra), a concise, rigorous notation that unifies linear and nonlinear mathematics in the image processing domain. IA has been implemented on a variety of parallel computers, with preprocessors available for the Ada and FORTRAN languages. An image algebra C++ class library has recently been made available. Thus, our algorithms are both feasible implementationally and portable to numerous machines. Analyses emphasize the aspects of image algebra that aid the design of multispectral vision algorithms, such as parameterized templates that facilitate the flexible specification of ATR filters.
Characterization of redox conditions in groundwater contaminant plumes
NASA Astrophysics Data System (ADS)
Christensen, Thomas H.; Bjerg, Poul L.; Banwart, Steven A.; Jakobsen, Rasmus; Heron, Gorm; Albrechtsen, Hans-Jørgen
2000-10-01
Evaluation of redox conditions in groundwater pollution plumes is often a prerequisite for understanding the behaviour of the pollutants in the plume and for selecting remediation approaches. Measuring of redox conditions in pollution plumes is, however, a fairly recent issue and yet relative few cases have been reported. No standardised or generally accepted approach exists. Slow electrode kinetics and the common lack of internal equilibrium of redox processes in pollution plumes make, with a few exceptions, direct electrochemical measurement and rigorous interpretation of redox potentials dubious, if not erroneous. Several other approaches have been used in addressing redox conditions in pollution plumes: redox-sensitive compounds in groundwater samples, hydrogen concentrations in groundwater, concentrations of volatile fatty acids in groundwater, sediment characteristics and microbial tools, such as MPN counts, PLFA biomarkers and redox bioassays. This paper reviews the principles behind the different approaches, summarizes methods used and evaluates the approaches based on the experience from the reported applications.
Developing Analogy Cost Estimates for Space Missions
NASA Technical Reports Server (NTRS)
Shishko, Robert
2004-01-01
The analogy approach in cost estimation combines actual cost data from similar existing systems, activities, or items with adjustments for a new project's technical, physical or programmatic differences to derive a cost estimate for the new system. This method is normally used early in a project cycle when there is insufficient design/cost data to use as a basis for (or insufficient time to perform) a detailed engineering cost estimate. The major limitation of this method is that it relies on the judgment and experience of the analyst/estimator. The analyst must ensure that the best analogy or analogies have been selected, and that appropriate adjustments have been made. While analogy costing is common, there is a dearth of advice in the literature on the 'adjustment methodology', especially for hardware projects. This paper discusses some potential approaches that can improve rigor and repeatability in the analogy costing process.
From screening to synthesis: using nvivo to enhance transparency in qualitative evidence synthesis.
Houghton, Catherine; Murphy, Kathy; Meehan, Ben; Thomas, James; Brooker, Dawn; Casey, Dympna
2017-03-01
To explore the experiences and perceptions of healthcare staff caring for people with dementia in the acute setting. This article focuses on the methodological process of conducting framework synthesis using nvivo for each stage of the review: screening, data extraction, synthesis and critical appraisal. Qualitative evidence synthesis brings together many research findings in a meaningful way that can be used to guide practice and policy development. For this purpose, synthesis must be conducted in a comprehensive and rigorous way. There has been previous discussion on how using nvivo can assist in enhancing and illustrate the rigorous processes involved. Qualitative framework synthesis. Twelve documents, or research reports, based on nine studies, were included for synthesis. The benefits of using nvivo are outlined in terms of facilitating teams of researchers to systematically and rigorously synthesise findings. nvivo functions were used to conduct a sensitivity analysis. Some valuable lessons were learned, and these are presented to assist and guide researchers who wish to use similar methods in future. Ultimately, good qualitative evidence synthesis will provide practitioners and policymakers with significant information that will guide decision-making on many aspects of clinical practice. The example provided explored how people with dementia are cared for acute settings. © 2016 The Authors. Journal of Clinical Nursing Published by John Wiley & Sons Ltd.
Kim, Huiyong; Hwang, Sung June; Lee, Kwang Soon
2015-02-03
Among various CO2 capture processes, the aqueous amine-based absorption process is considered the most promising for near-term deployment. However, the performance evaluation of newly developed solvents still requires complex and time-consuming procedures, such as pilot plant tests or the development of a rigorous simulator. Absence of accurate and simple calculation methods for the energy performance at an early stage of process development has lengthened and increased expense of the development of economically feasible CO2 capture processes. In this paper, a novel but simple method to reliably calculate the regeneration energy in a standard amine-based carbon capture process is proposed. Careful examination of stripper behaviors and exploitation of energy balance equations around the stripper allowed for calculation of the regeneration energy using only vapor-liquid equilibrium and caloric data. Reliability of the proposed method was confirmed by comparing to rigorous simulations for two well-known solvents, monoethanolamine (MEA) and piperazine (PZ). The proposed method can predict the regeneration energy at various operating conditions with greater simplicity, greater speed, and higher accuracy than those proposed in previous studies. This enables faster and more precise screening of various solvents and faster optimization of process variables and can eventually accelerate the development of economically deployable CO2 capture processes.
Use of old antibiotics now and in the future from a pharmacokinetic/pharmacodynamic perspective.
Muller, A E; Theuretzbacher, U; Mouton, J W
2015-10-01
Because of the increase in bacterial resistance to commonly used antibacterial drugs, old antibiotics are being 'revived' and, once again, are attracting interest. Many of these old antibiotics were approved long ago, in an era when there was no clear process for development, and requirements for efficacy to be demonstrated in rigorous clinical trials did not exist. At the time of these approvals, pharmacokinetic and pharmacodynamic principles were largely unknown, and did not inform the dose-finding process or recommendations for optimal usage. Indeed, the task of generating basic vital information for these old antibiotics remains to be performed. In this review, we provide a brief overview of the most essential data needed for dose justification and optimization. An overview of the shortage of data for selected old antibiotics illustrates the scope of the problem. In order to prevent harming patients with clinical decisions based on inadequate evidence, a redevelopment procedure for old antibiotics is urgently needed, including a regulatory framework. Copyright © 2015 European Society of Clinical Microbiology and Infectious Diseases. Published by Elsevier Ltd. All rights reserved.
Geophysical information for teachers: Wave tanks, homemade clouds, glacial goo, and more!
NASA Astrophysics Data System (ADS)
Adamec, Bethany Holm
2012-02-01
AGU is deeply committed to fostering the next generation of Earth and space scientists. Union activities contribute to this effort in many ways, one of which is partnering with the National Earth Science Teacher's Association (NESTA) to hold the Annual Geophysical Information for Teachers (GIFT) workshop at AGU's annual Fall Meeting. GIFT allows K-12 science teachers to hear about the latest geoscience research from the scientists making the discoveries, explore new classroom resources for their students, and visit exhibits and technical sessions of the AGU meeting for free. In 2011 AGU worked with NESTA to develop an improved rigorous and open application process for scientists and education professionals who wished to work as a team and present their Earth and space science work to teachers, as well as lead the educators in a hands-on, classroom- ready activity. Twenty-four applications were received for five slots, so the selected presentations (on tsunamis, clouds, field campaigns, glaciers, and volcanoes), chosen through a peer- review process, truly represented the best ways of getting cutting-edge science into the classroom.
Chirality-induced spin polarization places symmetry constraints on biomolecular interactions.
Kumar, Anup; Capua, Eyal; Kesharwani, Manoj K; Martin, Jan M L; Sitbon, Einat; Waldeck, David H; Naaman, Ron
2017-03-07
Noncovalent interactions between molecules are key for many biological processes. Necessarily, when molecules interact, the electronic charge in each of them is redistributed. Here, we show experimentally that, in chiral molecules, charge redistribution is accompanied by spin polarization. We describe how this spin polarization adds an enantioselective term to the forces, so that homochiral interaction energies differ from heterochiral ones. The spin polarization was measured by using a modified Hall effect device. An electric field that is applied along the molecules causes charge redistribution, and for chiral molecules, a Hall voltage is measured that indicates the spin polarization. Based on this observation, we conjecture that the spin polarization enforces symmetry constraints on the biorecognition process between two chiral molecules, and we describe how these constraints can lead to selectivity in the interaction between enantiomers based on their handedness. Model quantum chemistry calculations that rigorously enforce these constraints show that the interaction energy for methyl groups on homochiral molecules differs significantly from that found for heterochiral molecules at van der Waals contact and shorter (i.e., ∼0.5 kcal/mol at 0.26 nm).
DOE Office of Scientific and Technical Information (OSTI.GOV)
Rios, Orlando; Radhakrishnan, Balasubramaniam; Caravias, George
2015-03-11
Grid Logic Inc. is developing a method for sintering and melting fine metallic powders for additive manufacturing using spatially-compact, high-frequency magnetic fields called Micro-Induction Sintering (MIS). One of the challenges in advancing MIS technology for additive manufacturing is in understanding the power transfer to the particles in a powder bed. This knowledge is important to achieving efficient power transfer, control, and selective particle heating during the MIS process needed for commercialization of the technology. The project s work provided a rigorous physics-based model for induction heating of fine spherical particles as a function of frequency and particle size. This simulationmore » improved upon Grid Logic s earlier models and provides guidance that will make the MIS technology more effective. The project model will be incorporated into Grid Logic s power control circuit of the MIS 3D printer product and its diagnostics technology to optimize the sintering process for part quality and energy efficiency.« less
Experimental Evaluation of Processing Time for the Synchronization of XML-Based Business Objects
NASA Astrophysics Data System (ADS)
Ameling, Michael; Wolf, Bernhard; Springer, Thomas; Schill, Alexander
Business objects (BOs) are data containers for complex data structures used in business applications such as Supply Chain Management and Customer Relationship Management. Due to the replication of application logic, multiple copies of BOs are created which have to be synchronized and updated. This is a complex and time consuming task because BOs rigorously vary in their structure according to the distribution, number and size of elements. Since BOs are internally represented as XML documents, the parsing of XML is one major cost factor which has to be considered for minimizing the processing time during synchronization. The prediction of the parsing time for BOs is an significant property for the selection of an efficient synchronization mechanism. In this paper, we present a method to evaluate the influence of the structure of BOs on their parsing time. The results of our experimental evaluation incorporating four different XML parsers examine the dependencies between the distribution of elements and the parsing time. Finally, a general cost model will be validated and simplified according to the results of the experimental setup.
NASA Astrophysics Data System (ADS)
Porter, Sophia; Strolger, Louis-Gregory; Lagerstrom, Jill; Weissman, Sarah
2016-01-01
The Space Telescope Science Institute annually receives more than one thousand formal proposals for Hubble Space Telescope time, exceeding the available time with the observatory by a factor of over four. With JWST, the proposal pressure will only increase, straining our ability to provide rigorous peer review of each proposal's scientific merit. Significant hurdles in this process include the proper categorization of proposals, to ensure Time Allocation Committees (TACs) have the required and desired expertise to fairly and appropriately judge each proposal, and the selection of reviewers themselves, to establish diverse and well-qualified TACs. The Panel Auto-Categorizer and Manager (PACMan; a naive Bayesian classifier) was developed to automatically sort new proposals into their appropriate science categories and, similarly, to appoint panel reviewers with the best qualifications to serve on the corresponding TACs. We will provide an overview of PACMan and present the results of its testing on five previous cycles of proposals. PACMan will be implemented in upcoming cycles to support and eventually replace the process for constructing the time allocation reviews.
Methodology for balancing design and process tradeoffs for deep-subwavelength technologies
NASA Astrophysics Data System (ADS)
Graur, Ioana; Wagner, Tina; Ryan, Deborah; Chidambarrao, Dureseti; Kumaraswamy, Anand; Bickford, Jeanne; Styduhar, Mark; Wang, Lee
2011-04-01
For process development of deep-subwavelength technologies, it has become accepted practice to use model-based simulation to predict systematic and parametric failures. Increasingly, these techniques are being used by designers to ensure layout manufacturability, as an alternative to, or complement to, restrictive design rules. The benefit of model-based simulation tools in the design environment is that manufacturability problems are addressed in a design-aware way by making appropriate trade-offs, e.g., between overall chip density and manufacturing cost and yield. The paper shows how library elements and the full ASIC design flow benefit from eliminating hot spots and improving design robustness early in the design cycle. It demonstrates a path to yield optimization and first time right designs implemented in leading edge technologies. The approach described herein identifies those areas in the design that could benefit from being fixed early, leading to design updates and avoiding later design churn by careful selection of design sensitivities. This paper shows how to achieve this goal by using simulation tools incorporating various models from sparse to rigorously physical, pattern detection and pattern matching, checking and validating failure thresholds.
Inelastic electron tunneling mediated by a molecular quantum rotator
NASA Astrophysics Data System (ADS)
Sugimoto, Toshiki; Kunisada, Yuji; Fukutani, Katsuyuki
2017-12-01
Inelastic electron tunneling (IET) accompanying nuclear motion is not only of fundamental physical interest but also has strong impacts on chemical and biological processes in nature. Although excitation of rotational motion plays an important role in enhancing electric conductance at a low bias, the mechanism of rotational excitation remains veiled. Here, we present a basic theoretical framework of IET that explicitly takes into consideration quantum angular momentum, focusing on a molecular H2 rotator trapped in a nanocavity between two metallic electrodes as a model system. It is shown that orientationally anisotropic electrode-rotator coupling is the origin of angular-momentum exchange between the electron and molecule; we found that the anisotropic coupling imposes rigorous selection rules in rotational excitation. In addition, rotational symmetry breaking induced by the anisotropic potential lifts the degeneracy of the energy level of the degenerated rotational state of the quantum rotator and tunes the threshold bias voltage that triggers rotational IET. Our theoretical results provide a paradigm for physical understanding of the rotational IET process and spectroscopy, as well as molecular-level design of electron-rotation coupling in nanoelectronics.
Evidence-based nursing leadership: Evaluation of a Joint Academic-Service Journal Club.
Duffy, Joanne R; Thompson, Diane; Hobbs, Terry; Niemeyer-Hackett, Nancy Lee; Elpers, Susan
2011-10-01
This article describes the importance of evidence-based nursing leadership in the development and evaluation of a joint academic-service nursing leadership journal club. The use of scientific evidence and the embracing of an environment of continuous learning are essential to quality practice; however, nursing leadership has been slow to apply evidence-based practice to their own work. A noontime monthly meeting schedule, incentivized by lunch, was organized as a nursing leadership journal club. Articles were selected and reviewed monthly, and the process was formally evaluated using a written evaluation at the end of year 1. Eighteen articles were appraised by the group with 6 topics identified. Positive results included increased knowledge, competence of the leader, and attainment of goals. Recommendations include revision of goals, plans to share leadership of the group, development of a rigorous evaluation of outcomes, and dissemination of findings. The journal club was valuable in increasing awareness of nursing leadership research, promoting leadership development, and improving competence in the performance of research appraisals. Process improvement and further study are needed to increase understanding regarding the benefits of leadership journal clubs.
Social Networking Phenomena in the First-Year Experience
ERIC Educational Resources Information Center
Corwin, Jay R.; Cintron, Rosa
2011-01-01
The transition from high school to college is never an easy process. New freedoms and new independence provides for an exciting first year. There is no debate as to the importance of starting college off well. However, juggling these new opportunities with increased academic rigor is not an easy process. Several authors have described the…
Case Studies Add Value to a Diverse Teaching Portfolio in Science Courses
ERIC Educational Resources Information Center
Camill, Philip
2006-01-01
Together with lectures and labs, case studies assist students in acquiring content knowledge, process skills, and an understanding of the context and application of science to their daily lives. Cases make the process of scientific learning more genuine and rigorous, bringing alive classroom learning and helping students apply concepts to…
The Leverage of National Board Candidacy: An Exploration of Teacher Learning
ERIC Educational Resources Information Center
Hunzicker, Jana
2008-01-01
The vast majority of teachers who engage in the process of National Board certification describe it as the best professional development they have ever experienced - even when they do not achieve the certification. Learning leverage, an interactive dynamic characterized by rigor, reward, and risk, is what makes the certification process such a…
School Psychology as a Relational Enterprise: The Role and Process of Qualitative Methodology
ERIC Educational Resources Information Center
Newman, Daniel S.; Clare, Mary M.
2016-01-01
The purpose of this article is to explore the application of qualitative research to establishing a more complete understanding of relational processes inherent in school psychology practice. We identify the building blocks of rigorous qualitative research design through a conceptual overview of qualitative paradigms, methodologies, methods (i.e.,…
A Methodological Self-Study of Quantitizing: Negotiating Meaning and Revealing Multiplicity
ERIC Educational Resources Information Center
Seltzer-Kelly, Deborah; Westwood, Sean J.; Pena-Guzman, David M.
2012-01-01
This inquiry developed during the process of "quantitizing" qualitative data the authors had gathered for a mixed methods curriculum efficacy study. Rather than providing the intended rigor to their data coding process, their use of an intercoder reliability metric prompted their investigation of the multiplicity and messiness that, as they…
Meeting the needs of an ever-demanding market.
Rigby, Richard
2002-04-01
Balancing cost and performance in packaging is critical. This article outlines techniques to assist in this whilst delivering added value and product differentiation. The techniques include a rigorous statistical process capable of delivering cost reduction and improved quality and a computer modelling process that can save time when validating new packaging options.
ERIC Educational Resources Information Center
Follette, William C.; Bonow, Jordan T.
2009-01-01
Whether explicitly acknowledged or not, behavior-analytic principles are at the heart of most, if not all, empirically supported therapies. However, the change process in psychotherapy is only now being rigorously studied. Functional analytic psychotherapy (FAP; Kohlenberg & Tsai, 1991; Tsai et al., 2009) explicitly identifies behavioral-change…
Development Of Simulation Model For Fluid Catalytic Cracking
NASA Astrophysics Data System (ADS)
Ghosh, Sobhan
2010-10-01
Fluid Catalytic Cracking (FCC) is the most widely used secondary conversion process in the refining industry, for producing gasoline, olefins, and middle distillate from heavier petroleum fractions. There are more than 500 units in the world with a total processing capacity of about 17 to 20% of the crude capacity. FCC catalyst is the highest consumed catalyst in the process industry. On one hand, FCC is quite flexible with respect to it's ability to process wide variety of crudes with a flexible product yield pattern, and on the other hand, the interdependence of the major operating parameters makes the process extremely complex. An operating unit is self balancing and some fluctuations in the independent parameters are automatically adjusted by changing the temperatures and flow rates at different sections. However, a good simulation model is very useful to the refiner to get the best out of the process, in terms of selection of the best catalyst, to cope up with the day to day changing of the feed quality and the demands of different products from FCC unit. In addition, a good model is of great help in designing the process units and peripherals. A simple empirical model is often adequate to monitor the day to day operations, but they are not of any use in handling the other problems such as, catalyst selection or, design / modification of the plant. For this, a kinetic based rigorous model is required. Considering the complexity of the process, large number of chemical species undergoing "n" number of parallel and consecutive reactions, it is virtually impossible to develop a simulation model based on the kinetic parameters. The most common approach is to settle for a semi empirical model. We shall take up the key issues for developing a FCC model and the contribution of such models in the optimum operation of the plant.
Receiver psychology turns 20: is it time for a broader approach?
Miller, Cory T.; Bee, Mark A.
2013-01-01
Twenty years ago, a new conceptual paradigm known as ‘receiver psychology’ was introduced to explain the evolution of animal communication systems. This paradigm advanced the idea that psychological processes in the receiver's nervous system influence a signal's detectability, discriminability and memorability, and thereby serve as powerful sources of selection shaping signal design. While advancing our understanding of signal diversity, more recent studies make clear that receiver psychology, as a paradigm, has been structured too narrowly and does not incorporate many of the perceptual and cognitive processes of signal reception that operate between sensory transduction and a receiver's response. Consequently, the past two decades of research on receiver psychology have emphasized considerations of signal evolution but failed to ask key questions about the mechanisms of signal reception and their evolution. The primary aim of this essay is to advocate for a broader receiver psychology paradigm that more explicitly includes a research focus on receivers' psychological landscapes. We review recent experimental studies of hearing and sound communication to illustrate how considerations of several general perceptual and cognitive processes will facilitate future research on animal signalling systems. We also emphasize how a rigorous comparative approach to receiver psychology is critical to explicating the full range of perceptual and cognitive processes involved in receiving and responding to signals. PMID:24013277
Effects of well-boat transportation on the muscle pH and onset of rigor mortis in Atlantic salmon.
Gatica, M C; Monti, G; Gallo, C; Knowles, T G; Warriss, P D
2008-07-26
During the transport of salmon (Salmo salar), in a well-boat, 10 fish were sampled at each of six stages: in cages after crowding at the farm (stage 1), in the well-boat after loading (stage 2), in the well-boat after eight hours transport and before unloading (stage 3), in the resting cages immediately after finishing unloading (stage 4), after 24 hours resting in cages, (stage 5) and in the processing plant after pumping from the resting cages (stage 6). The water in the well-boat was at ambient temperature with recirculation to the sea. At each stage the fish were stunned percussively and bled by gill cutting. Immediately after death, and then every three hours for 18 hours, the muscle pH and rigor index of the fish were measured. At successive stages the initial muscle pH of the fish decreased, except for a slight gain in stage 5, after they had been rested for 24 hours. The lowest initial muscle pH was observed at stage 6. The fishes' rigor index showed that rigor developed more quickly at each successive stage, except for a slight decrease in rate at stage 5, attributable to the recovery of muscle reserves.
Rigorous diffraction analysis using geometrical theory of diffraction for future mask technology
NASA Astrophysics Data System (ADS)
Chua, Gek S.; Tay, Cho J.; Quan, Chenggen; Lin, Qunying
2004-05-01
Advanced lithographic techniques such as phase shift masks (PSM) and optical proximity correction (OPC) result in a more complex mask design and technology. In contrast to the binary masks, which have only transparent and nontransparent regions, phase shift masks also take into consideration transparent features with a different optical thickness and a modified phase of the transmitted light. PSM are well-known to show prominent diffraction effects, which cannot be described by the assumption of an infinitely thin mask (Kirchhoff approach) that is used in many commercial photolithography simulators. A correct prediction of sidelobe printability, process windows and linearity of OPC masks require the application of rigorous diffraction theory. The problem of aerial image intensity imbalance through focus with alternating Phase Shift Masks (altPSMs) is performed and compared between a time-domain finite-difference (TDFD) algorithm (TEMPEST) and Geometrical theory of diffraction (GTD). Using GTD, with the solution to the canonical problems, we obtained a relationship between the edge on the mask and the disturbance in image space. The main interest is to develop useful formulations that can be readily applied to solve rigorous diffraction for future mask technology. Analysis of rigorous diffraction effects for altPSMs using GTD approach will be discussed.
ERIC Educational Resources Information Center
Gold, Anne U.; Ledley, Tamara Shapiro; Buhr, Susan M.; Fox, Sean; McCaffrey, Mark; Niepold, Frank; Manduca, Cathy; Lynds, Susan E.
2012-01-01
Educators seek to develop 21st century skills in the classroom by incorporating educational materials other than textbooks into their lessons, such as digitally available activities, videos, and visualizations. A problem that educators face is that no review process similar to the formal adoption processes used for K-12 textbooks or the…
Use of software engineering techniques in the design of the ALEPH data acquisition system
NASA Astrophysics Data System (ADS)
Charity, T.; McClatchey, R.; Harvey, J.
1987-08-01
The SASD methodology is being used to provide a rigorous design framework for various components of the ALEPH data acquisition system. The Entity-Relationship data model is used to describe the layout and configuration of the control and acquisition systems and detector components. State Transition Diagrams are used to specify control applications such as run control and resource management and Data Flow Diagrams assist in decomposing software tasks and defining interfaces between processes. These techniques encourage rigorous software design leading to enhanced functionality and reliability. Improved documentation and communication ensures continuity over the system life-cycle and simplifies project management.
Teichmann, Marc; Lesoil, Constance; Godard, Juliette; Vernet, Marine; Bertrand, Anne; Levy, Richard; Dubois, Bruno; Lemoine, Laurie; Truong, Dennis Q; Bikson, Marom; Kas, Aurélie; Valero-Cabré, Antoni
2016-11-01
Noninvasive brain stimulation in primary progressive aphasia (PPA) is a promising approach. Yet, applied to single cases or insufficiently controlled small-cohort studies, it has not clarified its therapeutic value. We here address the effectiveness of transcranial direct current stimulation (tDCS) on the semantic PPA variant (sv-PPA), applying a rigorous study design to a large, homogeneous sv-PPA cohort. Using a double-blind, sham-controlled counterbalanced cross-over design, we applied three tDCS conditions targeting the temporal poles of 12 sv-PPA patients. Efficiency was assessed by a semantic matching task orthogonally manipulating "living"/"nonliving" categories and verbal/visual modalities. Conforming to predominantly left-lateralized damage in sv-PPA and accounts of interhemispheric inhibition, we applied left hemisphere anodal-excitatory and right hemisphere cathodal-inhibitory tDCS, compared to sham stimulation. Prestimulation data, compared to 15 healthy controls, showed that patients had semantic disorders predominating with living categories in the verbal modality. Stimulation selectively impacted these most impaired domains: Left-excitatory and right-inhibitory tDCS improved semantic accuracy in verbal modality, and right-inhibitory tDCS improved processing speed with living categories and accuracy and processing speed in the combined verbal × living condition. Our findings demonstrate the efficiency of tDCS in sv-PPA by generating highly specific intrasemantic effects. They provide "proof of concept" for future applications of tDCS in therapeutic multiday regimes, potentially driving sustained improvement of semantic processing. Our data also support the hotly debated existence of a left temporal-pole network for verbal semantics selectively modulated through both left-excitatory and right-inhibitory brain stimulation. Ann Neurol 2016;80:693-707. © 2016 American Neurological Association.
Schaffelke, Britta; Fabricius, Katharina; Kroon, Frederieke; Brodie, Jon; De'ath, Glenn; Shaw, Roger; Tarte, Diane; Warne, Michael; Thorburn, Peter
2018-04-01
This is a response to the published Viewpoint by Larcombe and Ridd (2018). We agree with Larcombe and Ridd (2018) that scientific merit goes hand in hand with rigorous quality control. However, we are responding here to several points raised by Larcombe and Ridd (2018) which in our view were misrepresented. We describe the formal and effective science review, synthesis and advice processes that are in place for science supporting decision-making in the Great Barrier Reef. We also respond in detail to critiques of selected publications that were used by Larcombe and Ridd (2018) as a case study to illustrate shortcomings in science quality control. We provide evidence that their representation of the published research and arguments to support the statement that "many (…) conclusions are demonstrably incorrect" is based on misinterpretation, selective use of data and over-simplification, and also ignores formal responses to previously published critiques. Copyright © 2018 The Authors. Published by Elsevier Ltd.. All rights reserved.
Polarization-selective infrared bandpass filter based on a two-layer subwavelength metallic grating
NASA Astrophysics Data System (ADS)
Hohne, Andrew J.; Moon, Benjamin; Baumbauer, Carol L.; Gray, Tristan; Dilts, James; Shaw, Joseph A.; Dickensheets, David L.; Nakagawa, Wataru
2017-08-01
We present the design, fabrication, and characterization of a polarization-selective infrared bandpass filter based on a two-layer subwavelength metallic grating for use in polarimetric imaging. Gold nanowires were deposited via physical vapor deposition (PVD) onto a silicon surface relief grating that was patterned using electron beam lithography (EBL) and fabricated using standard silicon processing techniques. Optical characterization with a broad-spectrum tungsten halogen light source and a grating spectrometer showed normalized peak TM transmission of 53% with a full-width at half-maximum (FWHM) of 122 nm, which was consistent with rigorous coupled-wave analysis (RCWA) simulations. Simulation results suggested that device operation relied on suppression of the TM transmission caused by surface plasmon polariton (SPP) excitation at the gold-silicon interface and an increase in TM transmission caused by a Fabry-Perot (FP) resonance in the cavity between the gratings. TE rejection occurred at the initial air/gold interface. We also present simulation results of an improved design based on a two-dielectric grating where two different SPP resonances allowed us to improve the shape of the passband by suppressing the side lobes. This newer design resulted in improved side-band performance and increased peak TM transmission.
Mignot, Mélanie; Schammé, Benjamin; Tognetti, Vincent; Joubert, Laurent; Cardinael, Pascal; Peulon-Agasse, Valérie
2017-10-13
New polar embedded aromatic stationary phases (mono- and trifunctional versions) that contain an amide-embedded group coupled with a tricyclic aromatic moiety were developed for chromatographic applications and described in the first paper of this series. These phases offered better separation performance for PAHs than for alkylbenzene homologues, and an enhanced ability to differentiate aromatic planarity to aromatic tridimensional conformation, especially for the trifunctional version and when using methanol instead of acetonitrile. In this second paper, a density functional theory study of the retention process is reported. In particular, it was shown that the selection of the suitable computational protocol allowed for describing rigorously the interactions that could take place, the solvent effects, and the structural changes for the monofunctional and the trifunctional versions. For the first time, the experimental data coupled with these DFT results provided a better understanding of the interaction mechanisms and highlighted the importance of the multimodal character of the designed stationary phases: alkyl spacers for interactions with hydrophobic solutes, amide embedded groups for dipole-dipole and hydrogen-bond interactions, and aromatic terminal groups for π-π interactions. Copyright © 2017 Elsevier B.V. All rights reserved.
Inhibiting Metal Oxide Atomic Layer Deposition: Beyond Zinc Oxide
Sampson, Matthew D.; Emery, Jonathan D.; Pellin, Michael J.; ...
2017-04-05
The atomic layer deposition (ALD) of several metal oxides is selectivity inhibited on alkanethiol self-assembled monolayers (SAMs) on Au and the eventual nucleation mechanism is investigated. The inhibition ability of the SAM is significantly improved by the in situ H 2-plasma pretreatment of the Au substrate prior to gas-phase deposition of a long-chain alkanethiol, 1-dodecanethiol (DDT). This more rigorous surface preparation inhibits even aggressive oxide ALD precursors, including trimethylaluminum and water, for at least 20 cycles. We study the effect that ALD precursor purge times, growth temperature, alkanethiol chain length, alkanethiol deposition time, and plasma treatment time have on Almore » 2O 3 ALD inhibition. This is the first example of Al 2O 3 ALD inhibition from a vapor-deposited SAM. Inhibition of Al 2O 3, ZnO, and MnO ALD processes are compared, revealing the versatility of this selective surface treatment. As a result, atomic force microscopy (AFM) and grazing incidence x-ray fluorescence (GIXRF) further reveals insight into the mechanism by which the well-defined surface chemistry of ALD may eventually be circumvented to allow metal oxide nucleation and growth on SAM-modified surfaces.« less
Inhibiting Metal Oxide Atomic Layer Deposition: Beyond Zinc Oxide
DOE Office of Scientific and Technical Information (OSTI.GOV)
Sampson, Matthew D.; Emery, Jonathan D.; Pellin, Michael J.
The atomic layer deposition (ALD) of several metal oxides is selectivity inhibited on alkanethiol self-assembled monolayers (SAMs) on Au and the eventual nucleation mechanism is investigated. The inhibition ability of the SAM is significantly improved by the in situ H 2-plasma pretreatment of the Au substrate prior to gas-phase deposition of a long-chain alkanethiol, 1-dodecanethiol (DDT). This more rigorous surface preparation inhibits even aggressive oxide ALD precursors, including trimethylaluminum and water, for at least 20 cycles. We study the effect that ALD precursor purge times, growth temperature, alkanethiol chain length, alkanethiol deposition time, and plasma treatment time have on Almore » 2O 3 ALD inhibition. This is the first example of Al 2O 3 ALD inhibition from a vapor-deposited SAM. Inhibition of Al 2O 3, ZnO, and MnO ALD processes are compared, revealing the versatility of this selective surface treatment. As a result, atomic force microscopy (AFM) and grazing incidence x-ray fluorescence (GIXRF) further reveals insight into the mechanism by which the well-defined surface chemistry of ALD may eventually be circumvented to allow metal oxide nucleation and growth on SAM-modified surfaces.« less
Climate change vulnerability assessment for the Chugach National Forest and the Kenai Peninsula
Gregory H. Hayward; Steve Colt; Monica L. McTeague; Teresa N. Hollingsworth
2017-01-01
This assessment evaluates the effects of future climate change on a select set of ecological systems and ecosystem services in Alaskaâs Kenai Peninsula and Chugach National Forest regions. The focus of the assessment was established during a multi-agency/organization workshop that established the goal to conduct a rigorous evaluation of a limited range of topics rather...
Impact of insects on multiple-use values of north-central forests: an experimental rating scheme.
Norton D. Addy; Harold O. Batzer; William J. Mattson; William E. Miller
1971-01-01
Ranking or assigning priorities to problems is an essential step in research problem selection. Up to now, no rigorous basis for ranking forest insects has been available. We evaluate and rank forest insects with a systematic numerical scheme that considers insect impact on the multiple-use values of timber, wildlife, recreation, and water. The result is a better...
Communicating the rigor behind science results
NASA Astrophysics Data System (ADS)
Jackson, R.; Callery, S.
2015-12-01
Communicating the rigor behind science resultsNASA's Global Climate Change website and companion Facebook page have an exceptionally large reach. Moderating the vast quantity of questions, feedback and comments from these public platforms has provided a unique perspective on the way the public views science, the scientific method and how science is funded. Email feedback and social media interactions reveal widespread misperceptions about how science is carried out: There is considerable criticism and suspicion surrounding methods of funding, and the difficulty of obtaining grants is underestimated. There appears to be limited public awareness of the peer review process. This talk will highlight the need for better communication not only of science results, but the process of science--from proposal writing and getting funded to peer-review and fundamental science terminology. As a community of science communicators, we also need to highlight the inaccuracies sometimes introduced by media reports of peer-reviewed science papers.
Eslick, John C.; Ng, Brenda; Gao, Qianwen; ...
2014-12-31
Under the auspices of the U.S. Department of Energy’s Carbon Capture Simulation Initiative (CCSI), a Framework for Optimization and Quantification of Uncertainty and Sensitivity (FOQUS) has been developed. This tool enables carbon capture systems to be rapidly synthesized and rigorously optimized, in an environment that accounts for and propagates uncertainties in parameters and models. FOQUS currently enables (1) the development of surrogate algebraic models utilizing the ALAMO algorithm, which can be used for superstructure optimization to identify optimal process configurations, (2) simulation-based optimization utilizing derivative free optimization (DFO) algorithms with detailed black-box process models, and (3) rigorous uncertainty quantification throughmore » PSUADE. FOQUS utilizes another CCSI technology, the Turbine Science Gateway, to manage the thousands of simulated runs necessary for optimization and UQ. Thus, this computational framework has been demonstrated for the design and analysis of a solid sorbent based carbon capture system.« less
NASA Astrophysics Data System (ADS)
Määttä, A.; Laine, M.; Tamminen, J.; Veefkind, J. P.
2014-05-01
Satellite instruments are nowadays successfully utilised for measuring atmospheric aerosol in many applications as well as in research. Therefore, there is a growing need for rigorous error characterisation of the measurements. Here, we introduce a methodology for quantifying the uncertainty in the retrieval of aerosol optical thickness (AOT). In particular, we concentrate on two aspects: uncertainty due to aerosol microphysical model selection and uncertainty due to imperfect forward modelling. We apply the introduced methodology for aerosol optical thickness retrieval of the Ozone Monitoring Instrument (OMI) on board NASA's Earth Observing System (EOS) Aura satellite, launched in 2004. We apply statistical methodologies that improve the uncertainty estimates of the aerosol optical thickness retrieval by propagating aerosol microphysical model selection and forward model error more realistically. For the microphysical model selection problem, we utilise Bayesian model selection and model averaging methods. Gaussian processes are utilised to characterise the smooth systematic discrepancies between the measured and modelled reflectances (i.e. residuals). The spectral correlation is composed empirically by exploring a set of residuals. The operational OMI multi-wavelength aerosol retrieval algorithm OMAERO is used for cloud-free, over-land pixels of the OMI instrument with the additional Bayesian model selection and model discrepancy techniques introduced here. The method and improved uncertainty characterisation is demonstrated by several examples with different aerosol properties: weakly absorbing aerosols, forest fires over Greece and Russia, and Sahara desert dust. The statistical methodology presented is general; it is not restricted to this particular satellite retrieval application.
Rigorous force field optimization principles based on statistical distance minimization
Vlcek, Lukas; Chialvo, Ariel A.
2015-10-12
We use the concept of statistical distance to define a measure of distinguishability between a pair of statistical mechanical systems, i.e., a model and its target, and show that its minimization leads to general convergence of the model’s static measurable properties to those of the target. Here we exploit this feature to define a rigorous basis for the development of accurate and robust effective molecular force fields that are inherently compatible with coarse-grained experimental data. The new model optimization principles and their efficient implementation are illustrated through selected examples, whose outcome demonstrates the higher robustness and predictive accuracy of themore » approach compared to other currently used methods, such as force matching and relative entropy minimization. We also discuss relations between the newly developed principles and established thermodynamic concepts, which include the Gibbs-Bogoliubov inequality and the thermodynamic length.« less
Specifying the behavior of concurrent systems
NASA Technical Reports Server (NTRS)
Furtek, F. C.
1984-01-01
A framework for rigorously specifying the behavior of concurrent systems is proposed. It is based on the view of a concurrent system as a collection of interacting processes but no assumptions are made about the mechanisms for process synchronization and communication. A formal language is described that permits the expression of a broad range of logical and timing dependencies.
ERIC Educational Resources Information Center
Dhirathiti, Nopraenue S.; Pichitpatja, Pojjana
2018-01-01
The study examined the process of policy implementation of lifelong learning for the elderly in Thailand, covering four main regions within the country. The study empirically compared inputs, processes, outputs, and outcomes of policy implementation in the north, south, northeast, and central regions of Thailand and captured the rigor of policy…
Using the Inquiry Process to Motivate and Engage All (Including Struggling) Readers
ERIC Educational Resources Information Center
Savitz, Rachelle S.; Wallace, Kelly
2016-01-01
With increasingly rigorous standards and mounting high stakes testing, it seems harder than ever to motivate and engage struggling readers. In this article the authors provide an overview of the inquiry learning process, which details how providing students with choice and opportunities to collaborate with peers can keep students invested in their…
Academic Excellence: A Commentary and Reflections on the Inherent Value of Peer Review
ERIC Educational Resources Information Center
Roberts, Thomas J.; Shambrook, Jennifer
2012-01-01
Academic peer review is widely viewed as fair, equitable, and essential to academic quality. Successfully completing the process through publication or award is widely deemed as one of the most rigorous and prestigious forms of scholarly accomplishment. Despite this sentiment the academic peer review process is not without fault. It is criticized…
Considerations for the Systematic Analysis and Use of Single-Case Research
ERIC Educational Resources Information Center
Horner, Robert H.; Swaminathan, Hariharan; Sugai, George; Smolkowski, Keith
2012-01-01
Single-case research designs provide a rigorous research methodology for documenting experimental control. If single-case methods are to gain wider application, however, a need exists to define more clearly (a) the logic of single-case designs, (b) the process and decision rules for visual analysis, and (c) an accepted process for integrating…
Bringing scientific rigor to community-developed programs in Hong Kong.
Fabrizio, Cecilia S; Hirschmann, Malia R; Lam, Tai Hing; Cheung, Teresa; Pang, Irene; Chan, Sophia; Stewart, Sunita M
2012-12-31
This paper describes efforts to generate evidence for community-developed programs to enhance family relationships in the Chinese culture of Hong Kong, within the framework of community-based participatory research (CBPR). The CBPR framework was applied to help maximize the development of the intervention and the public health impact of the studies, while enhancing the capabilities of the social service sector partners. Four academic-community research teams explored the process of designing and implementing randomized controlled trials in the community. In addition to the expected cultural barriers between teams of academics and community practitioners, with their different outlooks, concerns and languages, the team navigated issues in utilizing the principles of CBPR unique to this Chinese culture. Eventually the team developed tools for adaptation, such as an emphasis on building the relationship while respecting role delineation and an iterative process of defining the non-negotiable parameters of research design while maintaining scientific rigor. Lessons learned include the risk of underemphasizing the size of the operational and skills shift between usual agency practices and research studies, the importance of minimizing non-negotiable parameters in implementing rigorous research designs in the community, and the need to view community capacity enhancement as a long term process. The four pilot studies under the FAMILY Project demonstrated that nuanced design adaptations, such as wait list controls and shorter assessments, better served the needs of the community and led to the successful development and vigorous evaluation of a series of preventive, family-oriented interventions in the Chinese culture of Hong Kong.
Gentles, Stephen J; Charles, Cathy; Nicholas, David B; Ploeg, Jenny; McKibbon, K Ann
2016-10-11
Overviews of methods are potentially useful means to increase clarity and enhance collective understanding of specific methods topics that may be characterized by ambiguity, inconsistency, or a lack of comprehensiveness. This type of review represents a distinct literature synthesis method, although to date, its methodology remains relatively undeveloped despite several aspects that demand unique review procedures. The purpose of this paper is to initiate discussion about what a rigorous systematic approach to reviews of methods, referred to here as systematic methods overviews, might look like by providing tentative suggestions for approaching specific challenges likely to be encountered. The guidance offered here was derived from experience conducting a systematic methods overview on the topic of sampling in qualitative research. The guidance is organized into several principles that highlight specific objectives for this type of review given the common challenges that must be overcome to achieve them. Optional strategies for achieving each principle are also proposed, along with discussion of how they were successfully implemented in the overview on sampling. We describe seven paired principles and strategies that address the following aspects: delimiting the initial set of publications to consider, searching beyond standard bibliographic databases, searching without the availability of relevant metadata, selecting publications on purposeful conceptual grounds, defining concepts and other information to abstract iteratively, accounting for inconsistent terminology used to describe specific methods topics, and generating rigorous verifiable analytic interpretations. Since a broad aim in systematic methods overviews is to describe and interpret the relevant literature in qualitative terms, we suggest that iterative decision making at various stages of the review process, and a rigorous qualitative approach to analysis are necessary features of this review type. We believe that the principles and strategies provided here will be useful to anyone choosing to undertake a systematic methods overview. This paper represents an initial effort to promote high quality critical evaluations of the literature regarding problematic methods topics, which have the potential to promote clearer, shared understandings, and accelerate advances in research methods. Further work is warranted to develop more definitive guidance.
FORMAL SCENARIO DEVELOPMENT FOR ENVIRONMENTAL IMPACT ASSESSMENT STUDIES
Scenario analysis is a process of evaluating possible future events through the consideration of alternative plausible (though not equally likely) outcomes (scenarios). The analysis is designed to enable improved decision-making and assessment through a more rigorous evaluation o...
Evaluation of reference genes for insect olfaction studies.
Omondi, Bonaventure Aman; Latorre-Estivalis, Jose Manuel; Rocha Oliveira, Ivana Helena; Ignell, Rickard; Lorenzo, Marcelo Gustavo
2015-04-22
Quantitative reverse transcription PCR (qRT-PCR) is a robust and accessible method to assay gene expression and to infer gene regulation. Being a chain of procedures, this technique is subject to systematic error due to biological and technical limitations mainly set by the starting material and downstream procedures. Thus, rigorous data normalization is critical to grant reliability and repeatability of gene expression quantification by qRT-PCR. A number of 'housekeeping genes', involved in basic cellular functions, have been commonly used as internal controls for this normalization process. However, these genes could themselves be regulated and must therefore be tested a priori. We evaluated eight potential reference genes for their stability as internal controls for RT-qPCR studies of olfactory gene expression in the antennae of Rhodnius prolixus, a Chagas disease vector. The set of genes included were: α-tubulin; β-actin; Glyceraldehyde-3-phosphate dehydrogenase; Eukaryotic initiation factor 1A; Glutathione-S-transferase; Serine protease; Succinate dehydrogenase; and Glucose-6-phosphate dehydrogenase. Five experimental conditions, including changes in age,developmental stage and feeding status were tested in both sexes. We show that the evaluation of candidate reference genes is necessary for each combination of sex, tissue and physiological condition analyzed in order to avoid inconsistent results and conclusions. Although, Normfinder and geNorm software yielded different results between males and females, five genes (SDH, Tub, GAPDH, Act and G6PDH) appeared in the first positions in all rankings obtained. By using gene expression data of a single olfactory coreceptor gene as an example, we demonstrated the extent of changes expected using different internal standards. This work underlines the need for a rigorous selection of internal standards to grant the reliability of normalization processes in qRT-PCR studies. Furthermore, we show that particular physiological or developmental conditions require independent evaluation of a diverse set of potential reference genes.
NASA Astrophysics Data System (ADS)
Hubbell, Jody M.
This study explored three selected phases of Rogers' (1995) Diffusion of Innovations Theory to examine the diffusion process of the distance Entomology Master's Degree program at the University of Nebraska, Lincoln. A qualitative descriptive case study approach incorporated semi-structured interviews with individuals involved in one or more of the three stages: Development, Implementation, and Institutionalization. Documents and archival evidence were used to triangulate findings. This research analyzed descriptions of the program as it moved from the Development, to the Implementation, and finally, the Institutionalization stages of diffusion. Each respective stage was examined through open and axial coding. Process coding identified themes common to two or more diffusion stages, and explored the evolution of themes from one diffusion stage to the next. At a time of significant budget constraints, many departments were faced with the possibility of merger or dissolution. The Entomology Master's Degree Program evolved from being an entrepreneurial means to prevent departmental dissolution to eventually being viewed as a model for the development of similar programs across this university and other institutions of higher education. During this evolution, the program was reinvented to meet the broader needs of industry and a global student market. One finding not consistent with Rogers' model was that smaller, rather than larger, departmental size contributed to the success of the program. Within this small department, faculty members were able to share their experiences and knowledge with each other on a regular basis, which promoted greater acceptance of the distance program. How quality and rigor may be defined and measured was a key issue in each respective stage. In this specific case, quality and rigor was initially a comparison of on-campus and distance course content and then moved to program-based assessment and measures of student outcomes such as job placement rates.
Development of a software safety process and a case study of its use
NASA Technical Reports Server (NTRS)
Knight, John C.
1993-01-01
The goal of this research is to continue the development of a comprehensive approach to software safety and to evaluate the approach with a case study. The case study is a major part of the project, and it involves the analysis of a specific safety-critical system from the medical equipment domain. The particular application being used was selected because of the availability of a suitable candidate system. We consider the results to be generally applicable and in no way particularly limited by the domain. The research is concentrating on issues raised by the specification and verification phases of the software lifecycle since they are central to our previously-developed rigorous definitions of software safety. The theoretical research is based on our framework of definitions for software safety. In the area of specification, the main topics being investigated are the development of techniques for building system fault trees that correctly incorporate software issues and the development of rigorous techniques for the preparation of software safety specifications. The research results are documented. Another area of theoretical investigation is the development of verification methods tailored to the characteristics of safety requirements. Verification of the correct implementation of the safety specification is central to the goal of establishing safe software. The empirical component of this research is focusing on a case study in order to provide detailed characterizations of the issues as they appear in practice, and to provide a testbed for the evaluation of various existing and new theoretical results, tools, and techniques. The Magnetic Stereotaxis System is summarized.
Maelstrom Research guidelines for rigorous retrospective data harmonization
Fortier, Isabel; Raina, Parminder; Van den Heuvel, Edwin R; Griffith, Lauren E; Craig, Camille; Saliba, Matilda; Doiron, Dany; Stolk, Ronald P; Knoppers, Bartha M; Ferretti, Vincent; Granda, Peter; Burton, Paul
2017-01-01
Abstract Background: It is widely accepted and acknowledged that data harmonization is crucial: in its absence, the co-analysis of major tranches of high quality extant data is liable to inefficiency or error. However, despite its widespread practice, no formalized/systematic guidelines exist to ensure high quality retrospective data harmonization. Methods: To better understand real-world harmonization practices and facilitate development of formal guidelines, three interrelated initiatives were undertaken between 2006 and 2015. They included a phone survey with 34 major international research initiatives, a series of workshops with experts, and case studies applying the proposed guidelines. Results: A wide range of projects use retrospective harmonization to support their research activities but even when appropriate approaches are used, the terminologies, procedures, technologies and methods adopted vary markedly. The generic guidelines outlined in this article delineate the essentials required and describe an interdependent step-by-step approach to harmonization: 0) define the research question, objectives and protocol; 1) assemble pre-existing knowledge and select studies; 2) define targeted variables and evaluate harmonization potential; 3) process data; 4) estimate quality of the harmonized dataset(s) generated; and 5) disseminate and preserve final harmonization products. Conclusions: This manuscript provides guidelines aiming to encourage rigorous and effective approaches to harmonization which are comprehensively and transparently documented and straightforward to interpret and implement. This can be seen as a key step towards implementing guiding principles analogous to those that are well recognised as being essential in securing the foundational underpinning of systematic reviews and the meta-analysis of clinical trials. PMID:27272186
Lenas, Petros; Moos, Malcolm; Luyten, Frank P
2009-12-01
The field of tissue engineering is moving toward a new concept of "in vitro biomimetics of in vivo tissue development." In Part I of this series, we proposed a theoretical framework integrating the concepts of developmental biology with those of process design to provide the rules for the design of biomimetic processes. We named this methodology "developmental engineering" to emphasize that it is not the tissue but the process of in vitro tissue development that has to be engineered. To formulate the process design rules in a rigorous way that will allow a computational design, we should refer to mathematical methods to model the biological process taking place in vitro. Tissue functions cannot be attributed to individual molecules but rather to complex interactions between the numerous components of a cell and interactions between cells in a tissue that form a network. For tissue engineering to advance to the level of a technologically driven discipline amenable to well-established principles of process engineering, a scientifically rigorous formulation is needed of the general design rules so that the behavior of networks of genes, proteins, or cells that govern the unfolding of developmental processes could be related to the design parameters. Now that sufficient experimental data exist to construct plausible mathematical models of many biological control circuits, explicit hypotheses can be evaluated using computational approaches to facilitate process design. Recent progress in systems biology has shown that the empirical concepts of developmental biology that we used in Part I to extract the rules of biomimetic process design can be expressed in rigorous mathematical terms. This allows the accurate characterization of manufacturing processes in tissue engineering as well as the properties of the artificial tissues themselves. In addition, network science has recently shown that the behavior of biological networks strongly depends on their topology and has developed the necessary concepts and methods to describe it, allowing therefore a deeper understanding of the behavior of networks during biomimetic processes. These advances thus open the door to a transition for tissue engineering from a substantially empirical endeavor to a technology-based discipline comparable to other branches of engineering.
[Charles Darwin and the problem of evolutionary progress].
Iordanskiĭ, N N
2010-01-01
According to Ch. Darwin's evolutionary theory, evolutionary progress (interpreted as morpho-physiological progress or arogenesis in recent terminology) is one of logical results of natural selection. At the same time, natural selection does not hold any factors especially promoting evolutionary progress. Darwin emphasized that the pattern of evolutionary changes depends on organism nature more than on the pattern of environment changes. Arogenesis specificity is determined by organization of rigorous biological systems - integral organisms. Onward progressive development is determined by fundamental features of living organisms: metabolism and homeostasis. The concept of social Darwinism differs fundamentally from Darwin's ideas about the most important role of social instincts in progress of mankind. Competition and selection play secondary role in socio-cultural progress of human society.
Thiros, Susan A.; Bender, David A.; Mueller, David K.; Rose, Donna L.; Olsen, Lisa D.; Martin, Jeffrey D.; Bernard, Bruce; Zogorski, John S.
2011-01-01
The Field Contamination Study (FCS) was designed to determine the field processes that tend to result in clean field blanks and to identify potential sources of contamination to blanks collected in the field from selected volatile organic compounds (VOCs) and wastewater-indicator compounds (WICs). The VOCs and WICs analyzed in the FCS were detected in blanks collected by the U.S. Geological Survey (USGS) National Water-Quality Assessment (NAWQA) Program during 1996–2008 and 2002–08, respectively. To minimize the number of variables, the study required ordering of supplies just before sampling, storage of supplies and equipment in clean areas, and use of adequate amounts of purge-and-trap volatile-grade methanol and volatile pesticide-grade blank water (VPBW) to clean sampling equipment and to collect field blanks.Blanks and groundwater samples were collected during 2008–09 at 16 sites, which were a mix of water-supply and monitoring wells, located in 9 States. Five different sample types were collected for the FCS at each site: (1) a source-solution blank collected at the USGS National Water Quality Laboratory (NWQL) using laboratory-purged VPBW, (2) source-solution blanks collected in the field using laboratory-purged VPBW, (3) source-solution blanks collected in the field using field-purged VPBW, (4) a field blank collected using field-purged VPBW, and (5) a groundwater sample collected from a well. The source-solution blank and field-blank analyses were used to identify, quantify, and document extrinsic contamination and to help determine the sources and causes of data-quality problems that can affect groundwater samples.Concentrations of compounds detected in FCS analyses were quantified and results were stored in the USGS National Water Information System database after meeting rigorous identification and quantification criteria. The study also utilized information provided by laboratory analysts about evidence indicating the presence of selected compounds, using less rigorous identification criteria than is required for reporting data to the National Water Information System database. For the FCS, these data are considered adequate to indicate "evidence of presence," and were used only for diagnostic purposes. Evidence of VOCs and WICs at low concentrations near or less than the long-term method detection level can indicate a contamination problem that could affect future datasets if method detection levels were ever to be lowered.
ERIC Educational Resources Information Center
Touchstone, Allison J. L.
2010-01-01
Dual credit programs have become increasingly popular with 71% U.S. public high schools offering dual credit courses in 2002-2003. As this popularity has grown, so have concerns regarding academic rigor, course quality, parity with college courses, and effects on higher education. Determining actual dual credit course equivalent in higher…
High-Temperature Alloys for Automotive Stirling Engines
NASA Technical Reports Server (NTRS)
Stephens, J. R.; Titran, R. H.
1986-01-01
Stirling engine is external-combustion engine that offers fuel economy, low emissions, low noise, and low vibrations. One of most critical areas in engine development concerns material selection for component parts. Alloys CG-27 and XF-818 identified capable of withstanding rigorous requirements of automotive Stirling engine. Alloys chosen for availability, performance, and manufacturability. Advanced iron-base alloys have potential for variety of applications, including stationary solar-power systems.
Krompecher, T
1981-01-01
Objective measurements were carried out to study the evolution of rigor mortis on rats at various temperatures. Our experiments showed that: (1) at 6 degrees C rigor mortis reaches full development between 48 and 60 hours post mortem, and is resolved at 168 hours post mortem; (2) at 24 degrees C rigor mortis reaches full development at 5 hours post mortem, and is resolved at 16 hours post mortem; (3) at 37 degrees C rigor mortis reaches full development at 3 hours post mortem, and is resolved at 6 hours post mortem; (4) the intensity of rigor mortis grows with increase in temperature (difference between values obtained at 24 degrees C and 37 degrees C); and (5) and 6 degrees C a "cold rigidity" was found, in addition to and independent of rigor mortis.
Graphical Descriptives: A Way to Improve Data Transparency and Methodological Rigor in Psychology.
Tay, Louis; Parrigon, Scott; Huang, Qiming; LeBreton, James M
2016-09-01
Several calls have recently been issued to the social sciences for enhanced transparency of research processes and enhanced rigor in the methodological treatment of data and data analytics. We propose the use of graphical descriptives (GDs) as one mechanism for responding to both of these calls. GDs provide a way to visually examine data. They serve as quick and efficient tools for checking data distributions, variable relations, and the potential appropriateness of different statistical analyses (e.g., do data meet the minimum assumptions for a particular analytic method). Consequently, we believe that GDs can promote increased transparency in the journal review process, encourage best practices for data analysis, and promote a more inductive approach to understanding psychological data. We illustrate the value of potentially including GDs as a step in the peer-review process and provide a user-friendly online resource (www.graphicaldescriptives.org) for researchers interested in including data visualizations in their research. We conclude with suggestions on how GDs can be expanded and developed to enhance transparency. © The Author(s) 2016.
Coupling of Rigor Mortis and Intestinal Necrosis during C. elegans Organismal Death.
Galimov, Evgeniy R; Pryor, Rosina E; Poole, Sarah E; Benedetto, Alexandre; Pincus, Zachary; Gems, David
2018-03-06
Organismal death is a process of systemic collapse whose mechanisms are less well understood than those of cell death. We previously reported that death in C. elegans is accompanied by a calcium-propagated wave of intestinal necrosis, marked by a wave of blue autofluorescence (death fluorescence). Here, we describe another feature of organismal death, a wave of body wall muscle contraction, or death contraction (DC). This phenomenon is accompanied by a wave of intramuscular Ca 2+ release and, subsequently, of intestinal necrosis. Correlation of directions of the DC and intestinal necrosis waves implies coupling of these death processes. Long-lived insulin/IGF-1-signaling mutants show reduced DC and delayed intestinal necrosis, suggesting possible resistance to organismal death. DC resembles mammalian rigor mortis, a postmortem necrosis-related process in which Ca 2+ influx promotes muscle hyper-contraction. In contrast to mammals, DC is an early rather than a late event in C. elegans organismal death. VIDEO ABSTRACT. Copyright © 2018 The Author(s). Published by Elsevier Inc. All rights reserved.
Rimington, Rowan P; Capel, Andrew J; Player, Darren J; Bibb, Richard J; Christie, Steven D R; Lewis, Mark P
2018-06-13
The integration of additive manufacturing (AM) technology within biological systems holds significant potential, specifically when refining the methods utilized for the creation of in vitro models. Therefore, examination of cellular interaction with the physical/physicochemical properties of 3D-printed polymers is critically important. In this work, skeletal muscle (C 2 C 12 ), neuronal (SH-SY5Y) and hepatic (HepG2) cell lines are utilized to ascertain critical evidence of cellular behavior in response to 3D-printed candidate polymers: Clear-FL (stereolithography, SL), PA-12 (laser sintering, LS), and VeroClear (PolyJet). This research outlines initial critical evidence for a framework of polymer/AM process selection when 3D printing biologically receptive scaffolds, derived from industry standard, commercially available AM instrumentation. C 2 C 12 , SH-SY5Y, and HepG2 cells favor LS polymer PA-12 for applications in which cellular adherence is necessitated. However, cell type specific responses are evident when cultured in the chemical leachate of photopolymers (Clear-FL and VeroClear). With the increasing prevalence of 3D-printed biointerfaces, the development of rigorous cell type specific biocompatibility data is imperative. Supplementing the currently limited database of functional 3D-printed biomaterials affords the opportunity for experiment-specific AM process and polymer selection, dependent on biological application and intricacy of design features required. © 2018 WILEY-VCH Verlag GmbH & Co. KGaA, Weinheim.
ERIC Educational Resources Information Center
Francis, Clay
2018-01-01
Historic notions of academic rigor usually follow from critiques of the system--we often define our goals for academically rigorous work through the lens of our shortcomings. This chapter discusses how the Truman Commission in 1947 and the Spellings Commission in 2006 shaped the way we think about academic rigor in today's context.
Considerations for designing chemical screening strategies in plant biology
Serrano, Mario; Kombrink, Erich; Meesters, Christian
2015-01-01
Traditionally, biologists regularly used classical genetic approaches to characterize and dissect plant processes. However, this strategy is often impaired by redundancy, lethality or pleiotropy of gene functions, which prevent the isolation of viable mutants. The chemical genetic approach has been recognized as an alternative experimental strategy, which has the potential to circumvent these problems. It relies on the capacity of small molecules to modify biological processes by specific binding to protein target(s), thereby conditionally modifying protein function(s), which phenotypically resemble mutation(s) of the encoding gene(s). A successful chemical screening campaign comprises three equally important elements: (1) a reliable, robust, and quantitative bioassay, which allows to distinguish between potent and less potent compounds, (2) a rigorous validation process for candidate compounds to establish their selectivity, and (3) an experimental strategy for elucidating a compound's mode of action and molecular target. In this review we will discuss details of this general strategy and additional aspects that deserve consideration in order to take full advantage of the power provided by the chemical approach to plant biology. In addition, we will highlight some success stories of recent chemical screenings in plant systems, which may serve as teaching examples for the implementation of future chemical biology projects. PMID:25904921
Modelling Evolutionary Algorithms with Stochastic Differential Equations.
Heredia, Jorge Pérez
2017-11-20
There has been renewed interest in modelling the behaviour of evolutionary algorithms (EAs) by more traditional mathematical objects, such as ordinary differential equations or Markov chains. The advantage is that the analysis becomes greatly facilitated due to the existence of well established methods. However, this typically comes at the cost of disregarding information about the process. Here, we introduce the use of stochastic differential equations (SDEs) for the study of EAs. SDEs can produce simple analytical results for the dynamics of stochastic processes, unlike Markov chains which can produce rigorous but unwieldy expressions about the dynamics. On the other hand, unlike ordinary differential equations (ODEs), they do not discard information about the stochasticity of the process. We show that these are especially suitable for the analysis of fixed budget scenarios and present analogues of the additive and multiplicative drift theorems from runtime analysis. In addition, we derive a new more general multiplicative drift theorem that also covers non-elitist EAs. This theorem simultaneously allows for positive and negative results, providing information on the algorithm's progress even when the problem cannot be optimised efficiently. Finally, we provide results for some well-known heuristics namely Random Walk (RW), Random Local Search (RLS), the (1+1) EA, the Metropolis Algorithm (MA), and the Strong Selection Weak Mutation (SSWM) algorithm.
CMMI Level 5 and the Team Software Process
2007-04-01
could meet the rigors of a CMMI assessment and achieve their group’s goal of Level 5. Watts Humphrey , who is widely acknowledged as the founder of the...Capability Maturity Model® (CMM®) approach to improvement and who later created the Personal Software Process ( PSP )SM and TSP, has noted that one of the...intents of PSP and TSP is to be an operational process enactment of CMM Level 5 processes at the personal and pro- ject levels respectively [1]. CMM
Emergency cricothyrotomy for trismus caused by instantaneous rigor in cardiac arrest patients.
Lee, Jae Hee; Jung, Koo Young
2012-07-01
Instantaneous rigor as muscle stiffening occurring in the moment of death (or cardiac arrest) can be confused with rigor mortis. If trismus is caused by instantaneous rigor, orotracheal intubation is impossible and a surgical airway should be secured. Here, we report 2 patients who had emergency cricothyrotomy for trismus caused by instantaneous rigor. This case report aims to help physicians understand instantaneous rigor and to emphasize the importance of securing a surgical airway quickly on the occurrence of trismus. Copyright © 2012 Elsevier Inc. All rights reserved.
Investigating the Utility of a GPA Institutional Adjustment Index
ERIC Educational Resources Information Center
Didier, Thomas; Kreiter, Clarence D.; Buri, Russell; Solow, Catherine
2006-01-01
Background: Grading standards vary widely across undergraduate institutions. If, during the medical school admissions process, GPA is considered without reference to the institution attended, it will disadvantage applicants from undergraduate institutions employing rigorous grading standards. Method: A regression-based GPA institutional equating…
2004-09-01
excluded) do not apply a rigorous and disciplined process of game development , especially in the area of objective definition. Without clear...prior to the game, in controlled circumstances. The actual “game” becomes a discussion of the very orchestrated results of the pre- game “ development ” efforts
Telecourses: 20 Myths, 21 Realities.
ERIC Educational Resources Information Center
Luskin, Bernard J.
1983-01-01
Refutes 20 myths about telecourses, e.g., telecourses require only television viewing, they are easy, they will replace classroom teachers and all courses on campus, they dehumanize learning, and they do not undergo a rigorous curriculum process, and are too expensive to produce and offer. (DMM)
Modelling Tradeoffs Evolution in Multipurpose Water Systems Operation in Response to Extreme Events
NASA Astrophysics Data System (ADS)
Mason, E.; Gazzotti, P.; Amigoni, F.; Giuliani, M.; Castelletti, A.
2015-12-01
Multipurpose water resource systems are usually operated on a tradeoff of the operating objectives, which - under steady state climatic and socio-economic boundary conditions - is supposed to ensure a fair and/or efficient balance among the conflicting interests. Extreme variability in the system's drivers might affect operators' risk aversion and force a change in the tradeoff. Properly accounting for these shifts is key to any rigorous retrospective assessment of operators' behavior and the associated system's performance. In this study, we explore how the selection of different optimal tradeoffs among the operating objectives is linked to the variations of the boundary conditions, such as, for example, drifting rainfall season or remarkable changes in crop and energy prices. We argue that tradeoff selection is driven by recent, extreme variations in system performance: underperforming on one of the operating objective target value should push the tradeoff toward the disadvantaged objective. To test this assumption, we developed a rational procedure to simulate the operators' tradeoff selection process. We map the selection onto a multi lateral negotiation process, where different multiple, virtual agents optimize different operating objectives. The agents periodically negotiate a compromise on the operating policy. The agent's rigidity in each negotiation round is determined by the recent system performances according to the specific objective it represents. The negotiation follows a set-based egocentric monotonic concession protocol: at each negotiation step an agent incrementally adds some options to the set of its acceptable compromises and (possibly) accepts lower and lower satisfying policies until an agreement is achieved. We apply this reiterated negotiation framework on the regulated Lake Como, Italy, simulating the lake dam operation and its recurrent updates over the last 50 years. The operation aims to balance shoreline flood prevention and irrigation deficit control in the downstream irrigated areas. The results of our simulated negotiations are able to accurately capture the operator's risk aversion changes as driven by extreme wet and dry situations, and to well reproduce the observational release data.
Loescher, Lois J; Rains, Stephen A; Kramer, Sandra S; Akers, Chelsie; Moussa, Renee
2018-05-01
To systematically review healthy lifestyle interventions targeted to adolescents and delivered using text messaging (TM). PubMed, Embase, CINAHL, PsycINFO, and Web of Science databases. Study Inclusion Criteria: Research articles published during 2011 to 2014; analyses focused on intervention targeting adolescents (10-19 years), with healthy lifestyle behaviors as main variables, delivered via mobile phone-based TM. The authors extracted data from 27 of 281 articles using the Preferred Reporting Items for Systematic Reviews and Meta-Analyses method. Adolescent and setting characteristics, study design and rigor, intervention effectiveness, challenges, and risk of bias. Across studies, 16 (59.3%) of 27 included non-Caucasians. The gender was split for 22 (81.5%) of 27 studies. Thirteen studies were randomized controlled trials. There was heterogeneity among targeted conditions, rigor of methods, and intervention effects. Interventions for monitoring/adherence (n = 8) reported more positive results than those for health behavior change (n = 19). Studies that only included message delivered via TM (n = 14) reported more positive effects than studies integrating multiple intervention components. Interventions delivered using TM presented minimal challenges, but selection and performance bias were observed across studies. Interventions delivered using TM have the potential, under certain conditions, to improve healthy lifestyle behaviors in adolescents. However, the rigor of studies varies, and established theory and validated measures have been inconsistently incorporated.
Kim, Hyun-Wook; Hwang, Ko-Eun; Song, Dong-Heon; Kim, Yong-Jae; Ham, Youn-Kyung; Yeo, Eui-Joo; Jeong, Tae-Jun; Choi, Yun-Sang; Kim, Cheon-Jei
2015-01-01
This study was conducted to evaluate the effect of pre-rigor salting level (0-4% NaCl concentration) on physicochemical and textural properties of pre-rigor chicken breast muscles. The pre-rigor chicken breast muscles were de-boned 10 min post-mortem and salted within 25 min post-mortem. An increase in pre-rigor salting level led to the formation of high ultimate pH of chicken breast muscles at post-mortem 24 h. The addition of minimum of 2% NaCl significantly improved water holding capacity, cooking loss, protein solubility, and hardness when compared to the non-salting chicken breast muscle (p<0.05). On the other hand, the increase in pre-rigor salting level caused the inhibition of myofibrillar protein degradation and the acceleration of lipid oxidation. However, the difference in NaCl concentration between 3% and 4% had no great differences in the results of physicochemical and textural properties due to pre-rigor salting effects (p>0.05). Therefore, our study certified the pre-rigor salting effect of chicken breast muscle salted with 2% NaCl when compared to post-rigor muscle salted with equal NaCl concentration, and suggests that the 2% NaCl concentration is minimally required to ensure the definite pre-rigor salting effect on chicken breast muscle.
Choi, Yun-Sang
2015-01-01
This study was conducted to evaluate the effect of pre-rigor salting level (0-4% NaCl concentration) on physicochemical and textural properties of pre-rigor chicken breast muscles. The pre-rigor chicken breast muscles were de-boned 10 min post-mortem and salted within 25 min post-mortem. An increase in pre-rigor salting level led to the formation of high ultimate pH of chicken breast muscles at post-mortem 24 h. The addition of minimum of 2% NaCl significantly improved water holding capacity, cooking loss, protein solubility, and hardness when compared to the non-salting chicken breast muscle (p<0.05). On the other hand, the increase in pre-rigor salting level caused the inhibition of myofibrillar protein degradation and the acceleration of lipid oxidation. However, the difference in NaCl concentration between 3% and 4% had no great differences in the results of physicochemical and textural properties due to pre-rigor salting effects (p>0.05). Therefore, our study certified the pre-rigor salting effect of chicken breast muscle salted with 2% NaCl when compared to post-rigor muscle salted with equal NaCl concentration, and suggests that the 2% NaCl concentration is minimally required to ensure the definite pre-rigor salting effect on chicken breast muscle. PMID:26761884
Robert M. Scheller; James B. Domingo; Brian R. Sturtevant; Jeremy S. Williams; Arnold Rudy; Eric J. Gustafson; David J. Mladenoff
2007-01-01
We introduce LANDIS-II, a landscape model designed to simulate forest succession and disturbances. LANDIS-II builds upon and preserves the functionality of previous LANDIS forest landscape simulation models. LANDIS-II is distinguished by the inclusion of variable time steps for different ecological processes; our use of a rigorous development and testing process used...
A New Look at the Code of Conduct
1974-04-01
often kept apart from the cthers in solitary confinement. Prisoners were fed two meals each day, usually consisting of pumpkin or cab- bage soup, rice...capture, other "early release" prisoners appeared to have been summarily selected, and they surmised that they were re- 163leased for purely political...the routine rigors of confinement or for other purely personal relief. A parole of this nature may be authorized for example, topermit a prisoner to
Zhang, Zheshen; Voss, Paul L
2009-07-06
We propose a continuous variable based quantum key distribution protocol that makes use of discretely signaled coherent light and reverse error reconciliation. We present a rigorous security proof against collective attacks with realistic lossy, noisy quantum channels, imperfect detector efficiency, and detector electronic noise. This protocol is promising for convenient, high-speed operation at link distances up to 50 km with the use of post-selection.
Malsby, Robert F; Quesada, Jose; Powell-Dunford, Nicole; Kinoshita, Ren; Kurtz, John; Gehlen, William; Adams, Colleen; Martin, Dustin; Shackelford, Stacy
2013-07-01
U.S. Army flight medics performed a process improvement initiative of 15 blood product transfusions on select Category A (Urgent) helicopter evacuation casualties meeting approved clinical indications for transfusion. These transfusions were initiated from point of injury locations aboard MEDEVAC aircraft originating from one of two locations in southern Afghanistan. All flight medics executing the transfusions were qualified through a standardized and approved program of instruction, which included day and night skills validation, and a 90% or higher written examination score. There was no adverse reaction or out-of-standard blood product temperature despite hazardous conditions and elevated cabin temperatures. All casualties within a 10-minute flight time who met clinical indications were transfused. Utilization of a standard operating procedure with strict handling and administration parameters, a rigorous training and qualification program, an elaborate cold chain system, and redundant documentation of blood product units ensured that flight medic initiated transfusions were safe and effective. Research study is needed to refine the indications for prehospital blood transfusion and to determine the effect on outcomes in severely injured trauma patients. Reprint & Copyright © 2013 Association of Military Surgeons of the U.S.
O'Hare, Esther Marie; Wood, Angela; Fiske, Elizabeth
2013-01-01
Forms of human milk banking and donation have been present for more than a century worldwide, but, since 1985, the Human Milk Banking Association of North America (HM BANA) has established guidelines to make the use of donor's breast milk safe and the second best form of feeding to maternal breast milk for a neonatal intensive care unit (NICU) infant. The Indiana Mother's Human Milk Bank provides an extensive and meticulous process of selecting breast milk donors. The process begins with a phone interview with a potential donor and includes the review of the donor's medical records, blood laboratory screening, medication and dietary intake, as well as consent from the donor's pediatrician. The milk bank follows steps of collecting, storing, and receiving the breast milk in accordance with the guidelines of the HM BANA. Pasteurization is the method used to ensure the proper heating and cooling of breast milk. Despite the rigorous pasteurization method, the donor's breast milk will not lose most of the important beneficial components needed for sick or ill NICU infants. Every batch of pasteurized breast milk will be cultured for any possible contamination and shipped to NICUs after it has been cleared by laboratory testing.
Practical protocols for fast histopathology by Fourier transform infrared spectroscopic imaging
NASA Astrophysics Data System (ADS)
Keith, Frances N.; Reddy, Rohith K.; Bhargava, Rohit
2008-02-01
Fourier transform infrared (FT-IR) spectroscopic imaging is an emerging technique that combines the molecular selectivity of spectroscopy with the spatial specificity of optical microscopy. We demonstrate a new concept in obtaining high fidelity data using commercial array detectors coupled to a microscope and Michelson interferometer. Next, we apply the developed technique to rapidly provide automated histopathologic information for breast cancer. Traditionally, disease diagnoses are based on optical examinations of stained tissue and involve a skilled recognition of morphological patterns of specific cell types (histopathology). Consequently, histopathologic determinations are a time consuming, subjective process with innate intra- and inter-operator variability. Utilizing endogenous molecular contrast inherent in vibrational spectra, specially designed tissue microarrays and pattern recognition of specific biochemical features, we report an integrated algorithm for automated classifications. The developed protocol is objective, statistically significant and, being compatible with current tissue processing procedures, holds potential for routine clinical diagnoses. We first demonstrate that the classification of tissue type (histology) can be accomplished in a manner that is robust and rigorous. Since data quality and classifier performance are linked, we quantify the relationship through our analysis model. Last, we demonstrate the application of the minimum noise fraction (MNF) transform to improve tissue segmentation.
The Capabilities of Chaos and Complexity
Abel, David L.
2009-01-01
To what degree could chaos and complexity have organized a Peptide or RNA World of crude yet necessarily integrated protometabolism? How far could such protolife evolve in the absence of a heritable linear digital symbol system that could mutate, instruct, regulate, optimize and maintain metabolic homeostasis? To address these questions, chaos, complexity, self-ordered states, and organization must all be carefully defined and distinguished. In addition their cause-and-effect relationships and mechanisms of action must be delineated. Are there any formal (non physical, abstract, conceptual, algorithmic) components to chaos, complexity, self-ordering and organization, or are they entirely physicodynamic (physical, mass/energy interaction alone)? Chaos and complexity can produce some fascinating self-ordered phenomena. But can spontaneous chaos and complexity steer events and processes toward pragmatic benefit, select function over non function, optimize algorithms, integrate circuits, produce computational halting, organize processes into formal systems, control and regulate existing systems toward greater efficiency? The question is pursued of whether there might be some yet-to-be discovered new law of biology that will elucidate the derivation of prescriptive information and control. “System” will be rigorously defined. Can a low-informational rapid succession of Prigogine’s dissipative structures self-order into bona fide organization? PMID:19333445
Theories of State Analyzing the Policy Process,
1973-11-01
values and goals - which is the heart of the rational process-- in reality cannot be separated from the actor’s empirical analysis of the situation...rigorous and objective in analysis . How different would our foreign policy actually be? Would it necessarily be better? In fact, would one even need...State, but the fact is that much of the outside research and analysis of policy process is pointed at the 6 As Robert Rothstein says in his valuable
A novel methodology for litho-to-etch pattern fidelity correction for SADP process
NASA Astrophysics Data System (ADS)
Chen, Shr-Jia; Chang, Yu-Cheng; Lin, Arthur; Chang, Yi-Shiang; Lin, Chia-Chi; Lai, Jun-Cheng
2017-03-01
For 2x nm node semiconductor devices and beyond, more aggressive resolution enhancement techniques (RETs) such as source-mask co-optimization (SMO), litho-etch-litho-etch (LELE) and self-aligned double patterning (SADP) are utilized for the low k1 factor lithography processes. In the SADP process, the pattern fidelity is extremely critical since a slight photoresist (PR) top-loss or profile roughness may impact the later core trim process, due to its sensitivity to environment. During the subsequent sidewall formation and core removal processes, the core trim profile weakness may worsen and induces serious defects that affect the final electrical performance. To predict PR top-loss, a rigorous lithography simulation can provide a reference to modify mask layouts; but it takes a much longer run time and is not capable of full-field mask data preparation. In this paper, we first brought out an algorithm which utilizes multi-intensity levels from conventional aerial image simulation to assess the physical profile through lithography to core trim etching steps. Subsequently, a novel correction method was utilized to improve the post-etch pattern fidelity without the litho. process window suffering. The results not only matched PR top-loss in rigorous lithography simulation, but also agreed with post-etch wafer data. Furthermore, this methodology can also be incorporated with OPC and post-OPC verification to improve core trim profile and final pattern fidelity at an early stage.
Krompecher, T; Bergerioux, C; Brandt-Casadevall, C; Gujer, H R
1983-07-01
The evolution of rigor mortis was studied in cases of nitrogen asphyxia, drowning and strangulation, as well as in fatal intoxications due to strychnine, carbon monoxide and curariform drugs, using a modified method of measurement. Our experiments demonstrated that: (1) Strychnine intoxication hastens the onset and passing of rigor mortis. (2) CO intoxication delays the resolution of rigor mortis. (3) The intensity of rigor may vary depending upon the cause of death. (4) If the stage of rigidity is to be used to estimate the time of death, it is necessary: (a) to perform a succession of objective measurements of rigor mortis intensity; and (b) to verify the eventual presence of factors that could play a role in the modification of its development.
NASA Technical Reports Server (NTRS)
Pulkkinen, A.; Rastaetter, L.; Kuznetsova, M.; Singer, H.; Balch, C.; Weimer, D.; Toth, G.; Ridley, A.; Gombosi, T.; Wiltberger, M.;
2013-01-01
In this paper we continue the community-wide rigorous modern space weather model validation efforts carried out within GEM, CEDAR and SHINE programs. In this particular effort, in coordination among the Community Coordinated Modeling Center (CCMC), NOAA Space Weather Prediction Center (SWPC), modelers, and science community, we focus on studying the models' capability to reproduce observed ground magnetic field fluctuations, which are closely related to geomagnetically induced current phenomenon. One of the primary motivations of the work is to support NOAA SWPC in their selection of the next numerical model that will be transitioned into operations. Six geomagnetic events and 12 geomagnetic observatories were selected for validation.While modeled and observed magnetic field time series are available for all 12 stations, the primary metrics analysis is based on six stations that were selected to represent the high-latitude and mid-latitude locations. Events-based analysis and the corresponding contingency tables were built for each event and each station. The elements in the contingency table were then used to calculate Probability of Detection (POD), Probability of False Detection (POFD) and Heidke Skill Score (HSS) for rigorous quantification of the models' performance. In this paper the summary results of the metrics analyses are reported in terms of POD, POFD and HSS. More detailed analyses can be carried out using the event by event contingency tables provided as an online appendix. An online interface built at CCMC and described in the supporting information is also available for more detailed time series analyses.
Model-based analysis of pattern motion processing in mouse primary visual cortex
Muir, Dylan R.; Roth, Morgane M.; Helmchen, Fritjof; Kampa, Björn M.
2015-01-01
Neurons in sensory areas of neocortex exhibit responses tuned to specific features of the environment. In visual cortex, information about features such as edges or textures with particular orientations must be integrated to recognize a visual scene or object. Connectivity studies in rodent cortex have revealed that neurons make specific connections within sub-networks sharing common input tuning. In principle, this sub-network architecture enables local cortical circuits to integrate sensory information. However, whether feature integration indeed occurs locally in rodent primary sensory areas has not been examined directly. We studied local integration of sensory features in primary visual cortex (V1) of the mouse by presenting drifting grating and plaid stimuli, while recording the activity of neuronal populations with two-photon calcium imaging. Using a Bayesian model-based analysis framework, we classified single-cell responses as being selective for either individual grating components or for moving plaid patterns. Rather than relying on trial-averaged responses, our model-based framework takes into account single-trial responses and can easily be extended to consider any number of arbitrary predictive models. Our analysis method was able to successfully classify significantly more responses than traditional partial correlation (PC) analysis, and provides a rigorous statistical framework to rank any number of models and reject poorly performing models. We also found a large proportion of cells that respond strongly to only one stimulus class. In addition, a quarter of selectively responding neurons had more complex responses that could not be explained by any simple integration model. Our results show that a broad range of pattern integration processes already take place at the level of V1. This diversity of integration is consistent with processing of visual inputs by local sub-networks within V1 that are tuned to combinations of sensory features. PMID:26300738
Bayesian evidence computation for model selection in non-linear geoacoustic inference problems.
Dettmer, Jan; Dosso, Stan E; Osler, John C
2010-12-01
This paper applies a general Bayesian inference approach, based on Bayesian evidence computation, to geoacoustic inversion of interface-wave dispersion data. Quantitative model selection is carried out by computing the evidence (normalizing constants) for several model parameterizations using annealed importance sampling. The resulting posterior probability density estimate is compared to estimates obtained from Metropolis-Hastings sampling to ensure consistent results. The approach is applied to invert interface-wave dispersion data collected on the Scotian Shelf, off the east coast of Canada for the sediment shear-wave velocity profile. Results are consistent with previous work on these data but extend the analysis to a rigorous approach including model selection and uncertainty analysis. The results are also consistent with core samples and seismic reflection measurements carried out in the area.
Television camera as a scientific instrument
NASA Technical Reports Server (NTRS)
Smokler, M. I.
1970-01-01
Rigorous calibration program, coupled with a sophisticated data-processing program that introduced compensation for system response to correct photometry, geometric linearity, and resolution, converted a television camera to a quantitative measuring instrument. The output data are in the forms of both numeric printout records and photographs.
RIGOR MORTIS AND THE INFLUENCE OF CALCIUM AND MAGNESIUM SALTS UPON ITS DEVELOPMENT.
Meltzer, S J; Auer, J
1908-01-01
Calcium salts hasten and magnesium salts retard the development of rigor mortis, that is, when these salts are administered subcutaneously or intravenously. When injected intra-arterially, concentrated solutions of both kinds of salts cause nearly an immediate onset of a strong stiffness of the muscles which is apparently a contraction, brought on by a stimulation caused by these salts and due to osmosis. This contraction, if strong, passes over without a relaxation into a real rigor. This form of rigor may be classed as work-rigor (Arbeitsstarre). In animals, at least in frogs, with intact cords, the early contraction and the following rigor are stronger than in animals with destroyed cord. If M/8 solutions-nearly equimolecular to "physiological" solutions of sodium chloride-are used, even when injected intra-arterially, calcium salts hasten and magnesium salts retard the onset of rigor. The hastening and retardation in this case as well as in the cases of subcutaneous and intravenous injections, are ion effects and essentially due to the cations, calcium and magnesium. In the rigor hastened by calcium the effects of the extensor muscles mostly prevail; in the rigor following magnesium injection, on the other hand, either the flexor muscles prevail or the muscles become stiff in the original position of the animal at death. There seems to be no difference in the degree of stiffness in the final rigor, only the onset and development of the rigor is hastened in the case of the one salt and retarded in the other. Calcium hastens also the development of heat rigor. No positive facts were obtained with regard to the effect of magnesium upon heat vigor. Calcium also hastens and magnesium retards the onset of rigor in the left ventricle of the heart. No definite data were gathered with regard to the effects of these salts upon the right ventricle.
RIGOR MORTIS AND THE INFLUENCE OF CALCIUM AND MAGNESIUM SALTS UPON ITS DEVELOPMENT
Meltzer, S. J.; Auer, John
1908-01-01
Calcium salts hasten and magnesium salts retard the development of rigor mortis, that is, when these salts are administered subcutaneously or intravenously. When injected intra-arterially, concentrated solutions of both kinds of salts cause nearly an immediate onset of a strong stiffness of the muscles which is apparently a contraction, brought on by a stimulation caused by these salts and due to osmosis. This contraction, if strong, passes over without a relaxation into a real rigor. This form of rigor may be classed as work-rigor (Arbeitsstarre). In animals, at least in frogs, with intact cords, the early contraction and the following rigor are stronger than in animals with destroyed cord. If M/8 solutions—nearly equimolecular to "physiological" solutions of sodium chloride—are used, even when injected intra-arterially, calcium salts hasten and magnesium salts retard the onset of rigor. The hastening and retardation in this case as well as in the cases of subcutaneous and intravenous injections, are ion effects and essentially due to the cations, calcium and magnesium. In the rigor hastened by calcium the effects of the extensor muscles mostly prevail; in the rigor following magnesium injection, on the other hand, either the flexor muscles prevail or the muscles become stiff in the original position of the animal at death. There seems to be no difference in the degree of stiffness in the final rigor, only the onset and development of the rigor is hastened in the case of the one salt and retarded in the other. Calcium hastens also the development of heat rigor. No positive facts were obtained with regard to the effect of magnesium upon heat vigor. Calcium also hastens and magnesium retards the onset of rigor in the left ventricle of the heart. No definite data were gathered with regard to the effects of these salts upon the right ventricle. PMID:19867124
NASA Astrophysics Data System (ADS)
Hidayat, D.; Nurlaelah, E.; Dahlan, J. A.
2017-09-01
The ability of mathematical creative and critical thinking are two abilities that need to be developed in the learning of mathematics. Therefore, efforts need to be made in the design of learning that is capable of developing both capabilities. The purpose of this research is to examine the mathematical creative and critical thinking ability of students who get rigorous mathematical thinking (RMT) approach and students who get expository approach. This research was quasi experiment with control group pretest-posttest design. The population were all of students grade 11th in one of the senior high school in Bandung. The result showed that: the achievement of mathematical creative and critical thinking abilities of student who obtain RMT is better than students who obtain expository approach. The use of Psychological tools and mediation with criteria of intentionality, reciprocity, and mediated of meaning on RMT helps students in developing condition in critical and creative processes. This achievement contributes to the development of integrated learning design on students’ critical and creative thinking processes.
Revisiting classical design in engineering from a perspective of frugality.
Rao, Balkrishna C
2017-05-01
The conservative nature of design in engineering has typically unleashed products fabricated with generous amounts of raw materials. This is epitomized by the factor of safety whose values higher than unity suggests various uncertainties of design that are tackled through material padding. This effort proposes a new factor of safety called the factor of frugality that could be used in ecodesign and which addresses both rigors of the classical design process and quantification of savings in materials going into a product. An example of frugal shaft design together with some other cases has been presented to explain the working of the factor of frugality . Adoption of the frugality factor would entail a change in design philosophy whereby designers would constantly make avail of a rigorous design process coupled with material-saving schemes for realizing products that are benign to the environment. Such a change in the foundations of design would abet the stewardship of earth in avoiding planetary boundaries since engineering influences a significant proportion of human endeavors.
Hanning, Brian; Predl, Nicolle
2015-09-01
Traditional overnight rehabilitation payment models in the private sector are not based on a rigorous classification system and vary greatly between contracts with no consideration of patient complexity. The payment rates are not based on relative cost and the length-of-stay (LOS) point at which a reduced rate applies (step downs) varies markedly. The rehabilitation Australian National Sub-Acute and Non-Acute Patient (AN-SNAP) model (RAM), which has been in place for over 2 years in some private hospitals, bases payment on a rigorous classification system, relative cost and industry LOS. RAM is in the process of being rolled out more widely. This paper compares and contrasts RAM with traditional overnight rehabilitation payment models. It considers the advantages of RAM for hospitals and Australian Health Service Alliance. It also considers payment model changes in the context of maintaining industry consistency with Electronic Claims Lodgement and Information Processing System Environment (ECLIPSE) and health reform generally.
Derivation of rigorous conditions for high cell-type diversity by algebraic approach.
Yoshida, Hiroshi; Anai, Hirokazu; Horimoto, Katsuhisa
2007-01-01
The development of a multicellular organism is a dynamic process. Starting with one or a few cells, the organism develops into different types of cells with distinct functions. We have constructed a simple model by considering the cell number increase and the cell-type order conservation, and have assessed conditions for cell-type diversity. This model is based on a stochastic Lindenmayer system with cell-to-cell interactions for three types of cells. In the present model, we have successfully derived complex but rigorous algebraic relations between the proliferation and transition rates for cell-type diversity by using a symbolic method: quantifier elimination (QE). Surprisingly, three modes for the proliferation and transition rates have emerged for large ratios of the initial cells to the developed cells. The three modes have revealed that the equality between the development rates for the highest cell-type diversity is reduced during the development process of multicellular organisms. Furthermore, we have found that the highest cell-type diversity originates from order conservation.
Qualitative adaptation of child behaviour problem instruments in a developing-country setting.
Khan, B; Avan, B I
2014-07-08
A key barrier to epidemiological research on child behaviour problems in developing countries is the lack of culturally relevant, internationally recognized psychometric instruments. This paper proposes a model for the qualitative adaptation of psychometric instruments in developing-country settings and presents a case study of the adaptation of 3 internationally recognized instruments in Pakistan: the Child Behavior Checklist, the Youth Self-Report and the Teacher's Report Form. This model encompassed a systematic procedure with 6 distinct phases to minimize bias and ensure equivalence with the original instruments: selection, deliberation, alteration, feasibility, testing and formal approval. The process was conducted in collaboration with the instruments' developer. A multidisciplinary working group of experts identified equivalence issues and suggested modifications. Focus group discussions with informants highlighted comprehension issues. Subsequently modified instruments were thoroughly tested. Finally, the instruments' developer approval further validated the qualitative adaptation. The study proposes a rigorous and systematic model to effectively achieve cultural adaptation of psychometric instruments.
NASA Astrophysics Data System (ADS)
Zimmermann, Judith; von Davier, Alina A.; Buhmann, Joachim M.; Heinimann, Hans R.
2018-01-01
Graduate admission has become a critical process in tertiary education, whereby selecting valid admissions instruments is key. This study assessed the validity of Graduate Record Examination (GRE) General Test scores for admission to Master's programmes at a technical university in Europe. We investigated the indicative value of GRE scores for the Master's programme grade point average (GGPA) with and without the addition of the undergraduate GPA (UGPA) and the TOEFL score, and of GRE scores for study completion and Master's thesis performance. GRE scores explained 20% of the variation in the GGPA, while additional 7% were explained by the TOEFL score and 3% by the UGPA. Contrary to common belief, the GRE quantitative reasoning score showed only little explanatory power. GRE scores were also weakly related to study progress but not to thesis performance. Nevertheless, GRE and TOEFL scores were found to be sensible admissions instruments. Rigorous methodology was used to obtain highly reliable results.
Olutoye, M A; Hameed, B H
2011-02-01
Fatty acid methyl ester was produced from used vegetable cooking oil using Mg(1-)(x) Zn(1+)(x)O(2) solid catalyst and the performance monitored in terms of ester content obtained. Used vegetable cooking oil was employed to reduce operation cost of biodiesel. The significant operating parameters which affect the overall yield of the process were studied. The highest ester content, 80%, was achieved with the catalyst during 4h 15 min reaction at 188°C with methanol to oil ratio of 9:1 and catalyst loading of 2.55 wt% oil. Also, transesterification of virgin oil gave higher yield with the heterogeneous catalyst and showed high selectivity towards ester production. The used vegetable cooking oil did not require any rigorous pretreatment. Catalyst stability was examined and there was no leaching of the active components, and its performance was as good at the fourth as at the first cycle. Copyright © 2010 Elsevier Ltd. All rights reserved.
OLED emission zone measurement with high accuracy
NASA Astrophysics Data System (ADS)
Danz, N.; MacCiarnain, R.; Michaelis, D.; Wehlus, T.; Rausch, A. F.; Wächter, C. A.; Reusch, T. C. G.
2013-09-01
Highly efficient state of the art organic light-emitting diodes (OLED) comprise thin emitting layers with thicknesses in the order of 10 nm. The spatial distribution of the photon generation rate, i.e. the profile of the emission zone, inside these layers is of interest for both device efficiency analysis and characterization of charge recombination processes. It can be accessed experimentally by reverse simulation of far-field emission pattern measurements. Such a far-field pattern is the sum of individual emission patterns associated with the corresponding positions inside the active layer. Based on rigorous electromagnetic theory the relation between far-field pattern and emission zone is modeled as a linear problem. This enables a mathematical analysis to be applied to the cases of single and double emitting layers in the OLED stack as well as to pattern measurements in air or inside the substrate. From the results, guidelines for optimum emitter - cathode separation and for selecting the best experimental approach are obtained. Limits for the maximum spatial resolution can be derived.
Wildlife conservation and solar energy development in the Desert Southwest, United States
Lovich, Jeffrey E.; Ennen, Josua R.
2011-01-01
Large areas of public land are currently being permitted or evaluated for utility-scale solar energy development (USSED) in the southwestern United States, including areas with high biodiversity and protected species. However, peer-reviewed studies of the effects of USSED on wildlife are lacking. The potential effects of the construction and the eventual decommissioning of solar energy facilities include the direct mortality of wildlife; environmental impacts of fugitive dust and dust suppressants; destruction and modification of habitat, including the impacts of roads; and off-site impacts related to construction material acquisition, processing, and transportation. The potential effects of the operation and maintenance of the facilities include habitat fragmentation and barriers to gene flow, increased noise, electromagnetic field generation, microclimate alteration, pollution, water consumption, and fire. Facility design effects, the efficacy of site-selection criteria, and the cumulative effects of USSED on regional wildlife populations are unknown. Currently available peer-reviewed data are insufficient to allow a rigorous assessment of the impact of USSED on wildlife.
El-Houjeiri, Hassan M; Brandt, Adam R; Duffy, James E
2013-06-04
Existing transportation fuel cycle emissions models are either general and calculate nonspecific values of greenhouse gas (GHG) emissions from crude oil production, or are not available for public review and auditing. We have developed the Oil Production Greenhouse Gas Emissions Estimator (OPGEE) to provide open-source, transparent, rigorous GHG assessments for use in scientific assessment, regulatory processes, and analysis of GHG mitigation options by producers. OPGEE uses petroleum engineering fundamentals to model emissions from oil and gas production operations. We introduce OPGEE and explain the methods and assumptions used in its construction. We run OPGEE on a small set of fictional oil fields and explore model sensitivity to selected input parameters. Results show that upstream emissions from petroleum production operations can vary from 3 gCO2/MJ to over 30 gCO2/MJ using realistic ranges of input parameters. Significant drivers of emissions variation are steam injection rates, water handling requirements, and rates of flaring of associated gas.
2017-01-01
This work investigates the design of alternative monitoring tools based on state estimators for industrial crystallization systems with nucleation, growth, and agglomeration kinetics. The estimation problem is regarded as a structure design problem where the estimation model and the set of innovated states have to be chosen; the estimator is driven by the available measurements of secondary variables. On the basis of Robust Exponential estimability arguments, it is found that the concentration is distinguishable with temperature and solid fraction measurements while the crystal size distribution (CSD) is not. Accordingly, a state estimator structure is selected such that (i) the concentration (and other distinguishable states) are innovated by means of the secondary measurements processed with the geometric estimator (GE), and (ii) the CSD is estimated by means of a rigorous model in open loop mode. The proposed estimator has been tested through simulations showing good performance in the case of mismatch in the initial conditions, parametric plant-model mismatch, and noisy measurements. PMID:28890604
Porru, Marcella; Özkan, Leyla
2017-08-30
This work investigates the design of alternative monitoring tools based on state estimators for industrial crystallization systems with nucleation, growth, and agglomeration kinetics. The estimation problem is regarded as a structure design problem where the estimation model and the set of innovated states have to be chosen; the estimator is driven by the available measurements of secondary variables. On the basis of Robust Exponential estimability arguments, it is found that the concentration is distinguishable with temperature and solid fraction measurements while the crystal size distribution (CSD) is not. Accordingly, a state estimator structure is selected such that (i) the concentration (and other distinguishable states) are innovated by means of the secondary measurements processed with the geometric estimator (GE), and (ii) the CSD is estimated by means of a rigorous model in open loop mode. The proposed estimator has been tested through simulations showing good performance in the case of mismatch in the initial conditions, parametric plant-model mismatch, and noisy measurements.
Radiative transfer code SHARM for atmospheric and terrestrial applications
NASA Astrophysics Data System (ADS)
Lyapustin, A. I.
2005-12-01
An overview of the publicly available radiative transfer Spherical Harmonics code (SHARM) is presented. SHARM is a rigorous code, as accurate as the Discrete Ordinate Radiative Transfer (DISORT) code, yet faster. It performs simultaneous calculations for different solar zenith angles, view zenith angles, and view azimuths and allows the user to make multiwavelength calculations in one run. The Δ-M method is implemented for calculations with highly anisotropic phase functions. Rayleigh scattering is automatically included as a function of wavelength, surface elevation, and the selected vertical profile of one of the standard atmospheric models. The current version of the SHARM code does not explicitly include atmospheric gaseous absorption, which should be provided by the user. The SHARM code has several built-in models of the bidirectional reflectance of land and wind-ruffled water surfaces that are most widely used in research and satellite data processing. A modification of the SHARM code with the built-in Mie algorithm designed for calculations with spherical aerosols is also described.
Radiative transfer code SHARM for atmospheric and terrestrial applications.
Lyapustin, A I
2005-12-20
An overview of the publicly available radiative transfer Spherical Harmonics code (SHARM) is presented. SHARM is a rigorous code, as accurate as the Discrete Ordinate Radiative Transfer (DISORT) code, yet faster. It performs simultaneous calculations for different solar zenith angles, view zenith angles, and view azimuths and allows the user to make multiwavelength calculations in one run. The Delta-M method is implemented for calculations with highly anisotropic phase functions. Rayleigh scattering is automatically included as a function of wavelength, surface elevation, and the selected vertical profile of one of the standard atmospheric models. The current version of the SHARM code does not explicitly include atmospheric gaseous absorption, which should be provided by the user. The SHARM code has several built-in models of the bidirectional reflectance of land and wind-ruffled water surfaces that are most widely used in research and satellite data processing. A modification of the SHARM code with the built-in Mie algorithm designed for calculations with spherical aerosols is also described.
[Evidence based medicine. A new paradigm for medical practice].
Carneiro, A V
1998-01-01
Modern medical practice is an ever-changing process, and the doctor's need for information has been partially met by continuous medical education (CME) activities. It has been shown that CME activities have not prevented clinical knowledge, as well as medical practice, from deteriorating with time. When faced with the need to get the most recent and relevant information possible, the busy clinician has two major problems: most of the published medical literature is either irrelevant or not useful; and there is little time to read it. Evidence-based medicine constitutes a new paradigm for medical practice in the sense that it tries to transform clinical problems into well formulated clinical questions, selecting and critically appraising scientific evidence with predefined and rigorous rules. It combines the expertise of the individual clinician with the best external evidence from clinical research for rational, ethical and efficacious practice. Evidence-based medicine can be taught and practiced by physicians with different degrees of autonomy, with several subspecialties, working in the hospital or in outpatient clinics, alone or in groups.
McKee, S R; Sams, A R
1998-01-01
Development of rigor mortis at elevated post-mortem temperatures may contribute to turkey meat characteristics that are similar to those found in pale, soft, exudative pork. To evaluate this effect, 36 Nicholas tom turkeys were processed at 19 wk of age and placed in water at 40, 20, and 0 C immediately after evisceration. Pectoralis muscle samples were taken at 15 min, 30 min, 1 h, 2 h, and 4 h post-mortem and analyzed for R-value (an indirect measure of adenosine triphosphate), glycogen, pH, color, and sarcomere length. At 4 h, the remaining intact Pectoralis muscle was harvested, and aged on ice 23 h, and analyzed for drip loss, cook loss, shear values, and sarcomere length. By 15 min post-mortem, the 40 C treatment had higher R-values, which persisted through 4 h. By 1 h, the 40 C treatment pH and glycogen levels were lower than the 0 C treatment; however, they did not differ from those of the 20 C treatment. Increased L* values indicated that color became more pale by 2 h post-mortem in the 40 C treatment when compared to the 20 and 0 C treatments. Drip loss, cook loss, and shear value were increased whereas sarcomere lengths were decreased as a result of the 40 C treatment. These findings suggested that elevated post-mortem temperatures during processing resulted in acceleration of rigor mortis and biochemical changes in the muscle that produced pale, exudative meat characteristics in turkey.
Fleisher, Linda; Wen, Kuang Yi; Miller, Suzanne M; Diefenbach, Michael; Stanton, Annette L; Ropka, Mary; Morra, Marion; Raich, Peter C
2015-11-01
Cancer patients and survivors are assuming active roles in decision-making and digital patient support tools are widely used to facilitate patient engagement. As part of Cancer Information Service Research Consortium's randomized controlled trials focused on the efficacy of eHealth interventions to promote informed treatment decision-making for newly diagnosed prostate and breast cancer patients, and post-treatment breast cancer, we conducted a rigorous process evaluation to examine the actual use of and perceived benefits of two complementary communication channels -- print and eHealth interventions. The three Virtual Cancer Information Service (V-CIS) interventions were developed through a rigorous developmental process, guided by self-regulatory theory, informed decision-making frameworks, and health communications best practices. Control arm participants received NCI print materials; experimental arm participants received the additional V-CIS patient support tool. Actual usage data from the web-based V-CIS was also obtained and reported. Print materials were highly used by all groups. About 60% of the experimental group reported using the V-CIS. Those who did use the V-CIS rated it highly on improvements in knowledge, patient-provider communication and decision-making. The findings show that how patients actually use eHealth interventions either singularly or within the context of other communication channels is complex. Integrating rigorous best practices and theoretical foundations is essential and multiple communication approaches should be considered to support patient preferences.
Long persistence of rigor mortis at constant low temperature.
Varetto, Lorenzo; Curto, Ombretta
2005-01-06
We studied the persistence of rigor mortis by using physical manipulation. We tested the mobility of the knee on 146 corpses kept under refrigeration at Torino's city mortuary at a constant temperature of +4 degrees C. We found a persistence of complete rigor lasting for 10 days in all the cadavers we kept under observation; and in one case, rigor lasted for 16 days. Between the 11th and the 17th days, a progressively increasing number of corpses showed a change from complete into partial rigor (characterized by partial bending of the articulation). After the 17th day, all the remaining corpses showed partial rigor and in the two cadavers that were kept under observation "à outrance" we found the absolute resolution of rigor mortis occurred on the 28th day. Our results prove that it is possible to find a persistence of rigor mortis that is much longer than the expected when environmental conditions resemble average outdoor winter temperatures in temperate zones. Therefore, this datum must be considered when a corpse is found in those environmental conditions so that when estimating the time of death, we are not misled by the long persistence of rigor mortis.
Psychiatric diagnoses in a group of astronaut applicants
NASA Technical Reports Server (NTRS)
Santy, Patricia A.; Faulk, Dean M.; Holland, Al W.
1991-01-01
Between 1959 and 1987, the psychiatric evaluation of astronaut candidates evolved from a 30-h intensive examination evaluating applicants for psychopathology, and studying their performance under stress, to a 2-h clinical interview whose structure and contents were determined by the individual examiner. Evaluations done during these years applied both psychiatric (or, 'select-out') criteria and psychological (or, 'select-in') criteria. In an attempt to more rigorously define the psychiatric, 'select-out' component, a standardized, semistructured clinical interview was developed to identify the presence or history of psychiatric disorders listed in the Diagnostic and Statistical Manual of Mental Disorders, 3rd Ed. ('DSM-III'). A total of 117 astronaut applicants underwent this clinical interview as part of a comprehensive medical evaluation during a recent astronaut selection. Of the 117 applicants, 9 (7.7 percent) met DSM-III criteria for a variety of Axis I and Axis II diagnoses, including V-code diagnoses.
Good, Andrew C; Hermsmeier, Mark A
2007-01-01
Research into the advancement of computer-aided molecular design (CAMD) has a tendency to focus on the discipline of algorithm development. Such efforts are often wrought to the detriment of the data set selection and analysis used in said algorithm validation. Here we highlight the potential problems this can cause in the context of druglikeness classification. More rigorous efforts are applied to the selection of decoy (nondruglike) molecules from the ACD. Comparisons are made between model performance using the standard technique of random test set creation with test sets derived from explicit ontological separation by drug class. The dangers of viewing druglike space as sufficiently coherent to permit simple classification are highlighted. In addition the issues inherent in applying unfiltered data and random test set selection to (Q)SAR models utilizing large and supposedly heterogeneous databases are discussed.
Anthropic selection and the habitability of planets orbiting M and K dwarfs
NASA Astrophysics Data System (ADS)
Waltham, Dave
2011-10-01
The Earth may have untypical characteristics which were necessary preconditions for the emergence of life and, ultimately, intelligent observers. This paper presents a rigorous procedure for quantifying such "anthropic selection" effects by comparing Earth's properties to those of exoplanets. The hypothesis that there is anthropic selection for stellar mass (i.e. planets orbiting stars with masses within a particular range are more favourable for the emergence of observers) is then tested. The results rule out the expected strong selection for low mass stars which would result, all else being equal, if the typical timescale for the emergence of intelligent observers is very long. This indicates that the habitable zone of small stars may be less hospitable for intelligent life than the habitable zone of solar-mass stars. Additional planetary properties can also be analyzed, using the approach introduced here, once relatively complete and unbiased statistics are made available by current and planned exoplanet characterization projects.
Lin, Chun-Chih; Han, Chin-Yen; Pan, I-Ju
2015-03-01
Professional hemodialysis (HD) nursing tends to be task-oriented and lack consideration of the client's viewpoint. This study aims to interpret the process of psychosocial adaptation to dealing with HD in people with end-stage renal disease (ESRD). A grounded theory guided this study. Theoretical sampling included 15 people receiving HD at the HD center of a hospital from July to November 2010. Participants received an information sheet in writing, a verbal invitation, and informed consent forms before interviews were conducted. A constant comparative data analysis was analyzed using open, axial and selective coding. The computer software ATLAS.ti assisted data management. Credibility, transferability, dependability, and confirmability ensured the rigor of study process. This study identified "adopting life with hemodialysis", which captures the process of the psychosocial adaptation in people with ESRD as one transformation. Four categories that evolved from "adopting HD life" are (a) slipping into, (b) restricted to a renal world, (c) losing self control, and (d) stuck in an endless process. The findings of this investigation indicate the multidimensional requirements of people receiving maintenance dialysis, with an emphasis on the deficiency in psychosocial and emotional care. The study's findings contribute to clinical practice by increasing the understanding of the experience of chronic HD treatment from the recipient's viewpoint. The better our understanding, the better the care provided will meet the needs of the people receiving HD. Copyright © 2015. Published by Elsevier B.V.
When Assessment Data Are Words: Validity Evidence for Qualitative Educational Assessments.
Cook, David A; Kuper, Ayelet; Hatala, Rose; Ginsburg, Shiphra
2016-10-01
Quantitative scores fail to capture all important features of learner performance. This awareness has led to increased use of qualitative data when assessing health professionals. Yet the use of qualitative assessments is hampered by incomplete understanding of their role in forming judgments, and lack of consensus in how to appraise the rigor of judgments therein derived. The authors articulate the role of qualitative assessment as part of a comprehensive program of assessment, and translate the concept of validity to apply to judgments arising from qualitative assessments. They first identify standards for rigor in qualitative research, and then use two contemporary assessment validity frameworks to reorganize these standards for application to qualitative assessment.Standards for rigor in qualitative research include responsiveness, reflexivity, purposive sampling, thick description, triangulation, transparency, and transferability. These standards can be reframed using Messick's five sources of validity evidence (content, response process, internal structure, relationships with other variables, and consequences) and Kane's four inferences in validation (scoring, generalization, extrapolation, and implications). Evidence can be collected and evaluated for each evidence source or inference. The authors illustrate this approach using published research on learning portfolios.The authors advocate a "methods-neutral" approach to assessment, in which a clearly stated purpose determines the nature of and approach to data collection and analysis. Increased use of qualitative assessments will necessitate more rigorous judgments of the defensibility (validity) of inferences and decisions. Evidence should be strategically sought to inform a coherent validity argument.
Gatica, M C; Monti, G E; Knowles, T G; Gallo, C B
2010-01-09
Two systems for transporting live salmon (Salmo salar) were compared in terms of their effects on blood variables, muscle pH and rigor index: an 'open system' well-boat with recirculated sea water at 13.5 degrees C and a stocking density of 107 kg/m3 during an eight-hour journey, and a 'closed system' well-boat with water chilled from 16.7 to 2.1 degrees C and a stocking density of 243.7 kg/m3 during a seven-hour journey. Groups of 10 fish were sampled at each of four stages: in cages at the farm, in the well-boat after loading, in the well-boat after the journey and before unloading, and in the processing plant after they were pumped from the resting cages. At each sampling, the fish were stunned and bled by gill cutting. Blood samples were taken to measure lactate, osmolality, chloride, sodium, cortisol and glucose, and their muscle pH and rigor index were measured at death and three hours later. In the open system well-boat, the initial muscle pH of the fish decreased at each successive stage, and at the final stage they had a significantly lower initial muscle pH and more rapid onset of rigor than the fish transported on the closed system well-boat. At the final stage all the blood variables except glucose were significantly affected in the fish transported on both types of well-boat.
Kunz, Regina; Fretheim, Atle; Cluzeau, Françoise; Wilt, Timothy J; Qaseem, Amir; Lelgemann, Monika; Kelson, Marcia; Guyatt, Gordon; Schünemann, Holger J
2012-12-01
Professional societies, like many other organizations around the world, have recognized the need to use more rigorous processes to ensure that health care recommendations are informed by the best available research evidence. This is the third of a series of 14 articles that were prepared to advise guideline developers in respiratory and other diseases on considerations for group compositions and group processes in guideline development, and how this can be effectively integrated in the context of respiratory disease guidelines on a national and international level. We updated a review of the literature addressing group composition and group process, focusing on the following questions: 1. How to compose a functioning and representative guideline group; Who should be included in a guideline panel?; How to select organizations, groups, and individuals; What expertise is needed?; Consultation with non-included groups. 2. How to assure a functioning group process; How to make the process constructive; Balancing participation and finding agreement; Administrative support; What constitutes sufficient resources? Our conclusions are based on available evidence from published literature, experience from guideline developers, and workshop discussions. Formal studies addressing optimal processes in developing guidelines are limited, and experience from guideline organizations supplement the formal studies. When resources are available, guideline development groups should aim for multidisciplinary groups, including patients. Prerequisites for a multidisciplinary group include: a strong chair experienced in group facilitation with broad acceptance in the group, training the group in guideline methodology, and professional technical support. Formal consensus developing methods have proved effective in reaching agreement on the final recommendations.
Rigor Made Easy: Getting Started
ERIC Educational Resources Information Center
Blackburn, Barbara R.
2012-01-01
Bestselling author and noted rigor expert Barbara Blackburn shares the secrets to getting started, maintaining momentum, and reaching your goals. Learn what rigor looks like in the classroom, understand what it means for your students, and get the keys to successful implementation. Learn how to use rigor to raise expectations, provide appropriate…
Close Early Learning Gaps with Rigorous DAP
ERIC Educational Resources Information Center
Brown, Christopher P.; Mowry, Brian
2015-01-01
Rigorous DAP (developmentally appropriate practices) is a set of 11 principles of instruction intended to help close early childhood learning gaps. Academically rigorous learning environments create the conditions for children to learn at high levels. While academic rigor focuses on one dimension of education--academic--DAP considers the whole…
Lenzenweger, Mark F
2015-01-01
During World War II, the Office of Strategic Services (OSS), the forerunner of the Central Intelligence Agency, sought the assistance of clinical psychologists and psychiatrists to establish an assessment program for evaluating candidates for the OSS. The assessment team developed a novel and rigorous program to evaluate OSS candidates. It is described in Assessment of Men: Selection of Personnel for the Office of Strategic Services (OSS Assessment Staff, 1948). This study examines the sole remaining multivariate data matrix that includes all final ratings for a group of candidates (n = 133) assessed near the end of the assessment program. It applies the modern statistical methods of both exploratory and confirmatory factor analysis to this rich and highly unique data set. An exploratory factor analysis solution suggested 3 factors underlie the OSS assessment staff ratings. Confirmatory factor analysis results of multiple plausible substantive models reveal that a 3-factor model provides the best fit to these data. The 3 factors are emotional/interpersonal factors (social relations, emotional stability, security), intelligence processing (effective IQ, propaganda skills, observing and reporting), and agency/surgency (motivation, energy and initiative, leadership, physical ability). These factors are discussed in terms of their potential utility for personnel selection within the intelligence community.
Writing a Research Proposal to The Research Council of Oman.
Al-Shukaili, Ahmed; Al-Maniri, Abdullah
2017-05-01
Writing a research proposal can be a challenging task for young researchers. This article explains how to write a strong research proposal to apply for funding, specifically, a proposal for The Research Council (TRC) of Oman. Three different research proposal application forms are currently used in TRC, including Open Research Grant (ORG), Graduate Research Support Program (GRSP), and Faculty-mentored Undergraduate Research Award Program (FURAP). The application forms are filled and submitted electronically on TRC website. Each of the proposals submitted to TRC is selected through a rigorous reviewing and screening process. Novelty and originality of the research idea is the most crucial element in writing a research proposal. Performing an in-depth review of the literature will assist you to compose a good researchable question and generate a strong hypothesis. The development of a good hypothesis will offer insight into the specific objectives of a study. Research objectives should be focused, measurable, and achievable by a specific time using the most appropriate methodology. Moreover, it is essential to select a proper study design in-line with the purpose of the study and the hypothesis. Furthermore, social/economic impact and reasonable budget of proposed research are important criteria in research proposal evaluation by TRC. Finally, ethical principles should be observed before writing a research proposal involving human or animal subjects.
Waste in health information systems: a systematic review.
Awang Kalong, Nadia; Yusof, Maryati
2017-05-08
Purpose The purpose of this paper is to discuss a systematic review on waste identification related to health information systems (HIS) in Lean transformation. Design/methodology/approach A systematic review was conducted on 19 studies to evaluate Lean transformation and tools used to remove waste related to HIS in clinical settings. Findings Ten waste categories were identified, along with their relationships and applications of Lean tool types related to HIS. Different Lean tools were used at the early and final stages of Lean transformation; the tool selection depended on the waste characteristic. Nine studies reported a positive impact from Lean transformation in improving daily work processes. The selection of Lean tools should be made based on the timing, purpose and characteristics of waste to be removed. Research limitations/implications Overview of waste and its category within HIS and its analysis from socio-technical perspectives enabled the identification of its root cause in a holistic and rigorous manner. Practical implications Understanding waste types, their root cause and review of Lean tools could subsequently lead to the identification of mitigation approach to prevent future error occurrence. Originality/value Specific waste models for HIS settings are yet to be developed. Hence, the identification of the waste categories could guide future implementation of Lean transformations in HIS settings.
Velasco, Veronica; Griffin, Kenneth W; Antichi, Mariella; Celata, Corrado
2015-10-01
Across developed countries, experimentation with alcohol, tobacco, and other drugs often begins in the early adolescent years. Several evidence-based programs have been developed to prevent adolescent substance use. Many of the most rigorously tested and empirically supported prevention programs were initially developed and tested in the United States. Increasingly, these interventions are being adopted for use in Europe and throughout the world. This paper reports on a large-scale comprehensive initiative designed to select, adapt, implement, and sustain an evidence-based drug abuse prevention program in Italy. As part of a large-scale regionally funded collaboration in the Lombardy region of Italy, we report on processes through which a team of stakeholders selected, translated and culturally adapted, planned, implemented and evaluated the Life Skills Training (LST) school-based drug abuse prevention program, an evidence-based intervention developed in the United States. We discuss several challenges and lessons learned and implications for prevention practitioners and researchers attempting to undertake similar international dissemination projects. We review several published conceptual models designed to promote the replication and widespread dissemination of effective programs, and discuss their strengths and limitations in the context of planning and implementing a complex, large-scale real-world dissemination effort. Copyright © 2015 Elsevier Ltd. All rights reserved.
Rigorous Science: a How-To Guide.
Casadevall, Arturo; Fang, Ferric C
2016-11-08
Proposals to improve the reproducibility of biomedical research have emphasized scientific rigor. Although the word "rigor" is widely used, there has been little specific discussion as to what it means and how it can be achieved. We suggest that scientific rigor combines elements of mathematics, logic, philosophy, and ethics. We propose a framework for rigor that includes redundant experimental design, sound statistical analysis, recognition of error, avoidance of logical fallacies, and intellectual honesty. These elements lead to five actionable recommendations for research education. Copyright © 2016 Casadevall and Fang.
Evaluation, Instruction and Policy Making. IIEP Seminar Paper: 9.
ERIC Educational Resources Information Center
Bloom, Benjamin S.
Recently, educational evaluation has attempted to use the precision, objectivity, and mathematical rigor of the psychological measurement field as well as to find ways in which instrumentation and data utilization could more directly be related to educational institutions, educational processes, and educational purposes. The linkages between…
Caution--Praise Can Be Dangerous.
ERIC Educational Resources Information Center
Dweck, Carol S.
1999-01-01
Reviews research into the effects of praise on students. Suggests an approach that gets students to focus on their potential to learn, to value challenge, and to concentrate on effort and learning processes in the face of obstacles. This can all be done while holding students to rigorous standards. (SLD)
A Commentary on Hughes: The Tension between Pragmatics and Rigor.
ERIC Educational Resources Information Center
Doll, Beth
2000-01-01
Argues against Hughes' recommendation for a comprehensive clinical case planning process since it cannot displace consumers' need for psychological practice guidelines which can be applied autonomously. Challenges the notion that theory has been omitted from current research on empirically supported treatments. Concludes the problem facing…
Aesthetic Criticism in the Music Classroom.
ERIC Educational Resources Information Center
Sibbald, Mary Jo
1993-01-01
Contends that a number of music educators believe that music should be taught in the same rigorous, sequential, and comprehensive way that other subjects are taught. Asserts that critical thinking skills and interpretation are essential factors in this process. Recommends cooperative learning and portfolios as appropriate instructional strategies.…
ACM TOMS replicated computational results initiative
Heroux, Michael Allen
2015-06-03
In this study, the scientific community relies on the peer review process for assuring the quality of published material, the goal of which is to build a body of work we can trust. Computational journals such as The ACM Transactions on Mathematical Software (TOMS) use this process for rigorously promoting the clarity and completeness of content, and citation of prior work. At the same time, it is unusual to independently confirm computational results.
Using process groups to implement failure detection in asynchronous environments
NASA Technical Reports Server (NTRS)
Ricciardi, Aleta M.; Birman, Kenneth P.
1991-01-01
Agreement on the membership of a group of processes in a distributed system is a basic problem that arises in a wide range of applications. Such groups occur when a set of processes cooperate to perform some task, share memory, monitor one another, subdivide a computation, and so forth. The group membership problems is discussed as it relates to failure detection in asynchronous, distributed systems. A rigorous, formal specification for group membership is presented under this interpretation. A solution is then presented for this problem.
Wong, Michelle; Bejarano, Esther; Carvlin, Graeme; Fellows, Katie; King, Galatea; Lugo, Humberto; Jerrett, Michael; Meltzer, Dan; Northcross, Amanda; Olmedo, Luis; Seto, Edmund; Wilkie, Alexa; English, Paul
2018-03-15
Air pollution continues to be a global public health threat, and the expanding availability of small, low-cost air sensors has led to increased interest in both personal and crowd-sourced air monitoring. However, to date, few low-cost air monitoring networks have been developed with the scientific rigor or continuity needed to conduct public health surveillance and inform policy. In Imperial County, California, near the U.S./Mexico border, we used a collaborative, community-engaged process to develop a community air monitoring network that attains the scientific rigor required for research, while also achieving community priorities. By engaging community residents in the project design, monitor siting processes, data dissemination, and other key activities, the resulting air monitoring network data are relevant, trusted, understandable, and used by community residents. Integration of spatial analysis and air monitoring best practices into the network development process ensures that the data are reliable and appropriate for use in research activities. This combined approach results in a community air monitoring network that is better able to inform community residents, support research activities, guide public policy, and improve public health. Here we detail the monitor siting process and outline the advantages and challenges of this approach.
Effect of Pore Clogging on Kinetics of Lead Uptake by Clinoptilolite.
Inglezakis; Diamandis; Loizidou; Grigoropoulou
1999-07-01
The kinetics of lead-sodium ion exchange using pretreated natural clinoptilolite are investigated, more specifically the influence of agitation (0, 210, and 650 rpm) on the limiting step of the overall process, for particle sizes of 0.63-0.8 and 0.8-1 mm at ambient temperature and initial lead solutions of 500 mg l-1 without pH adjustment. The isotopic exchange model is found to fit the ion exchange process. Particle diffusion is shown to be the controlling step for both particle sizes under agitation, while in the absence of agitation film diffusion is shown to control. The ion exchange process effective diffusion coefficients are calculated and found to depend strongly on particle size in the case of agitation at 210 rpm and only slightly on particle size at 650 rpm. Lead uptake rates are higher for smaller particles only at rigorous agitation, while at mild agitation the results are reversed. These facts are due to partial clogging of the pores of the mineral during the grinding process. This is verified through comparison of lead uptake rates for two samples of the same particle size, one of which is rigorously washed for a certain time before being exposed to the ion exchange. Copyright 1999 Academic Press.
Wong, Michelle; Bejarano, Esther; Carvlin, Graeme; King, Galatea; Lugo, Humberto; Jerrett, Michael; Northcross, Amanda; Olmedo, Luis; Seto, Edmund; Wilkie, Alexa; English, Paul
2018-01-01
Air pollution continues to be a global public health threat, and the expanding availability of small, low-cost air sensors has led to increased interest in both personal and crowd-sourced air monitoring. However, to date, few low-cost air monitoring networks have been developed with the scientific rigor or continuity needed to conduct public health surveillance and inform policy. In Imperial County, California, near the U.S./Mexico border, we used a collaborative, community-engaged process to develop a community air monitoring network that attains the scientific rigor required for research, while also achieving community priorities. By engaging community residents in the project design, monitor siting processes, data dissemination, and other key activities, the resulting air monitoring network data are relevant, trusted, understandable, and used by community residents. Integration of spatial analysis and air monitoring best practices into the network development process ensures that the data are reliable and appropriate for use in research activities. This combined approach results in a community air monitoring network that is better able to inform community residents, support research activities, guide public policy, and improve public health. Here we detail the monitor siting process and outline the advantages and challenges of this approach. PMID:29543726
Reviewing the methodology of an integrative review.
Hopia, Hanna; Latvala, Eila; Liimatainen, Leena
2016-12-01
Whittemore and Knafl's updated description of methodological approach for integrative review was published in 2005. Since then, the five stages of the approach have been regularly used as a basic conceptual structure of the integrative reviews conducted by nursing researchers. However, this methodological approach is seldom examined from the perspective of how systematically and rigorously the stages are implemented in the published integrative reviews. To appraise the selected integrative reviews on the basis of the methodological approach according to the five stages published by Whittemore and Knafl in 2005. A literature review was used in this study. CINAHL (Cumulative Index to Nursing and Allied Health), PubMed, OVID (Journals@Ovid) and the Cochrane Library databases were searched for integrative reviews published between 2002 and 2014. Papers were included if they used the methodological approach described by Whittemore and Knafl, were published in English and were focused on nursing education or nursing expertise. A total of 259 integrative review publications for potential inclusion were identified. Ten integrative reviews fulfilled the inclusion criteria. Findings from the studies were extracted and critically examined according to the five methodological stages. The reviews assessed followed the guidelines of the stated methodology approach to different extents. The stages of literature search, data evaluation and data analysis were fairly poorly formulated and only partially implemented in the studies included in the sample. The other two stages, problem identification and presentation, followed those described in the methodological approach quite well. Increasing use of research in clinical practice is inevitable, and therefore, integrative reviews can play a greater role in developing evidence-based nursing practices. Because of this, nurse researchers should pay more attention to sound integrative nursing research to systematise the review process and make it more rigorous. © 2016 Nordic College of Caring Science.
Krompecher, T; Bergerioux, C
1988-01-01
The influence of electrocution on the evolution of rigor mortis was studied on rats. Our experiments showed that: (1) Electrocution hastens the onset of rigor mortis. After an electrocution of 90 s, a complete rigor develops already 1 h post-mortem (p.m.) compared to 5 h p.m. for the controls. (2) Electrocution hastens the passing of rigor mortis. After an electrocution of 90 s, the first significant decrease occurs at 3 h p.m. (8 h p.m. in the controls). (3) These modifications in rigor mortis evolution are less pronounced in the limbs not directly touched by the electric current. (4) In case of post-mortem electrocution, the changes are slightly less pronounced, the resistance is higher and the absorbed energy is lower as compared with the ante-mortem electrocution cases. The results are completed by two practical observations on human electrocution cases.
Eremenco, Sonya; Pease, Sheryl; Mann, Sarah; Berry, Pamela
2017-01-01
This paper describes the rationale and goals of the Patient-Reported Outcome (PRO) Consortium's instrument translation process. The PRO Consortium has developed a number of novel PRO measures which are in the process of qualification by the U.S. Food and Drug Administration (FDA) for use in clinical trials where endpoints based on these measures would support product labeling claims. Given the importance of FDA qualification of these measures, the PRO Consortium's Process Subcommittee determined that a detailed linguistic validation (LV) process was necessary to ensure that all translations of Consortium-developed PRO measures are performed using a standardized approach with the rigor required to meet regulatory and pharmaceutical industry expectations, as well as having a clearly defined instrument translation process that the translation industry can support. The consensus process involved gathering information about current best practices from 13 translation companies with expertise in LV, consolidating the findings to generate a proposed process, and obtaining iterative feedback from the translation companies and PRO Consortium member firms on the proposed process in two rounds of review in order to update existing principles of good practice in LV and to provide sufficient detail for the translation process to ensure consistency across PRO Consortium measures, sponsors, and translation companies. The consensus development resulted in a 12-step process that outlines universal and country-specific new translation approaches, as well as country-specific adaptations of existing translations. The PRO Consortium translation process will play an important role in maintaining the validity of the data generated through these measures by ensuring that they are translated by qualified linguists following a standardized and rigorous process that reflects best practice.
Rigorous Schools and Classrooms: Leading the Way
ERIC Educational Resources Information Center
Williamson, Ronald; Blackburn, Barbara R.
2010-01-01
Turn your school into a student-centered learning environment, where rigor is at the heart of instruction in every classroom. From the bestselling author of "Rigor is Not a Four-Letter Word," Barbara Blackburn, and award-winning educator Ronald Williamson, this comprehensive guide to establishing a schoolwide culture of rigor is for principals and…
Rigor Revisited: Scaffolding College Student Learning by Incorporating Their Lived Experiences
ERIC Educational Resources Information Center
Castillo-Montoya, Milagros
2018-01-01
This chapter explores how students' lived experiences contribute to the rigor of their thinking. Insights from research indicate faculty can enhance rigor by accounting for the many ways it may surface in the classroom. However, to see this type of rigor, we must revisit the way we conceptualize it for higher education.
Mungure, Tanyaradzwa E; Bekhit, Alaa El-Din A; Birch, E John; Stewart, Ian
2016-04-01
The effects of rigor temperature (5, 15, 20 and 25°C), ageing (3, 7, 14, and 21 days) and display time on meat quality and lipid oxidative stability of hot boned beef M. Semimembranosus (SM) muscle were investigated. Ultimate pH (pH(u)) was rapidly attained at higher rigor temperatures. Electrical conductivity increased with rigor temperature (p<0.001). Tenderness, purge and cooking losses were not affected by rigor temperature; however purge loss and tenderness increased with ageing (p<0.01). Lightness (L*) and redness (a*) of the SM increased as rigor temperature increased (p<0.01). Lipid oxidation was assessed using (1)H NMR where changes in aliphatic to olefinic (R(ao)) and diallylmethylene (R(ad)) proton ratios can be rapidly monitored. R(ad), R(ao), PUFA and TBARS were not affected by rigor temperature, however ageing and display increased lipid oxidation (p<0.05). This study shows that rigor temperature manipulation of hot boned beef SM muscle does not have adverse effects on lipid oxidation. Copyright © 2016 Elsevier Ltd. All rights reserved.
Meeting the information system demands of the future through outsourcing.
Goldman, S J
1994-05-01
As managed care organizations work to meet the rigorous data and information requirements of a rapidly evolving health care system, many are recognizing the need to out-source their computer operations. Developing a cost-effective, efficient approach to outsourcing is a challenge to many organizations. This article offers an in-depth view of outsourcing as it relates to the managed health care industry as well as criteria for selecting an outsourcing consultant or vendor.
When Does Frequency-Independent Selection Maintain Genetic Variation?
Novak, Sebastian; Barton, Nicholas H
2017-10-01
Frequency-independent selection is generally considered as a force that acts to reduce the genetic variation in evolving populations, yet rigorous arguments for this idea are scarce. When selection fluctuates in time, it is unclear whether frequency-independent selection may maintain genetic polymorphism without invoking additional mechanisms. We show that constant frequency-independent selection with arbitrary epistasis on a well-mixed haploid population eliminates genetic variation if we assume linkage equilibrium between alleles. To this end, we introduce the notion of frequency-independent selection at the level of alleles, which is sufficient to prove our claim and contains the notion of frequency-independent selection on haploids. When selection and recombination are weak but of the same order, there may be strong linkage disequilibrium; numerical calculations show that stable equilibria are highly unlikely. Using the example of a diallelic two-locus model, we then demonstrate that frequency-independent selection that fluctuates in time can maintain stable polymorphism if linkage disequilibrium changes its sign periodically. We put our findings in the context of results from the existing literature and point out those scenarios in which the possible role of frequency-independent selection in maintaining genetic variation remains unclear. Copyright © 2017 by the Genetics Society of America.
Lyon, Aaron R; Connors, Elizabeth; Jensen-Doss, Amanda; Landes, Sara J; Lewis, Cara C; McLeod, Bryce D; Rutt, Christopher; Stanick, Cameo; Weiner, Bryan J
2017-09-01
The advancement of implementation science is dependent on identifying assessment strategies that can address implementation and clinical outcome variables in ways that are valid, relevant to stakeholders, and scalable. This paper presents a measurement agenda for implementation science that integrates the previously disparate assessment traditions of idiographic and nomothetic approaches. Although idiographic and nomothetic approaches are both used in implementation science, a review of the literature on this topic suggests that their selection can be indiscriminate, driven by convenience, and not explicitly tied to research study design. As a result, they are not typically combined deliberately or effectively. Thoughtful integration may simultaneously enhance both the rigor and relevance of assessments across multiple levels within health service systems. Background on nomothetic and idiographic assessment is provided as well as their potential to support research in implementation science. Drawing from an existing framework, seven structures (of various sequencing and weighting options) and five functions (Convergence, Complementarity, Expansion, Development, Sampling) for integrating conceptually distinct research methods are articulated as they apply to the deliberate, design-driven integration of nomothetic and idiographic assessment approaches. Specific examples and practical guidance are provided to inform research consistent with this framework. Selection and integration of idiographic and nomothetic assessments for implementation science research designs can be improved. The current paper argues for the deliberate application of a clear framework to improve the rigor and relevance of contemporary assessment strategies.
Invariant polarimetric contrast parameters of coherent light.
Réfrégier, Philippe; Goudail, François
2002-06-01
Many applications use an active coherent illumination and analyze the variation of the polarization state of optical signals. However, as a result of the use of coherent light, these signals are generally strongly perturbed with speckle noise. This is the case, for example, for active polarimetric imaging systems that are useful for enhancing contrast between different elements in a scene. We propose a rigorous definition of the minimal set of parameters that characterize the difference between two coherent and partially polarized states. Indeed, two states of partially polarized light are a priori defined by eight parameters, for example, their two Stokes vectors. We demonstrate that the processing performance for such signal processing tasks as detection, localization, or segmentation of spatial or temporal polarization variations is uniquely determined by two scalar functions of these eight parameters. These two scalar functions are the invariant parameters that define the polarimetric contrast between two polarized states of coherent light. Different polarization configurations with the same invariant contrast parameters will necessarily lead to the same performance for a given task, which is a desirable quality for a rigorous contrast measure. The definition of these polarimetric contrast parameters simplifies the analysis and the specification of processing techniques for coherent polarimetric signals.
Initiation reactions in acetylene pyrolysis
Zador, Judit; Fellows, Madison D.; Miller, James A.
2017-05-10
In gas-phase combustion systems the interest in acetylene stems largely from its role in molecular weight growth processes. The consensus is that above 1500 K acetylene pyrolysis starts mainly with the homolytic fission of the C–H bond creating an ethynyl radical and an H atom. However, below ~1500 K this reaction is too slow to initiate the chain reaction. It has been hypothesized that instead of dissociation, self-reaction initiates this process. Nevertheless, rigorous theoretical or direct experimental evidence is lacking, to an extent that even the molecular mechanism is debated in the literature. In this work we use rigorous abmore » initio transition-state theory master equation methods to calculate pressure- and temperature-dependent rate coefficients for the association of two acetylene molecules and related reactions. We establish the role of vinylidene, the high-energy isomer of acetylene in this process, compare our results with available experimental data, and assess the competition between the first-order and second-order initiation steps. As a result, we also show the effect of the rapid isomerization among the participating wells and highlight the need for time-scale analysis when phenomenological rate coefficients are compared to observed time scales in certain experiments.« less
High and low rigor temperature effects on sheep meat tenderness and ageing.
Devine, Carrick E; Payne, Steven R; Peachey, Bridget M; Lowe, Timothy E; Ingram, John R; Cook, Christian J
2002-02-01
Immediately after electrical stimulation, the paired m. longissimus thoracis et lumborum (LT) of 40 sheep were boned out and wrapped tightly with a polyethylene cling film. One of the paired LT's was chilled in 15°C air to reach a rigor mortis (rigor) temperature of 18°C and the other side was placed in a water bath at 35°C and achieved rigor at this temperature. Wrapping reduced rigor shortening and mimicked meat left on the carcass. After rigor, the meat was aged at 15°C for 0, 8, 26 and 72 h and then frozen. The frozen meat was cooked to 75°C in an 85°C water bath and shear force values obtained from a 1×1 cm cross-section. The shear force values of meat for 18 and 35°C rigor were similar at zero ageing, but as ageing progressed, the 18 rigor meat aged faster and became more tender than meat that went into rigor at 35°C (P<0.001). The mean sarcomere length values of meat samples for 18 and 35°C rigor at each ageing time were significantly different (P<0.001), the samples at 35°C being shorter. When the short sarcomere length values and corresponding shear force values were removed for further data analysis, the shear force values for the 35°C rigor were still significantly greater. Thus the toughness of 35°C meat was not a consequence of muscle shortening and appears to be due to both a faster rate of tenderisation and the meat tenderising to a greater extent at the lower temperature. The cook loss at 35°C rigor (30.5%) was greater than that at 18°C rigor (28.4%) (P<0.01) and the colour Hunter L values were higher at 35°C (P<0.01) compared with 18°C, but there were no significant differences in a or b values.
Medications during pregnancy: antihypertensives and immunosuppressives.
Umans, Jason G
2007-04-01
Use of prescription and nonprescription medications is common during pregnancy and is required in many women with underlying kidney disease or hypertension and in all with solid-organ allografts. Systematic assessment of drug safety during pregnancy is lacking, as are rigorous and comprehensive studies of pharmacokinetics and pharmacodynamics to guide drug selection and dosing across pregnancy. Renal and hepatic clearances of many drugs change markedly during pregnancy and pitfalls may complicate routine therapeutic monitoring of some drugs during pregnancy. However, available data and clinical experience allow reasonable strategies for selection and dosing of immunosuppressive agents in pregnant transplant recipients and of antihypertensive agents in women with mild or more severe hypertension complicating their pregnancies.
Neoliberalism, Policy Reforms and Higher Education in Bangladesh
ERIC Educational Resources Information Center
Kabir, Ariful Haq
2013-01-01
Bangladesh has introduced neoliberal policies since the 1970s. Military regimes, since the dramatic political changes in 1975, accelerated the process. A succession of military rulers made rigorous changes in policy-making in various sectors. This article uses a critical approach to document analysis and examines the perceptions of key…
Automating Nuclear-Safety-Related SQA Procedures with Custom Applications
Freels, James D.
2016-01-01
Nuclear safety-related procedures are rigorous for good reason. Small design mistakes can quickly turn into unwanted failures. Researchers at Oak Ridge National Laboratory worked with COMSOL to define a simulation app that automates the software quality assurance (SQA) verification process and provides results in less than 24 hours.
Effects of Juvenile Court Exposure on Crime in Young Adulthood
ERIC Educational Resources Information Center
Petitclerc, Amelie; Gatti, Uberto; Vitaro, Frank; Tremblay, Richard E.
2013-01-01
Background: The juvenile justice system's interventions are expected to help reduce recidivism. However, previous studies suggest that official processing in juvenile court fails to reduce adolescents' criminal behavior in the following year. Longer term effects have not yet been investigated with a rigorous method. This study used propensity…
ERIC Educational Resources Information Center
Boohan, Richard
2014-01-01
This article describes an approach to teaching about the energy concept that aims to be accessible to students starting in early secondary school, while being scientifically rigorous and forming the foundation for later work. It discusses how exploring thermal processes is a good starting point for a more general consideration of the ways that…
ERIC Educational Resources Information Center
Murphy, Kevin; Ratey, Nancy; Maynard, Sandy; Sussman, Susan; Wright, Sarah D.
2010-01-01
Despite limited scientific study on ADHD coaching as an intervention for adults with ADHD, the field of ADHD coaching has grown significantly and gained popularity in recent years. ADHD coaching is becoming a bona fide profession where one must advance through a rigorous training process, in order to be certified as a professional ADHD coach.…
Management Information System Based on the Balanced Scorecard
ERIC Educational Resources Information Center
Kettunen, Juha; Kantola, Ismo
2005-01-01
Purpose: This study seeks to describe the planning and implementation in Finland of a campus-wide management information system using a rigorous planning methodology. Design/methodology/approach: The structure of the management information system is planned on the basis of the management process, where strategic management and the balanced…
ERIC Educational Resources Information Center
Harris, Rick
1995-01-01
In a partnership between several tool companies and vocational high schools, students in construction technology classes give new products a fair and rigorous workout at a fraction of the cost of focus groups. The process allows companies to expose their products to students who, in turn, provide critical evaluation of the tools. (JOW)
The Achiever. Volume 6, Number 5
ERIC Educational Resources Information Center
Ashby, Nicole, Ed.
2007-01-01
"The Achiever" is a monthly newsletter designed expressly for parents and community leaders. Each issue contains news and information about and from public and private organizations about school improvement in the United States. Highlights of this issue include: (1) New Online Tool Simplifies Financial Aid Process; (2) Rigor in K-6:…
Language Supports for Journal Abstract Writing across Disciplines
ERIC Educational Resources Information Center
Liou, H.-C.; Yang, P.-C.; Chang, J. S.
2012-01-01
Various writing assistance tools have been developed through efforts in the areas of natural language processing with different degrees of success of curriculum integration depending on their functional rigor and pedagogical designs. In this paper, we developed a system, WriteAhead, that provides six types of suggestions when non-native graduate…
Assessing Students' Understanding of Fraction Multiplication
ERIC Educational Resources Information Center
Rumsey, Chepina; Guarino, Jody; Beltramini, Jennie; Cole, Shelbi; Farmer, Alicia; Gray, Kristin; Saxby, Morgan
2016-01-01
In this article the authors describe a project during which they unpacked fraction standards, created rigorous tasks and lesson plans, and developed formative and summative assessments to analyze students' thinking about fraction multiplication. The purpose of this article is to (1) illustrate a process that can be replicated by educators…
Making Visionary Design Work at Policy Level and in Practice
ERIC Educational Resources Information Center
Burgdorff, Frauke
2011-01-01
Good architecture responds to contemporary challenges, irrespective of fashionable trends, and is characterised by rigorous building processes. Office buildings, housing, stadiums and theatres offer outstanding examples of this principle. But do school buildings figure among them? Happily, in Germany they do, although they are still too rare. This…
Tenure Troubles and Equity Matters in Canadian Academe
ERIC Educational Resources Information Center
Acker, Sandra; Webber, Michelle; Smyth, Elizabeth
2012-01-01
The focus of this article is the tenure review process in Canadian universities, a rigorous and high-stakes evaluation of junior academics that serves as a prime exemplar of "disciplining academics", our project's title. In-depth interviews in seven Ontario universities with 30 knowledgeable informants such as senior managers and faculty…
Financial Aid's Role in Meeting State College Completion Goals
ERIC Educational Resources Information Center
Hillman, Nicholas W.; Orians, Erica Lee
2013-01-01
This brief utilizes the most recent and rigorous financial aid research to inform state higher education leaders about innovative and effective financial aid practices. By simplifying aid eligibility requirements, improving the aid application process, and engaging in early awareness efforts, states could improve the effectiveness of existing aid…
ERIC Educational Resources Information Center
DiMeo, Michelle A.; Moore, G. Kurt; Lichtenstein, Carolyn
2012-01-01
Evidence-based treatments (EBTs) are "interventions" that have been proven effective through rigorous research methodologies. Evidence-based practice (EBP), however, refers to a "decision-making process" that integrates the best available research, clinician expertise, and client characteristics. This study examined community mental health service…
Cash on Demand: A Framework for Managing a Cash Liquidity Position.
ERIC Educational Resources Information Center
Augustine, John H.
1995-01-01
A well-run college or university will seek to accumulate and maintain an appropriate cash reserve or liquidity position. A rigorous analytic process for estimating the size and cost of a liquidity position, based on judgments about the institution's operating risks and opportunities, is outlined. (MSE)
ARCHITECTURAL PROGRAMMING--STATE OF THE ART.
ERIC Educational Resources Information Center
EVANS, BENJAMIN H.
IN RESPONSE TO A NEED FOR A MORE THOROUGH AND RIGOROUS STUDY AND ANALYSIS PROCESS IN ENVIRONMENTAL FUNCTIONS PRIOR TO THE DESIGN OF NEW BUILDINGS, A STUDY WAS UNDERTAKEN TO IDENTIFY THE EMERGING TECHNIQUES OF ARCHITECTURAL PROGRAMING PRACTICE. THE STUDY INCLUDED CORRESPONDENCE AND REVIEW OF PERIODICALS, QUESTIONNAIRES AND VISITATIONS, AND A…
New Perils for the Contract Ethnographer.
ERIC Educational Resources Information Center
Fetterman, David M.
1981-01-01
Conditions of contract research may lead some workers to ignore publication rights of colleagues whose reports are of limited circulation. The author presents a case example of how this process occurred with the use of his own work and argues for rigorous ethical standards in the publication of contract research results. (Author/GC)
Exploring Student Perceptions of Rigor Online: Toward a Definition of Rigorous Learning
ERIC Educational Resources Information Center
Duncan, Heather E.; Range, Bret; Hvidston, David
2013-01-01
Technological advances in the last decade have impacted delivery methods of university courses. More and more courses are offered in a variety of formats. While academic rigor is a term often used, its definition is less clear. This mixed-methods study explored graduate student conceptions of rigor in the online learning environment embedded…
Methodological rigor and citation frequency in patient compliance literature.
Bruer, J T
1982-01-01
An exhaustive bibliography which assesses the methodological rigor of the patient compliance literature, and citation data from the Science Citation Index (SCI) are combined to determine if methodologically rigorous papers are used with greater frequency than substandard articles by compliance investigators. There are low, but statistically significant, correlations between methodological rigor and citation indicators for 138 patient compliance papers published in SCI source journals during 1975 and 1976. The correlation is not strong enough to warrant use of citation measures as indicators of rigor on a paper-by-paper basis. The data do suggest that citation measures might be developed as crude indicators of methodological rigor. There is no evidence that randomized trials are cited more frequently than studies that employ other experimental designs. PMID:7114334
Phillips, Christine B; Dwan, Kathryn; Hepworth, Julie; Pearce, Christopher; Hall, Sally
2014-11-19
The primary health care sector delivers the majority of health care in western countries through small, community-based organizations. However, research into these healthcare organizations is limited by the time constraints and pressure facing them, and the concern by staff that research is peripheral to their work. We developed Q-RARA-Qualitative Rapid Appraisal, Rigorous Analysis-to study small, primary health care organizations in a way that is efficient, acceptable to participants and methodologically rigorous. Q-RARA comprises a site visit, semi-structured interviews, structured and unstructured observations, photographs, floor plans, and social scanning data. Data were collected over the course of one day per site and the qualitative analysis was integrated and iterative. We found Q-RARA to be acceptable to participants and effective in collecting data on organizational function in multiple sites without disrupting the practice, while maintaining a balance between speed and trustworthiness. The Q-RARA approach is capable of providing a richly textured, rigorous understanding of the processes of the primary care practice while also allowing researchers to develop an organizational perspective. For these reasons the approach is recommended for use in small-scale organizations both within and outside the primary health care sector.
Mann, Courtney M; Ward, Dianne S; Vaughn, Amber; Benjamin Neelon, Sara E; Long Vidal, Lenita J; Omar, Sakinah; Namenek Brouwer, Rebecca J; Østbye, Truls
2015-12-10
Many families rely on child care outside the home, making these settings important influences on child development. Nearly 1.5 million children in the U.S. spend time in family child care homes (FCCHs), where providers care for children in their own residences. There is some evidence that children in FCCHs are heavier than those cared for in centers. However, few interventions have targeted FCCHs for obesity prevention. This paper will describe the application of the Intervention Mapping (IM) framework to the development of a childhood obesity prevention intervention for FCCHs Following the IM protocol, six steps were completed in the planning and development of an intervention targeting FCCHs: needs assessment, formulation of change objectives matrices, selection of theory-based methods and strategies, creation of intervention components and materials, adoption and implementation planning, and evaluation planning Application of the IM process resulted in the creation of the Keys to Healthy Family Child Care Homes program (Keys), which includes three modules: Healthy You, Healthy Home, and Healthy Business. Delivery of each module includes a workshop, educational binder and tool-kit resources, and four coaching contacts. Social Cognitive Theory and Self-Determination Theory helped guide development of change objective matrices, selection of behavior change strategies, and identification of outcome measures. The Keys program is currently being evaluated through a cluster-randomized controlled trial The IM process, while time-consuming, enabled rigorous and systematic development of intervention components that are directly tied to behavior change theory and may increase the potential for behavior change within the FCCHs.
Beratarrechea, Andrea; Lee, Allison G; Willner, Jonathan M; Jahangir, Eiman; Ciapponi, Agustín; Rubinstein, Adolfo
2014-01-01
Rates of chronic diseases will continue to rise in developing countries unless effective and cost-effective interventions are implemented. This review aims to discuss the impact of mobile health (m-health) on chronic disease outcomes in low- and middle-income countries (LMIC). Systematic literature searches were performed using CENTRAL, MEDLINE, EMBASE, and LILACS databases and gray literature. Scientific literature was searched to identify controlled studies evaluating cell phone voice and text message interventions to address chronic diseases in adults in low- or middle-income countries. Outcomes measured included morbidity, mortality, hospitalization rates, behavioral or lifestyle changes, process of care improvements, clinical outcomes, costs, patient-provider satisfaction, compliance, and health-related quality of life (HRQoL). From the 1,709 abstracts retrieved, 163 articles were selected for full text review, including 9 randomized controlled trials with 4,604 participants. Most of the studies addressed more than one outcome. Of the articles selected, six studied clinical outcomes, six studied processes of care, three examined healthcare costs, and two examined HRQoL. M-health positively impacted on chronic disease outcomes, improving attendance rates, clinical outcomes, and HRQoL, and was cost-effective. M-health is emerging as a promising tool to address access, coverage, and equity gaps in developing countries and low-resource settings. The results for m-health interventions showed a positive impact on chronic diseases in LMIC. However, a limiting factor of this review was the relatively small number of studies and patients enrolled, highlighting the need for more rigorous research in this area in developing countries.
Evidence for adaptive radiation from a phylogenetic study of plant defenses
Agrawal, Anurag A.; Fishbein, Mark; Halitschke, Rayko; Hastings, Amy P.; Rabosky, Daniel L.; Rasmann, Sergio
2009-01-01
One signature of adaptive radiation is a high level of trait change early during the diversification process and a plateau toward the end of the radiation. Although the study of the tempo of evolution has historically been the domain of paleontologists, recently developed phylogenetic tools allow for the rigorous examination of trait evolution in a tremendous diversity of organisms. Enemy-driven adaptive radiation was a key prediction of Ehrlich and Raven's coevolutionary hypothesis [Ehrlich PR, Raven PH (1964) Evolution 18:586–608], yet has remained largely untested. Here we examine patterns of trait evolution in 51 North American milkweed species (Asclepias), using maximum likelihood methods. We study 7 traits of the milkweeds, ranging from seed size and foliar physiological traits to defense traits (cardenolides, latex, and trichomes) previously shown to impact herbivores, including the monarch butterfly. We compare the fit of simple random-walk models of trait evolution to models that incorporate stabilizing selection (Ornstein-Ulenbeck process), as well as time-varying rates of trait evolution. Early bursts of trait evolution were implicated for 2 traits, while stabilizing selection was implicated for several others. We further modeled the relationship between trait change and species diversification while allowing rates of trait evolution to vary during the radiation. Species-rich lineages underwent a proportionately greater decline in latex and cardenolides relative to species-poor lineages, and the rate of trait change was most rapid early in the radiation. An interpretation of this result is that reduced investment in defensive traits accelerated diversification, and disproportionately so, early in the adaptive radiation of milkweeds. PMID:19805160
Microbial Communities as Experimental Units
DAY, MITCH D.; BECK, DANIEL; FOSTER, JAMES A.
2011-01-01
Artificial ecosystem selection is an experimental technique that treats microbial communities as though they were discrete units by applying selection on community-level properties. Highly diverse microbial communities associated with humans and other organisms can have significant impacts on the health of the host. It is difficult to find correlations between microbial community composition and community-associated diseases, in part because it may be impossible to define a universal and robust species concept for microbes. Microbial communities are composed of potentially thousands of unique populations that evolved in intimate contact, so it is appropriate in many situations to view the community as the unit of analysis. This perspective is supported by recent discoveries using metagenomics and pangenomics. Artificial ecosystem selection experiments can be costly, but they bring the logical rigor of biological model systems to the emerging field of microbial community analysis. PMID:21731083
Berger, Lawrence M; Bruch, Sarah K; Johnson, Elizabeth I; James, Sigrid; Rubin, David
2009-01-01
This study used data on 2,453 children aged 4-17 from the National Survey of Child and Adolescent Well-Being and 5 analytic methods that adjust for selection factors to estimate the impact of out-of-home placement on children's cognitive skills and behavior problems. Methods included ordinary least squares (OLS) regressions and residualized change, simple change, difference-in-difference, and fixed effects models. Models were estimated using the full sample and a matched sample generated by propensity scoring. Although results from the unmatched OLS and residualized change models suggested that out-of-home placement is associated with increased child behavior problems, estimates from models that more rigorously adjust for selection bias indicated that placement has little effect on children's cognitive skills or behavior problems.
NASA Astrophysics Data System (ADS)
Nucciotti, V.; Stringari, C.; Sacconi, L.; Vanzi, F.; Linari, M.; Piazzesi, G.; Lombardi, V.; Pavone, F. S.
2009-02-01
Recently, the use of Second Harmonic Generation (SHG) for imaging biological samples has been explored with regard to intrinsic SHG in highly ordered biological samples. As shown by fractional extraction of proteins, myosin is the source of SHG signal in skeletal muscle. SHG is highly dependent on symmetries and provides selective information on the structural order and orientation of the emitting proteins and the dynamics of myosin molecules responsible for the mechano-chemical transduction during contraction. We characterise the polarization-dependence of SHG intensity in three different physiological states: resting, rigor and isometric tetanic contraction in a sarcomere length range between 2.0 μm and 4.0 μm. The orientation of motor domains of the myosin molecules is dependent on their physiological states and modulate the SHG signal. We can discriminate the orientation of the emitting dipoles in four different molecular conformations of myosin heads in intact fibers during isometric contraction, in resting and rigor. We estimate the contribution of the myosin motor domain to the total second order bulk susceptibility from its molecular structure and its functional conformation. We demonstrate that SHG is sensitive to the fraction of ordered myosin heads by disrupting the order of myosin heads in rigor with an ATP analog. We estimate the fraction of myosin motors generating the isometric force in the active muscle fiber from the dependence of the SHG modulation on the degree of overlap between actin and myosin filaments during an isometric contraction.
Increasing rigor in NMR-based metabolomics through validated and open source tools
Eghbalnia, Hamid R; Romero, Pedro R; Westler, William M; Baskaran, Kumaran; Ulrich, Eldon L; Markley, John L
2016-01-01
The metabolome, the collection of small molecules associated with an organism, is a growing subject of inquiry, with the data utilized for data-intensive systems biology, disease diagnostics, biomarker discovery, and the broader characterization of small molecules in mixtures. Owing to their close proximity to the functional endpoints that govern an organism’s phenotype, metabolites are highly informative about functional states. The field of metabolomics identifies and quantifies endogenous and exogenous metabolites in biological samples. Information acquired from nuclear magnetic spectroscopy (NMR), mass spectrometry (MS), and the published literature, as processed by statistical approaches, are driving increasingly wider applications of metabolomics. This review focuses on the role of databases and software tools in advancing the rigor, robustness, reproducibility, and validation of metabolomics studies. PMID:27643760
Increasing rigor in NMR-based metabolomics through validated and open source tools.
Eghbalnia, Hamid R; Romero, Pedro R; Westler, William M; Baskaran, Kumaran; Ulrich, Eldon L; Markley, John L
2017-02-01
The metabolome, the collection of small molecules associated with an organism, is a growing subject of inquiry, with the data utilized for data-intensive systems biology, disease diagnostics, biomarker discovery, and the broader characterization of small molecules in mixtures. Owing to their close proximity to the functional endpoints that govern an organism's phenotype, metabolites are highly informative about functional states. The field of metabolomics identifies and quantifies endogenous and exogenous metabolites in biological samples. Information acquired from nuclear magnetic spectroscopy (NMR), mass spectrometry (MS), and the published literature, as processed by statistical approaches, are driving increasingly wider applications of metabolomics. This review focuses on the role of databases and software tools in advancing the rigor, robustness, reproducibility, and validation of metabolomics studies. Copyright © 2016. Published by Elsevier Ltd.
The Rigor Mortis of Education: Rigor Is Required in a Dying Educational System
ERIC Educational Resources Information Center
Mixon, Jason; Stuart, Jerry
2009-01-01
In an effort to answer the "Educational Call to Arms", our national public schools have turned to Advanced Placement (AP) courses as the predominate vehicle used to address the lack of academic rigor in our public high schools. Advanced Placement is believed by many to provide students with the rigor and work ethic necessary to…
2014-05-01
propagations CoCs Contaminants of concern GC Gas chromatography DNAPL Dense nonaqueous phase liquid ISCO In situ chemical oxidation HCA...used for the design and scale-up of air strippers, ion exchange systems, precipitation reactors , and many other treatment processes. Such treatability...studies provide definitive data on system dimensions and reagent dosages using linear or non -linear scale-up. Designing these processes without the
Air Force Project Risk Management - The Impact of Inconsistent Processes
2016-09-15
grave development of a project. The Challenger disaster drove rigor in process based risk analysis which began to emerge formally in the early 1990’s...is a natural tendency to be aggressive with assumptions early in a program to make the program appear attractive” (p. 30). The literature dovetails...and having an unwanted pregnancy (Burger & Burns, 1988). Investigations revealed that the siting of the Fukushima Daiichi reactors and a 5.7
The US Army War College: Gearing Up for the 21st Century
1988-12-01
modality. The Committee also identified a need for increased emphasis on active learning as well as greater academic rigor and challenge...and to involve them more directly in an active learning process, case studies, exercises, gaming, and analytical discussions have been increased...in the active learning process; that most challenging test which occurs when the student is perform- ing or reciting in front of his or her peers as
A case study of resources management planning with multiple objectives and projects
NASA Astrophysics Data System (ADS)
Peterson, David L.; Silsbee, David G.; Schmoldt, Daniel L.
1994-09-01
Each National Park Service unit in the United States produces a resources management plan (RMP) every four years or less. The plans commit budgets and personnel to specific projects for four years, but they are prepared with little quantitative and analytical rigor and without formal decision-making tools. We have previously described a multiple objective planning process for inventory and monitoring programs (Schmoldt and others 1994). To test the applicability of that process for the more general needs of resources management planning, we conducted an exercise on the Olympic National Park (NP) in Washington State, USA. Eight projects were selected as typical of those considered in RMPs and five members of the Olympic NP staff used the analytic hierarchy process (AHP) to prioritize the eight projects with respect to their implicit management objectives. By altering management priorities for the park, three scenarios were generated. All three contained some similarities in rankings for the eight projects, as well as some differences. Mathematical allocations of money and people differed among these scenarios and differed substantially from what the actual 1990 Olympic NP RMP contains. Combining subjective priority measures with budget dollars and personnel time into an objective function creates a subjective economic metric for comparing different RMP’s. By applying this planning procedure, actual expenditures of budget and personnel in Olympic NP can agree more closely with the staff’s management objectives for the park.
Putting people on the map through an approach that integrates social data in conservation planning.
Stephanson, Sheri L; Mascia, Michael B
2014-10-01
Conservation planning is integral to strategic and effective operations of conservation organizations. Drawing upon biological sciences, conservation planning has historically made limited use of social data. We offer an approach for integrating data on social well-being into conservation planning that captures and places into context the spatial patterns and trends in human needs and capacities. This hierarchical approach provides a nested framework for characterizing and mapping data on social well-being in 5 domains: economic well-being, health, political empowerment, education, and culture. These 5 domains each have multiple attributes; each attribute may be characterized by one or more indicators. Through existing or novel data that display spatial and temporal heterogeneity in social well-being, conservation scientists, planners, and decision makers may measure, benchmark, map, and integrate these data within conservation planning processes. Selecting indicators and integrating these data into conservation planning is an iterative, participatory process tailored to the local context and planning goals. Social well-being data complement biophysical and threat-oriented social data within conservation planning processes to inform decisions regarding where and how to conserve biodiversity, provide a structure for exploring socioecological relationships, and to foster adaptive management. Building upon existing conservation planning methods and insights from multiple disciplines, this approach to putting people on the map can readily merge with current planning practices to facilitate more rigorous decision making. © 2014 Society for Conservation Biology.
Can power-law scaling and neuronal avalanches arise from stochastic dynamics?
Touboul, Jonathan; Destexhe, Alain
2010-02-11
The presence of self-organized criticality in biology is often evidenced by a power-law scaling of event size distributions, which can be measured by linear regression on logarithmic axes. We show here that such a procedure does not necessarily mean that the system exhibits self-organized criticality. We first provide an analysis of multisite local field potential (LFP) recordings of brain activity and show that event size distributions defined as negative LFP peaks can be close to power-law distributions. However, this result is not robust to change in detection threshold, or when tested using more rigorous statistical analyses such as the Kolmogorov-Smirnov test. Similar power-law scaling is observed for surrogate signals, suggesting that power-law scaling may be a generic property of thresholded stochastic processes. We next investigate this problem analytically, and show that, indeed, stochastic processes can produce spurious power-law scaling without the presence of underlying self-organized criticality. However, this power-law is only apparent in logarithmic representations, and does not survive more rigorous analysis such as the Kolmogorov-Smirnov test. The same analysis was also performed on an artificial network known to display self-organized criticality. In this case, both the graphical representations and the rigorous statistical analysis reveal with no ambiguity that the avalanche size is distributed as a power-law. We conclude that logarithmic representations can lead to spurious power-law scaling induced by the stochastic nature of the phenomenon. This apparent power-law scaling does not constitute a proof of self-organized criticality, which should be demonstrated by more stringent statistical tests.
Cost analysis of advanced turbine blade manufacturing processes
NASA Technical Reports Server (NTRS)
Barth, C. F.; Blake, D. E.; Stelson, T. S.
1977-01-01
A rigorous analysis was conducted to estimate relative manufacturing costs for high technology gas turbine blades prepared by three candidate materials process systems. The manufacturing costs for the same turbine blade configuration of directionally solidified eutectic alloy, an oxide dispersion strengthened superalloy, and a fiber reinforced superalloy were compared on a relative basis to the costs of the same blade currently in production utilizing the directional solidification process. An analytical process cost model was developed to quantitatively perform the cost comparisons. The impact of individual process yield factors on costs was also assessed as well as effects of process parameters, raw materials, labor rates and consumable items.
Real-time radar signal processing using GPGPU (general-purpose graphic processing unit)
NASA Astrophysics Data System (ADS)
Kong, Fanxing; Zhang, Yan Rockee; Cai, Jingxiao; Palmer, Robert D.
2016-05-01
This study introduces a practical approach to develop real-time signal processing chain for general phased array radar on NVIDIA GPUs(Graphical Processing Units) using CUDA (Compute Unified Device Architecture) libraries such as cuBlas and cuFFT, which are adopted from open source libraries and optimized for the NVIDIA GPUs. The processed results are rigorously verified against those from the CPUs. Performance benchmarked in computation time with various input data cube sizes are compared across GPUs and CPUs. Through the analysis, it will be demonstrated that GPGPUs (General Purpose GPU) real-time processing of the array radar data is possible with relatively low-cost commercial GPUs.
Effect of muscle restraint on sheep meat tenderness with rigor mortis at 18°C.
Devine, Carrick E; Payne, Steven R; Wells, Robyn W
2002-02-01
The effect on shear force of skeletal restraint and removing muscles from lamb m. longissimus thoracis et lumborum (LT) immediately after slaughter and electrical stimulation was undertaken at a rigor temperature of 18°C (n=15). The temperature of 18°C was achieved through chilling of electrically stimulated sheep carcasses in air at 12°C, air flow 1-1.5 ms(-2). In other groups, the muscle was removed at 2.5 h post-mortem and either wrapped or left non-wrapped before being placed back on the carcass to follow carcass cooling regimes. Following rigor mortis, the meat was aged for 0, 16, 40 and 65 h at 15°C and frozen. For the non-stimulated samples, the meat was aged for 0, 12, 36 and 60 h before being frozen. The frozen meat was cooked to 75°C in an 85°C water bath and shear force values obtained from a 1 × 1 cm cross-section. Commencement of ageing was considered to take place at rigor mortis and this was taken as zero aged meat. There were no significant differences in the rate of tenderisation and initial shear force for all treatments. The 23% cook loss was similar for all wrapped and non-wrapped situations and the values decreased slightly with longer ageing durations. Wrapping was shown to mimic meat left intact on the carcass, as it prevented significant prerigor shortening. Such techniques allows muscles to be removed and placed in a controlled temperature environment to enable precise studies of ageing processes.
NASA Technical Reports Server (NTRS)
Leon, Gloria R.; Koscheyev, Victor S.; Coca, Aitor; List, Nathan
2004-01-01
The aim of this study was to compare the effectiveness of different cooling regime intensities to maintain physiological and subjective comfort during physical exertion levels comparable to that engaged in during extravehicular activities (EVA) in space. We studied eight subjects (six males, two females) donned in our newly developed physiologically based shortened liquid cooling/warming garment (SLCWG). Rigorous (condition 1) and mild (condition 2) water temperature cooling regimes were compared at physical exertion levels comparable to that performed during EVA to ascertain the effectiveness of a lesser intensity of cooling in maintaining thermal comfort, thus reducing energy consumption in the portable life support system. Exercise intensity was varied across stages of the session. Finger temperature, rectal temperature, and subjective perception of overall body and hand comfort were assessed. Finger temperature was significantly higher in the rigorous cooling condition and showed a consistent increase across exercise stages, likely due to the restriction of heat extraction because of the intensive cold. In the mild cooling condition, finger temperature exhibited an overall decline with cooling, indicating greater heat extraction from the body. Rectal temperature was not significantly different between conditions, and showed a steady increase over exercise stages in both rigorous and mild cooling conditions. Ratings of overall comfort were 30% higher (more positive) and more stable in mild cooling (p<0.001). The mild cooling regime was more effective than rigorous cooling in allowing the process of heat exchange to occur, thus maintaining thermal homeostasis and subjective comfort during physical exertion.
NASA Astrophysics Data System (ADS)
Di, K.; Liu, Y.; Liu, B.; Peng, M.
2012-07-01
Chang'E-1(CE-1) and Chang'E-2(CE-2) are the two lunar orbiters of China's lunar exploration program. Topographic mapping using CE-1 and CE-2 images is of great importance for scientific research as well as for preparation of landing and surface operation of Chang'E-3 lunar rover. In this research, we developed rigorous sensor models of CE-1 and CE-2 CCD cameras based on push-broom imaging principle with interior and exterior orientation parameters. Based on the rigorous sensor model, the 3D coordinate of a ground point in lunar body-fixed (LBF) coordinate system can be calculated by space intersection from the image coordinates of con-jugate points in stereo images, and the image coordinates can be calculated from 3D coordinates by back-projection. Due to uncer-tainties of the orbit and the camera, the back-projected image points are different from the measured points. In order to reduce these inconsistencies and improve precision, we proposed two methods to refine the rigorous sensor model: 1) refining EOPs by correcting the attitude angle bias, 2) refining the interior orientation model by calibration of the relative position of the two linear CCD arrays. Experimental results show that the mean back-projection residuals of CE-1 images are reduced to better than 1/100 pixel by method 1 and the mean back-projection residuals of CE-2 images are reduced from over 20 pixels to 0.02 pixel by method 2. Consequently, high precision DEM (Digital Elevation Model) and DOM (Digital Ortho Map) are automatically generated.
Application of high speed machining technology in aviation
NASA Astrophysics Data System (ADS)
Bałon, Paweł; Szostak, Janusz; Kiełbasa, Bartłomiej; Rejman, Edward; Smusz, Robert
2018-05-01
Aircraft structures are exposed to many loads during their working lifespan. Every particular action made during a flight is composed of a series of air movements which generate various aircraft loads. The most rigorous requirement which modern aircraft structures must fulfill is to maintain their high durability and reliability. This requirement involves taking many restrictions into account during the aircraft design process. The most important factor is the structure's overall mass, which has a crucial impact on both utility properties and cost-effectiveness. This makes aircraft one of the most complex results of modern technology. Additionally, there is currently an increasing utilization of high strength aluminum alloys, which requires the implementation of new manufacturing processes. High Speed Machining technology (HSM) is currently one of the most important machining technologies used in the aviation industry, especially in the machining of aluminium alloys. The primary difference between HSM and other milling techniques is the ability to select cutting parameters - depth of the cut layer, feed rate, and cutting speed in order to simultaneously ensure high quality, precision of the machined surface, and high machining efficiency, all of which shorten the manufacturing process of the integral components. In this paper, the authors explain the implementation of the HSM method in integral aircraft constructions. It presents the method of the airframe manufacturing method, and the final results. The HSM method is compared to the previous method where all subcomponents were manufactured by bending and forming processes, and then, they were joined by riveting.
Weixler, Dietmar; Roider-Schur, Sophie; Likar, Rudolf; Bozzaro, Claudia; Daniczek, Thomas; Feichtner, Angelika; Gabl, Christoph; Hammerl-Ferrari, Bernhard; Kletecka-Pulker, Maria; Körtner, Ulrich H J; Kössler, Hilde; Meran, Johannes G; Miksovsky, Aurelia; Pusswald, Bettina; Wienerroither, Thomas; Watzke, Herbert
2017-02-01
Palliative sedation therapy (PST) is an important and ethically accepted therapy in the care of selected palliative care patients with otherwise unbearable suffering from refractory distress. PST is increasingly used in end-of-life care. Austria does not have a standardized ethical guideline for this exceptional practice near end of life, but there is evidence that practice varies throughout the country. The Austrian Palliative Society (OPG) nominated a multidisciplinary working group of 16 palliative care experts and ethicists who established the national guideline on the basis of recent review work with the aim to adhere to the Europeans Association of Palliative Care's (EAPC) framework on palliative sedation therapy respecting Austrians legal, structural and cultural background. Consensus was achieved by a four-step sequential Delphi process. The Delphi-process was strictly orientated to the recently published EUROIMPACT-sedation-study-checklist and to the AGREE-2-tool. Additionally national stakeholders participated in the reflection of the results. As a result of a rigorous consensus process the long version of the Austrian National Palliative Sedation Guideline contains 112 statements within eleven domains and is supplemented by a philosophers excursus on suffering. By establishing a national guideline for palliative sedation therapy using the Delphi technique for consensus and stakeholder involvement the Austrian Palliative Society aims to ensure nationwide good practice of palliative sedation therapy. Screening for the practicability and efficacy of this guideline will be a future task.
2001-01-01
selection is made which ensure further injury is not incurred during evacuation. As a result, the basic task identified is carry. Demanding Physical Tasks...greatest amount of force a muscle or muscle group can exert in a single effort.3 An example, in very simple terms, would be an Olympic weightlifter ...Olympic weightlifters lift as much as possible in one lift. This requires a great amount of strength. A need for strength in light infantry soldiers is
Using constraints and their value for optimization of large ODE systems
Domijan, Mirela; Rand, David A.
2015-01-01
We provide analytical tools to facilitate a rigorous assessment of the quality and value of the fit of a complex model to data. We use this to provide approaches to model fitting, parameter estimation, the design of optimization functions and experimental optimization. This is in the context where multiple constraints are used to select or optimize a large model defined by differential equations. We illustrate the approach using models of circadian clocks and the NF-κB signalling system. PMID:25673300
Diverse Molecular Targets for Chalcones with Varied Bioactivities
Zhou, Bo; Xing, Chengguo
2015-01-01
Natural or synthetic chalcones with different substituents have revealed a variety of biological activities that may benefit human health. The underlying mechanisms of action, particularly with respect to the direct cellular targets and the modes of interaction with the targets, have not been rigorously characterized, which imposes challenges to structure-guided rational development of therapeutic agents or chemical probes with acceptable target-selectivity profile. This review summarizes literature evidence on chalcones’ direct molecular targets in the context of their biological activities. PMID:26798565
2014-09-01
The NATO Science and Technology Organization Science & Technology (S& T ) in the NATO context is defined as the selective and rigorous...generation and application of state-of-the-art, validated knowledge for defence and security purposes. S& T activities embrace scientific research...engineering, operational research and analysis, synthesis, integration and validation of knowledge derived through the scientific method. In NATO, S& T is
Selection of nest-site habitat by interior least terns in relation to sandbar construction
Sherfy, M.H.; Stucker, J.H.; Buhl, D.A.
2012-01-01
Federally endangered interior least terns (Sternula antillarum) nest on bare or sparsely vegetated sandbars on midcontinent river systems. Loss of nesting habitat has been implicated as a cause of population declines, and managing these habitats is a major initiative in population recovery. One such initiative involves construction of mid-channel sandbars on the Missouri River, where natural sandbar habitat has declined in quantity and quality since the late 1990s. We evaluated nest-site habitat selection by least terns on constructed and natural sandbars by comparing vegetation, substrate, and debris variables at nest sites (na =a 798) and random points (na =a 1,113) in bare or sparsely vegetated habitats. Our logistic regression models revealed that a broader suite of habitat features was important in nest-site selection on constructed than on natural sandbars. Odds ratios for habitat variables indicated that avoidance of habitat features was the dominant nest-site selection process on both sandbar types, with nesting terns being attracted to nest-site habitat features (gravel and debris) and avoiding vegetation only on constructed sandbars, and avoiding silt and leaf litter on both sandbar types. Despite the seemingly uniform nature of these habitats, our results suggest that a complex suite of habitat features influences nest-site choice by least terns. However, nest-site selection in this social, colonially nesting species may be influenced by other factors, including spatial arrangement of bare sand habitat, proximity to other least terns, and prior habitat occupancy by piping plovers (Charadrius melodus). We found that nest-site selection was sensitive to subtle variation in habitat features, suggesting that rigor in maintaining habitat condition will be necessary in managing sandbars for the benefit of least terns. Further, management strategies that reduce habitat features that are avoided by least terns may be the most beneficial to nesting least terns. ?? 2011 The Wildlife Society.
Selection of nest-site habitat by interior least terns in relation to sandbar construction
Sherfy, Mark H.; Stucker, Jennifer H.; Buhl, Deborah A.
2012-01-01
Federally endangered interior least terns (Sternula antillarum) nest on bare or sparsely vegetated sandbars on midcontinent river systems. Loss of nesting habitat has been implicated as a cause of population declines, and managing these habitats is a major initiative in population recovery. One such initiative involves construction of mid-channel sandbars on the Missouri River, where natural sandbar habitat has declined in quantity and quality since the late 1990s. We evaluated nest-site habitat selection by least terns on constructed and natural sandbars by comparing vegetation, substrate, and debris variables at nest sites (n = 798) and random points (n = 1,113) in bare or sparsely vegetated habitats. Our logistic regression models revealed that a broader suite of habitat features was important in nest-site selection on constructed than on natural sandbars. Odds ratios for habitat variables indicated that avoidance of habitat features was the dominant nest-site selection process on both sandbar types, with nesting terns being attracted to nest-site habitat features (gravel and debris) and avoiding vegetation only on constructed sandbars, and avoiding silt and leaf litter on both sandbar types. Despite the seemingly uniform nature of these habitats, our results suggest that a complex suite of habitat features influences nest-site choice by least terns. However, nest-site selection in this social, colonially nesting species may be influenced by other factors, including spatial arrangement of bare sand habitat, proximity to other least terns, and prior habitat occupancy by piping plovers (Charadrius melodus). We found that nest-site selection was sensitive to subtle variation in habitat features, suggesting that rigor in maintaining habitat condition will be necessary in managing sandbars for the benefit of least terns. Further, management strategies that reduce habitat features that are avoided by least terns may be the most beneficial to nesting least terns.
Krompecher, T; Fryc, O
1978-01-01
The use of new methods and an appropriate apparatus has allowed us to make successive measurements of rigor mortis and a study of its evolution in the rat. By a comparative examination on the front and hind limbs, we have determined the following: (1) The muscular mass of the hind limbs is 2.89 times greater than that of the front limbs. (2) In the initial phase rigor mortis is more pronounced in the front limbs. (3) The front and hind limbs reach maximum rigor mortis at the same time and this state is maintained for 2 hours. (4) Resolution of rigor mortis is accelerated in the front limbs during the initial phase, but both front and hind limbs reach complete resolution at the same time.
Onset of rigor mortis is earlier in red muscle than in white muscle.
Kobayashi, M; Takatori, T; Nakajima, M; Sakurada, K; Hatanaka, K; Ikegaya, H; Matsuda, Y; Iwase, H
2000-01-01
Rigor mortis is thought to be related to falling ATP levels in muscles postmortem. We measured rigor mortis as tension determined isometrically in three rat leg muscles in liquid paraffin kept at 37 degrees C or 25 degrees C--two red muscles, red gastrocnemius (RG) and soleus (SO) and one white muscle, white gastrocnemius (WG). Onset, half and full rigor mortis occurred earlier in RG and SO than in WG both at 37 degrees C and at 25 degrees C even though RG and WG were portions of the same muscle. This suggests that rigor mortis directly reflects the postmortem intramuscular ATP level, which decreases more rapidly in red muscle than in white muscle after death. Rigor mortis was more retarded at 25 degrees C than at 37 degrees C in each type of muscle.
How to Conduct Surveys: A Step-by-Step Guide. Sixth Edition
ERIC Educational Resources Information Center
Fink, Arlene
2016-01-01
Packed with new topics that reflect today's challenges, the Sixth Edition of the bestselling "How to Conduct Surveys" guides readers through the process of developing their own rigorous surveys and evaluating the credibility and transparency of surveys created by others. Offering practical, step-by-step advice and written in the same…
Redesigning a Special Education Teacher-Preparation Program: The Rationale, Process, and Outcomes
ERIC Educational Resources Information Center
Fuchs, Wendy W.; Fahsl, Allison J.; James, Susanne M.
2014-01-01
Teacher-preparation programs across the nation are faced with multiple mandates that aim to increase rigor and effectiveness of future educators. In the state of Illinois, all teacher-preparation programs must revise and redesign their coursework to align with the new professional teaching standards and licensure requirements. This article will…
Ocean Profile Measurements during the Seasonal Ice Zone Reconnaissance Surveys
2012-09-30
physical processes that occur within the BCSIZ that require data from all components of SIZRS, and improve predictive models of the SIZ through model ...the IABP (Ignatius Rigor) are approved by the USCG for operation from the ADA aircraft, but we anticipate being informed of any Safety of Flight Test
ERIC Educational Resources Information Center
Young, Jason C.
2017-01-01
There is broad consensus amongst scholars across a wide range of disciplines that digital technologies are having profound effects on micro- and macropolitical processes across the world. However, research into digital geographies has not rigorously examined the role of the Internet in bridging epistemological difference. Rather, most of this…
AP Courses Get Audited for Quality
ERIC Educational Resources Information Center
Ashford, Ellie
2007-01-01
As the college admissions process has gotten much more competitive, the number of high school students taking Advanced Placement (AP) courses has soared. At the same time, policymakers and education leaders seek to get more minorities and students not on the college track to sign up for AP and other rigorous classes. But as high schools have…
Replication of Special Education Research: Necessary but Far Too Rare
ERIC Educational Resources Information Center
Makel, Matthew C.; Plucker, Jonathan A.; Freeman, Jennifer; Lombardi, Allison; Simonsen, Brandi; Coyne, Michael
2016-01-01
Increased calls for rigor in special education have often revolved around the use of experimental research design. However, the replicability of research results is also a central tenet to the scientific research process. To assess the prevalence, success rate, and authorship history of replications in special education, we investigated the…
The Art of Learning: A Guide to Outstanding North Carolina Arts in Education Programs.
ERIC Educational Resources Information Center
Herman, Miriam L.
The Arts in Education programs delineated in this guide complement the rigorous arts curriculum taught by arts specialists in North Carolina schools and enable students to experience the joy of the creative process while reinforcing learning in other curricula: language arts, mathematics, social studies, science, and physical education. Programs…
A Replication of Failure, Not a Failure to Replicate
ERIC Educational Resources Information Center
Holden, Gary; Barker, Kathleen; Kuppens, Sofie; Rosenberg, Gary; LeBreton, Jonathan
2015-01-01
Purpose: The increasing role of systematic reviews in knowledge production demands greater rigor in the literature search process. The performance of the Social Work Abstracts (SWA) database has been examined multiple times over the past three decades. The current study is a replication within this line of research. Method: Issue-level coverage…
Student Perceptions of Dual-Listed Courses
ERIC Educational Resources Information Center
Balassiano, Katia; Rosentrater, Kurt A.; Marcketti, Sara B.
2014-01-01
Academic departments regularly offer dual-listed courses in which one course has two course numbers, yet are taught in the same place, at the same time, by a single instructor, and in one department to undergraduate and graduate students. While universities discourage their use by subjecting such courses to more rigorous approval processes,…
"NAEYC's Key Attributes of Quality Preschool Programs" Applied to the Jordanian Kindergarten Context
ERIC Educational Resources Information Center
Taleb, Tagreed Fathi Abu
2013-01-01
Jordan's education system is currently undergoing rigorous and comprehensive reform processes that focus on improving the status of educational policies and experiences for students in kindergarten through 12th grade. Kindergarten education in Jordan has been dominated until recently, by the private-sector. For the past decade, the Ministry of…
2002-05-16
which could actually increase the decision-making capacity of a military leader. In a recent article entitled, “The Coming of the Cyborgs ,” the...Coming of the Cyborgs ,” Fantasy & Science Fiction, January 2002, 107. 13 through the extensive reading of history and the rigorous conduct of
ERIC Educational Resources Information Center
Moss, Leah; Brown, Andy
2014-01-01
Recognition of Acquired Competencies (RAC) as it is known in Quebec, Canada, or Prior Learning Assessment (PLA), requires a learner to engage in retrospective thought about their learning path, their learning style and their experiential knowledge. This process of critical self-reflection and rigorous analysis by the learner of their prior…
A Narrative Review of Generic Intervention Fidelity Measures
ERIC Educational Resources Information Center
Di Rezze, Briano; Law, Mary; Gorter, Jan Willem; Eva, Kevin; Pollock, Nancy
2012-01-01
To increase the rigor of pediatric rehabilitation research, there is a need to evaluate the degree to which an intervention is conducted as planned (i.e., fidelity). Generic fidelity measures evaluate more than one intervention and often include nonspecific attributes of the therapy process common to both interventions. The objective of this study…
The Small College Imperative: From Survival to Transformation. An AGB White Paper
ERIC Educational Resources Information Center
Marcy, Mary B.
2017-01-01
Small colleges and universities serve a profoundly important role in American society. They provide the rigorous, personalized experience that is the hallmark of the best of higher education. In the process, they build opportunity and encourage civic engagement among a new generation of citizens. The landscape for these institutions has changed…
ERIC Educational Resources Information Center
Wright, Douglas; Moles, Ollie
A preliminary review of early responses to a questionnaire sent to secondary school principals across the United States revealed that most administrators felt more rigorous due process procedures should be followed in discipline cases than those required by federal regulations and school policies. The principals also tended to believe that…
A Bold Move: Reframing Composition through Assessment Design
ERIC Educational Resources Information Center
Condran, Jeffrey
2010-01-01
This article discusses the decision of the Art Institute of Pittsburgh (AiP) in Pittsburgh, Pennsylvania to implement a rigorous writing program assessment in order to obtain the Middle States accreditation, and it describes the process of determining which assessment model would be the most appropriate for AiP's needs. The use of a quantitative…
Spectral Signatures in the Classroom
ERIC Educational Resources Information Center
Huber, Thomas P.
2004-01-01
Ensuring that students understand the basis behind their geography/science courses is an essential part of their education. This article looks at an inexpensive and rigorous way of teaching students how to develop the needed data for remote sensing work. The procedure shows instructors how to build a system to teach students the process of…
ERIC Educational Resources Information Center
Spell, Rachelle M.; Guinan, Judith A.; Miller, Kristen R.; Beck, Christopher W.
2014-01-01
Incorporating authentic research experiences in introductory biology laboratory classes would greatly expand the number of students exposed to the excitement of discovery and the rigor of the scientific process. However, the essential components of an authentic research experience and the barriers to their implementation in laboratory classes are…
Reaching for Rigor: Identifying Practices of Effective High Schools. Research Report
ERIC Educational Resources Information Center
Cannata, Marisa; Haynes, Katherine Taylor; Smith, Thomas M.
2013-01-01
What distinguishes high schools that "beat the odds" for students from traditionally lower-performing groups from schools that struggle to improve the achievement and graduation rates of these student populations? What types of programs, practices, and processes support better than expected outcomes for students at risk of failure? How…
Practice as Research: A Fine Art Contextual Study
ERIC Educational Resources Information Center
Adams, Suze
2014-01-01
This paper examines the dynamic interplay between practice and theory in practice-led research in the visual arts. Building on recent debate around the issue and following appropriately rigorous models, the importance of locating a suitable methodology to adequately reflect the integrated process of research practice in written as well as visual…
Improving Communication Textbooks through Rigorous Processes of Revision and Review.
ERIC Educational Resources Information Center
Reppert, James E.
How relevant are some textbooks when professors, even in the same department, teach similar courses in different ways? Are there methods by which textbook companies can receive realistic, substantive feedback concerning the quality of communication textbook offerings? This paper details the methods by which its author/educator became a textbook…
The Professors behind the MOOC Hype
ERIC Educational Resources Information Center
Kolowich, Steve
2013-01-01
The largest-ever survey of professors who have taught MOOCs, or massive open online courses, shows that the process is time-consuming, but, according to the instructors, often successful. Nearly half of the professors felt their online courses were as rigorous academically as the versions they taught in the classroom. The survey, conducted by "The…
Engaging Minds in the Common Core: Integrating Standards for Student Engagement
ERIC Educational Resources Information Center
Howard, Christy
2016-01-01
With the implementation of Common Core State Standards (CCSS) many teachers continue to search for ways to engage students in the learning process while meeting the rigorous demands of the standards. Researchers suggest that by providing opportunities for higher order thinking, student choice, and creative ways to showcase knowledge, students will…
High School Health and Physical Education: Reinforcing the 3Rs
ERIC Educational Resources Information Center
Moore, John
2009-01-01
The ultimate goal of the education process should be to improve instruction and increase student learning. To effectively accomplish this would truly result in education reform. Therefore, the first step in bringing about education reform is to provide academic rigor, vocational relevance and curricula relationships in programs that students see…
Is the 'driving test' a robust quality indicator of colonoscopy performance?
Kelly, Nicholas M; Moorehead, John; Tham, Tony
2010-04-16
Colorectal cancer is a major cause of death in the western world and is currently the second commonest cause of death from malignant disease in the UK. Recently a "driving test" for colonoscopists wishing to take part in the National Health Service Bowel Cancer Screening Program has been introduced, with the aim of improving quality in colonoscopy. We describe the accreditation process and have reviewed the published evidence for its use. We compared this method of assessment to what occurs in other developed countries. To the authors' knowledge no other countries have similar methods of assessment of practicing colonoscopists, and instead use critical evaluation of key quality criteria. The UK appears to have one of the most rigorous accreditation processes, although this still has flaws. The published evidence suggests that the written part of the accreditation is not a good discriminating test and it needs to be improved or abandoned. Further work is needed on the best methods of assessing polypectomy skills. Rigorous systems need to be in place for the colonoscopist who fails the assessment.
Pre rigor processing, ageing and freezing on tenderness and colour stability of lamb loins.
Kim, Yuan H Brad; Luc, Genevieve; Rosenvold, Katja
2013-10-01
Forty eight lamb carcasses with temperature and pH monitored were obtained from two commercial plants. At 24h post mortem both loins (M. longissimus) from each carcass were randomly allocated to a) unaged frozen at -18°C, (b) aged at -1.5°C for 2weeks before freezing, (c) aged for 3 weeks before freezing and (d) aged for 9 weeks without freezing. Shear force, colour stability and proteolysis were analyzed. Carcasses with a slower temperature and more rapid pH decline had more calpain autolysis, slightly higher shear force and less colour stable compared to that counterpart in general (P<0.05). However, the shear force values of the loins were all acceptable (<6 kgF) regardless of different pre rigor processing and ageing/freezing treatments. Furthermore, the loins aged for 2 weeks-then-frozen/thawed had a similar shear force to the loins aged only for 9 weeks suggesting that ageing-then-freezing would result in equivalent tenderness compared to aged only loins for the long-term storage. Copyright © 2013 Elsevier Ltd. All rights reserved.
Machkovech, Heather M.; Bedford, Trevor; Suchard, Marc A.
2015-01-01
ABSTRACT Numerous experimental studies have demonstrated that CD8+ T cells contribute to immunity against influenza by limiting viral replication. It is therefore surprising that rigorous statistical tests have failed to find evidence of positive selection in the epitopes targeted by CD8+ T cells. Here we use a novel computational approach to test for selection in CD8+ T-cell epitopes. We define all epitopes in the nucleoprotein (NP) and matrix protein (M1) with experimentally identified human CD8+ T-cell responses and then compare the evolution of these epitopes in parallel lineages of human and swine influenza viruses that have been diverging since roughly 1918. We find a significant enrichment of substitutions that alter human CD8+ T-cell epitopes in NP of human versus swine influenza virus, consistent with the idea that these epitopes are under positive selection. Furthermore, we show that epitope-altering substitutions in human influenza virus NP are enriched on the trunk versus the branches of the phylogenetic tree, indicating that viruses that acquire these mutations have a selective advantage. However, even in human influenza virus NP, sites in T-cell epitopes evolve more slowly than do nonepitope sites, presumably because these epitopes are under stronger inherent functional constraint. Overall, our work demonstrates that there is clear selection from CD8+ T cells in human influenza virus NP and illustrates how comparative analyses of viral lineages from different hosts can identify positive selection that is otherwise obscured by strong functional constraint. IMPORTANCE There is a strong interest in correlates of anti-influenza immunity that are protective against diverse virus strains. CD8+ T cells provide such broad immunity, since they target conserved viral proteins. An important question is whether T-cell immunity is sufficiently strong to drive influenza virus evolution. Although many studies have shown that T cells limit viral replication in animal models and are associated with decreased symptoms in humans, no studies have proven with statistical significance that influenza virus evolves under positive selection to escape T cells. Here we use comparisons of human and swine influenza viruses to rigorously demonstrate that human influenza virus evolves under pressure to fix mutations in the nucleoprotein that promote escape from T cells. We further show that viruses with these mutations have a selective advantage since they are preferentially located on the “trunk” of the phylogenetic tree. Overall, our results show that CD8+ T cells targeting nucleoprotein play an important role in shaping influenza virus evolution. PMID:26311880
How we see others: the psychobiology of schemas and transference.
Stein, Dan J
2009-01-01
Social cognition involves automatic and stimulus-driven processes; these may be important in mediating stereotypes in the community and schemas and transference in the clinic setting. Significant differences in self-related processing and other-related processing may also lead to important biases in our view of the other. The psychobiology of social cognition is gradually being delineated, and may be useful in understanding these phenomena, and in responding appropriately. In the clinic, schemas can be rigorously assessed, and schema-focused psychotherapy may be useful in a number of indications.
Warriss, P D; Brown, S N; Knowles, T G
2003-12-13
The degree of development of rigor mortis in the carcases of slaughter pigs was assessed subjectively on a three-point scale 35 minutes after they were exsanguinated, and related to the levels of cortisol, lactate and creatine kinase in blood collected at exsanguination. Earlier rigor development was associated with higher concentrations of these stress indicators in the blood. This relationship suggests that the mean rigor score, and the frequency distribution of carcases that had or had not entered rigor, could be used as an index of the degree of stress to which the pigs had been subjected.
Accuracy and performance of 3D mask models in optical projection lithography
NASA Astrophysics Data System (ADS)
Agudelo, Viviana; Evanschitzky, Peter; Erdmann, Andreas; Fühner, Tim; Shao, Feng; Limmer, Steffen; Fey, Dietmar
2011-04-01
Different mask models have been compared: rigorous electromagnetic field (EMF) modeling, rigorous EMF modeling with decomposition techniques and the thin mask approach (Kirchhoff approach) to simulate optical diffraction from different mask patterns in projection systems for lithography. In addition, each rigorous model was tested for two different formulations for partially coherent imaging: The Hopkins assumption and rigorous simulation of mask diffraction orders for multiple illumination angles. The aim of this work is to closely approximate results of the rigorous EMF method by the thin mask model enhanced with pupil filtering techniques. The validity of this approach for different feature sizes, shapes and illumination conditions is investigated.
NASA Astrophysics Data System (ADS)
Ngampitipan, Tritos; Boonserm, Petarpa; Chatrabhuti, Auttakit; Visser, Matt
2016-06-01
Hawking radiation is the evidence for the existence of black hole. What an observer can measure through Hawking radiation is the transmission probability. In the laboratory, miniature black holes can successfully be generated. The generated black holes are, most commonly, Myers-Perry black holes. In this paper, we will derive the rigorous bounds on the transmission probabilities for massless scalar fields of non-negative-angular-momentum modes emitted from a generated Myers-Perry black hole in six, seven, and eight dimensions. The results show that for low energy, the rigorous bounds increase with the increase in the energy of emitted particles. However, for high energy, the rigorous bounds decrease with the increase in the energy of emitted particles. When the black holes spin faster, the rigorous bounds decrease. For dimension dependence, the rigorous bounds also decrease with the increase in the number of extra dimensions. Furthermore, as comparison to the approximate transmission probability, the rigorous bound is proven to be useful.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Ngampitipan, Tritos, E-mail: tritos.ngampitipan@gmail.com; Particle Physics Research Laboratory, Department of Physics, Faculty of Science, Chulalongkorn University, Phayathai Road, Patumwan, Bangkok 10330; Boonserm, Petarpa, E-mail: petarpa.boonserm@gmail.com
Hawking radiation is the evidence for the existence of black hole. What an observer can measure through Hawking radiation is the transmission probability. In the laboratory, miniature black holes can successfully be generated. The generated black holes are, most commonly, Myers-Perry black holes. In this paper, we will derive the rigorous bounds on the transmission probabilities for massless scalar fields of non-negative-angular-momentum modes emitted from a generated Myers-Perry black hole in six, seven, and eight dimensions. The results show that for low energy, the rigorous bounds increase with the increase in the energy of emitted particles. However, for high energy,more » the rigorous bounds decrease with the increase in the energy of emitted particles. When the black holes spin faster, the rigorous bounds decrease. For dimension dependence, the rigorous bounds also decrease with the increase in the number of extra dimensions. Furthermore, as comparison to the approximate transmission probability, the rigorous bound is proven to be useful.« less
Madi, Banyana Cecilia; Hussein, Julia; Hounton, Sennen; D'Ambruoso, Lucia; Achadi, Endang; Arhinful, Daniel Kojo
2007-09-01
A participatory approach to priority setting in programme evaluation may help improve the allocation and more efficient use of scarce resources especially in low-income countries. Research agendas that are the result of collaboration between researchers, programme managers, policy makers and other stakeholders have the potential to ensure rigorous studies are conducted on matters of local priority, based on local, expert knowledge. This paper describes a process involving key stakeholders to elicit and prioritise evaluation needs for safe motherhood in three developing countries. A series of reiterative consultations with safe motherhood stakeholders from each country was conducted over a period of 36 months. In each country, the consultation process consisted of a series of participatory workshops; firstly, stakeholder's views on evaluation were elicited with parallel descriptive work on the contexts. Secondly, priorities for evaluation were identified from stakeholders; thirdly, the evaluation-priorities were refined; and finally, the evaluation research questions, reflecting the identified priorities, were agreed and finalised. Three evaluation-questions were identified in each country, and one selected, on which a full scale evaluation was undertaken. While there is a great deal written about the importance of transparent and participatory priority setting in evaluation; few examples of how such processes could be implemented exist, particularly for maternal health programmes. Our experience demonstrates that the investment in a participatory priority-setting effort is high but the process undertaken resulted in both globally and contextually-relevant priorities for evaluation. This experience provides useful lessons for public health practitioners committed to bridging the research-policy interface.
Investigating outliers to improve conceptual models of bedrock aquifers
NASA Astrophysics Data System (ADS)
Worthington, Stephen R. H.
2018-06-01
Numerical models play a prominent role in hydrogeology, with simplifying assumptions being inevitable when implementing these models. However, there is a risk of oversimplification, where important processes become neglected. Such processes may be associated with outliers, and consideration of outliers can lead to an improved scientific understanding of bedrock aquifers. Using rigorous logic to investigate outliers can help to explain fundamental scientific questions such as why there are large variations in permeability between different bedrock lithologies.
Toward Real-Time Infoveillance of Twitter Health Messages.
Colditz, Jason B; Chu, Kar-Hai; Emery, Sherry L; Larkin, Chandler R; James, A Everette; Welling, Joel; Primack, Brian A
2018-06-21
There is growing interest in conducting public health research using data from social media. In particular, Twitter "infoveillance" has demonstrated utility across health contexts. However, rigorous and reproducible methodologies for using Twitter data in public health are not yet well articulated, particularly those related to content analysis, which is a highly popular approach. In 2014, we gathered an interdisciplinary team of health science researchers, computer scientists, and methodologists to begin implementing an open-source framework for real-time infoveillance of Twitter health messages (RITHM). Through this process, we documented common challenges and novel solutions to inform future work in real-time Twitter data collection and subsequent human coding. The RITHM framework allows researchers and practitioners to use well-planned and reproducible processes in retrieving, storing, filtering, subsampling, and formatting data for health topics of interest. Further considerations for human coding of Twitter data include coder selection and training, data representation, codebook development and refinement, and monitoring coding accuracy and productivity. We illustrate methodological considerations through practical examples from formative work related to hookah tobacco smoking, and we reference essential methods literature related to understanding and using Twitter data. (Am J Public Health. Published online ahead of print June 21, 2018: e1-e6. doi:10.2105/AJPH.2018.304497).
Schlägel, Ulrike E; Lewis, Mark A
2016-12-01
Discrete-time random walks and their extensions are common tools for analyzing animal movement data. In these analyses, resolution of temporal discretization is a critical feature. Ideally, a model both mirrors the relevant temporal scale of the biological process of interest and matches the data sampling rate. Challenges arise when resolution of data is too coarse due to technological constraints, or when we wish to extrapolate results or compare results obtained from data with different resolutions. Drawing loosely on the concept of robustness in statistics, we propose a rigorous mathematical framework for studying movement models' robustness against changes in temporal resolution. In this framework, we define varying levels of robustness as formal model properties, focusing on random walk models with spatially-explicit component. With the new framework, we can investigate whether models can validly be applied to data across varying temporal resolutions and how we can account for these different resolutions in statistical inference results. We apply the new framework to movement-based resource selection models, demonstrating both analytical and numerical calculations, as well as a Monte Carlo simulation approach. While exact robustness is rare, the concept of approximate robustness provides a promising new direction for analyzing movement models.
Fitts Cochrane, Jean; Lonsdorf, Eric; Allison, Taber D; Sanders-Reed, Carol A
2015-09-01
Challenges arise when renewable energy development triggers "no net loss" policies for protected species, such as where wind energy facilities affect Golden Eagles in the western United States. When established mitigation approaches are insufficient to fully avoid or offset losses, conservation goals may still be achievable through experimental implementation of unproven mitigation methods provided they are analyzed within a framework that deals transparently and rigorously with uncertainty. We developed an approach to quantify and analyze compensatory mitigation that (1) relies on expert opinion elicited in a thoughtful and structured process to design the analysis (models) and supplement available data, (2) builds computational models as hypotheses about cause-effect relationships, (3) represents scientific uncertainty in stochastic model simulations, (4) provides probabilistic predictions of "relative" mortality with and without mitigation, (5) presents results in clear formats useful to applying risk management preferences (regulatory standards) and selecting strategies and levels of mitigation for immediate action, and (6) defines predictive parameters in units that could be monitored effectively, to support experimental adaptive management and reduction in uncertainty. We illustrate the approach with a case study characterized by high uncertainty about underlying biological processes and high conservation interest: estimating the quantitative effects of voluntary strategies to abate lead poisoning in Golden Eagles in Wyoming due to ingestion of spent game hunting ammunition.
Reframing Rigor: A Modern Look at Challenge and Support in Higher Education
ERIC Educational Resources Information Center
Campbell, Corbin M.; Dortch, Deniece; Burt, Brian A.
2018-01-01
This chapter describes the limitations of the traditional notions of academic rigor in higher education, and brings forth a new form of rigor that has the potential to support student success and equity.
The Importance of Method Selection in Determining Product Integrity for Nutrition Research1234
Mudge, Elizabeth M; Brown, Paula N
2016-01-01
The American Herbal Products Association estimates that there as many as 3000 plant species in commerce. The FDA estimates that there are about 85,000 dietary supplement products in the marketplace. The pace of product innovation far exceeds that of analytical methods development and validation, with new ingredients, matrixes, and combinations resulting in an analytical community that has been unable to keep up. This has led to a lack of validated analytical methods for dietary supplements and to inappropriate method selection where methods do exist. Only after rigorous validation procedures to ensure that methods are fit for purpose should they be used in a routine setting to verify product authenticity and quality. By following systematic procedures and establishing performance requirements for analytical methods before method development and validation, methods can be developed that are both valid and fit for purpose. This review summarizes advances in method selection, development, and validation regarding herbal supplement analysis and provides several documented examples of inappropriate method selection and application. PMID:26980823
The Importance of Method Selection in Determining Product Integrity for Nutrition Research.
Mudge, Elizabeth M; Betz, Joseph M; Brown, Paula N
2016-03-01
The American Herbal Products Association estimates that there as many as 3000 plant species in commerce. The FDA estimates that there are about 85,000 dietary supplement products in the marketplace. The pace of product innovation far exceeds that of analytical methods development and validation, with new ingredients, matrixes, and combinations resulting in an analytical community that has been unable to keep up. This has led to a lack of validated analytical methods for dietary supplements and to inappropriate method selection where methods do exist. Only after rigorous validation procedures to ensure that methods are fit for purpose should they be used in a routine setting to verify product authenticity and quality. By following systematic procedures and establishing performance requirements for analytical methods before method development and validation, methods can be developed that are both valid and fit for purpose. This review summarizes advances in method selection, development, and validation regarding herbal supplement analysis and provides several documented examples of inappropriate method selection and application. © 2016 American Society for Nutrition.
Trends in Study Methods Used in Undergraduate Medical Education Research, 1969–2007
Baernstein, Amy; Liss, Hillary K.; Carney, Patricia A.; Elmore, Joann G.
2011-01-01
Context Evidence-based medical education requires rigorous studies appraising educational efficacy. Objectives To assess trends over time in methods used to evaluate undergraduate medical education interventions and to identify whether participation of medical education departments or centers is associated with more rigorous methods. Data Sources The PubMed, Cochrane Controlled Trials Registry, Campbell Collaboration, and ERIC databases (January 1966–March 2007) were searched using terms equivalent to students, medical and education, medical crossed with all relevant study designs. Study Selection We selected publications in all languages from every fifth year, plus the most recent 12 months, that evaluated an educational intervention for undergraduate medical students. Four hundred seventy-two publications met criteria for review. Data Extraction Data were abstracted on number of participants; types of comparison groups; whether outcomes assessed were objective, subjective, and/or validated; timing of outcome assessments; funding; and participation of medical education departments and centers. Ten percent of publications were independently abstracted by 2 authors to assess validity of the data abstraction. Results The annual number of publications increased over time from 1 (1969–1970) to 147 (2006–2007). In the most recent year, there was a mean of 145 medical student participants; 9 (6%) recruited participants from multiple institutions; 80 (54%) used comparison groups; 37 (25%) used randomized control groups; 91 (62%) had objective outcomes; 23 (16%) had validated outcomes; 35 (24%) assessed an outcome more than 1 month later; 21 (14%) estimated statistical power; and 66 (45%) reported funding. In 2006–2007, medical education department or center participation, reported in 46 (31%) of the recent publications, was associated only with enrolling more medical student participants (P = .04); for all studies from 1969 to 2007, it was associated only with measuring an objective outcome (P = .048). Between 1969 and 2007, the percentage of publications reporting statistical power and funding increased; percentages did not change for other study features. Conclusions The annual number of published studies of undergraduate medical education interventions demonstrating methodological rigor has been increasing. However, considerable opportunities for improvement remain. PMID:17785648
Tingey, Lauren; Chambers, Rachel; Goklish, Novalene; Larzelere, Francene; Lee, Angelita; Suttle, Rosemarie; Rosenstock, Summer; Lake, Kristin; Barlow, Allison
2017-02-27
American Indian adolescents have one of the highest rates of teen pregnancy and repeat teen births in the US. Substance use is a significant risk factor for unprotected sex, and American Indian adolescents have the highest substance use-related morbidity and mortality of any US racial group. Despite these disparities, there are no existing, evidence-based programs for pregnancy prevention that have been rigorously evaluated among American Indian teens. The proposed study is a randomized controlled trial to test the efficacy of a comprehensive sexual and reproductive health program developed in partnership with an American Indian community. Participants will be American Indians ages 11-19 and their parent or trusted adult, randomized to receive the control condition or intervention called Respecting the Circle of Life: Mind, Body and Spirit. The intervention includes eight lessons delivered to self-selected peer groups during a summer basketball camp and one lesson delivered to the youth and parent/trusted adult together within 3 months after camp. All lessons are administered by trained community health workers from the participating American Indian community. Youth and parent/trusted adult participants will complete assessments at baseline, 3, 9, 12, 24 and 36 months post-intervention completion. The primary outcome variables are sexual/reproductive health knowledge, sexual initiation, condom use self-efficacy and intent to use a condom at next sex as changed from baseline to post-intervention between intervention and control participants. Selected primary outcomes are applicable to all study participants. Currently there are no sexual and reproductive health programs designed specifically for American Indian youth that have been rigorously evaluated and found to have an evidence base. Respecting the Circle of Life is highly innovative by incorporating lesson delivery into a summer basketball camp and involving parents or other trusted adults in curriculum administration. If found successful, it will be the first evidence-based program for teen pregnancy prevention for American Indian youth and adolescents. Clinicaltrials.gov, NCT02904629 . Retrospectively registered on 23 September 2016.
Vecchio, Jennifer; Kearney, Christopher A
2009-12-01
Selective mutism is a severe childhood disorder involving failure to speak in public situations in which speaking is expected. The present study examined 9 youths with selective mutism treated with child-focused, exposure-based practices and parent-focused contingency management via an alternating treatments design. Broadband measures of functioning were employed, but particular focus was made on behavioral assessment of words spoken audibly and daily in public situations. Treatment ranged from 8 to 32 sessions and resulted in positive end-state functioning for 8 of 9 participants. Broader analyses indicated greater effectiveness for exposure-based practice than contingency management. The results support recent case reports of behavioral treatment for this population but in more rigorous fashion. Clinical and research challenges are discussed, including caveats about length and intensity of treatment for this population and need to develop standardized daily measures.
Neville, Helen J.; Stevens, Courtney; Pakulak, Eric; Bell, Theodore A.; Fanning, Jessica; Klein, Scott; Isbell, Elif
2013-01-01
Using information from research on the neuroplasticity of selective attention and on the central role of successful parenting in child development, we developed and rigorously assessed a family-based training program designed to improve brain systems for selective attention in preschool children. One hundred forty-one lower socioeconomic status preschoolers enrolled in a Head Start program were randomly assigned to the training program, Head Start alone, or an active control group. Electrophysiological measures of children’s brain functions supporting selective attention, standardized measures of cognition, and parent-reported child behaviors all favored children in the treatment program relative to both control groups. Positive changes were also observed in the parents themselves. Effect sizes ranged from one-quarter to half of a standard deviation. These results lend impetus to the further development and broader implementation of evidence-based education programs that target at-risk families. PMID:23818591
Optimisation algorithms for ECG data compression.
Haugland, D; Heber, J G; Husøy, J H
1997-07-01
The use of exact optimisation algorithms for compressing digital electrocardiograms (ECGs) is demonstrated. As opposed to traditional time-domain methods, which use heuristics to select a small subset of representative signal samples, the problem of selecting the subset is formulated in rigorous mathematical terms. This approach makes it possible to derive algorithms guaranteeing the smallest possible reconstruction error when a bounded selection of signal samples is interpolated. The proposed model resembles well-known network models and is solved by a cubic dynamic programming algorithm. When applied to standard test problems, the algorithm produces a compressed representation for which the distortion is about one-half of that obtained by traditional time-domain compression techniques at reasonable compression ratios. This illustrates that, in terms of the accuracy of decoded signals, existing time-domain heuristics for ECG compression may be far from what is theoretically achievable. The paper is an attempt to bridge this gap.
Monitoring programs to assess reintroduction efforts: A critical component in recovery
Muths, E.; Dreitz, V.
2008-01-01
Reintroduction is a powerful tool in our conservation toolbox. However, the necessary follow-up, i.e. long-term monitoring, is not commonplace and if instituted may lack rigor. We contend that valid monitoring is possible, even with sparse data. We present a means to monitor based on demographic data and a projection model using the Wyoming toad (Bufo baxten) as an example. Using an iterative process, existing data is built upon gradually such that demographic estimates and subsequent inferences increase in reliability. Reintroduction and defensible monitoring may become increasingly relevant as the outlook for amphibians, especially in tropical regions, continues to deteriorate and emergency collection, captive breeding, and reintroduction become necessary. Rigorous use of appropriate modeling and an adaptive approach can validate the use of reintroduction and substantially increase its value to recovery programs. ?? 2008 Museu de Cie??ncies Naturals.
Testability of evolutionary game dynamics based on experimental economics data
NASA Astrophysics Data System (ADS)
Wang, Yijia; Chen, Xiaojie; Wang, Zhijian
2017-11-01
Understanding the dynamic processes of a real game system requires an appropriate dynamics model, and rigorously testing a dynamics model is nontrivial. In our methodological research, we develop an approach to testing the validity of game dynamics models that considers the dynamic patterns of angular momentum and speed as measurement variables. Using Rock-Paper-Scissors (RPS) games as an example, we illustrate the geometric patterns in the experiment data. We then derive the related theoretical patterns from a series of typical dynamics models. By testing the goodness-of-fit between the experimental and theoretical patterns, we show that the validity of these models can be evaluated quantitatively. Our approach establishes a link between dynamics models and experimental systems, which is, to the best of our knowledge, the most effective and rigorous strategy for ascertaining the testability of evolutionary game dynamics models.
Control group design: enhancing rigor in research of mind-body therapies for depression.
Kinser, Patricia Anne; Robins, Jo Lynne
2013-01-01
Although a growing body of research suggests that mind-body therapies may be appropriate to integrate into the treatment of depression, studies consistently lack methodological sophistication particularly in the area of control groups. In order to better understand the relationship between control group selection and methodological rigor, we provide a brief review of the literature on control group design in yoga and tai chi studies for depression, and we discuss challenges we have faced in the design of control groups for our recent clinical trials of these mind-body complementary therapies for women with depression. To address the multiple challenges of research about mind-body therapies, we suggest that researchers should consider 4 key questions: whether the study design matches the research question; whether the control group addresses performance, expectation, and detection bias; whether the control group is ethical, feasible, and attractive; and whether the control group is designed to adequately control for nonspecific intervention effects. Based on these questions, we provide specific recommendations about control group design with the goal of minimizing bias and maximizing validity in future research.
Control Group Design: Enhancing Rigor in Research of Mind-Body Therapies for Depression
Kinser, Patricia Anne; Robins, Jo Lynne
2013-01-01
Although a growing body of research suggests that mind-body therapies may be appropriate to integrate into the treatment of depression, studies consistently lack methodological sophistication particularly in the area of control groups. In order to better understand the relationship between control group selection and methodological rigor, we provide a brief review of the literature on control group design in yoga and tai chi studies for depression, and we discuss challenges we have faced in the design of control groups for our recent clinical trials of these mind-body complementary therapies for women with depression. To address the multiple challenges of research about mind-body therapies, we suggest that researchers should consider 4 key questions: whether the study design matches the research question; whether the control group addresses performance, expectation, and detection bias; whether the control group is ethical, feasible, and attractive; and whether the control group is designed to adequately control for nonspecific intervention effects. Based on these questions, we provide specific recommendations about control group design with the goal of minimizing bias and maximizing validity in future research. PMID:23662111
A square-force cohesion model and its extraction from bulk measurements
NASA Astrophysics Data System (ADS)
Liu, Peiyuan; Lamarche, Casey; Kellogg, Kevin; Hrenya, Christine
2017-11-01
Cohesive particles remain poorly understood, with order of magnitude differences exhibited for prior, physical predictions of agglomerate size. A major obstacle lies in the absence of robust models of particle-particle cohesion, thereby precluding accurate prediction of the behavior of cohesive particles. Rigorous cohesion models commonly contain parameters related to surface roughness, to which cohesion shows extreme sensitivity. However, both roughness measurement and its distillation into these model parameters are challenging. Accordingly, we propose a ``square-force'' model, where cohesive force remains constant until a cut-off separation. Via DEM simulations, we demonstrate validity of the square-force model as surrogate of more rigorous models, when its two parameters are selected to match the two key quantities governing dense and dilute granular flows, namely maximum cohesive force and critical cohesive energy, respectively. Perhaps more importantly, we establish a method to extract the parameters in the square-force model via defluidization, due to its ability to isolate the effects of the two parameters. Thus, instead of relying on complicated scans of individual grains, determination of particle-particle cohesion from simple bulk measurements becomes feasible. Dow Corning Corporation.
When is good, good enough? Methodological pragmatism for sustainable guideline development.
Browman, George P; Somerfield, Mark R; Lyman, Gary H; Brouwers, Melissa C
2015-03-06
Continuous escalation in methodological and procedural rigor for evidence-based processes in guideline development is associated with increasing costs and production delays that threaten sustainability. While health research methodologists are appropriately responsible for promoting increasing rigor in guideline development, guideline sponsors are responsible for funding such processes. This paper acknowledges that other stakeholders in addition to methodologists should be more involved in negotiating trade-offs between methodological procedures and efficiency in guideline production to produce guidelines that are 'good enough' to be trustworthy and affordable under specific circumstances. The argument for reasonable methodological compromise to meet practical circumstances is consistent with current implicit methodological practice. This paper proposes a conceptual tool as a framework to be used by different stakeholders in negotiating, and explicitly reporting, reasonable compromises for trustworthy as well as cost-worthy guidelines. The framework helps fill a transparency gap in how methodological choices in guideline development are made. The principle, 'when good is good enough' can serve as a basis for this approach. The conceptual tool 'Efficiency-Validity Methodological Continuum' acknowledges trade-offs between validity and efficiency in evidence-based guideline development and allows for negotiation, guided by methodologists, of reasonable methodological compromises among stakeholders. Collaboration among guideline stakeholders in the development process is necessary if evidence-based guideline development is to be sustainable.
High pressure processing of fresh meat--is it worth it?
Ma, Hanjun; Ledward, D A
2013-12-01
When subjected to high pressures at ambient temperatures, the shelf life of fresh meat is increased; however the meat takes on a cooked appearance as the actomyosin denatures at about 200MPa and the myoglobin denatures and/or co-precipitates with other proteins at about 400MPa. In addition, at pressures of 400MPa and above the unsaturated lipids in the meat become more susceptible to oxidation, probably due to the release of iron from complexes present in meat (haemosiderin and ferritin?) and/or changes in the lipid membrane. Thus, even if legislation allowed it, it is unlikely that many consumers would be prepared to buy such meat. However if pre-rigor meat is subjected to pressures of about 100-150MPa, below those necessary to cause colour changes, it becomes significantly more tender than its untreated counterpart and this may now be a commercially viable process, given the decreasing cost of high pressure rigs. When treated at 100-200MPa while the temperature is raised from ambient to around 60°C post-rigor meat also yields a tender product and this may also be a commercially attractive process to parts of the food industry, for example those involved in catering. Copyright © 2013 Elsevier Ltd. All rights reserved.
NASA Astrophysics Data System (ADS)
Vassena, G.; Clerici, A.
2018-05-01
The state of the art of 3D surveying technologies, if correctly applied, allows to obtain 3D coloured models of large open pit mines using different technologies as terrestrial laser scanner (TLS), with images, combined with UAV based digital photogrammetry. GNSS and/or total station are also currently used to geo reference the model. The University of Brescia has been realised a project to map in 3D an open pit mine located in Botticino, a famous location of marble extraction close to Brescia in North Italy. Terrestrial Laser Scanner 3D point clouds combined with RGB images and digital photogrammetry from UAV have been used to map a large part of the cave. By rigorous and well know procedures a 3D point cloud and mesh model have been obtained using an easy and rigorous approach. After the description of the combined mapping process, the paper describes the innovative process proposed for the daily/weekly update of the model itself. To realize this task a SLAM technology approach is described, using an innovative approach based on an innovative instrument capable to run an automatic localization process and real time on the field change detection analysis.
Conducting a multicentre and multinational qualitative study on patient transitions.
Johnson, Julie K; Barach, Paul; Vernooij-Dassen, Myrra
2012-12-01
A multicentre, multinational research study requires careful planning and coordination to accomplish the aims of the study and to ensure systematic and rigorous examination of all project methods and data collected. The aim of this paper is to describe the approach we used during the HANDOVER Project to develop a multicentre, multinational research project for studying transitions of patient care while creating a community of practice for the researchers. We highlight the process used to assure the quality of a multicentre qualitative study and to create a codebook for data analysis as examples of attending to the community of practice while conducting rigorous qualitative research. Essential elements for the success of this multinational, multilanguage research project included recruiting a strong research team, explicit planning for decision-making processes to be used throughout the project, acknowledging the differences among the study settings and planning the protocols to capitalise upon those differences. Although not commonly discussed in reports of large research projects, there is an underlying, concurrent stream of activities to develop a cohesive team that trusts and respects one another's skills and that engage independent researchers in a group process that contributes to achieving study goals. We discuss other lessons learned and offer recommendations for other teams planning multicentre research.
Rigor or mortis: best practices for preclinical research in neuroscience.
Steward, Oswald; Balice-Gordon, Rita
2014-11-05
Numerous recent reports document a lack of reproducibility of preclinical studies, raising concerns about potential lack of rigor. Examples of lack of rigor have been extensively documented and proposals for practices to improve rigor are appearing. Here, we discuss some of the details and implications of previously proposed best practices and consider some new ones, focusing on preclinical studies relevant to human neurological and psychiatric disorders. Copyright © 2014 Elsevier Inc. All rights reserved.
Forster, B; Ropohl, D; Raule, P
1977-07-05
The manual examination of rigor mortis as currently used and its often subjective evaluation frequently produced highly incorrect deductions. It is therefore desirable that such inaccuracies should be replaced by the objective measuring of rigor mortis at the extremities. To that purpose a method is described which can also be applied in on-the-spot investigations and a new formula for the determination of rigor mortis--indices (FRR) is introduced.
High blood pressure and visual sensitivity
NASA Astrophysics Data System (ADS)
Eisner, Alvin; Samples, John R.
2003-09-01
The study had two main purposes: (1) to determine whether the foveal visual sensitivities of people treated for high blood pressure (vascular hypertension) differ from the sensitivities of people who have not been diagnosed with high blood pressure and (2) to understand how visual adaptation is related to standard measures of systemic cardiovascular function. Two groups of middle-aged subjects-hypertensive and normotensive-were examined with a series of test/background stimulus combinations. All subjects met rigorous inclusion criteria for excellent ocular health. Although the visual sensitivities of the two subject groups overlapped extensively, the age-related rate of sensitivity loss was, for some measures, greater for the hypertensive subjects, possibly because of adaptation differences between the two groups. Overall, the degree of steady-state sensitivity loss resulting from an increase of background illuminance (for 580-nm backgrounds) was slightly less for the hypertensive subjects. Among normotensive subjects, the ability of a bright (3.8-log-td), long-wavelength (640-nm) adapting background to selectively suppress the flicker response of long-wavelength-sensitive (LWS) cones was related inversely to the ratio of mean arterial blood pressure to heart rate. The degree of selective suppression was also related to heart rate alone, and there was evidence that short-term changes of cardiovascular response were important. The results suggest that (1) vascular hypertension, or possibly its treatment, subtly affects visual function even in the absence of eye disease and (2) changes in blood flow affect retinal light-adaptation processes involved in the selective suppression of the flicker response from LWS cones caused by bright, long-wavelength backgrounds.
Writing a Research Proposal to The Research Council of Oman
Al-Shukaili, Ahmed; Al-Maniri, Abdullah
2017-01-01
Writing a research proposal can be a challenging task for young researchers. This article explains how to write a strong research proposal to apply for funding, specifically, a proposal for The Research Council (TRC) of Oman. Three different research proposal application forms are currently used in TRC, including Open Research Grant (ORG), Graduate Research Support Program (GRSP), and Faculty-mentored Undergraduate Research Award Program (FURAP). The application forms are filled and submitted electronically on TRC website. Each of the proposals submitted to TRC is selected through a rigorous reviewing and screening process. Novelty and originality of the research idea is the most crucial element in writing a research proposal. Performing an in-depth review of the literature will assist you to compose a good researchable question and generate a strong hypothesis. The development of a good hypothesis will offer insight into the specific objectives of a study. Research objectives should be focused, measurable, and achievable by a specific time using the most appropriate methodology. Moreover, it is essential to select a proper study design in-line with the purpose of the study and the hypothesis. Furthermore, social/economic impact and reasonable budget of proposed research are important criteria in research proposal evaluation by TRC. Finally, ethical principles should be observed before writing a research proposal involving human or animal subjects. PMID:28584597
Families as full research partners: what's in it for us?
Vander Stoep, A; Williams, M; Jones, R; Green, L; Trupin, E
1999-08-01
This article describes a children's managed mental health care program that incorporates both a family participation service model and a family-initiated evaluation model. The authors begin by tracing the evolution of the family support and the participatory research movements leading to current developments in children's mental health services research. In the King County Blended Funding Project, three service systems pool funds that are spent flexibly by child and family teams. Family advocates have led efforts to design and implement the project evaluation. During this process, many tensions have arisen between meeting the demands of both scientific rigor and multiple community stakeholders. Examples are given of the issues raised by family advocates and research scientists as together they established a theory of change, identified meaningful outcomes, selected measurement tools, and implemented the evaluation protocol. Guidelines are given for how services research partnerships can be successfully built to better address community needs. This article was selected as a distinguished research paper based on a presentation by the authors at the 1998 11th annual research conference titled "A system of Care for Children's Mental Health: Expanding the Research Base." This conference is sponsored annually by the Research and Training Center for Children's Mental Health, Department of Child and Family Studies, Louis de la Parte Florida Mental Health Institute, University of South Florida, Tampa.
Brodusch, Nicolas; Demers, Hendrix; Gauvin, Raynald
2015-01-01
Dark-field (DF) images were acquired in the scanning electron microscope with an offline procedure based on electron backscatter diffraction (EBSD) patterns (EBSPs). These EBSD-DF images were generated by selecting a particular reflection on the electron backscatter diffraction pattern and by reporting the intensity of one or several pixels around this point at each pixel of the EBSD-DF image. Unlike previous studies, the diffraction information of the sample is the basis of the final image contrast with a pixel scale resolution at the EBSP providing DF imaging in the scanning electron microscope. The offline facility of this technique permits the selection of any diffraction condition available in the diffraction pattern and displaying the corresponding image. The high number of diffraction-based images available allows a better monitoring of deformation structures compared to electron channeling contrast imaging (ECCI) which is generally limited to a few images of the same area. This technique was applied to steel and iron specimens and showed its high capability in describing more rigorously the deformation structures around micro-hardness indents. Due to the offline relation between the reference EBSP and the EBSD-DF images, this new technique will undoubtedly greatly improve our knowledge of deformation mechanism and help to improve our understanding of the ECCI contrast mechanisms. Copyright © 2014 Elsevier B.V. All rights reserved.
Resource Utilization and Site Selection for a Self-Sufficient Martian Outpost
NASA Technical Reports Server (NTRS)
Barker, Donald; Chamitoff, Gregory; James, George
1998-01-01
As a planet with striking similarities to Earth, Mars is an important focus for scientific research aimed at understanding the processes of planetary evolution and the formation of our solar system. Fortunately, Mars is also a planet with abundant natural resources, including assessible materials that can be used to support human life and to sustain a self-sufficient martian outpost. Resources required include water, breathable air, food, shelter, energy, and fuel. Through a mission design based on in situ resource development, we can establish a permanent outpost on Mars beginning with the first manned mission. This paper examines the potential for supporting the first manned mission with the objective of achieving self-sufficiency through well-understood resource development and a program of rigorous scientific research aimed at extending that capability. We examine the potential for initially extracting critical resources from the martian environment, and discuss the scientific investigations required to identify additional resources in the atmosphere, on the surface, and within the subsurface. We also discuss our current state of knowledge of Mars, technical considerations of resource utilization, and using unmanned missions' data for selecting an optimal site. The primary goal of achieving self-sufficiency on Mars would accelerate the development of human colonization beyond Earth, while providing a robust and permanent martian base from which humans can explore and conduct long-term research on planetary evolution, the solar system, and life itself.
Doyle, D M; Dauterive, R; Chuang, K H; Ellrodt, A G
2001-11-01
There are many challenges to effectively and efficiently translating evidence into practice. Potential strategies include (1) training more evidence-based practitioners in the art and science of evidence-based medicine, (2) enhancing the quality and availability of systematic reviews, and (3) more effectively linking evidence-based practitioners and evidence users through comprehensive behavioral change initiatives. Herein we explore the third strategy and highlight the key elements of success for a program using behavioral change strategies. We present a clinical model based on clear understanding of the "problem," a systematic approach to diagnosis, selection of scientifically sound treatment options, and effective evaluation with appropriate modification of the treatment plan. A successful program begins with effective team leadership, the expression of a clinically compelling case for change, and commitment to the pursuit of perfection in the delivery of key evidence-based interventions. The team must then diagnose behavioral barriers to change, using a systematic approach based on a published rigorous differential diagnosis framework. This diagnostic step provides the foundation for selection of effective dissemination and implementation strategies (treatments) proven to improve processes of care and clinical outcomes. Finally the team must evaluate progress toward perfection, reviewing interim data and adjusting the treatment regimen to newly diagnosed barriers. We then present a specific project (improving pneumococcal immunization rates in our rural community) and interim results to demonstrate the use of the framework in the real world.
NASA Astrophysics Data System (ADS)
Moine, Edouard; Privat, Romain; Sirjean, Baptiste; Jaubert, Jean-Noël
2017-09-01
The Gibbs energy of solvation measures the affinity of a solute for its solvent and is thus a key property for the selection of an appropriate solvent for a chemical synthesis or a separation process. More fundamentally, Gibbs energies of solvation are choice data for developing and benchmarking molecular models predicting solvation effects. The Comprehensive Solvation—CompSol—database was developed with the ambition to propose very large sets of new experimental solvation chemical-potential, solvation entropy, and solvation enthalpy data of pure and mixed components, covering extended temperature ranges. For mixed compounds, the solvation quantities were generated in infinite-dilution conditions by combining experimental values of pure-component and binary-mixture thermodynamic properties. Three types of binary-mixture properties were considered: partition coefficients, activity coefficients at infinite dilution, and Henry's-law constants. A rigorous methodology was implemented with the aim to select data at appropriate conditions of temperature, pressure, and concentration for the estimation of solvation data. Finally, our comprehensive CompSol database contains 21 671 data associated with 1969 pure species and 70 062 data associated with 14 102 binary mixtures (including 760 solvation data related to the ionic-liquid class of solvents). On the basis of the very large amount of experimental data contained in the CompSol database, it is finally discussed how solvation energies are influenced by hydrogen-bonding association effects.
NASA Technical Reports Server (NTRS)
Nall, Marsha M.; Barna, Gerald J.
2009-01-01
The John Glenn Biomedical Engineering Consortium was established by NASA in 2002 to formulate and implement an integrated, interdisciplinary research program to address risks faced by astronauts during long-duration space missions. The consortium is comprised of a preeminent team of Northeast Ohio institutions that include Case Western Reserve University, the Cleveland Clinic, University Hospitals Case Medical Center, The National Center for Space Exploration Research, and the NASA Glenn Research Center. The John Glenn Biomedical Engineering Consortium research is focused on fluid physics and sensor technology that addresses the critical risks to crew health, safety, and performance. Effectively utilizing the unique skills, capabilities and facilities of the consortium members is also of prime importance. Research efforts were initiated with a general call for proposals to the consortium members. The top proposals were selected for funding through a rigorous, peer review process. The review included participation from NASA's Johnson Space Center, which has programmatic responsibility for NASA's Human Research Program. The projects range in scope from delivery of prototype hardware to applied research that enables future development of advanced technology devices. All of the projects selected for funding have been completed and the results are summarized. Because of the success of the consortium, the member institutions have extended the original agreement to continue this highly effective research collaboration through 2011.
Hong, Ka L; Battistella, Luisa; Salva, Alysia D; Williams, Ryan M; Sooter, Letha J
2015-01-27
Alpha toxin is one of the major virulence factors secreted by Staphylococcus aureus, a bacterium that is responsible for a wide variety of infections in both community and hospital settings. Due to the prevalence of S. aureus related infections and the emergence of methicillin-resistant S. aureus, rapid and accurate diagnosis of S. aureus infections is crucial in benefiting patient health outcomes. In this study, a rigorous Systematic Evolution of Ligands by Exponential Enrichment (SELEX) variant previously developed by our laboratory was utilized to select a single-stranded DNA molecular recognition element (MRE) targeting alpha toxin with high affinity and specificity. At the end of the 12-round selection, the selected MRE had an equilibrium dissociation constant (Kd) of 93.7 ± 7.0 nM. Additionally, a modified sandwich enzyme-linked immunosorbent assay (ELISA) was developed by using the selected ssDNA MRE as the toxin-capturing element and a sensitive detection of 200 nM alpha toxin in undiluted human serum samples was achieved.
Parent Management Training-Oregon Model: Adapting Intervention with Rigorous Research.
Forgatch, Marion S; Kjøbli, John
2016-09-01
Parent Management Training-Oregon Model (PMTO(®) ) is a set of theory-based parenting programs with status as evidence-based treatments. PMTO has been rigorously tested in efficacy and effectiveness trials in different contexts, cultures, and formats. Parents, the presumed agents of change, learn core parenting practices, specifically skill encouragement, limit setting, monitoring/supervision, interpersonal problem solving, and positive involvement. The intervention effectively prevents and ameliorates children's behavior problems by replacing coercive interactions with positive parenting practices. Delivery format includes sessions with individual families in agencies or families' homes, parent groups, and web-based and telehealth communication. Mediational models have tested parenting practices as mechanisms of change for children's behavior and found support for the theory underlying PMTO programs. Moderating effects include children's age, maternal depression, and social disadvantage. The Norwegian PMTO implementation is presented as an example of how PMTO has been tailored to reach diverse populations as delivered by multiple systems of care throughout the nation. An implementation and research center in Oslo provides infrastructure and promotes collaboration between practitioners and researchers to conduct rigorous intervention research. Although evidence-based and tested within a wide array of contexts and populations, PMTO must continue to adapt to an ever-changing world. © 2016 Family Process Institute.
Painter, Thomas M; Ngalame, Paulyne M; Lucas, Basil; Lauby, Jennifer L; Herbst, Jeffrey H
2010-10-01
Community-based organizations (CBOs) play an important role in health promotion efforts and the delivery of HIV prevention interventions for at-risk minority populations. CBOs may also develop their own interventions but often lack the capacity or funds to rigorously evaluate them. The Innovative Interventions project of the Centers for Disease Control and Prevention (CDC) funded three CBOs to rigorously evaluate the efficacy of interventions they had developed and were delivering to Black women, Black men who have sex with men (MSM), and adolescent males in juvenile justice settings, respectively. The evaluation results have been reported elsewhere. This article describes operational issues that the CBOs identified as being particularly salient to their evaluations and the strategies they developed to address the issues and successfully complete their evaluations. These issues included the development of organizational capacity to conduct a rigorous outcome evaluation, difficulties with recruitment and retention of evaluation participants, and the use of process monitoring data to improve intervention delivery. The strategies described in this article can be used by CBOs when evaluating their locally developed HIV prevention interventions and may be of interest to funding agencies and researchers that collaborate with CBOs to evaluate their interventions.
NASA Astrophysics Data System (ADS)
Tewksbury, B. J.; MacDonald, R. H.
2004-12-01
As part of a professional development program for faculty in the geosciences, the NSF-funded program On the Cutting Edge (http://serc.carleton.edu/NAGTWorkshops/) has developed and offered workshops for geoscience faculty that guide participants through a stimulating process designed to help faculty members articulate goals and design effective and innovative courses that both meet those goals and assess outcomes. Of approximately 150 faculty members who have participated in the workshops, more than 120 have designed introductory courses in topics ranging from physical geology to Earth systems to historical geology to oceanography. The method of course design taught through these workshops leads to the development of rigorous, student-centered introductory courses. Our method of course design begins, not with a list of content items, but with setting goals by answering the question, "What do I want my students to be able to do on their own when they are done with my class?", rather than the question, "What do I want my students to know in this subject?" Focusing on what faculty members want students to be able to do, rather than on what topics should be covered by the faculty member, promotes designing courses in which students are actively engaged in doing geoscience. This course design method emphasizes setting goals for students involving higher order thinking skills (e.g., analysis, synthesis, design, formulation, prediction, interpretation, evaluation), rather than lower order thinking skills (e.g., identification, description, recognition, classification). For example, the goal of having students be able to evaluate the geologic hazards in an unfamiliar region involves higher order thinking skills and engages the student in deeper analysis than simply asking students to recall and describe examples of geologic hazards covered in class. This goal also has imbedded in it many lower order thinking skills tasks (e.g., identification, description). Rigor comes in having the students involved in doing significant and meaningful geologic tasks. Long-term value comes from improving students' abilities for future challenges, rather focusing on having students pass the final exam. Goals for courses of many different types can be found in the Cutting Edge goals/syllabus database at http://serc.carleton.edu/NAGTWorkshops/coursedesign/browse.html. The workshop also introduces participants to a wide range of teaching and assessment tools so that faculty members will leave the workshop with a larger toolbox of techniques to choose from when deciding how to give students practice during the semester in tasks relevant to the goals of the course and how to evaluate students' progress toward the goals. Most of the techniques emphasize student engagement, which promotes development of more rigorous courses. Over 100 institutions now have introductory geoscience courses designed by faculty members who have participated in our course design workshops. Participants have stated repeatedly in evaluations that the workshop transformed their views of the course design process and that they will carry the focus on goals, student engagement, and rigor into designing other courses and assessing curricula in their departments. On the Cutting Edge is in the process of developing a web-based course design workshop so that faculty who cannot attend our face-to-face workshops can go through our course design process.
Walker, Sarah Cusworth; Bumbarger, Brian K; Phillippi, Stephen W
2015-10-01
Evidence-based programs (EBPs) are an increasingly visible aspect of the treatment landscape in juvenile justice. Research demonstrates that such programs yield positive returns on investment and are replacing more expensive, less effective options. However, programs are unlikely to produce expected benefits when they are not well-matched to community needs, not sustained and do not reach sufficient reach and scale. We argue that achieving these benchmarks for successful implementation will require states and county governments to invest in data-driven decision infrastructure in order to respond in a rigorous and flexible way to shifting political and funding climates. We conceptualize this infrastructure as diagnostic capacity and evaluative capacity: Diagnostic capacity is defined as the process of selecting appropriate programing and evaluative capacity is defined as the ability to monitor and evaluate progress. Policy analyses of Washington State, Pennsylvania and Louisiana's program implementation successes are used to illustrate the benefits of diagnostic and evaluate capacity as a critical element of EBP implementation. Copyright © 2015 Elsevier Ltd. All rights reserved.
NASA Technical Reports Server (NTRS)
Prudhomme, C.; Rovas, D. V.; Veroy, K.; Machiels, L.; Maday, Y.; Patera, A. T.; Turinici, G.; Zang, Thomas A., Jr. (Technical Monitor)
2002-01-01
We present a technique for the rapid and reliable prediction of linear-functional outputs of elliptic (and parabolic) partial differential equations with affine parameter dependence. The essential components are (i) (provably) rapidly convergent global reduced basis approximations, Galerkin projection onto a space W(sub N) spanned by solutions of the governing partial differential equation at N selected points in parameter space; (ii) a posteriori error estimation, relaxations of the error-residual equation that provide inexpensive yet sharp and rigorous bounds for the error in the outputs of interest; and (iii) off-line/on-line computational procedures, methods which decouple the generation and projection stages of the approximation process. The operation count for the on-line stage, in which, given a new parameter value, we calculate the output of interest and associated error bound, depends only on N (typically very small) and the parametric complexity of the problem; the method is thus ideally suited for the repeated and rapid evaluations required in the context of parameter estimation, design, optimization, and real-time control.
Economic Evaluation of Health IT.
Luzi, Daniela; Pecoraro, Fabrizio; Tamburis, Oscar
2016-01-01
Economic evaluation in health care supports decision makers in prioritizing interventions and maximizing the available limited resources for social benefits. Health Information Technology (health IT) constitutes a promising strategy to improve the quality and delivery of health care. However, to determine whether the appropriate health IT solution has been selected in a specific health context, its impact on the clinical and organizational process, on costs, on user satisfaction as well as on patient outcomes, a rigorous and multidimensional evaluation analysis is necessary. Starting from the principles of evaluation introduced since the mid-1980s within the Health Technology Assessment (HTA) guidelines, this contribution provides an overview of the main challenging issues related to the complex task of performing an economic evaluation of health IT. A set of necessary key principles to deliver a proper design and implementation of a multidimensional economic evaluation study is described, focusing in particular on the classification of costs and outcomes as well as on the type of economic analysis to be performed. A case study is eventually described to show how the key principles introduced are applied.
Improving health, safety, and profits in extended hours operations (shiftwork).
Kerin, Alex; Aguirre, Acacia
2005-01-01
Circadian Technologies, Inc. (CIRCADIAN) is an international consulting firm that for over 20 yr has helped employees better cope with shiftwork, while ensuring their employers reduce costs and maximize profits. This paper describes two of the methods employed to reduce accidents, absenteeism, and turnover, while improving morale and productivity in extended hour operations (shiftwork). Shiftworkers rarely receive on-the-job training to help them cope better with the rigors of working nights, evenings, and early mornings. A pre- and post-training study by CIRCADIAN demonstrated improvements in health and fatigue indices, and an increase in daytime sleep length. Companies who provide shiftwork lifestyle training also have lower rates of absenteeism and turnover compared to companies that do not provide training. The second intervention described in this paper is a novel method of shift scheduling. There are thousands of potential schedules, so how does a company choose the best schedule for their facility? The answer lies in allowing employees to be involved in the process of selection. This results in a workforce that experiences less fatigue, turnover and absenteeism, and higher morale.
NASA Astrophysics Data System (ADS)
Leem, J. W.; Song, Y. M.; Lee, Y. T.; Yu, J. S.
2010-09-01
Silicon (Si) subwavelength grating (SWG) structures were fabricated on Si substrates by holographic lithography and subsequent inductively coupled plasma (ICP) etching process using SiCl4 with or without Ar addition for solar cell applications. To ensure a good nanosized pattern transfer into the underlying Si layer, the etch selectivity of Si over the photoresist mask is optimized by varying the etching parameters, thus improving antireflection characteristics. For antireflection analysis of Si SWG surfaces, the optical reflectivity is measured experimentally and it is also calculated theoretically by a rigorous coupled-wave analysis. The reflectance depends on the height, period, and shape of two-dimensional periodic Si subwavelength structures, correlated with ICP etching parameters. The optimized Si SWG structure exhibits a dramatic decrease in optical reflection of the Si surface over a wide angle of incident light ( θ i ), i.e. less than 5% at wavelengths of 300-1100 nm, leading to good wide-angle antireflection characteristics (i.e. solar-weighted reflection of 1.7-4.9% at θ i <50°) of Si solar cells.
High-order asynchrony-tolerant finite difference schemes for partial differential equations
NASA Astrophysics Data System (ADS)
Aditya, Konduri; Donzis, Diego A.
2017-12-01
Synchronizations of processing elements (PEs) in massively parallel simulations, which arise due to communication or load imbalances between PEs, significantly affect the scalability of scientific applications. We have recently proposed a method based on finite-difference schemes to solve partial differential equations in an asynchronous fashion - synchronization between PEs is relaxed at a mathematical level. While standard schemes can maintain their stability in the presence of asynchrony, their accuracy is drastically affected. In this work, we present a general methodology to derive asynchrony-tolerant (AT) finite difference schemes of arbitrary order of accuracy, which can maintain their accuracy when synchronizations are relaxed. We show that there are several choices available in selecting a stencil to derive these schemes and discuss their effect on numerical and computational performance. We provide a simple classification of schemes based on the stencil and derive schemes that are representative of different classes. Their numerical error is rigorously analyzed within a statistical framework to obtain the overall accuracy of the solution. Results from numerical experiments are used to validate the performance of the schemes.
CULTURAL ADAPTATIONS OF EVIDENCE-BASED HOME-VISITATION MODELS IN TRIBAL COMMUNITIES.
Hiratsuka, Vanessa Y; Parker, Myra E; Sanchez, Jenae; Riley, Rebecca; Heath, Debra; Chomo, Julianna C; Beltangady, Moushumi; Sarche, Michelle
2018-05-01
The Tribal Maternal, Infant, and Early Childhood Home Visiting (Tribal MIECHV) Program provides federal grants to tribes, tribal consortia, tribal organizations, and urban Indian organizations to implement evidence-based home-visiting services for American Indian and Alaska Native (AI/AN) families. To date, only one evidence-based home-visiting program has been developed for use in AI/AN communities. The purpose of this article is to describe the steps that four Tribal MIECHV Programs took to assess community needs, select a home-visiting model, and culturally adapt the model for use in AI/AN communities. In these four unique Tribal MIECHV Program settings, each program employed a rigorous needs-assessment process and developed cultural modifications in accordance with community strengths and needs. Adaptations occurred in consultation with model developers, with consideration of the conceptual rationale for the program, while grounding new content in indigenous cultures. Research is needed to improve measurement of home-visiting outcomes in tribal and urban AI/AN settings, develop culturally grounded home-visiting interventions, and assess the effectiveness of home visiting in AI/AN communities. © 2018 Michigan Association for Infant Mental Health.
Monitoring amphibians in Great Smoky Mountains National Park
Dodd, C. Kenneth
2003-01-01
This report provides an overview of the Park’s amphibians, the factors affecting their distribution, a review of important areas of biodiversity, and a summary of amphibian life history in the Southern Appalachians. In addition, survey techniques are described as well as examples of how the techniques are set up, a critique of what the results tell the observer, and a discussion of the limitations of the techniques and the data. The report reviews considerations for site selection, outlines steps for biosecurity and for processing diseased or dying animals, and provides resource managers with a decision tree on how to monitor the Park’s amphibians based on different levels of available resources. It concludes with an extensive list of references for inventorying and monitoring amphibians. USGS and Great Smoky Mountains National Park biologists need to establish cooperative efforts and training to ensure that congressionally mandated amphibian surveys are performed in a statistically rigorous and biologically meaningful manner, and that amphibian populations on Federal lands are monitored to ensure their long-term survival. The research detailed in this report will aid these cooperative efforts.
A Fast Reduced Kernel Extreme Learning Machine.
Deng, Wan-Yu; Ong, Yew-Soon; Zheng, Qing-Hua
2016-04-01
In this paper, we present a fast and accurate kernel-based supervised algorithm referred to as the Reduced Kernel Extreme Learning Machine (RKELM). In contrast to the work on Support Vector Machine (SVM) or Least Square SVM (LS-SVM), which identifies the support vectors or weight vectors iteratively, the proposed RKELM randomly selects a subset of the available data samples as support vectors (or mapping samples). By avoiding the iterative steps of SVM, significant cost savings in the training process can be readily attained, especially on Big datasets. RKELM is established based on the rigorous proof of universal learning involving reduced kernel-based SLFN. In particular, we prove that RKELM can approximate any nonlinear functions accurately under the condition of support vectors sufficiency. Experimental results on a wide variety of real world small instance size and large instance size applications in the context of binary classification, multi-class problem and regression are then reported to show that RKELM can perform at competitive level of generalized performance as the SVM/LS-SVM at only a fraction of the computational effort incurred. Copyright © 2015 Elsevier Ltd. All rights reserved.
Strategies to promote a climate of academic integrity and minimize student cheating and plagiarism.
Scanlan, Craig L
2006-01-01
Student academic misconduct is a growing problem for colleges and universities, including those responsible for preparing health professionals. Although the implementation of honor codes has had a positive impact on this problem, further reduction in student cheating and plagiarism can be achieved only via a comprehensive strategy that promotes an institutional culture of academic integrity. Such a strategy must combine efforts both to deter and detect academic misconduct, along with fair but rigorous application of sanctions against such behaviors. Methods useful in preventing or deterring dishonest behaviors among students include early integrity training complemented with course-level reinforcement, faculty role-modeling, and the application of selected testing/assignment preventive strategies, including honor pledges and honesty declarations. Giving students more responsibility for oversight of academic integrity also may help address this problem and better promote the culture needed to uphold its principles. Successful enforcement requires that academic administration provide strong and visible support for upholding academic integrity standards, including the provision of a clear and fair process and the consistent application of appropriate sanctions against those whose conduct is found to violate these standards.
Conserve, Donaldson F; Jennings, Larissa; Aguiar, Carolina; Shin, Grace; Handler, Lara; Maman, Suzanne
2017-02-01
Introduction This systematic narrative review examined the empirical evidence on the effectiveness of mobile health (mHealth) behavioural interventions designed to increase the uptake of HIV testing among vulnerable and key populations. Methods MEDLINE/PubMed, Embase, Web of Science, and Global Health electronic databases were searched. Studies were eligible for inclusion if they were published between 2005 and 2015, evaluated an mHealth intervention, and reported an outcome relating to HIV testing. We also reviewed the bibliographies of retrieved studies for other relevant citations. The methodological rigor of selected articles was assessed, and narrative analyses were used to synthesize findings from mixed methodologies. Results A total of seven articles met the inclusion criteria. Most mHealth interventions employed a text-messaging feature and were conducted in middle- and high-income countries. The methodological rigor was moderate among studies. The current literature suggests that mHealth interventions can have significant positive effects on HIV testing initiation among vulnerable and key populations, as well as the general public. In some cases, null results were observed. Qualitative themes relating to the use of mobile technologies to increase HIV testing included the benefits of having low-cost, confidential, and motivational communication. Reported barriers included cellular network restrictions, poor linkages with physical testing services, and limited knowledge of appropriate text-messaging dose. Discussion MHealth interventions may prove beneficial in reducing the proportion of undiagnosed persons living with HIV, particularly among vulnerable and key populations. However, more rigorous and tailored interventions are needed to assess the effectiveness of widespread use.
Conserve, Donaldson F.; Jennings, Larissa; Aguiar, Carolina; Shin, Grace; Handler, Lara; Maman, Suzanne
2016-01-01
Objective This systematic narrative review examined the empirical evidence on the effectiveness of mobile health (mHealth) behavioral interventions designed to increase uptake of HIV testing among vulnerable and key populations. Methods MEDLINE/PubMed, Embase, Web of Science, and Global Health electronic databases were searched. Studies were eligible for inclusion if they were published between 2005 and 2015, evaluated an mHealth intervention, and reported an outcome relating to HIV testing. We also reviewed the bibliographies of retrieved studies for other relevant citations. The methodological rigor of selected articles was assessed, and narrative analyses were used to synthesize findings from mixed methodologies. Results A total of seven articles met the inclusion criteria. Most mHealth interventions employed a text-messaging feature and were conducted in middle- and high-income countries. The methodological rigor was moderate among studies. The current literature suggests that mHealth interventions can have significant positive effects on HIV testing initiation among vulnerable and key populations, as well as the general public. In some cases, null results were observed. Qualitative themes relating to use of mobile technologies to increase HIV testing included the benefits of having low-cost, confidential, and motivational communication. Reported barriers included cellular network restrictions, poor linkages with physical testing services, and limited knowledge of appropriate text-messaging dose. Conclusions MHealth interventions may prove beneficial in reducing the proportion of undiagnosed persons living with HIV, particularly among vulnerable and key populations. However, more rigorous and tailored intervention trials are needed to assess the effectiveness of widespread use. PMID:27056905
Kline, Joshua C.
2014-01-01
Over the past four decades, various methods have been implemented to measure synchronization of motor-unit firings. In this work, we provide evidence that prior reports of the existence of universal common inputs to all motoneurons and the presence of long-term synchronization are misleading, because they did not use sufficiently rigorous statistical tests to detect synchronization. We developed a statistically based method (SigMax) for computing synchronization and tested it with data from 17,736 motor-unit pairs containing 1,035,225 firing instances from the first dorsal interosseous and vastus lateralis muscles—a data set one order of magnitude greater than that reported in previous studies. Only firing data, obtained from surface electromyographic signal decomposition with >95% accuracy, were used in the study. The data were not subjectively selected in any manner. Because of the size of our data set and the statistical rigor inherent to SigMax, we have confidence that the synchronization values that we calculated provide an improved estimate of physiologically driven synchronization. Compared with three other commonly used techniques, ours revealed three types of discrepancies that result from failing to use sufficient statistical tests necessary to detect synchronization. 1) On average, the z-score method falsely detected synchronization at 16 separate latencies in each motor-unit pair. 2) The cumulative sum method missed one out of every four synchronization identifications found by SigMax. 3) The common input assumption method identified synchronization from 100% of motor-unit pairs studied. SigMax revealed that only 50% of motor-unit pairs actually manifested synchronization. PMID:25210152
Systematic review of the quality of prognosis studies in systemic lupus erythematosus.
Lim, Lily S H; Lee, Senq J; Feldman, Brian M; Gladman, Dafna D; Pullenayegum, Eleanor; Uleryk, Elizabeth; Silverman, Earl D
2014-10-01
Prognosis studies examine outcomes and/or seek to identify predictors or factors associated with outcomes. Many prognostic factors have been identified in systemic lupus erythematosus (SLE), but few have been consistently found across studies. We hypothesized that this is due to a lack of rigor of study designs. This study aimed to systematically assess the methodologic quality of prognosis studies in SLE. A search of prognosis studies in SLE was performed using MEDLINE and Embase, from January 1990 to June 2011. A representative sample of 150 articles was selected using a random number generator and assessed by 2 reviewers. Each study was assessed by a risk of bias tool according to 6 domains: study participation, study attrition, measurement of prognostic factors, measurement of outcomes, measurement/adjustment for confounders, and appropriateness of statistical analysis. Information about missing data was also collected. A cohort design was used in 71% of studies. High risk of bias was found in 65% of studies for confounders, 57% for study participation, 56% for attrition, 36% for statistical analyses, 20% for prognostic factors, and 18% for outcome. Missing covariate or outcome information was present in half of the studies. Only 6 studies discussed reasons for missing data and 2 imputed missing data. Lack of rigorous study design, especially in addressing confounding, study participation and attrition, and inadequately handled missing data, has limited the quality of prognosis studies in SLE. Future prognosis studies should be designed with consideration of these factors to improve methodologic rigor. Copyright © 2014 by the American College of Rheumatology.
Qian, Ma; Ma, Jie
2009-06-07
Fletcher's spherical substrate model [J. Chem. Phys. 29, 572 (1958)] is a basic model for understanding the heterogeneous nucleation phenomena in nature. However, a rigorous thermodynamic formulation of the model has been missing due to the significant complexities involved. This has not only left the classical model deficient but also likely obscured its other important features, which would otherwise have helped to better understand and control heterogeneous nucleation on spherical substrates. This work presents a rigorous thermodynamic formulation of Fletcher's model using a novel analytical approach and discusses the new perspectives derived. In particular, it is shown that the use of an intermediate variable, a selected geometrical angle or pseudocontact angle between the embryo and spherical substrate, revealed extraordinary similarities between the first derivatives of the free energy change with respect to embryo radius for nucleation on spherical and flat substrates. Enlightened by the discovery, it was found that there exists a local maximum in the difference between the equivalent contact angles for nucleation on spherical and flat substrates due to the existence of a local maximum in the difference between the shape factors for nucleation on spherical and flat substrate surfaces. This helps to understand the complexity of the heterogeneous nucleation phenomena in a practical system. Also, it was found that the unfavorable size effect occurs primarily when R<5r( *) (R: radius of substrate and r( *): critical embryo radius) and diminishes rapidly with increasing value of R/r( *) beyond R/r( *)=5. This finding provides a baseline for controlling the size effects in heterogeneous nucleation.
Greenwood, Eleni A; Cedars, Marcelle I; Santoro, Nanette; Eisenberg, Esther; Kao, Chia-Ning; Haisenleder, Daniel J; Diamond, Michael P; Huddleston, Heather G
2017-12-01
To test the hypothesis that women with unexplained infertility demonstrate evidence of diminished ovarian reserve when compared with a population of community controls. Cross-sectional study. Multicenter university-based clinical practices. Study participants included 277 healthy, normo-ovulatory female partners with rigorously defined unexplained infertility randomly selected from a multicenter trial (Assessment of Multiple Intrauterine Gestations from Ovarian Stimulation). Controls included 226 healthy, normo-ovulatory women not seeking treatment for fertility from a community-based cohort (Ovarian Aging study). Serum antimüllerian hormone (AMH) assay at a central laboratory, FSH, fasting serum metabolic testing, transvaginal ultrasonography for antral follicle counts (AFCs), anthropometric measurements. Average AMH, AFC, and AMH/AFC were compared between infertile and control women by age. Analyses of covariance compared these outcomes while controlling for confounders, including age, race, body mass index, smoking history, and study site. In our models, AMH, AFC, and AMH/AFC ovarian reserve indices did not differ between infertile women and community-based controls, after controlling for age, race, body mass index, smoking history, and study site. Currently utilized predictors of ovarian reserve do not discriminate women with rigorously defined unexplained infertility from healthy community-based women of similar demographic characteristics. Contrary to our hypothesis, among women with FSH in the normal range (≤12 IU/L), women with unexplained infertility did not show evidence of decreased ovarian reserve as measured by AMH and AFC. Ovarian reserve markers in isolation may not serve as predictors of future fertility. Copyright © 2017 American Society for Reproductive Medicine. All rights reserved.
IRREVERSIBLE PROCESSES IN A PLASMA
DOE Office of Scientific and Technical Information (OSTI.GOV)
Balescu, R.
1959-04-01
ABS>The characteristic divergences caused by long-range phenomena in gases can be eliminated in equilibrium situations by partial summations of terms individually divergent but whose sum converges. It is shown how the recently developed diagram technique enables treatment of non-equilibrium cases by a rigorous asymptotic method. The general ideas underlying the approach are briefly indicated. (T.R. H.)
National Aeronautics and Space Administration Exploration Systems Interim Strategy
NASA Technical Reports Server (NTRS)
2004-01-01
Contents include the following: 1. The Exploration Systems Mission Directorate within NASA. Enabling the Vision for Space Exploration. The Role of the Directorate. 2. Strategic Context and Approach. Corporate Focus. Focused, Prioritized Requirements. Spiral Transformation. Management Rigor. 3. Achieving Directorate Objectives. Strategy to Task Process. Capability Development. Research and Technology Development. 4. Beyond the Horizon. Appendices.
ERIC Educational Resources Information Center
Levinson, Bradley A.
1998-01-01
The theoretical insight and ethnographic rigor of this collection of essays from participants at Departamento de Investigaciones Educativas (DIE) of the National Polytechnic Institute about the role of the public school in Mexican social and political life promote understanding of educational processes in different contexts, including rural and…
The Relative Costs of New York City's New Small Public High Schools of Choice
ERIC Educational Resources Information Center
Bifulco, Robert; Unterman, Rebecca; Bloom, Howard S.
2014-01-01
Building on prior research by two of the present authors, which uses lottery-like features in New York City's high school admissions process to rigorously demonstrate that new small public high schools in the district are markedly improving graduation prospects for disadvantaged students, the present paper demonstrates that these graduation…