The Computer as a Teaching Aid for Eleventh Grade Mathematics: A Comparison Study.
ERIC Educational Resources Information Center
Kieren, Thomas Ervin
To determine the effect of learning computer programming and the use of a computer on mathematical achievement of eleventh grade students, for each of two years, average and above average students were randomly assigned to an experimental and control group. The experimental group wrote computer programs and used the output from the computer in…
ASC FY17 Implementation Plan, Rev. 1
DOE Office of Scientific and Technical Information (OSTI.GOV)
Hamilton, P. G.
The Stockpile Stewardship Program (SSP) is an integrated technical program for maintaining the safety, surety, and reliability of the U.S. nuclear stockpile. The SSP uses nuclear test data, computational modeling and simulation, and experimental facilities to advance understanding of nuclear weapons. It includes stockpile surveillance, experimental research, development and engineering programs, and an appropriately scaled production capability to support stockpile requirements. This integrated national program requires the continued use of experimental facilities and programs, and the computational capabilities to support these programs. The Advanced Simulation and Computing Program (ASC) is a cornerstone of the SSP, providing simulation capabilities and computationalmore » resources that support annual stockpile assessment and certification, study advanced nuclear weapons design and manufacturing processes, analyze accident scenarios and weapons aging, and provide the tools to enable stockpile Life Extension Programs (LEPs) and the resolution of Significant Finding Investigations (SFIs). This requires a balance of resources, including technical staff, hardware, simulation software, and computer science solutions.« less
Advanced Simulation and Computing Fiscal Year 2016 Implementation Plan, Version 0
DOE Office of Scientific and Technical Information (OSTI.GOV)
McCoy, M.; Archer, B.; Hendrickson, B.
2015-08-27
The Stockpile Stewardship Program (SSP) is an integrated technical program for maintaining the safety, surety, and reliability of the U.S. nuclear stockpile. The SSP uses nuclear test data, computational modeling and simulation, and experimental facilities to advance understanding of nuclear weapons. It includes stockpile surveillance, experimental research, development and engineering programs, and an appropriately scaled production capability to support stockpile requirements. This integrated national program requires the continued use of experimental facilities and programs, and the computational capabilities to support these programs. The purpose of this IP is to outline key work requirements to be performed and to control individualmore » work activities within the scope of work. Contractors may not deviate from this plan without a revised WA or subsequent IP.« less
Practical application of computer programs for supersonic combustion
NASA Technical Reports Server (NTRS)
Groves, F. R., Jr.
1972-01-01
Experimental data were interpreted using two supersonic combustion computer programs. The P1 program is based on a conventional boundary layer treatment of the mixing of concentric gas streams and complete combustion chemistry. The H1 program is based on a modified boundary layer approach which accounts for radial pressure gradients in the flow and also incorporates a finite rate chemistry calculation. The objective of the investigation was to compare the experimental data with theoretical predictions of the two programs with special emphasis on the prediction of radial pressure gradients by the H1 program. A test of the H1 program was also desired through comparison with the experimental data and with the P1 program.
NASA Technical Reports Server (NTRS)
Tabakoff, W.
1975-01-01
A two-dimensional finite difference numerical technique is presented to determine the temperature distribution in a solid blade of a radial turbine guide vane. A computer program is written in FORTRAN 4 for the IBM 370/165 computer. The computer results obtained from these programs have a similar behavior and trend as those obtained by experimental results.
Composite Failures: A Comparison of Experimental Test Results and Computational Analysis Using XFEM
2016-09-30
NUWC-NPT Technical Report 12,218 30 September 2016 Composite Failures: A Comparison of Experimental Test Results and Computational Analysis...A Comparison of Experimental Test Results and Computational Analysis Using XFEM 5a. CONTRACT NUMBER 5b. GRANT NUMBER 5c. PROGRAM ELEMENT...availability of measurement techniques, experimental testing of composite materials has largely outpaced the computational modeling ability, forcing
Advanced Simulation & Computing FY15 Implementation Plan Volume 2, Rev. 0.5
DOE Office of Scientific and Technical Information (OSTI.GOV)
McCoy, Michel; Archer, Bill; Matzen, M. Keith
2014-09-16
The Stockpile Stewardship Program (SSP) is a single, highly integrated technical program for maintaining the surety and reliability of the U.S. nuclear stockpile. The SSP uses nuclear test data, computational modeling and simulation, and experimental facilities to advance understanding of nuclear weapons. It includes stockpile surveillance, experimental research, development and engineering programs, and an appropriately scaled production capability to support stockpile requirements. This integrated national program requires the continued use of experimental facilities and programs, and the computational enhancements to support these programs. The Advanced Simulation and Computing Program (ASC) is a cornerstone of the SSP, providing simulation capabilities andmore » computational resources that support annual stockpile assessment and certification, study advanced nuclear weapons design and manufacturing processes, analyze accident scenarios and weapons aging, and provide the tools to enable stockpile Life Extension Programs (LEPs) and the resolution of Significant Finding Investigations (SFIs). This requires a balance of resource, including technical staff, hardware, simulation software, and computer science solutions. As the program approaches the end of its second decade, ASC is intently focused on increasing predictive capabilities in a three-dimensional (3D) simulation environment while maintaining support to the SSP. The program continues to improve its unique tools for solving progressively more difficult stockpile problems (sufficient resolution, dimensionality, and scientific details), quantify critical margins and uncertainties, and resolve increasingly difficult analyses needed for the SSP. Where possible, the program also enables the use of high-performance simulation and computing tools to address broader national security needs, such as foreign nuclear weapon assessments and counternuclear terrorism.« less
Computer as a Tool in SAT Preparation.
ERIC Educational Resources Information Center
Coffin, Gregory C.
Two experimental programs, designed to increase Scholastic Aptitude Test (SAT) scores of inner city, low achieving students by using computer-assisted SAT preparation, produced differing results. Forty volunteers from a nearby high school were assigned to two groups of 20 each--one experimental and one control group. The first program provided six…
von Arnim, Albrecht G.; Missra, Anamika
2017-01-01
Leading voices in the biological sciences have called for a transformation in graduate education leading to the PhD degree. One area commonly singled out for growth and innovation is cross-training in computational science. In 1998, the University of Tennessee (UT) founded an intercollegiate graduate program called the UT-ORNL Graduate School of Genome Science and Technology in partnership with the nearby Oak Ridge National Laboratory. Here, we report outcome data that attest to the program’s effectiveness in graduating computationally enabled biologists for diverse careers. Among 77 PhD graduates since 2003, the majority came with traditional degrees in the biological sciences, yet two-thirds moved into computational or hybrid (computational–experimental) positions. We describe the curriculum of the program and how it has changed. We also summarize how the program seeks to establish cohesion between computational and experimental biologists. This type of program can respond flexibly and dynamically to unmet training needs. In conclusion, this study from a flagship, state-supported university may serve as a reference point for creating a stable, degree-granting, interdepartmental graduate program in computational biology and allied areas. PMID:29167223
The Detection of Nonplanar Surfaces in Visual Space.
1984-03-01
involve quasi -dotted stimuli. For example, applications may be found in fields such as air traffic control ; geophysical surveys (e.g., to distinguish a...line microcomputers. The control program was initially loaded by the experimenter from the computer’s disk O memory into its randomly addressable... experimenter and the computer carried out certain initialization segments of the control program. Next, the observer signed on at the computer terminal with a
ERIC Educational Resources Information Center
McAllister, Deborah A.
1994-01-01
This dissertation investigates the effects of two different grouping strategies on secondary students concerning their preparation for, and performance at, a computer programming contest. Students were divided into two groups, experimental and control, for a 3-week treatment, followed by the contest. Students in the experimental group became…
NASA Astrophysics Data System (ADS)
Parker, Tehri Davenport
1997-09-01
This study designed, implemented, and evaluated an environmental education hypermedia program for use in a residential environmental education facility. The purpose of the study was to ascertain whether a hypermedia program could increase student knowledge and positive attitudes toward the environment and environmental education. A student/computer interface, based on the theory of social cognition, was developed to direct student interactions with the computer. A quasi-experimental research design was used. Students were randomly assigned to either the experimental or control group. The experimental group used the hypermedia program to learn about the topic of energy. The control group received the same conceptual information from a teacher/naturalist. An Environmental Awareness Quiz was administered to measure differences in the students' cognitive understanding of energy issues. Students participated in one on one interviews to discuss their attitudes toward the lesson and the overall environmental education experience. Additionally, members of the experimental group were tape recorded while they used the hypermedia program. These tapes were analyzed to identify aspects of the hypermedia program that promoted student learning. The findings of this study suggest that computers, and hypermedia programs, can be integrated into residential environmental education facilities, and can assist environmental educators in meeting their goals for students. The study found that the hypermedia program was as effective as the teacher/naturalist for teaching about environmental education material. Students who used the computer reported more positive attitudes toward the lesson on energy, and thought that they had learned more than the control group. Students in the control group stated that they did not learn as much as the computer group. The majority of students had positive attitudes toward the inclusion of computers in the camp setting, and stated that they were a good way to learn about environmental education material. This study also identified lack of social skills as a barrier to social cognition among mixed gender groups using the computer program.
1979-12-01
team progranming in reducing software dleveloup- ment costs relative to ad hoc approaches and improving software product quality relative to...are interpreted as demonstrating the advantages of disciplined team programming in reducing software development costs relative to ad hoc approaches...is due oartialty to the cost and imoracticality of a valiI experimental setup within a oroauct ion environment. Thus the question remains, are
NASA Technical Reports Server (NTRS)
Sawyer, W. C.; Allen, J. M.; Hernandez, G.; Dillenius, M. F. E.; Hemsch, M. J.
1982-01-01
This paper presents a survey of engineering computational methods and experimental programs used for estimating the aerodynamic characteristics of missile configurations. Emphasis is placed on those methods which are suitable for preliminary design of conventional and advanced concepts. An analysis of the technical approaches of the various methods is made in order to assess their suitability to estimate longitudinal and/or lateral-directional characteristics for different classes of missile configurations. Some comparisons between the predicted characteristics and experimental data are presented. These comparisons are made for a large variation in flow conditions and model attitude parameters. The paper also presents known experimental research programs developed for the specific purpose of validating analytical methods and extending the capability of data-base programs.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Turner, L.R.; Praeg, W.F.
1982-03-01
The experimental requirements, test-bed design, and computational requirements are reviewed and updated. Next, in Sections 3, 4 and 5, the experimental plan, instrumentation, and computer plan, respectively, are described. Finally, Section 6 treats other considerations, such as personnel, outside participation, and distribution of results.
ERIC Educational Resources Information Center
Alhassan, Riyadh
2017-01-01
The purpose of this study was to examine the effect of employing self-explanation learning strategy supported with Worked Examples on acquiring computer programing skills among freshmen high school students. The study adopted a quasi-experimental method, where an experimental group (n = 33) used the self-explanation strategy supported with worked…
Effects of experimental design on calibration curve precision in routine analysis
Pimentel, Maria Fernanda; Neto, Benício de Barros; Saldanha, Teresa Cristina B.
1998-01-01
A computational program which compares the effciencies of different experimental designs with those of maximum precision (D-optimized designs) is described. The program produces confidence interval plots for a calibration curve and provides information about the number of standard solutions, concentration levels and suitable concentration ranges to achieve an optimum calibration. Some examples of the application of this novel computational program are given, using both simulated and real data. PMID:18924816
ERIC Educational Resources Information Center
Psycharis, Sarantos; Kallia, Maria
2017-01-01
In this paper we investigate whether computer programming has an impact on high school student's reasoning skills, problem solving and self-efficacy in Mathematics. The quasi-experimental design was adopted to implement the study. The sample of the research comprised 66 high school students separated into two groups, the experimental and the…
ERIC Educational Resources Information Center
Mascaró, Maite; Sacristán, Ana Isabel; Rufino, Marta M.
2016-01-01
For the past 4 years, we have been involved in a project that aims to enhance the teaching and learning of experimental analysis and statistics, of environmental and biological sciences students, through computational programming activities (using R code). In this project, through an iterative design, we have developed sequences of R-code-based…
NASA Technical Reports Server (NTRS)
Pan, Y. S.; Drummond, J. P.; Mcclinton, C. R.
1978-01-01
Two parabolic flow computer programs, SHIP (a finite-difference program) and COMOC (a finite-element program), are used for predicting three-dimensional turbulent reacting flow fields in supersonic combustors. The theoretical foundation of the two computer programs are described, and then the programs are applied to a three-dimensional turbulent mixing experiment. The cold (nonreacting) flow experiment was performed to study the mixing of helium jets with a supersonic airstream in a rectangular duct. Surveys of the flow field at an upstream were used as the initial data by programs; surveys at a downstream station provided comparison to assess program accuracy. Both computer programs predicted the experimental results and data trends reasonably well. However, the comparison between the computations from the two programs indicated that SHIP was more accurate in computation and more efficient in both computer storage and computing time than COMOC.
DOE Office of Scientific and Technical Information (OSTI.GOV)
McCoy, Michel; Archer, Bill; Hendrickson, Bruce
The Stockpile Stewardship Program (SSP) is an integrated technical program for maintaining the safety, surety, and reliability of the U.S. nuclear stockpile. The SSP uses nuclear test data, computational modeling and simulation, and experimental facilities to advance understanding of nuclear weapons. It includes stockpile surveillance, experimental research, development and engineering programs, and an appropriately scaled production capability to support stockpile requirements. This integrated national program requires the continued use of experimental facilities and programs, and the computational capabilities to support these programs. The Advanced Simulation and Computing Program (ASC) is a cornerstone of the SSP, providing simulation capabilities and computationalmore » resources that support annual stockpile assessment and certification, study advanced nuclear weapons design and manufacturing processes, analyze accident scenarios and weapons aging, and provide the tools to enable stockpile Life Extension Programs (LEPs) and the resolution of Significant Finding Investigations (SFIs). This requires a balance of resource, including technical staff, hardware, simulation software, and computer science solutions. ASC is now focused on increasing predictive capabilities in a three-dimensional (3D) simulation environment while maintaining support to the SSP. The program continues to improve its unique tools for solving progressively more difficult stockpile problems (sufficient resolution, dimensionality, and scientific details), and quantifying critical margins and uncertainties. Resolving each issue requires increasingly difficult analyses because the aging process has progressively moved the stockpile further away from the original test base. Where possible, the program also enables the use of high performance computing (HPC) and simulation tools to address broader national security needs, such as foreign nuclear weapon assessments and counter nuclear terrorism.« less
A research program in empirical computer science
NASA Technical Reports Server (NTRS)
Knight, J. C.
1991-01-01
During the grant reporting period our primary activities have been to begin preparation for the establishment of a research program in experimental computer science. The focus of research in this program will be safety-critical systems. Many questions that arise in the effort to improve software dependability can only be addressed empirically. For example, there is no way to predict the performance of the various proposed approaches to building fault-tolerant software. Performance models, though valuable, are parameterized and cannot be used to make quantitative predictions without experimental determination of underlying distributions. In the past, experimentation has been able to shed some light on the practical benefits and limitations of software fault tolerance. It is common, also, for experimentation to reveal new questions or new aspects of problems that were previously unknown. A good example is the Consistent Comparison Problem that was revealed by experimentation and subsequently studied in depth. The result was a clear understanding of a previously unknown problem with software fault tolerance. The purpose of a research program in empirical computer science is to perform controlled experiments in the area of real-time, embedded control systems. The goal of the various experiments will be to determine better approaches to the construction of the software for computing systems that have to be relied upon. As such it will validate research concepts from other sources, provide new research results, and facilitate the transition of research results from concepts to practical procedures that can be applied with low risk to NASA flight projects. The target of experimentation will be the production software development activities undertaken by any organization prepared to contribute to the research program. Experimental goals, procedures, data analysis and result reporting will be performed for the most part by the University of Virginia.
Advanced Simulation and Computing Fiscal Year 14 Implementation Plan, Rev. 0.5
DOE Office of Scientific and Technical Information (OSTI.GOV)
Meisner, Robert; McCoy, Michel; Archer, Bill
2013-09-11
The Stockpile Stewardship Program (SSP) is a single, highly integrated technical program for maintaining the surety and reliability of the U.S. nuclear stockpile. The SSP uses nuclear test data, computational modeling and simulation, and experimental facilities to advance understanding of nuclear weapons. It includes stockpile surveillance, experimental research, development and engineering programs, and an appropriately scaled production capability to support stockpile requirements. This integrated national program requires the continued use of experimental facilities and programs, and the computational enhancements to support these programs. The Advanced Simulation and Computing Program (ASC) is a cornerstone of the SSP, providing simulation capabilities andmore » computational resources that support annual stockpile assessment and certification, study advanced nuclear weapons design and manufacturing processes, analyze accident scenarios and weapons aging, and provide the tools to enable stockpile Life Extension Programs (LEPs) and the resolution of Significant Finding Investigations (SFIs). This requires a balanced resource, including technical staff, hardware, simulation software, and computer science solutions. In its first decade, the ASC strategy focused on demonstrating simulation capabilities of unprecedented scale in three spatial dimensions. In its second decade, ASC is now focused on increasing predictive capabilities in a three-dimensional (3D) simulation environment while maintaining support to the SSP. The program continues to improve its unique tools for solving progressively more difficult stockpile problems (sufficient resolution, dimensionality, and scientific details), quantify critical margins and uncertainties, and resolve increasingly difficult analyses needed for the SSP. Moreover, ASC’s business model is integrated and focused on requirements-driven products that address long-standing technical questions related to enhanced predictive capability in the simulation tools.« less
An experimental and theoretical investigation of deposition patterns from an agricultural airplane
NASA Technical Reports Server (NTRS)
Morris, D. J.; Croom, C. C.; Vandam, C. P.; Holmes, B. J.
1984-01-01
A flight test program has been conducted with a representative agricultural airplane to provide data for validating a computer program model which predicts aerially applied particle deposition. Test procedures and the data from this test are presented and discussed. The computer program features are summarized, and comparisons of predicted and measured particle deposition are presented. Applications of the computer program for spray pattern improvement are illustrated.
von Arnim, Albrecht G; Missra, Anamika
2017-01-01
Leading voices in the biological sciences have called for a transformation in graduate education leading to the PhD degree. One area commonly singled out for growth and innovation is cross-training in computational science. In 1998, the University of Tennessee (UT) founded an intercollegiate graduate program called the UT-ORNL Graduate School of Genome Science and Technology in partnership with the nearby Oak Ridge National Laboratory. Here, we report outcome data that attest to the program's effectiveness in graduating computationally enabled biologists for diverse careers. Among 77 PhD graduates since 2003, the majority came with traditional degrees in the biological sciences, yet two-thirds moved into computational or hybrid (computational-experimental) positions. We describe the curriculum of the program and how it has changed. We also summarize how the program seeks to establish cohesion between computational and experimental biologists. This type of program can respond flexibly and dynamically to unmet training needs. In conclusion, this study from a flagship, state-supported university may serve as a reference point for creating a stable, degree-granting, interdepartmental graduate program in computational biology and allied areas. © 2017 A. G. von Arnim and A. Missra. CBE—Life Sciences Education © 2017 The American Society for Cell Biology. This article is distributed by The American Society for Cell Biology under license from the author(s). It is available to the public under an Attribution–Noncommercial–Share Alike 3.0 Unported Creative Commons License (http://creativecommons.org/licenses/by-nc-sa/3.0).
Computerized optimization of radioimmunoassays for hCG and estradiol: an experimental evaluation
DOE Office of Scientific and Technical Information (OSTI.GOV)
Yanagishita, M.; Rodbard, D.
1978-07-15
The mathematical and statistical theory of radioimmunoassays (RIAs) has been used to develop a series of computer programs to optimize sensitivity or precision at any desired dose level for either equilibrium or nonequilibrium assays. These computer programs provide for the calculation of the equilibrium constants of association and binding capacities for antisera (parameters of Scatchard plots), the association and dissociation rate constants, and prediction of optimum concentration of labeled ligand and antibody and optimum incubation times for the assay. This paper presents an experimental evaluation of the use of these computer programs applied to RIAs for human chorionic gonadotropin (hCG)more » and estradiol. The experimental results are in reasonable semiquantitative agreement with the predictions of the computer simulations (usually within a factor of two) and thus partially validate the use of computer techniques to optimize RIAs that are reasonably well behaved, as in the case of the hCG and estradiol RIAs. Further, these programs can provide insights into the nature of the RIA system, e.g., the general nature of the sensitivity and precision surfaces. This facilitates empirical optimization of conditions.« less
Experimental and computational surface and flow-field results for an all-body hypersonic aircraft
NASA Technical Reports Server (NTRS)
Lockman, William K.; Lawrence, Scott L.; Cleary, Joseph W.
1990-01-01
The objective of the present investigation is to establish a benchmark experimental data base for a generic hypersonic vehicle shape for validation and/or calibration of advanced computational fluid dynamics computer codes. This paper includes results from the comprehensive test program conducted in the NASA/Ames 3.5-foot Hypersonic Wind Tunnel for a generic all-body hypersonic aircraft model. Experimental and computational results on flow visualization, surface pressures, surface convective heat transfer, and pitot-pressure flow-field surveys are presented. Comparisons of the experimental results with computational results from an upwind parabolized Navier-Stokes code developed at Ames demonstrate the capabilities of this code.
Recursive computer architecture for VLSI
DOE Office of Scientific and Technical Information (OSTI.GOV)
Treleaven, P.C.; Hopkins, R.P.
1982-01-01
A general-purpose computer architecture based on the concept of recursion and suitable for VLSI computer systems built from replicated (lego-like) computing elements is presented. The recursive computer architecture is defined by presenting a program organisation, a machine organisation and an experimental machine implementation oriented to VLSI. The experimental implementation is being restricted to simple, identical microcomputers each containing a memory, a processor and a communications capability. This future generation of lego-like computer systems are termed fifth generation computers by the Japanese. 30 references.
On-line computer system for use with low- energy nuclear physics experiments is reported
NASA Technical Reports Server (NTRS)
Gemmell, D. S.
1969-01-01
Computer program handles data from low-energy nuclear physics experiments which utilize the ND-160 pulse-height analyzer and the PHYLIS computing system. The program allows experimenters to choose from about 50 different basic data-handling functions and to prescribe the order in which these functions will be performed.
Yang, Yea-Ru; Chen, Yi-Hua; Chang, Heng-Chih; Chan, Rai-Chi; Wei, Shun-Hwa; Wang, Ray-Yau
2015-10-01
We investigated the effects of a computer-generated interactive visual feedback training program on the recovery from pusher syndrome in stroke patients. Assessor-blinded, pilot randomized controlled study. A total of 12 stroke patients with pusher syndrome were randomly assigned to either the experimental group (N = 7, computer-generated interactive visual feedback training) or control group (N = 5, mirror visual feedback training). The scale for contraversive pushing for severity of pusher syndrome, the Berg Balance Scale for balance performance, and the Fugl-Meyer assessment scale for motor control were the outcome measures. Patients were assessed pre- and posttraining. A comparison of pre- and posttraining assessment results revealed that both training programs led to the following significant changes: decreased severity of pusher syndrome scores (decreases of 4.0 ± 1.1 and 1.4 ± 1.0 in the experimental and control groups, respectively); improved balance scores (increases of 14.7 ± 4.3 and 7.2 ± 1.6 in the experimental and control groups, respectively); and higher scores for lower extremity motor control (increases of 8.4 ± 2.2 and 5.6 ± 3.3 in the experimental and control groups, respectively). Furthermore, the computer-generated interactive visual feedback training program produced significantly better outcomes in the improvement of pusher syndrome (p < 0.01) and balance (p < 0.05) compared with the mirror visual feedback training program. Although both training programs were beneficial, the computer-generated interactive visual feedback training program more effectively aided recovery from pusher syndrome compared with mirror visual feedback training. © The Author(s) 2014.
NASA Technical Reports Server (NTRS)
Silva, Walter A.; Perry, Boyd, III; Florance, James R.; Sanetrik, Mark D.; Wieseman, Carol D.; Stevens, William L.; Funk, Christie J.; Hur, Jiyoung; Christhilf, David M.; Coulson, David A.
2011-01-01
A summary of computational and experimental aeroelastic and aeroservoelastic (ASE) results for the Semi-Span Super-Sonic Transport (S4T) wind-tunnel model is presented. A broad range of analyses and multiple ASE wind-tunnel tests of the S4T have been performed in support of the ASE element in the Supersonics Program, part of NASA's Fundamental Aeronautics Program. The computational results to be presented include linear aeroelastic and ASE analyses, nonlinear aeroelastic analyses using an aeroelastic CFD code, and rapid aeroelastic analyses using CFD-based reduced-order models (ROMs). Experimental results from two closed-loop wind-tunnel tests performed at NASA Langley's Transonic Dynamics Tunnel (TDT) will be presented as well.
Ryhänen, Anne M; Siekkinen, Mervi; Rankinen, Sirkku; Korvenranta, Heikki; Leino-Kilpi, Helena
2010-04-01
The aim of this systematic review was to analyze what kind of Internet or interactive computer-based patient education programs have been developed and to analyze the effectiveness of these programs in the field of breast cancer patient education. Patient education for breast cancer patients is an important intervention to empower the patient. However, we know very little about the effects and potential of Internet-based patient education in the empowerment of breast cancer patients. Complete databases were searched covering the period from the beginning of each database to November 2008. Studies were included if they concerned patient education for breast cancer patients with Internet or interactive computer programs and were based on randomized controlled, on clinical trials or quasi-experimental studies. We identified 14 articles involving 2374 participants. The design was randomized controlled trial in nine papers, in two papers clinical trial and in three quasi-experimental. Seven of the studies were randomized to experimental and control groups, in two papers participants were grouped by ethnic and racial differences and by mode of Internet use and three studies measured the same group pre- and post-tests after using a computer program. The interventions used were described as interactive computer or multimedia programs and use of the Internet. The methodological solutions of the studies varied. The effects of the studies were diverse except for knowledge-related issues. Internet or interactive computer-based patient education programs in the care of breast cancer patients may have positive effect increasing breast cancer knowledge. The results suggest a positive relationship between the Internet or computer-based patient education program use and the knowledge level of patients with breast cancer but a diverse relationship between patient's participation and other outcome measures. There is need to develop and research more Internet-based patient education. 2009 Elsevier Ireland Ltd. All rights reserved.
Integration of Major Computer Program Packages into Experimental Courses: A Freshman Experience.
ERIC Educational Resources Information Center
Lipschitz, Irving
1981-01-01
Describes the use of the Gaussian 70 computer programs to carry out quantum chemical calculations, including single calculations, geometry, optimization, and potential surface scans. Includes a summary of student activities and benefits for students in an honors freshman chemistry course. (SK)
Computer Aided Instruction in Teaching Program Evaluation.
ERIC Educational Resources Information Center
Dowell, David A.; Binette, Holly A. Lizotte
This paper reports the results of two semesters of experience using computer-assisted instruction (CAI) to teach topics in program evaluation to undergraduate and graduate psychology students at California State University, Long Beach. (The topics addressed are models of evaluation, evaluability assessment, needs assessment, experimental and…
Hamlet on the Macintosh: An Experimental Seminar That Worked.
ERIC Educational Resources Information Center
Strange, William C.
1987-01-01
Describes experimental college Shakespeare seminar that used Macintosh computers and software called ELIZA and ADVENTURE to develop character dialogs and adventure games based on Hamlet's characters and plots. Programming languages are examined, particularly their relationship to metaphor, and the use of computers in humanities is discussed. (LRW)
ERIC Educational Resources Information Center
Shacham, Mordechai; Cutlip, Michael B.; Brauner, Neima
2009-01-01
A continuing challenge to the undergraduate chemical engineering curriculum is the time-effective incorporation and use of computer-based tools throughout the educational program. Computing skills in academia and industry require some proficiency in programming and effective use of software packages for solving 1) single-model, single-algorithm…
Toward synthesizing executable models in biology.
Fisher, Jasmin; Piterman, Nir; Bodik, Rastislav
2014-01-01
Over the last decade, executable models of biological behaviors have repeatedly provided new scientific discoveries, uncovered novel insights, and directed new experimental avenues. These models are computer programs whose execution mechanistically simulates aspects of the cell's behaviors. If the observed behavior of the program agrees with the observed biological behavior, then the program explains the phenomena. This approach has proven beneficial for gaining new biological insights and directing new experimental avenues. One advantage of this approach is that techniques for analysis of computer programs can be applied to the analysis of executable models. For example, one can confirm that a model agrees with experiments for all possible executions of the model (corresponding to all environmental conditions), even if there are a huge number of executions. Various formal methods have been adapted for this context, for example, model checking or symbolic analysis of state spaces. To avoid manual construction of executable models, one can apply synthesis, a method to produce programs automatically from high-level specifications. In the context of biological modeling, synthesis would correspond to extracting executable models from experimental data. We survey recent results about the usage of the techniques underlying synthesis of computer programs for the inference of biological models from experimental data. We describe synthesis of biological models from curated mutation experiment data, inferring network connectivity models from phosphoproteomic data, and synthesis of Boolean networks from gene expression data. While much work has been done on automated analysis of similar datasets using machine learning and artificial intelligence, using synthesis techniques provides new opportunities such as efficient computation of disambiguating experiments, as well as the ability to produce different kinds of models automatically from biological data.
An Improved Version of the NASA-Lockheed Multielement Airfoil Analysis Computer Program
NASA Technical Reports Server (NTRS)
Brune, G. W.; Manke, J. W.
1978-01-01
An improved version of the NASA-Lockheed computer program for the analysis of multielement airfoils is described. The predictions of the program are evaluated by comparison with recent experimental high lift data including lift, pitching moment, profile drag, and detailed distributions of surface pressures and boundary layer parameters. The results of the evaluation show that the contract objectives of improving program reliability and accuracy have been met.
NASA Technical Reports Server (NTRS)
Rudder, F. F., Jr.
1971-01-01
Detailed information is presented for the analytical and the experimental programs described in the report on Effects of Design Details on Structural Response to Acoustic Excitation, NASA CR-1959. This is a supplement to the basic report and detailed experimental data from all the specimens described in that report are discussed and presented in tabular form.
A computer program to determine the possible daily release window for sky target experiments
NASA Technical Reports Server (NTRS)
Michaud, N. H.
1973-01-01
A computer program is presented which is designed to determine the daily release window for sky target experiments. Factors considered in the program include: (1) target illumination by the sun at release time and during the tracking period; (2) look angle elevation above local horizon from each tracking station to the target; (3) solar depression angle from the local horizon of each tracking station during the experimental period after target release; (4) lunar depression angle from the local horizon of each tracking station during the experimental period after target release; and (5) total sky background brightness as seen from each tracking station while viewing the target. Program output is produced in both graphic and data form. Output data can be plotted for a single calendar month or year. The numerical values used to generate the plots are furnished to permit a more detailed review of the computed daily release windows.
NASA Technical Reports Server (NTRS)
Howell, W. E.
1974-01-01
The structural performance of a boron-epoxy reinforced titanium drag strut, which contains a bonded scarf joint and was designed to the criteria of the Boeing 747 transport, was evaluated. An experimental and analytical investigation was conducted. The strut was exposed to two lifetimes of spectrum loading and was statically loaded to the tensile and compressive design ultimate loads. Throughout the test program no evidence of any damage in the drag strut was detected by strain gage measurements, ultrasonic inspection, or visual observation. An analytical study of the bonded joint was made using the NASA structural analysis computer program NASTRAN. A comparison of the strains predicted by the NASTRAN computer program with the experimentally determined values shows excellent agreement. The NASTRAN computer program is a viable tool for studying, in detail, the stresses and strains induced in a bonded joint.
Validation of Magnetic Resonance Thermometry by Computational Fluid Dynamics
NASA Astrophysics Data System (ADS)
Rydquist, Grant; Owkes, Mark; Verhulst, Claire M.; Benson, Michael J.; Vanpoppel, Bret P.; Burton, Sascha; Eaton, John K.; Elkins, Christopher P.
2016-11-01
Magnetic Resonance Thermometry (MRT) is a new experimental technique that can create fully three-dimensional temperature fields in a noninvasive manner. However, validation is still required to determine the accuracy of measured results. One method of examination is to compare data gathered experimentally to data computed with computational fluid dynamics (CFD). In this study, large-eddy simulations have been performed with the NGA computational platform to generate data for a comparison with previously run MRT experiments. The experimental setup consisted of a heated jet inclined at 30° injected into a larger channel. In the simulations, viscosity and density were scaled according to the local temperature to account for differences in buoyant and viscous forces. A mesh-independent study was performed with 5 mil-, 15 mil- and 45 mil-cell meshes. The program Star-CCM + was used to simulate the complete experimental geometry. This was compared to data generated from NGA. Overall, both programs show good agreement with the experimental data gathered with MRT. With this data, the validity of MRT as a diagnostic tool has been shown and the tool can be used to further our understanding of a range of flows with non-trivial temperature distributions.
ERIC Educational Resources Information Center
Smetana, Frederick O.; Phillips, Dennis M.
In an effort to increase dissemination of scientific and technological information, a program was undertaken whereby graduate students in science and engineering could request a computer-produced bibliography and/or abstracts of documents identified by the computer. The principal resource was the National Aeronautics and Space Administration…
NASA Technical Reports Server (NTRS)
Hosny, W. M.; Tabakoff, W.
1975-01-01
A two-dimensional finite difference numerical technique is presented to determine the temperature distribution in a solid blade of a radial guide vane. A computer program is written in Fortran IV for IBM 370/165 computer. The computer results obtained from these programs have a similar behavior and trend as those obtained by experimental results.
Computer-aided injection molding system
NASA Astrophysics Data System (ADS)
Wang, K. K.; Shen, S. F.; Cohen, C.; Hieber, C. A.; Isayev, A. I.
1982-10-01
Achievements are reported in cavity-filling simulation, modeling viscoelastic effects, measuring and predicting frozen-in birefringence in molded parts, measuring residual stresses and associated mechanical properties of molded parts, and developing an interactive mold-assembly design program and an automatic NC maching data generation and verification program. The Cornell Injection Molding Program (CIMP) consortium is discussed as are computer user manuals that have been published by the consortium. Major tasks which should be addressed in future efforts are listed, including: (1) predict and experimentally determine the post-fillin behavior of thermoplastics; (2) simulate and experimentally investigate the injection molding of thermosets and filled materials; and (3) further investigate residual stresses, orientation and mechanical properties.
Computations of soot and and NO sub x emissions from gas turbine combustors
NASA Technical Reports Server (NTRS)
Srivatsa, S. K.
1982-01-01
An analytical program was conducted to compute the soot and NOx emissions from a combustor and the radiation heat transfer to the combustor walls. The program involved the formulation of an emission and radiation model and the incorporation of this model into the Garrett 3-D Combustor Perfomance Computer Program. Computations were performed for the idle, cruise, and take-off conditions of a JT8D can combustor. The predicted soot and NOx emissions and the radiation heat transfer to the combustor walls agree reasonably well with the limited experimental data available.
NASA Technical Reports Server (NTRS)
Dayton, J. A., Jr.; Kosmahl, H. G.; Ramins, P.; Stankiewicz, N.
1979-01-01
Experimental and analytical results are compared for two high performance, octave bandwidth TWT's that use depressed collectors (MDC's) to improve the efficiency. The computations were carried out with advanced, multidimensional computer programs that are described here in detail. These programs model the electron beam as a series of either disks or rings of charge and follow their multidimensional trajectories from the RF input of the ideal TWT, through the slow wave structure, through the magnetic refocusing system, to their points of impact in the depressed collector. Traveling wave tube performance, collector efficiency, and collector current distribution were computed and the results compared with measurements for a number of TWT-MDC systems. Power conservation and correct accounting of TWT and collector losses were observed. For the TWT's operating at saturation, very good agreement was obtained between the computed and measured collector efficiencies. For a TWT operating 3 and 6 dB below saturation, excellent agreement between computed and measured collector efficiencies was obtained in some cases but only fair agreement in others. However, deviations can largely be explained by small differences in the computed and actual spent beam energy distributions. The analytical tools used here appear to be sufficiently refined to design efficient collectors for this class of TWT. However, for maximum efficiency, some experimental optimization (e.g., collector voltages and aperture sizes) will most likely be required.
Validation of an Active Gear, Flexible Aircraft Take-off and Landing analysis (AGFATL)
NASA Technical Reports Server (NTRS)
Mcgehee, J. R.
1984-01-01
The results of an analytical investigation using a computer program for active gear, flexible aircraft take off and landing analysis (AGFATL) are compared with experimental data from shaker tests, drop tests, and simulated landing tests to validate the AGFATL computer program. Comparison of experimental and analytical responses for both passive and active gears indicates good agreement for shaker tests and drop tests. For the simulated landing tests, the passive and active gears were influenced by large strut binding friction forces. The inclusion of these undefined forces in the analytical simulations was difficult, and consequently only fair to good agreement was obtained. An assessment of the results from the investigation indicates that the AGFATL computer program is a valid tool for the study and initial design of series hydraulic active control landing gear systems.
DOSESCREEN: a computer program to aid dose placement
Kimberly C. Smith; Jacqueline L. Robertson
1984-01-01
Careful selection of an experimental design for a bioassay substantially improves the precision of effective dose (ED) estimates. Design considerations typically include determination of sample size, dose selection, and allocation of subjects to doses. DOSESCREEN is a computer program written to help investigators select an efficient design for the estimation of an...
DOE Office of Scientific and Technical Information (OSTI.GOV)
Hills, Richard G.; Maniaci, David Charles; Naughton, Jonathan W.
2015-09-01
A Verification and Validation (V&V) framework is presented for the development and execution of coordinated modeling and experimental program s to assess the predictive capability of computational models of complex systems through focused, well structured, and formal processes.The elements of the framework are based on established V&V methodology developed by various organizations including the Department of Energy, National Aeronautics and Space Administration, the American Institute of Aeronautics and Astronautics, and the American Society of Mechanical Engineers. Four main topics are addressed: 1) Program planning based on expert elicitation of the modeling physics requirements, 2) experimental design for model assessment, 3)more » uncertainty quantification for experimental observations and computational model simulations, and 4) assessment of the model predictive capability. The audience for this document includes program planners, modelers, experimentalist, V &V specialist, and customers of the modeling results.« less
Flexible rotor dynamics analysis
NASA Technical Reports Server (NTRS)
Shen, F. A.
1973-01-01
A digital computer program was developed to analyze the general nonaxisymmetric and nonsynchronous transient and steady-state rotor dynamic performance of a bending- and shear-wise flexible rotor-bearing system under various operating conditions. The effects of rotor material mechanical hysteresis, rotor torsion flexibility, transverse effects of rotor axial and torsional loading and the anisotropic, in-phase and out-of-phase bearing stiffness and damping force and moment coefficients were included in the program to broaden its capability. An optimum solution method was found and incorporated in the computer program. Computer simulation of experimental data was made and qualitative agreements observed. The mathematical formulations, computer program verification, test data simulation, and user instruction was presented and discussed.
Flow induction by pressure forces
NASA Technical Reports Server (NTRS)
Garris, C. A.; Toh, K. H.; Amin, S.
1992-01-01
A dual experimental/computational approach to the fluid mechanics of complex interactions that take place in a rotary-jet ejector is presented. The long-range goal is to perform both detailed flow mapping and finite element computational analysis. The described work represents an initial finding on the experimental mapping program. Test results on the hubless rotary-jet are discussed.
Making Advanced Computer Science Topics More Accessible through Interactive Technologies
ERIC Educational Resources Information Center
Shao, Kun; Maher, Peter
2012-01-01
Purpose: Teaching advanced technical concepts in a computer science program to students of different technical backgrounds presents many challenges. The purpose of this paper is to present a detailed experimental pedagogy in teaching advanced computer science topics, such as computer networking, telecommunications and data structures using…
Interactive program for analysis and design problems in advanced composites technology
NASA Technical Reports Server (NTRS)
Cruse, T. A.; Swedlow, J. L.
1971-01-01
During the past year an experimental program in the fracture of advanced fiber composites has been completed. The experimental program has given direction to additional experimental and theoretical work. A synthesis program for designing low weight multifastener joints in composites is proposed, based on extensive analytical background. A number of failed joints have been thoroughly analyzed to evaluate the failure hypothesis used in the synthesis procedure. Finally, a new solution is reported for isotropic and anisotropic laminates using the boundary-integral method. The solution method offers significant savings of computer core and time for important problems.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Ahrens, J.P.; Shapiro, L.G.; Tanimoto, S.L.
1997-04-01
This paper describes a computing environment which supports computer-based scientific research work. Key features include support for automatic distributed scheduling and execution and computer-based scientific experimentation. A new flexible and extensible scheduling technique that is responsive to a user`s scheduling constraints, such as the ordering of program results and the specification of task assignments and processor utilization levels, is presented. An easy-to-use constraint language for specifying scheduling constraints, based on the relational database query language SQL, is described along with a search-based algorithm for fulfilling these constraints. A set of performance studies show that the environment can schedule and executemore » program graphs on a network of workstations as the user requests. A method for automatically generating computer-based scientific experiments is described. Experiments provide a concise method of specifying a large collection of parameterized program executions. The environment achieved significant speedups when executing experiments; for a large collection of scientific experiments an average speedup of 3.4 on an average of 5.5 scheduled processors was obtained.« less
Digital optical computers at the optoelectronic computing systems center
NASA Technical Reports Server (NTRS)
Jordan, Harry F.
1991-01-01
The Digital Optical Computing Program within the National Science Foundation Engineering Research Center for Opto-electronic Computing Systems has as its specific goal research on optical computing architectures suitable for use at the highest possible speeds. The program can be targeted toward exploiting the time domain because other programs in the Center are pursuing research on parallel optical systems, exploiting optical interconnection and optical devices and materials. Using a general purpose computing architecture as the focus, we are developing design techniques, tools and architecture for operation at the speed of light limit. Experimental work is being done with the somewhat low speed components currently available but with architectures which will scale up in speed as faster devices are developed. The design algorithms and tools developed for a general purpose, stored program computer are being applied to other systems such as optimally controlled optical communication networks.
Computation of transonic potential flow about 3 dimensional inlets, ducts, and bodies
NASA Technical Reports Server (NTRS)
Reyhner, T. A.
1982-01-01
An analysis was developed and a computer code, P465 Version A, written for the prediction of transonic potential flow about three dimensional objects including inlet, duct, and body geometries. Finite differences and line relaxation are used to solve the complete potential flow equation. The coordinate system used for the calculations is independent of body geometry. Cylindrical coordinates are used for the computer code. The analysis is programmed in extended FORTRAN 4 for the CYBER 203 vector computer. The programming of the analysis is oriented toward taking advantage of the vector processing capabilities of this computer. Comparisons of computed results with experimental measurements are presented to verify the analysis. Descriptions of program input and output formats are also presented.
ERIC Educational Resources Information Center
Korkmaz, Özgen
2016-01-01
The aim of this study was to investigate the effect of the Scratch and Lego Mindstorms Ev3 programming activities on academic achievement with respect to computer programming, and on the problem-solving and logical-mathematical thinking skills of students. This study was a semi-experimental, pretest-posttest study with two experimental groups and…
ERIC Educational Resources Information Center
Hung, Y.-C.
2012-01-01
This paper investigates the impact of combining self explaining (SE) with computer architecture diagrams to help novice students learn assembly language programming. Pre- and post-test scores for the experimental and control groups were compared and subjected to covariance (ANCOVA) statistical analysis. Results indicate that the SE-plus-diagram…
HOPI: on-line injection optimization program
DOE Office of Scientific and Technical Information (OSTI.GOV)
LeMaire, J L
1977-10-26
A method of matching the beam from the 200 MeV linac to the AGS without the necessity of making emittance measurements is presented. An on-line computer program written on the PDP10 computer performs the matching by modifying independently the horizontal and vertical emittance. Experimental results show success with this method, which can be applied to any matching section.
ERIC Educational Resources Information Center
Prokhorov, Alexander V.; Fouladi, Rachel T.; de Moor, Carl; Warneke, Carla L.; Luca, Mario; Jones, Mary Mullin; Rosenblum, Carol; Emmons, Karen M.; Hudmon, Karen Suchanek; Yost, Tracey E.; Gritz, Ellen R.
2007-01-01
This report presents the experimental approach and baseline findings from "Look at Your Health," an ongoing study to develop and evaluate a computer-assisted, counselor-delivered smoking cessation program for community college students. It describes the expert system software program used for data collection and for provision of tailored feedback,…
Experimental and computational flow-field results for an all-body hypersonic aircraft
NASA Technical Reports Server (NTRS)
Cleary, Joseph W.
1989-01-01
A comprehensive test program is defined which is being implemented in the NASA/Ames 3.5 foot Hypersonic Wind Tunnel for obtaining data on a generic all-body hypersonic vehicle for computational fluid dynamics (CFD) code validation. Computational methods (approximate inviscid methods and an upwind parabolized Navier-Stokes code) currently being applied to the all-body model are outlined. Experimental and computational results on surface pressure distributions and Pitot-pressure surveys for the basic sharp-nose model (without control surfaces) at a free-stream Mach number of 7 are presented.
An interactive program for pharmacokinetic modeling.
Lu, D R; Mao, F
1993-05-01
A computer program, PharmK, was developed for pharmacokinetic modeling of experimental data. The program was written in C computer language based on the high-level user-interface Macintosh operating system. The intention was to provide a user-friendly tool for users of Macintosh computers. An interactive algorithm based on the exponential stripping method is used for the initial parameter estimation. Nonlinear pharmacokinetic model fitting is based on the maximum likelihood estimation method and is performed by the Levenberg-Marquardt method based on chi 2 criterion. Several methods are available to aid the evaluation of the fitting results. Pharmacokinetic data sets have been examined with the PharmK program, and the results are comparable with those obtained with other programs that are currently available for IBM PC-compatible and other types of computers.
NASA Technical Reports Server (NTRS)
Smith, Tamara A.; Pavli, Albert J.; Kacynski, Kenneth J.
1987-01-01
The Joint Army, Navy, NASA, Air Force (JANNAF) rocket-engine performance-prediction procedure is based on the use of various reference computer programs. One of the reference programs for nozzle analysis is the Two-Dimensional Kinetics (TDK) Program. The purpose of this report is to calibrate the JANNAF procedure that has been incorporated into the December 1984 version of the TDK program for the high-area-ratio rocket-engine regime. The calibration was accomplished by modeling the performance of a 1030:1 rocket nozzle tested at NASA Lewis. A detailed description of the test conditions and TDK input parameters is given. The reuslts indicate that the computer code predicts delivered vacuum specific impulse to within 0.12 to 1.9 percent of the experimental data. Vacuum thrust coefficient predictions were within + or - 1.3 percent of experimental results. Predictions of wall static pressure were within approximately + or - 5 percent of the measured values.
NASA Technical Reports Server (NTRS)
Anderson, O. L.
1974-01-01
A finite-difference procedure for computing the turbulent, swirling, compressible flow in axisymmetric ducts is described. Arbitrary distributions of heat and mass transfer at the boundaries can be treated, and the effects of struts, inlet guide vanes, and flow straightening vanes can be calculated. The calculation procedure is programmed in FORTRAN 4 and has operated successfully on the UNIVAC 1108, IBM 360, and CDC 6600 computers. The analysis which forms the basis of the procedure, a detailed description of the computer program, and the input/output formats are presented. The results of sample calculations performed with the computer program are compared with experimental data.
Computer code for off-design performance analysis of radial-inflow turbines with rotor blade sweep
NASA Technical Reports Server (NTRS)
Meitner, P. L.; Glassman, A. J.
1983-01-01
The analysis procedure of an existing computer program was extended to include rotor blade sweep, to model the flow more accurately at the rotor exit, and to provide more detail to the loss model. The modeling changes are described and all analysis equations and procedures are presented. Program input and output are described and are illustrated by an example problem. Results obtained from this program and from a previous program are compared with experimental data.
Learning sorting algorithms through visualization construction
NASA Astrophysics Data System (ADS)
Cetin, Ibrahim; Andrews-Larson, Christine
2016-01-01
Recent increased interest in computational thinking poses an important question to researchers: What are the best ways to teach fundamental computing concepts to students? Visualization is suggested as one way of supporting student learning. This mixed-method study aimed to (i) examine the effect of instruction in which students constructed visualizations on students' programming achievement and students' attitudes toward computer programming, and (ii) explore how this kind of instruction supports students' learning according to their self-reported experiences in the course. The study was conducted with 58 pre-service teachers who were enrolled in their second programming class. They expect to teach information technology and computing-related courses at the primary and secondary levels. An embedded experimental model was utilized as a research design. Students in the experimental group were given instruction that required students to construct visualizations related to sorting, whereas students in the control group viewed pre-made visualizations. After the instructional intervention, eight students from each group were selected for semi-structured interviews. The results showed that the intervention based on visualization construction resulted in significantly better acquisition of sorting concepts. However, there was no significant difference between the groups with respect to students' attitudes toward computer programming. Qualitative data analysis indicated that students in the experimental group constructed necessary abstractions through their engagement in visualization construction activities. The authors of this study argue that the students' active engagement in the visualization construction activities explains only one side of students' success. The other side can be explained through the instructional approach, constructionism in this case, used to design instruction. The conclusions and implications of this study can be used by researchers and instructors dealing with computational thinking.
FORTRAN program for predicting off-design performance of radial-inflow turbines
NASA Technical Reports Server (NTRS)
Wasserbauer, C. A.; Glassman, A. J.
1975-01-01
The FORTRAN IV program uses a one-dimensional solution of flow conditions through the turbine along the mean streamline. The program inputs needed are the design-point requirements and turbine geometry. The output includes performance and velocity-diagram parameters over a range of speed and pressure ratio. Computed performance is compared with the experimental data from two radial-inflow turbines and with the performance calculated by a previous computer program. The flow equations, program listing, and input and output for a sample problem are given.
Hands-on work fine-tunes X-band PIN-diode duplexer
NASA Astrophysics Data System (ADS)
Schneider, P.
1985-06-01
Computer-aided design (CAD) programs for fabricating PIN-diode duplexers are useful in avoiding time-consuming cut-and-try techniques. Nevertheless, to attain minimum insertion loss, only experimentation yields the optimum microstrip circuitry. A PIN-diode duplexer, consisting of two SPST PIN-diode switches and a pair of 3-dB Lange microstrip couplers, designed for an X-band transmit/receive module exemplifies what is possible when computer-derived designs and experimentation are used together. Differences between the measured and computer-generated figures for insertion loss can be attributed to several factors not included in the CAD program - for example, radiation and connector losses. Mechanical tolerances of the microstrip PC board and variations in the SMA connector-to-microstrip transition contribute to the discrepancy.
Acoustic environmental accuracy requirements for response determination
NASA Technical Reports Server (NTRS)
Pettitt, M. R.
1983-01-01
A general purpose computer program was developed for the prediction of vehicle interior noise. This program, named VIN, has both modal and statistical energy analysis capabilities for structural/acoustic interaction analysis. The analytic models and their computer implementation were verified through simple test cases with well-defined experimental results. The model was also applied in a space shuttle payload bay launch acoustics prediction study. The computer program processes large and small problems with equal efficiency because all arrays are dynamically sized by program input variables at run time. A data base is built and easily accessed for design studies. The data base significantly reduces the computational costs of such studies by allowing the reuse of the still-valid calculated parameters of previous iterations.
Methods for design and evaluation of parallel computating systems (The PISCES project)
NASA Technical Reports Server (NTRS)
Pratt, Terrence W.; Wise, Robert; Haught, Mary JO
1989-01-01
The PISCES project started in 1984 under the sponsorship of the NASA Computational Structural Mechanics (CSM) program. A PISCES 1 programming environment and parallel FORTRAN were implemented in 1984 for the DEC VAX (using UNIX processes to simulate parallel processes). This system was used for experimentation with parallel programs for scientific applications and AI (dynamic scene analysis) applications. PISCES 1 was ported to a network of Apollo workstations by N. Fitzgerald.
THE COMPUTER AS AN AID TO INSTRUCTION AND GUIDANCE IN THE SCHOOL.
ERIC Educational Resources Information Center
IMPELLITTERI, JOSEPH T.
COMPUTER APPLICATIONS IN EDUCATION ARE DISCUSSED IN TERMS OF--(1) A DESCRIPTION OF COMPUTER-ASSISTED INSTRUCTION (CAI) AND COUNSELING, (2) THE NUMBER AND TYPES OF COMPUTER-ASSISTED DEVELOPMENTS, (3) THE NATURE OF THE PENN STATE UNIVERSITY PROGRAM, (4) TENTATIVE RESULTS OF EXPERIMENTATION USING CAI, AND (5) IMPLICATIONS AND PROJECTIONS FOR THE…
A primer for biomedical scientists on how to execute model II linear regression analysis.
Ludbrook, John
2012-04-01
1. There are two very different ways of executing linear regression analysis. One is Model I, when the x-values are fixed by the experimenter. The other is Model II, in which the x-values are free to vary and are subject to error. 2. I have received numerous complaints from biomedical scientists that they have great difficulty in executing Model II linear regression analysis. This may explain the results of a Google Scholar search, which showed that the authors of articles in journals of physiology, pharmacology and biochemistry rarely use Model II regression analysis. 3. I repeat my previous arguments in favour of using least products linear regression analysis for Model II regressions. I review three methods for executing ordinary least products (OLP) and weighted least products (WLP) regression analysis: (i) scientific calculator and/or computer spreadsheet; (ii) specific purpose computer programs; and (iii) general purpose computer programs. 4. Using a scientific calculator and/or computer spreadsheet, it is easy to obtain correct values for OLP slope and intercept, but the corresponding 95% confidence intervals (CI) are inaccurate. 5. Using specific purpose computer programs, the freeware computer program smatr gives the correct OLP regression coefficients and obtains 95% CI by bootstrapping. In addition, smatr can be used to compare the slopes of OLP lines. 6. When using general purpose computer programs, I recommend the commercial programs systat and Statistica for those who regularly undertake linear regression analysis and I give step-by-step instructions in the Supplementary Information as to how to use loss functions. © 2011 The Author. Clinical and Experimental Pharmacology and Physiology. © 2011 Blackwell Publishing Asia Pty Ltd.
Computers Can Make a Difference.
ERIC Educational Resources Information Center
Pressman, Harvey
1986-01-01
The impact of several experimental programs involving the use of microcomputers with mentally retarded, physically disabled, emotionally disabled, or developmentally disabled children is discussed. The continuing need for such programs, better access, and more dissemination of information on these programs is explored. (CB)
Hybrid, experimental and computational, investigation of mechanical components
NASA Astrophysics Data System (ADS)
Furlong, Cosme; Pryputniewicz, Ryszard J.
1996-07-01
Computational and experimental methodologies have unique features for the analysis and solution of a wide variety of engineering problems. Computations provide results that depend on selection of input parameters such as geometry, material constants, and boundary conditions which, for correct modeling purposes, have to be appropriately chosen. In addition, it is relatively easy to modify the input parameters in order to computationally investigate different conditions. Experiments provide solutions which characterize the actual behavior of the object of interest subjected to specific operating conditions. However, it is impractical to experimentally perform parametric investigations. This paper discusses the use of a hybrid, computational and experimental, approach for study and optimization of mechanical components. Computational techniques are used for modeling the behavior of the object of interest while it is experimentally tested using noninvasive optical techniques. Comparisons are performed through a fringe predictor program used to facilitate the correlation between both techniques. In addition, experimentally obtained quantitative information, such as displacements and shape, can be applied in the computational model in order to improve this correlation. The result is a validated computational model that can be used for performing quantitative analyses and structural optimization. Practical application of the hybrid approach is illustrated with a representative example which demonstrates the viability of the approach as an engineering tool for structural analysis and optimization.
Peck, Ronald F.; Colton, Shannon; Morris, Jennifer; Chaibub Neto, Elias; Kallio, Julie
2009-01-01
We conducted a controlled investigation to examine whether a combination of computer imagery and tactile tools helps introductory cell biology laboratory undergraduate students better learn about protein structure/function relationships as compared with computer imagery alone. In all five laboratory sections, students used the molecular imaging program, Protein Explorer (PE). In the three experimental sections, three-dimensional physical models were made available to the students, in addition to PE. Student learning was assessed via oral and written research summaries and videotaped interviews. Differences between the experimental and control group students were not found in our typical course assessments such as research papers, but rather were revealed during one-on-one interviews with students at the end of the semester. A subset of students in the experimental group produced superior answers to some higher-order interview questions as compared with students in the control group. During the interview, students in both groups preferred to use either the hand-held models alone or in combination with the PE imaging program. Students typically did not use any tools when answering knowledge (lower-level thinking) questions, but when challenged with higher-level thinking questions, students in both the control and experimental groups elected to use the models. PMID:19255134
ERIC Educational Resources Information Center
Pamula, F.; And Others
1995-01-01
Describes an interactive computer program written to provide accurate and immediate feedback to students while they are processing experimental data. Discusses the problems inherent in laboratory courses that led to the development of this program. Advantages of the software include allowing students to work at their own pace in a nonthreatening…
Computational structural mechanics for engine structures
NASA Technical Reports Server (NTRS)
Chamis, C. C.
1989-01-01
The computational structural mechanics (CSM) program at Lewis encompasses: (1) fundamental aspects for formulating and solving structural mechanics problems, and (2) development of integrated software systems to computationally simulate the performance/durability/life of engine structures. It is structured to mainly supplement, complement, and whenever possible replace, costly experimental efforts which are unavoidable during engineering research and development programs. Specific objectives include: investigate unique advantages of parallel and multiprocesses for: reformulating/solving structural mechanics and formulating/solving multidisciplinary mechanics and develop integrated structural system computational simulators for: predicting structural performances, evaluating newly developed methods, and for identifying and prioritizing improved/missing methods needed. Herein the CSM program is summarized with emphasis on the Engine Structures Computational Simulator (ESCS). Typical results obtained using ESCS are described to illustrate its versatility.
Improving robustness and computational efficiency using modern C++
DOE Office of Scientific and Technical Information (OSTI.GOV)
Paterno, M.; Kowalkowski, J.; Green, C.
2014-01-01
For nearly two decades, the C++ programming language has been the dominant programming language for experimental HEP. The publication of ISO/IEC 14882:2011, the current version of the international standard for the C++ programming language, makes available a variety of language and library facilities for improving the robustness, expressiveness, and computational efficiency of C++ code. However, much of the C++ written by the experimental HEP community does not take advantage of the features of the language to obtain these benefits, either due to lack of familiarity with these features or concern that these features must somehow be computationally inefficient. In thismore » paper, we address some of the features of modern C+-+, and show how they can be used to make programs that are both robust and computationally efficient. We compare and contrast simple yet realistic examples of some common implementation patterns in C, currently-typical C++, and modern C++, and show (when necessary, down to the level of generated assembly language code) the quality of the executable code produced by recent C++ compilers, with the aim of allowing the HEP community to make informed decisions on the costs and benefits of the use of modern C++.« less
V/STOL AND digital avionics system for UH-1H
NASA Technical Reports Server (NTRS)
Liden, S.
1978-01-01
A hardware and software system for the Bell UH-1H helicopter was developed that provides sophisticated navigation, guidance, control, display, and data acquisition capabilities for performing terminal area navigation, guidance and control research. Two Sperry 1819B general purpose digital computers were used. One contains the development software that performs all the specified system flight computations. The second computer is available to NASA for experimental programs that run simultaneously with the other computer programs and which may, at the push of a button, replace selected computer computations. Other features that provide research flexibility include keyboard selectable gains and parameters and software generated alphanumeric and CRT displays.
Understanding and Improving High-Performance I/O Subsystems
NASA Technical Reports Server (NTRS)
El-Ghazawi, Tarek A.; Frieder, Gideon; Clark, A. James
1996-01-01
This research program has been conducted in the framework of the NASA Earth and Space Science (ESS) evaluations led by Dr. Thomas Sterling. In addition to the many important research findings for NASA and the prestigious publications, the program has helped orienting the doctoral research program of two students towards parallel input/output in high-performance computing. Further, the experimental results in the case of the MasPar were very useful and helpful to MasPar with which the P.I. has had many interactions with the technical management. The contributions of this program are drawn from three experimental studies conducted on different high-performance computing testbeds/platforms, and therefore presented in 3 different segments as follows: 1. Evaluating the parallel input/output subsystem of a NASA high-performance computing testbeds, namely the MasPar MP- 1 and MP-2; 2. Characterizing the physical input/output request patterns for NASA ESS applications, which used the Beowulf platform; and 3. Dynamic scheduling techniques for hiding I/O latency in parallel applications such as sparse matrix computations. This study also has been conducted on the Intel Paragon and has also provided an experimental evaluation for the Parallel File System (PFS) and parallel input/output on the Paragon. This report is organized as follows. The summary of findings discusses the results of each of the aforementioned 3 studies. Three appendices, each containing a key scholarly research paper that details the work in one of the studies are included.
Computer Simulation in Undergraduate Instruction: A Symposium.
ERIC Educational Resources Information Center
Street, Warren R.; And Others
These symposium papers discuss the instructional use of computers in psychology, with emphasis on computer-produced simulations. The first, by Rich Edwards, briefly outlines LABSIM, a general purpose system of FORTRAN programs which simulate data collection in more than a dozen experimental models in psychology and are designed to train students…
Aeroelasticity Benchmark Assessment: Subsonic Fixed Wing Program
NASA Technical Reports Server (NTRS)
Florance, Jennifer P.; Chwalowski, Pawel; Wieseman, Carol D.
2010-01-01
The fundamental technical challenge in computational aeroelasticity is the accurate prediction of unsteady aerodynamic phenomena and the effect on the aeroelastic response of a vehicle. Currently, a benchmarking standard for use in validating the accuracy of computational aeroelasticity codes does not exist. Many aeroelastic data sets have been obtained in wind-tunnel and flight testing throughout the world; however, none have been globally presented or accepted as an ideal data set. There are numerous reasons for this. One reason is that often, such aeroelastic data sets focus on the aeroelastic phenomena alone (flutter, for example) and do not contain associated information such as unsteady pressures and time-correlated structural dynamic deflections. Other available data sets focus solely on the unsteady pressures and do not address the aeroelastic phenomena. Other discrepancies can include omission of relevant data, such as flutter frequency and / or the acquisition of only qualitative deflection data. In addition to these content deficiencies, all of the available data sets present both experimental and computational technical challenges. Experimental issues include facility influences, nonlinearities beyond those being modeled, and data processing. From the computational perspective, technical challenges include modeling geometric complexities, coupling between the flow and the structure, grid issues, and boundary conditions. The Aeroelasticity Benchmark Assessment task seeks to examine the existing potential experimental data sets and ultimately choose the one that is viewed as the most suitable for computational benchmarking. An initial computational evaluation of that configuration will then be performed using the Langley-developed computational fluid dynamics (CFD) software FUN3D1 as part of its code validation process. In addition to the benchmarking activity, this task also includes an examination of future research directions. Researchers within the Aeroelasticity Branch will examine other experimental efforts within the Subsonic Fixed Wing (SFW) program (such as testing of the NASA Common Research Model (CRM)) and other NASA programs and assess aeroelasticity issues and research topics.
ERIC Educational Resources Information Center
Ates, Alev; Altunay, Ugur; Altun, Eralp
2006-01-01
The aim of this research was to discern the effects of computer assisted English instruction on English language preparatory students' attitudes towards computers and English in a Turkish-medium high school with an intensive English program. A quasi-experimental time series research design, also called "before-after" or "repeated…
NASA Astrophysics Data System (ADS)
Lahti, Paul M.; Motyka, Eric J.; Lancashire, Robert J.
2000-05-01
A straightforward procedure is described to combine computation of molecular vibrational modes using commonly available molecular modeling programs with visualization of the modes using advanced features of the MDL Information Systems Inc. Chime World Wide Web browser plug-in. Minor editing of experimental spectra that are stored in the JCAMP-DX format allows linkage of IR spectral frequency ranges to Chime molecular display windows. The spectra and animation files can be combined by Hypertext Markup Language programming to allow interactive linkage between experimental spectra and computationally generated vibrational displays. Both the spectra and the molecular displays can be interactively manipulated to allow the user maximum control of the objects being viewed. This procedure should be very valuable not only for aiding students through visual linkage of spectra and various vibrational animations, but also by assisting them in learning the advantages and limitations of computational chemistry by comparison to experiment.
NASA Technical Reports Server (NTRS)
Jones, J. E.; Richmond, J. H.
1974-01-01
An integral equation formulation is applied to predict pitch- and roll-plane radiation patterns of a thin VHF/UHF (very high frequency/ultra high frequency) annular slot communications antenna operating at several locations in the nose region of the space shuttle orbiter. Digital computer programs used to compute radiation patterns are given and the use of the programs is illustrated. Experimental verification of computed patterns is given from measurements made on 1/35-scale models of the orbiter.
Prediction of overall and blade-element performance for axial-flow pump configurations
NASA Technical Reports Server (NTRS)
Serovy, G. K.; Kavanagh, P.; Okiishi, T. H.; Miller, M. J.
1973-01-01
A method and a digital computer program for prediction of the distributions of fluid velocity and properties in axial flow pump configurations are described and evaluated. The method uses the blade-element flow model and an iterative numerical solution of the radial equilbrium and continuity conditions. Correlated experimental results are used to generate alternative methods for estimating blade-element turning and loss characteristics. Detailed descriptions of the computer program are included, with example input and typical computed results.
Procedural Quantum Programming
NASA Astrophysics Data System (ADS)
Ömer, Bernhard
2002-09-01
While classical computing science has developed a variety of methods and programming languages around the concept of the universal computer, the typical description of quantum algorithms still uses a purely mathematical, non-constructive formalism which makes no difference between a hydrogen atom and a quantum computer. This paper investigates, how the concept of procedural programming languages, the most widely used classical formalism for describing and implementing algorithms, can be adopted to the field of quantum computing, and how non-classical features like the reversibility of unitary transformations, the non-observability of quantum states or the lack of copy and erase operations can be reflected semantically. It introduces the key concepts of procedural quantum programming (hybrid target architecture, operator hierarchy, quantum data types, memory management, etc.) and presents the experimental language QCL, which implements these principles.
NASA Technical Reports Server (NTRS)
Hazelton, R. C.; Yadlowsky, E. J.; Churchill, R. J.; Parker, L. W.; Sellers, B.
1981-01-01
The effect differential charging of spacecraft thermal control surfaces is assessed by studying the dynamics of the charging process. A program to experimentally validate a computer model of the charging process was established. Time resolved measurements of the surface potential were obtained for samples of Kapton and Teflon irradiated with a monoenergetic electron beam. Results indicate that the computer model and experimental measurements agree well and that for Teflon, secondary emission is the governing factor. Experimental data indicate that bulk conductivities play a significant role in the charging of Kapton.
NASA Technical Reports Server (NTRS)
Kleckner, R. J.; Rosenlieb, J. W.; Dyba, G.
1980-01-01
The results of a series of full scale hardware tests comparing predictions of the SPHERBEAN computer program with measured data are presented. The SPHERBEAN program predicts the thermomechanical performance characteristics of high speed lubricated double row spherical roller bearings. The degree of correlation between performance predicted by SPHERBEAN and measured data is demonstrated. Experimental and calculated performance data is compared over a range in speed up to 19,400 rpm (0.8 MDN) under pure radial, pure axial, and combined loads.
MEASUREMENT AND PRECISION, EXPERIMENTAL VERSION.
ERIC Educational Resources Information Center
Harvard Univ., Cambridge, MA. Harvard Project Physics.
THIS DOCUMENT IS AN EXPERIMENTAL VERSION OF A PROGRAMED TEXT ON MEASUREMENT AND PRECISION. PART I CONTAINS 24 FRAMES DEALING WITH PRECISION AND SIGNIFICANT FIGURES ENCOUNTERED IN VARIOUS MATHEMATICAL COMPUTATIONS AND MEASUREMENTS. PART II BEGINS WITH A BRIEF SECTION ON EXPERIMENTAL DATA, COVERING SUCH POINTS AS (1) ESTABLISHING THE ZERO POINT, (2)…
ERIC Educational Resources Information Center
Science Software Quarterly, 1984
1984-01-01
Provides extensive reviews of computer software, examining documentation, ease of use, performance, error handling, special features, and system requirements. Includes statistics, problem-solving (TK Solver), label printing, database management, experimental psychology, Encyclopedia Britannica biology, and DNA-sequencing programs. A program for…
SPECIFIC HEAT DATA ANALYSIS PROGRAM FOR THE IBM 704 DIGITAL COMPUTER
DOE Office of Scientific and Technical Information (OSTI.GOV)
Roach, P.R.
1962-01-01
A computer program was developed to calculate the specific heat of a substance in the temperature range from 0.3 to 4.2 deg K, given temperature calibration data for a carbon resistance thermometer, experimental temperature drift, and heating period data. The speciftc heats calculated from these data are then fitted by a curve by the methods of least squares and the specific heats are corrected for the effect of the curvature of the data. The method, operation, program details, and program stops are discussed. A program listing is included. (M.C.G.)
ERIC Educational Resources Information Center
Huffstetter, Mary; King, James R.; Onwuegbuzie, Anthony J.; Schneider, Jenifer J.; Powell-Smith, Kelly A.
2010-01-01
This study examined the effects of a computer-based early reading program (Headsprout Early Reading) on the oral language and early reading skills of at-risk preschool children. In a pretest-posttest control group design, 62 children were randomly assigned to receive supplemental instruction with Headsprout Early Reading (experimental group) or…
NASA Astrophysics Data System (ADS)
McCreery, Glenn Ernest
An experimental and analytical investigation of dispersed and dispersed-annular (rivulet or thin film) flow phase separation in tees has been successfully completed. The research was directed at, but is not specific to, determining flow conditions, following a loss of coolant accident, in the large rectangular passageways leading to vacuum buildings in the containment envelope of some CANDU nuclear reactors. The primary objectives of the research were to: (1) obtain experimental data to help formulate and test mechanistic analytical models of phase separation, and (2) develop the analytical models in computer programs which predict phase separation from upstream flow and pressure conditions and downstream and side branch pressure boundary conditions. To meet these objectives an air-water experimental apparatus was constructed, and consists of large air blowers attached to a long rectangular duct leading to a tee in the horizontal plane. A variety of phenomena was investigated including, for comparison with computer predictions, air streamlines and eddy boundary geometry, drop size spectra, macroscopic mass balances, liquid rivulet pathlines, and trajectories of drops of known size and velocity. Four separate computer programs were developed to analyze phase separation. Three of the programs are used sequentially to calculate dispersed mist phase separation in a tee. The fourth is used to calculate rivulet or thin film pathlines. Macroscopic mass balances are calculated from a summation of mass balances for drops with representative sizes (and masses) spaced across the drop size spectrum. The programs are tested against experimental data, and accurately predict gas flow fields, drop trajectories, rivulet pathlines and macroscopic mass balances. In addition to development of the computer programs, analysis was performed to specify the scaling of dispersed mist and rivulet or thin film flow, to investigate pressure losses in tees, and the inter-relationship of loss coefficients, contraction coefficients, and eddy geometry. The important transient effects of liquid storage in eddies were also analyzed.
ERIC Educational Resources Information Center
Journal of Chemical Education, 1988
1988-01-01
Reviews three computer software packages for chemistry education including "Osmosis and Diffusion" and "E.M.E. Titration Lab" for Apple II and "Simplex-V: An Interactive Computer Program for Experimental Optimization" for IBM PC. Summary ratings include ease of use, content, pedagogic value, student reaction, and…
Workflow of the Grover algorithm simulation incorporating CUDA and GPGPU
NASA Astrophysics Data System (ADS)
Lu, Xiangwen; Yuan, Jiabin; Zhang, Weiwei
2013-09-01
The Grover quantum search algorithm, one of only a few representative quantum algorithms, can speed up many classical algorithms that use search heuristics. No true quantum computer has yet been developed. For the present, simulation is one effective means of verifying the search algorithm. In this work, we focus on the simulation workflow using a compute unified device architecture (CUDA). Two simulation workflow schemes are proposed. These schemes combine the characteristics of the Grover algorithm and the parallelism of general-purpose computing on graphics processing units (GPGPU). We also analyzed the optimization of memory space and memory access from this perspective. We implemented four programs on CUDA to evaluate the performance of schemes and optimization. Through experimentation, we analyzed the organization of threads suited to Grover algorithm simulations, compared the storage costs of the four programs, and validated the effectiveness of optimization. Experimental results also showed that the distinguished program on CUDA outperformed the serial program of libquantum on a CPU with a speedup of up to 23 times (12 times on average), depending on the scale of the simulation.
NASA Technical Reports Server (NTRS)
Abbott, John M.; Anderson, Bernhard H.; Rice, Edward J.
1990-01-01
The internal fluid mechanics research program in inlets, ducts, and nozzles consists of a balanced effort between the development of computational tools (both parabolized Navier-Stokes and full Navier-Stokes) and the conduct of experimental research. The experiments are designed to better understand the fluid flow physics, to develop new or improved flow models, and to provide benchmark quality data sets for validation of the computational methods. The inlet, duct, and nozzle research program is described according to three major classifications of flow phenomena: (1) highly 3-D flow fields; (2) shock-boundary-layer interactions; and (3) shear layer control. Specific examples of current and future elements of the research program are described for each of these phenomenon. In particular, the highly 3-D flow field phenomenon is highlighted by describing the computational and experimental research program in transition ducts having a round-to-rectangular area variation. In the case of shock-boundary-layer interactions, the specific details of research for normal shock-boundary-layer interactions are described. For shear layer control, research in vortex generators and the use of aerodynamic excitation for enhancement of the jet mixing process are described.
Feasibility of MHD submarine propulsion
DOE Office of Scientific and Technical Information (OSTI.GOV)
Doss, E.D.; Sikes, W.C.
1992-09-01
This report describes the work performed during Phase 1 and Phase 2 of the collaborative research program established between Argonne National Laboratory (ANL) and Newport News Shipbuilding and Dry Dock Company (NNS). Phase I of the program focused on the development of computer models for Magnetohydrodynamic (MHD) propulsion. Phase 2 focused on the experimental validation of the thruster performance models and the identification, through testing, of any phenomena which may impact the attractiveness of this propulsion system for shipboard applications. The report discusses in detail the work performed in Phase 2 of the program. In Phase 2, a two Teslamore » test facility was designed, built, and operated. The facility test loop, its components, and their design are presented. The test matrix and its rationale are discussed. Representative experimental results of the test program are presented, and are compared to computer model predictions. In general, the results of the tests and their comparison with the predictions indicate that thephenomena affecting the performance of MHD seawater thrusters are well understood and can be accurately predicted with the developed thruster computer models.« less
New technology in turbine aerodynamics.
NASA Technical Reports Server (NTRS)
Glassman, A. J.; Moffitt, T. P.
1972-01-01
Cursory review of some recent work that has been done in turbine aerodynamic research. Topics discussed include the aerodynamic effect of turbine coolant, high work-factor (ratio of stage work to square of blade speed) turbines, and computer methods for turbine design and performance prediction. Experimental cooled-turbine aerodynamics programs using two-dimensional cascades, full annular cascades, and cold rotating turbine stage tests are discussed with some typical results presented. Analytically predicted results for cooled blade performance are compared to experimental results. The problems and some of the current programs associated with the use of very high work factors for fan-drive turbines of high-bypass-ratio engines are discussed. Computer programs have been developed for turbine design-point performance, off-design performance, supersonic blade profile design, and the calculation of channel velocities for subsonic and transonic flowfields. The use of these programs for the design and analysis of axial and radial turbines is discussed.
Gaming via Computer Simulation Techniques for Junior College Economics Education. Final Report.
ERIC Educational Resources Information Center
Thompson, Fred A.
A study designed to answer the need for more attractive and effective economics education involved the teaching of one junior college economics class by the conventional (lecture) method and an experimental class by computer simulation techniques. Econometric models approximating the "real world" were computer programed to enable the experimental…
Computational challenges in modeling gene regulatory events.
Pataskar, Abhijeet; Tiwari, Vijay K
2016-10-19
Cellular transcriptional programs driven by genetic and epigenetic mechanisms could be better understood by integrating "omics" data and subsequently modeling the gene-regulatory events. Toward this end, computational biology should keep pace with evolving experimental procedures and data availability. This article gives an exemplified account of the current computational challenges in molecular biology.
Instrumentino: An Open-Source Software for Scientific Instruments.
Koenka, Israel Joel; Sáiz, Jorge; Hauser, Peter C
2015-01-01
Scientists often need to build dedicated computer-controlled experimental systems. For this purpose, it is becoming common to employ open-source microcontroller platforms, such as the Arduino. These boards and associated integrated software development environments provide affordable yet powerful solutions for the implementation of hardware control of transducers and acquisition of signals from detectors and sensors. It is, however, a challenge to write programs that allow interactive use of such arrangements from a personal computer. This task is particularly complex if some of the included hardware components are connected directly to the computer and not via the microcontroller. A graphical user interface framework, Instrumentino, was therefore developed to allow the creation of control programs for complex systems with minimal programming effort. By writing a single code file, a powerful custom user interface is generated, which enables the automatic running of elaborate operation sequences and observation of acquired experimental data in real time. The framework, which is written in Python, allows extension by users, and is made available as an open source project.
NASA Astrophysics Data System (ADS)
Wilcox, William Edward, Jr.
1995-01-01
A computer program (LIDAR-PC) and associated atmospheric spectral databases have been developed which accurately simulate the laser remote sensing of the atmosphere and the system performance of a direct-detection Lidar or tunable Differential Absorption Lidar (DIAL) system. This simulation program allows, for the first time, the use of several different large atmospheric spectral databases to be coupled with Lidar parameter simulations on the same computer platform to provide a real-time, interactive, and easy to use design tool for atmospheric Lidar simulation and modeling. LIDAR -PC has been used for a range of different Lidar simulations and compared to experimental Lidar data. In general, the simulations agreed very well with the experimental measurements. In addition, the simulation offered, for the first time, the analysis and comparison of experimental Lidar data to easily determine the range-resolved attenuation coefficient of the atmosphere and the effect of telescope overlap factor. The software and databases operate on an IBM-PC or compatible computer platform, and thus are very useful to the research community for Lidar analysis. The complete Lidar and atmospheric spectral transmission modeling program uses the HITRAN database for high-resolution molecular absorption lines of the atmosphere, the BACKSCAT/LOWTRAN computer databases and models for the effects of aerosol and cloud backscatter and attenuation, and the range-resolved Lidar equation. The program can calculate the Lidar backscattered signal-to-noise for a slant path geometry from space and simulate the effect of high resolution, tunable, single frequency, and moderate line width lasers on the Lidar/DIAL signal. The program was used to model and analyze the experimental Lidar data obtained from several measurements. A fixed wavelength, Ho:YSGG aerosol Lidar (Sugimoto, 1990) developed at USF and a tunable Ho:YSGG DIAL system (Cha, 1991) for measuring atmospheric water vapor at 2.1 μm were analyzed. The simulations agreed very well with the measurements, and also yielded, for the first time, the ability to easily deduce the atmospheric attentuation coefficient, alpha, from the Lidar data. Simulations and analysis of other Lidar measurements included that of a 1.57 mu m OPO aerosol Lidar system developed at USF (Harrell, 1995) and of the NASA LITE (Laser-in-Space Technology Experiment) Lidar recently flown in the Space shuttle. Finally, an extensive series of laboratory experiments were made with the 1.57 μm OPO Lidar system to test calculations of the telescope/laser overlap and the effect of different telescope sizes and designs. The simulations agreed well with the experimental data for the telescope diameter and central obscuration test cases. The LIDAR-PC programs are available on the Internet from the USAF Lidar Home Page Web site, http://www.cas.usf.edu/physics/lidar.html/.
An Antibiotic Resource Program for Students of the Health Professions.
ERIC Educational Resources Information Center
Tritz, Gerald J.
1986-01-01
Provides a description of a computer program developed to supplement instruction in testing of antibiotics on clinical isolates of microorganisms. The program is a simulation and database for interpretation of experimental data designed to enhance laboratory learning and prepare future physicians to use computerized diagnostic instrumentation and…
NASA Technical Reports Server (NTRS)
Hopkins, Dale A.
1992-01-01
The presentation gives a partial overview of research and development underway in the Structures Division of LeRC, which collectively is referred to as the Computational Structures Technology Program. The activities in the program are diverse and encompass four major categories: (1) composite materials and structures; (2) probabilistic analysis and reliability; (3) design optimization and expert systems; and (4) computational methods and simulation. The approach of the program is comprehensive and entails exploration of fundamental theories of structural mechanics to accurately represent the complex physics governing engine structural performance, formulation, and implementation of computational techniques and integrated simulation strategies to provide accurate and efficient solutions of the governing theoretical models by exploiting the emerging advances in computer technology, and validation and verification through numerical and experimental tests to establish confidence and define the qualities and limitations of the resulting theoretical models and computational solutions. The program comprises both in-house and sponsored research activities. The remainder of the presentation provides a sample of activities to illustrate the breadth and depth of the program and to demonstrate the accomplishments and benefits that have resulted.
ERIC Educational Resources Information Center
Kern, Alfred
1983-01-01
Describes an experimental course at Allegheny College in computer-generated poetry, which required students to deal simultaneously with grammar and rhetoric, poetics, the computer and BASIC, logic and artificial intelligence in order to create programs that would generate poetry. Examples of verses produced by course participants are included.…
Robot computer problem solving system
NASA Technical Reports Server (NTRS)
Becker, J. D.; Merriam, E. W.
1974-01-01
The conceptual, experimental, and practical phases of developing a robot computer problem solving system are outlined. Robot intelligence, conversion of the programming language SAIL to run under the THNEX monitor, and the use of the network to run several cooperating jobs at different sites are discussed.
Analysis of Compression Pad Cavities for the Orion Heatshield
NASA Technical Reports Server (NTRS)
Thompson, Richard A.; Lessard, Victor R.; Jentink, Thomas N.; Zoby, Ernest V.
2009-01-01
Current results of a program for analysis of the compression pad cavities on the Orion heatshield are reviewed. The program was supported by experimental tests, engineering modeling, and applied computations with an emphasis on the latter presented in this paper. The computational tools and approach are described along with calculated results for wind tunnel and flight conditions. Correlations of the computed results are shown which can produce a credible prediction of heating augmentation due to cavity disturbances. The models developed for use in preliminary design of the Orion heatshield are presented.
Capillary device refilling. [liquid rocket propellant tank tests
NASA Technical Reports Server (NTRS)
Blatt, M. H.; Merino, F.; Symons, E. P.
1980-01-01
An analytical and experimental study was conducted dealing with refilling start baskets (capillary devices) with settled fluid. A computer program was written to include dynamic pressure, screen wicking, multiple-screen barriers, standpipe screens, variable vehicle mass for computing vehicle acceleration, and calculation of tank outflow rate and vapor pullthrough height. An experimental apparatus was fabricated and tested to provide data for correlation with the analytical model; the test program was conducted in normal gravity using a scale-model capillary device and ethanol as the test fluid. The test data correlated with the analytical model; the model is a versatile and apparently accurate tool for predicting start basket refilling under actual mission conditions.
Computational challenges in modeling gene regulatory events
Pataskar, Abhijeet; Tiwari, Vijay K.
2016-01-01
ABSTRACT Cellular transcriptional programs driven by genetic and epigenetic mechanisms could be better understood by integrating “omics” data and subsequently modeling the gene-regulatory events. Toward this end, computational biology should keep pace with evolving experimental procedures and data availability. This article gives an exemplified account of the current computational challenges in molecular biology. PMID:27390891
Apollo LM guidance computer software for the final lunar descent.
NASA Technical Reports Server (NTRS)
Eyles, D.
1973-01-01
In all manned lunar landings to date, the lunar module Commander has taken partial manual control of the spacecraft during the final stage of the descent, below roughly 500 ft altitude. This report describes programs developed at the Charles Stark Draper Laboratory, MIT, for use in the LM's guidance computer during the final descent. At this time computational demands on the on-board computer are at a maximum, and particularly close interaction with the crew is necessary. The emphasis is on the design of the computer software rather than on justification of the particular guidance algorithms employed. After the computer and the mission have been introduced, the current configuration of the final landing programs and an advanced version developed experimentally by the author are described.
Evaluation of the three-dimensional parabolic flow computer program SHIP
NASA Technical Reports Server (NTRS)
Pan, Y. S.
1978-01-01
The three-dimensional parabolic flow program SHIP designed for predicting supersonic combustor flow fields is evaluated to determine its capabilities. The mathematical foundation and numerical procedure are reviewed; simplifications are pointed out and commented upon. The program is then evaluated numerically by applying it to several subsonic and supersonic, turbulent, reacting and nonreacting flow problems. Computational results are compared with available experimental or other analytical data. Good agreements are obtained when the simplifications on which the program is based are justified. Limitations of the program and the needs for improvement and extension are pointed out. The present three dimensional parabolic flow program appears to be potentially useful for the development of supersonic combustors.
Investigation of air cleaning system response to accident conditions
DOE Office of Scientific and Technical Information (OSTI.GOV)
Andrae, R.W.; Bolstad, J.W.; Foster, R.D.
1980-01-01
Air cleaning system response to the stress of accident conditions are being investigated. A program overview and hghlight recent results of our investigation are presented. The program includes both analytical and experimental investigations. Computer codes for predicting effects of tornados, explosions, fires, and material transport are described. The test facilities used to obtain supportive experimental data to define structural integrity and confinement effectiveness of ventilation system components are described. Examples of experimental results for code verification, blower response to tornado transients, and filter response to tornado and explosion transients are reported.
ERIC Educational Resources Information Center
Korkmaz, Özgen
2016-01-01
The aim of the study was to present the effect of Lego Mindstorms Ev3 based design activities on students' attitudes towards learning computer programming, self-efficacy beliefs and levels of academic achievement. The research is a pretest-posttest with control group semi-experimental study. The study group of the research consisted of 53 students…
Designing for aircraft structural crashworthiness
NASA Technical Reports Server (NTRS)
Thomson, R. G.; Caiafa, C.
1981-01-01
This report describes structural aviation crash dynamics research activities being conducted on general aviation aircraft and transport aircraft. The report includes experimental and analytical correlations of load-limiting subfloor and seat configurations tested dynamically in vertical drop tests and in a horizontal sled deceleration facility. Computer predictions using a finite-element nonlinear computer program, DYCAST, of the acceleration time-histories of these innovative seat and subfloor structures are presented. Proposed application of these computer techniques, and the nonlinear lumped mass computer program KRASH, to transport aircraft crash dynamics is discussed. A proposed FAA full-scale crash test of a fully instrumented radio controlled transport airplane is also described.
HIFiRE-1 Turbulent Shock Boundary Layer Interaction - Flight Data and Computations
NASA Technical Reports Server (NTRS)
Kimmel, Roger L.; Prabhu, Dinesh
2015-01-01
The Hypersonic International Flight Research Experimentation (HIFiRE) program is a hypersonic flight test program executed by the Air Force Research Laboratory (AFRL) and Australian Defence Science and Technology Organisation (DSTO). This flight contained a cylinder-flare induced shock boundary layer interaction (SBLI). Computations of the interaction were conducted for a number of times during the ascent. The DPLR code used for predictions was calibrated against ground test data prior to exercising the code at flight conditions. Generally, the computations predicted the upstream influence and interaction pressures very well. Plateau pressures on the cylinder were predicted well at all conditions. Although the experimental heat transfer showed a large amount of scatter, especially at low heating levels, the measured heat transfer agreed well with computations. The primary discrepancy between the experiment and computation occurred in the pressures measured on the flare during second stage burn. Measured pressures exhibited large overshoots late in the second stage burn, the mechanism of which is unknown. The good agreement between flight measurements and CFD helps validate the philosophy of calibrating CFD against ground test, prior to exercising it at flight conditions.
NASA Technical Reports Server (NTRS)
Goradia, S. H.; Lilley, D. E.
1975-01-01
Theoretical and experimental studies are described which were conducted for the purpose of developing a new generalized method for the prediction of profile drag of single component airfoil sections with sharp trailing edges. This method aims at solution for the flow in the wake from the airfoil trailing edge to the large distance in the downstream direction; the profile drag of the given airfoil section can then easily be obtained from the momentum balance once the shape of velocity profile at a large distance from the airfoil trailing edge has been computed. Computer program subroutines have been developed for the computation of the profile drag and flow in the airfoil wake on CDC6600 computer. The required inputs to the computer program consist of free stream conditions and the characteristics of the boundary layers at the airfoil trailing edge or at the point of incipient separation in the neighborhood of airfoil trailing edge. The method described is quite generalized and hence can be extended to the solution of the profile drag for multi-component airfoil sections.
What Works: The Results of Evaluations on Two Interactive Multimedia Programs.
ERIC Educational Resources Information Center
Schmidt, Kathy J.
The American Institute for Learning (AIL) has developed two interactive multimedia drug-awareness programs for secondary students who have used substances experimentally or are being challenged by others to experiment with drugs. The first, "Addiction and Its Processes," is an awareness program with video and computer capabilities and an…
Testing a Wheeled Landing Gear System for the TH-57 Helicopter
1992-12-01
initial comparison was done using a structural analysis program, GIFTS , to simultaneously analyze an~i compare the gear systems. Experimental data was used...15 B. GIFTS PROGRAM RESULTS ............................ 15 1. Model...Element Total System ( GIFTS ) structural analysis program, which is resident oin the Aeiunauimia Euginme1ing Department computer system, an analysis
CHINESE GRAMMARS AND THE COMPUTER AT THE OHIO STATE UNIVERSITY. PRELIMINARY REPORT.
ERIC Educational Resources Information Center
MEYERS, L.F.; YANG, J.
SAMPLE OUTPUT SENTENCES OF VARIOUS COMIT AND SNOBOL PROGRAMS FOR TESTING A CHINESE GENERATIVE GRAMMAR ARE PRESENTED. THE GRAMMAR CHOSEN FOR EXPERIMENTATION IS A PRELIMINARY VERSION OF A TRANSFORMATIONAL GRAMMAR. ALL OF THE COMIT PROGRAMS AND ONE OF THE SNOBOL PROGRAMS USE A LINEARIZED REPRESENTATION OF TREE STRUCTURES, WITH ADDITIONAL NUMERICAL…
An Interactive Multimedia Program to Prevent HIV Transmission in Men with Intellectual Disability
ERIC Educational Resources Information Center
Wells, Jennifer; Clark, Khaya; Sarno, Karen
2014-01-01
The efficacy of a computer-based interactive multimedia HIV/AIDS prevention program for men with intellectual disability (ID) was examined using a quasi-experimental within-subjects design. Thirty-seven men with mild to moderate intellectual disability evaluated the program. The pretest and posttest instruments assessed HIV/AIDS knowledge…
Thomann, J M; Gasser, P; Bres, E F; Voegel, J C; Gramain, P
1990-02-01
An ion-selective electrode and microcomputer-based experimental setup for the study of ionic-exchange kinetics between a powdered solid and the solution is described. The equipment is composed of easily available commercial devices and a data acquisition and regularization computer program is presented. The system, especially developed to investigate the ionic adsorption, equilibrium attainment and dissolution of hard mineralized tissues, provides good reliable results by taking into account the volume changes of the reacting solution and the electrode behaviour under different experimental conditions, and by avoiding carbonation of the solution. A second computer program, using the regularized data and the experimental parameters, calculates the quantities of protons consumed and calcium released in the case of equilibrium attainment and dissolution of apatite-like compounds. Finally, typical examples of ion-exchange and dissolution kinetics under constant pH of enamel and synthetic hydroxyapatite are examined.
NASA Technical Reports Server (NTRS)
Rubesin, M. W.; Okuno, A. F.; Levy, L. L., Jr.; Mcdevitt, J. B.; Seegmiller, H. L.
1976-01-01
A combined experimental and computational research program is described for testing and guiding turbulence modeling within regions of separation induced by shock waves incident in turbulent boundary layers. Specifically, studies are made of the separated flow the rear portion of an 18%-thick circular-arc airfoil at zero angle of attack in high Reynolds number supercritical flow. The measurements include distributions of surface static pressure and local skin friction. The instruments employed include highfrequency response pressure cells and a large array of surface hot-wire skin-friction gages. Computations at the experimental flow conditions are made using time-dependent solutions of ensemble-averaged Navier-Stokes equations, plus additional equations for the turbulence modeling.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Doss, E.D.; Sikes, W.C.
1992-09-01
This report describes the work performed during Phase 1 and Phase 2 of the collaborative research program established between Argonne National Laboratory (ANL) and Newport News Shipbuilding and Dry Dock Company (NNS). Phase I of the program focused on the development of computer models for Magnetohydrodynamic (MHD) propulsion. Phase 2 focused on the experimental validation of the thruster performance models and the identification, through testing, of any phenomena which may impact the attractiveness of this propulsion system for shipboard applications. The report discusses in detail the work performed in Phase 2 of the program. In Phase 2, a two Teslamore » test facility was designed, built, and operated. The facility test loop, its components, and their design are presented. The test matrix and its rationale are discussed. Representative experimental results of the test program are presented, and are compared to computer model predictions. In general, the results of the tests and their comparison with the predictions indicate that thephenomena affecting the performance of MHD seawater thrusters are well understood and can be accurately predicted with the developed thruster computer models.« less
NASA Astrophysics Data System (ADS)
Maddrey, Elizabeth
Research in academia and industry continues to identify a decline in enrollment in computer science. One major component of this decline in enrollment is a shortage of female students. The primary reasons for the gender gap presented in the research include lack of computer experience prior to their first year in college, misconceptions about the field, negative cultural stereotypes, lack of female mentors and role models, subtle discriminations in the classroom, and lack of self-confidence (Pollock, McCoy, Carberry, Hundigopal, & You, 2004). Male students are also leaving the field due to misconceptions about the field, negative cultural stereotypes, and a lack of self-confidence. Analysis of first year attrition revealed that one of the major challenges faced by students of both genders is a lack of problem-solving skills (Beaubouef, Lucas & Howatt, 2001; Olsen, 2005; Paxton & Mumey, 2001). The purpose of this study was to investigate whether specific, non-mathematical problem-solving instruction as part of introductory programming courses significantly increased computer programming self-efficacy and achievement of students. The results of this study showed that students in the experimental group had significantly higher achievement than students in the control group. While this shows statistical significance, due to the effect size and disordinal nature of the data between groups, care has to be taken in its interpretation. The study did not show significantly higher programming self-efficacy among the experimental students. There was not enough data collected to statistically analyze the effect of the treatment on self-efficacy and achievement by gender. However, differences in means were observed between the gender groups, with females in the experimental group demonstrating a higher than average degree of self-efficacy when compared with males in the experimental group and both genders in the control group. These results suggest that the treatment from this study may provide a gender-based increase in self-efficacy and future research should focus on exploring this possibility.
Teaching Cardiovascular Integrations with Computer Laboratories.
ERIC Educational Resources Information Center
Peterson, Nils S.; Campbell, Kenneth B.
1985-01-01
Describes a computer-based instructional unit in cardiovascular physiology. The program (which employs simulated laboratory experimental techniques with a problem-solving format is designed to supplement an animal laboratory and to offer students an integrative approach to physiology through use of microcomputers. Also presents an overview of the…
Experimental and theoretical studies of near-ground acoustic radiation propagation in the atmosphere
NASA Astrophysics Data System (ADS)
Belov, Vladimir V.; Burkatovskaya, Yuliya B.; Krasnenko, Nikolai P.; Rakov, Aleksandr S.; Rakov, Denis S.; Shamanaeva, Liudmila G.
2017-11-01
Results of experimental and theoretical studies of the process of near-ground propagation of monochromatic acoustic radiation on atmospheric paths from a source to a receiver taking into account the contribution of multiple scattering on fluctuations of atmospheric temperature and wind velocity, refraction of sound on the wind velocity and temperature gradients, and its reflection by the underlying surface for different models of the atmosphere depending the sound frequency, coefficient of reflection from the underlying surface, propagation distance, and source and receiver altitudes are presented. Calculations were performed by the Monte Carlo method using the local estimation algorithm by the computer program developed by the authors. Results of experimental investigations under controllable conditions are compared with theoretical estimates and results of analytical calculations for the Delany-Bazley impedance model. Satisfactory agreement of the data obtained confirms the correctness of the suggested computer program.
NASA Astrophysics Data System (ADS)
Curtis, Joseph E.; Raghunandan, Sindhu; Nanda, Hirsh; Krueger, Susan
2012-02-01
A program to construct ensembles of biomolecular structures that are consistent with experimental scattering data are described. Specifically, we generate an ensemble of biomolecular structures by varying sets of backbone dihedral angles that are then filtered using experimentally determined restraints to rapidly determine structures that have scattering profiles that are consistent with scattering data. We discuss an application of these tools to predict a set of structures for the HIV-1 Gag protein, an intrinsically disordered protein, that are consistent with small-angle neutron scattering experimental data. We have assembled these algorithms into a program called SASSIE for structure generation, visualization, and analysis of intrinsically disordered proteins and other macromolecular ensembles using neutron and X-ray scattering restraints. Program summaryProgram title: SASSIE Catalogue identifier: AEKL_v1_0 Program summary URL:http://cpc.cs.qub.ac.uk/summaries/AEKL_v1_0.html Program obtainable from: CPC Program Library, Queen's University, Belfast, N. Ireland Licensing provisions: GNU General Public License v3 No. of lines in distributed program, including test data, etc.: 3 991 624 No. of bytes in distributed program, including test data, etc.: 826 Distribution format: tar.gz Programming language: Python, C/C++, Fortran Computer: PC/Mac Operating system: 32- and 64-bit Linux (Ubuntu 10.04, Centos 5.6) and Mac OS X (10.6.6) RAM: 1 GB Classification: 3 External routines: Python 2.6.5, numpy 1.4.0, swig 1.3.40, scipy 0.8.0, Gnuplot-py-1.8, Tcl 8.5, Tk 8.5, Mac installation requires aquaterm 1.0 (or X window system) and Xcode 3 development tools. Nature of problem: Open source software to generate structures of disordered biological molecules that subsequently allow for the comparison of computational and experimental results is limiting the use of scattering resources. Solution method: Starting with an all atom model of a protein, for example, users can input regions to vary dihedral angles, ensembles of structures can be generated. Additionally, simple two-body rigid-body rotations are supported with and without disordered regions. Generated structures can then be used to calculate small-angle scattering profiles which can then be filtered against experimentally determined data. Filtered structures can be visualized individually or as an ensemble using density plots. In the modular and expandable program framework the user can easily access our subroutines and structural coordinates can be easily obtained for study using other computational physics methods. Additional comments: The distribution file for this program is over 159 Mbytes and therefore is not delivered directly when download or Email is requested. Instead an html file giving details of how the program can be obtained is sent. Running time: Varies depending on application. Typically 10 minutes to 24 hours depending on the number of generated structures.
. "Experimental and Computational Investigation of Acetic Acid Deoxygenation over Oxophilic programmed reaction (TPRxn) at 200 to 500C. The right pane is a plot of experimental TPRxn data, collected Nanocrystal Films," The Journal of Physical Chemistry Letters (2013) "Non-aqueous Thermolytic Route
MEKS: A program for computation of inclusive jet cross sections at hadron colliders
NASA Astrophysics Data System (ADS)
Gao, Jun; Liang, Zhihua; Soper, Davison E.; Lai, Hung-Liang; Nadolsky, Pavel M.; Yuan, C.-P.
2013-06-01
EKS is a numerical program that predicts differential cross sections for production of single-inclusive hadronic jets and jet pairs at next-to-leading order (NLO) accuracy in a perturbative QCD calculation. We describe MEKS 1.0, an upgraded EKS program with increased numerical precision, suitable for comparisons to the latest experimental data from the Large Hadron Collider and Tevatron. The program integrates the regularized patron-level matrix elements over the kinematical phase space for production of two and three partons using the VEGAS algorithm. It stores the generated weighted events in finely binned two-dimensional histograms for fast offline analysis. A user interface allows one to customize computation of inclusive jet observables. Results of a benchmark comparison of the MEKS program and the commonly used FastNLO program are also documented. Program SummaryProgram title: MEKS 1.0 Catalogue identifier: AEOX_v1_0 Program summary URL:http://cpc.cs.qub.ac.uk/summaries/AEOX_v1_0.html Program obtainable from: CPC Program Library, Queen's University, Belfast, N. Ireland. Licensing provisions: Standard CPC licence, http://cpc.cs.qub.ac.uk/licence/licence.html No. of lines in distributed program, including test data, etc.: 9234 No. of bytes in distributed program, including test data, etc.: 51997 Distribution format: tar.gz Programming language: Fortran (main program), C (CUBA library and analysis program). Computer: All. Operating system: Any UNIX-like system. RAM: ˜300 MB Classification: 11.1. External routines: LHAPDF (https://lhapdf.hepforge.org/) Nature of problem: Computation of differential cross sections for inclusive production of single hadronic jets and jet pairs at next-to-leading order accuracy in perturbative quantum chromodynamics. Solution method: Upon subtraction of infrared singularities, the hard-scattering matrix elements are integrated over available phase space using an optimized VEGAS algorithm. Weighted events are generated and filled into a finely binned two-dimensional histogram, from which the final cross sections with typical experimental binning and cuts are computed by an independent analysis program. Monte Carlo sampling of event weights is tuned automatically to get better efficiency. Running time: Depends on details of the calculation and sought numerical accuracy. See benchmark performance in Section 4. The tests provided take approximately 27 min for the jetbin run and a few seconds for jetana.
NASA Technical Reports Server (NTRS)
Chang, I. C.
1984-01-01
A new computer program is presented for calculating the quasi-steady transonic flow past a helicopter rotor blade in hover as well as in forward flight. The program is based on the full potential equations in a blade attached frame of reference and is capable of treating a very general class of rotor blade geometries. Computed results show good agreement with available experimental data for both straight and swept tip blade geometries.
Numerical and experimental evaluations of the flow past nested chevrons
NASA Technical Reports Server (NTRS)
Foss, J. F.; Foss, J. K.; Spalart, P. R.
1989-01-01
An effort is made to contribute to the development of CFD by relating the successful use of vortex dynamics in the computation of the pressure drop past a planar array of chevron-shaped obstructions. An ensemble of results was used to compute the loss coefficient k, stimulating an experimental program for the assessment of the measured loss coefficient for the same geometry. The most provocative result of this study has been the representation of kinetic energy production in terms of vorticity source terms.
The Technologies of EXPER SIM.
ERIC Educational Resources Information Center
Hedberg, John G.
EXPER SIM has been translated into two basic software systems: the Michigan Experimental Simulation Supervisor (MESS) and Louisville Experiment Simulation Supervisor (LESS). MESS and LESS have been programed to facilitate student interaction with the computer for research purposes. The programs contain models for several statistical analyses, and…
An Interactive Multimedia Software Program for Exploring Electrochemical Cells.
ERIC Educational Resources Information Center
Greenbowe, Thomas J.
1994-01-01
Describes computer-animated sequences and interactive multimedia instructional programs for use in introductory chemistry which allow students to explore electrochemical cells. The workbench section enables students to manipulate the experimental apparatus, chemicals, and instruments in order to design and build an experiment. The interactive…
Computational modeling of RNA 3D structures, with the aid of experimental restraints
Magnus, Marcin; Matelska, Dorota; Łach, Grzegorz; Chojnowski, Grzegorz; Boniecki, Michal J; Purta, Elzbieta; Dawson, Wayne; Dunin-Horkawicz, Stanislaw; Bujnicki, Janusz M
2014-01-01
In addition to mRNAs whose primary function is transmission of genetic information from DNA to proteins, numerous other classes of RNA molecules exist, which are involved in a variety of functions, such as catalyzing biochemical reactions or performing regulatory roles. In analogy to proteins, the function of RNAs depends on their structure and dynamics, which are largely determined by the ribonucleotide sequence. Experimental determination of high-resolution RNA structures is both laborious and difficult, and therefore, the majority of known RNAs remain structurally uncharacterized. To address this problem, computational structure prediction methods were developed that simulate either the physical process of RNA structure formation (“Greek science” approach) or utilize information derived from known structures of other RNA molecules (“Babylonian science” approach). All computational methods suffer from various limitations that make them generally unreliable for structure prediction of long RNA sequences. However, in many cases, the limitations of computational and experimental methods can be overcome by combining these two complementary approaches with each other. In this work, we review computational approaches for RNA structure prediction, with emphasis on implementations (particular programs) that can utilize restraints derived from experimental analyses. We also list experimental approaches, whose results can be relatively easily used by computational methods. Finally, we describe case studies where computational and experimental analyses were successfully combined to determine RNA structures that would remain out of reach for each of these approaches applied separately. PMID:24785264
Computer Assisted Instruction for the Mentally Retarded.
ERIC Educational Resources Information Center
Providence Coll., RI.
Computer Assisted Instruction (CAI) for the mentally retarded is described; the advantages of CAI (which generally follows the pattern of programed instruction) are listed; and the roles of the teacher and the student are summarized. The coursewriter is explained, and its use as an experimental tool discussed. Guidelines are given covering…
Computer Assisted Teaching Comparisons with Handicapped. Final Report.
ERIC Educational Resources Information Center
Main, JoDell K.
A project was conducted to see if computer-assisted instruction could be used successfully with the low-level, non-reading adult. The experimental classroom group consisted of mentally handicapped and other educationally handicapped adults in adult basic education (ABE) programs. (Long-range implementation is aimed at ABE students who have a…
Nordstrom, M A; Mapletoft, E A; Miles, T S
1995-11-01
A solution is described for the acquisition on a personal computer of standard pulses derived from neuronal discharge, measurement of neuronal discharge times, real-time control of stimulus delivery based on specified inter-pulse interval conditions in the neuronal spike train, and on-line display and analysis of the experimental data. The hardware consisted of an Apple Macintosh IIci computer and a plug-in card (National Instruments NB-MIO16) that supports A/D, D/A, digital I/O and timer functions. The software was written in the object-oriented graphical programming language LabView. Essential elements of the source code of the LabView program are presented and explained. The use of the system is demonstrated in an experiment in which the reflex responses to muscle stretch are assessed for a single motor unit in the human masseter muscle.
Experimental aerothermodynamic research of hypersonic aircraft
NASA Technical Reports Server (NTRS)
Cleary, Joseph W.
1987-01-01
The 2-D and 3-D advance computer codes being developed for use in the design of such hypersonic aircraft as the National Aero-Space Plane require comparison of the computational results with a broad spectrum of experimental data to fully assess the validity of the codes. This is particularly true for complex flow fields with control surfaces present and for flows with separation, such as leeside flow. Therefore, the objective is to provide a hypersonic experimental data base required for validation of advanced computational fluid dynamics (CFD) computer codes and for development of more thorough understanding of the flow physics necessary for these codes. This is being done by implementing a comprehensive test program for a generic all-body hypersonic aircraft model in the NASA/Ames 3.5 foot Hypersonic Wind Tunnel over a broad range of test conditions to obtain pertinent surface and flowfield data. Results from the flow visualization portion of the investigation are presented.
Optomechanical study and optimization of cantilever plate dynamics
NASA Astrophysics Data System (ADS)
Furlong, Cosme; Pryputniewicz, Ryszard J.
1995-06-01
Optimum dynamic characteristics of an aluminum cantilever plate containing holes of different sizes and located at arbitrary positions on the plate are studied computationally and experimentally. The objective function of this optimization is the minimization/maximization of the natural frequencies of the plate in terms of such design variable s as the sizes and locations of the holes. The optimization process is performed using the finite element method and mathematical programming techniques in order to obtain the natural frequencies and the optimum conditions of the plate, respectively. The modal behavior of the resultant optimal plate layout is studied experimentally through the use of holographic interferometry techniques. Comparisons of the computational and experimental results show that good agreement between theory and test is obtained. The comparisons also show that the combined, or hybrid use of experimental and computational techniques complement each other and prove to be a very efficient tool for performing optimization studies of mechanical components.
Rotor-to-stator Partial Rubbing and Its Effects on Rotor Dynamic Response
NASA Technical Reports Server (NTRS)
Muszynska, Agnes; Franklin, Wesley D.; Hayashida, Robert D.
1991-01-01
Results from experimental and analytical studies on rotor to stationary element partial rubbings at several locations and their effects on rotor dynamic responses are presented. The mathematical model of a rubbing rotor is given. The computer program provides numerical results which agree with experimentally obtained rotor responses.
NASA Astrophysics Data System (ADS)
Kuan, Wen-Hsuan; Tseng, Chi-Hung; Chen, Sufen; Wong, Ching-Chang
2016-06-01
We propose an integrated curriculum to establish essential abilities of computer programming for the freshmen of a physics department. The implementation of the graphical-based interfaces from Scratch to LabVIEW then to LabVIEW for Arduino in the curriculum `Computer-Assisted Instrumentation in the Design of Physics Laboratories' brings rigorous algorithm and syntax protocols together with imagination, communication, scientific applications and experimental innovation. The effectiveness of the curriculum was evaluated via statistical analysis of questionnaires, interview responses, the increase in student numbers majoring in physics, and performance in a competition. The results provide quantitative support that the curriculum remove huge barriers to programming which occur in text-based environments, helped students gain knowledge of programming and instrumentation, and increased the students' confidence and motivation to learn physics and computer languages.
Mishra, Bud; Daruwala, Raoul-Sam; Zhou, Yi; Ugel, Nadia; Policriti, Alberto; Antoniotti, Marco; Paxia, Salvatore; Rejali, Marc; Rudra, Archisman; Cherepinsky, Vera; Silver, Naomi; Casey, William; Piazza, Carla; Simeoni, Marta; Barbano, Paolo; Spivak, Marina; Feng, Jiawu; Gill, Ofer; Venkatesh, Mysore; Cheng, Fang; Sun, Bing; Ioniata, Iuliana; Anantharaman, Thomas; Hubbard, E Jane Albert; Pnueli, Amir; Harel, David; Chandru, Vijay; Hariharan, Ramesh; Wigler, Michael; Park, Frank; Lin, Shih-Chieh; Lazebnik, Yuri; Winkler, Franz; Cantor, Charles R; Carbone, Alessandra; Gromov, Mikhael
2003-01-01
We collaborate in a research program aimed at creating a rigorous framework, experimental infrastructure, and computational environment for understanding, experimenting with, manipulating, and modifying a diverse set of fundamental biological processes at multiple scales and spatio-temporal modes. The novelty of our research is based on an approach that (i) requires coevolution of experimental science and theoretical techniques and (ii) exploits a certain universality in biology guided by a parsimonious model of evolutionary mechanisms operating at the genomic level and manifesting at the proteomic, transcriptomic, phylogenic, and other higher levels. Our current program in "systems biology" endeavors to marry large-scale biological experiments with the tools to ponder and reason about large, complex, and subtle natural systems. To achieve this ambitious goal, ideas and concepts are combined from many different fields: biological experimentation, applied mathematical modeling, computational reasoning schemes, and large-scale numerical and symbolic simulations. From a biological viewpoint, the basic issues are many: (i) understanding common and shared structural motifs among biological processes; (ii) modeling biological noise due to interactions among a small number of key molecules or loss of synchrony; (iii) explaining the robustness of these systems in spite of such noise; and (iv) cataloging multistatic behavior and adaptation exhibited by many biological processes.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Thrall, Brian D.; Minard, Kevin R.; Teeguarden, Justin G.
A Cooperative Research and Development Agreement (CRADA) was sponsored by Battelle Memorial Institute (Battelle, Columbus), to initiate a collaborative research program across multiple Department of Energy (DOE) National Laboratories aimed at developing a suite of new capabilities for predictive toxicology. Predicting the potential toxicity of emerging classes of engineered nanomaterials was chosen as one of two focusing problems for this program. PNNL’s focus toward this broader goal was to refine and apply experimental and computational tools needed to provide quantitative understanding of nanoparticle dosimetry for in vitro cell culture systems, which is necessary for comparative risk estimates for different nanomaterialsmore » or biological systems. Research conducted using lung epithelial and macrophage cell models successfully adapted magnetic particle detection and fluorescent microscopy technologies to quantify uptake of various forms of engineered nanoparticles, and provided experimental constraints and test datasets for benchmark comparison against results obtained using an in vitro computational dosimetry model, termed the ISSD model. The experimental and computational approaches developed were used to demonstrate how cell dosimetry is applied to aid in interpretation of genomic studies of nanoparticle-mediated biological responses in model cell culture systems. The combined experimental and theoretical approach provides a highly quantitative framework for evaluating relationships between biocompatibility of nanoparticles and their physical form in a controlled manner.« less
K-TIF: a two-fluid computer program for downcomer flow dynamics. [PWR
DOE Office of Scientific and Technical Information (OSTI.GOV)
Amsden, A.A.; Harlow, F.H.
1977-10-01
The K-TIF computer program has been developed for numerical solution of the time-varying dynamics of steam and water in a pressurized water reactor downcomer. The current status of physical and mathematical modeling is presented in detail. The report also contains a complete description of the numerical solution technique, a full description and listing of the computer program, instructions for its use, with a sample printout for a specific test problem. A series of calculations, performed with no change in the modeling parameters, shows consistent agreement with the experimental trends over a wide range of conditions, which gives confidence to themore » calculations as a basis for investigating the complicated physics of steam-water flows in the downcomer.« less
Comprehensive silicon solar-cell computer modeling
NASA Technical Reports Server (NTRS)
Lamorte, M. F.
1984-01-01
A comprehensive silicon solar cell computer modeling scheme was developed to perform the following tasks: (1) model and analysis of the net charge distribution in quasineutral regions; (2) experimentally determined temperature behavior of Spire Corp. n+pp+ solar cells where n+-emitter is formed by ion implantation of 75As or 31P; and (3) initial validation results of computer simulation program using Spire Corp. n+pp+ cells.
NASA Technical Reports Server (NTRS)
Bennett, Robert M.; Batina, John T.
1989-01-01
The application and assessment of a computer program called CAP-TSD (Computational Aeroelasticity Program - Transonic Small Disturbance) for flutter predictions are described. Flutter calculations are presented for two thin swept-and-tapered wing planforms with well-defined modal properties. One planform is a series of 45-degree swept wings and the other planform is a clipped delta wing. Comparisons are made between the results of CAP-TSD using the linear equation and no airfoil thickness and the results obtained from a subsonic kernel function analysis. The calculations cover a Mach number range from low subsonic to low supersonic values, including the transonic range, and are compared with subsonic linear theory and experimental data. It is noted that since both wings have very thin airfoil sections, the effects of thickness are minimal.
NASA Technical Reports Server (NTRS)
Brentner, K. S.
1986-01-01
A computer program has been developed at the Langley Research Center to predict the discrete frequency noise of conventional and advanced helicopter rotors. The program, called WOPWOP, uses the most advanced subsonic formulation of Farassat that is less sensitive to errors and is valid for nearly all helicopter rotor geometries and flight conditions. A brief derivation of the acoustic formulation is presented along with a discussion of the numerical implementation of the formulation. The computer program uses realistic helicopter blade motion and aerodynamic loadings, input by the user, for noise calculation in the time domain. A detailed definition of all the input variables, default values, and output data is included. A comparison with experimental data shows good agreement between prediction and experiment; however, accurate aerodynamic loading is needed.
NASA Technical Reports Server (NTRS)
Boyce, Lola; Bast, Callie C.; Trimble, Greg A.
1992-01-01
This report presents the results of a fourth year effort of a research program, conducted for NASA-LeRC by the University of Texas at San Antonio (UTSA). The research included on-going development of methodology that provides probabilistic lifetime strength of aerospace materials via computational simulation. A probabilistic material strength degradation model, in the form of a randomized multifactor interaction equation, is postulated for strength degradation of structural components of aerospace propulsion systems subject to a number of effects or primitive variables. These primitive variables may include high temperature, fatigue or creep. In most cases, strength is reduced as a result of the action of a variable. This multifactor interaction strength degradation equation has been randomized and is included in the computer program, PROMISS. Also included in the research is the development of methodology to calibrate the above-described constitutive equation using actual experimental materials data together with regression analysis of that data, thereby predicting values for the empirical material constants for each effect or primitive variable. This regression methodology is included in the computer program, PROMISC. Actual experimental materials data were obtained from industry and the open literature for materials typically for applications in aerospace propulsion system components. Material data for Inconel 718 has been analyzed using the developed methodology.
NASA Technical Reports Server (NTRS)
Boyce, Lola; Bast, Callie C.; Trimble, Greg A.
1992-01-01
The results of a fourth year effort of a research program conducted for NASA-LeRC by The University of Texas at San Antonio (UTSA) are presented. The research included on-going development of methodology that provides probabilistic lifetime strength of aerospace materials via computational simulation. A probabilistic material strength degradation model, in the form of a randomized multifactor interaction equation, is postulated for strength degradation of structural components of aerospace propulsion systems subjected to a number of effects or primitive variables. These primitive variables may include high temperature, fatigue, or creep. In most cases, strength is reduced as a result of the action of a variable. This multifactor interaction strength degradation equation was randomized and is included in the computer program, PROMISC. Also included in the research is the development of methodology to calibrate the above-described constitutive equation using actual experimental materials data together with regression analysis of that data, thereby predicting values for the empirical material constants for each effect or primitive variable. This regression methodology is included in the computer program, PROMISC. Actual experimental materials data were obtained from industry and the open literature for materials typically for applications in aerospace propulsion system components. Material data for Inconel 718 was analyzed using the developed methodology.
NASA Rotor 37 CFD Code Validation: Glenn-HT Code
NASA Technical Reports Server (NTRS)
Ameri, Ali A.
2010-01-01
In order to advance the goals of NASA aeronautics programs, it is necessary to continuously evaluate and improve the computational tools used for research and design at NASA. One such code is the Glenn-HT code which is used at NASA Glenn Research Center (GRC) for turbomachinery computations. Although the code has been thoroughly validated for turbine heat transfer computations, it has not been utilized for compressors. In this work, Glenn-HT was used to compute the flow in a transonic compressor and comparisons were made to experimental data. The results presented here are in good agreement with this data. Most of the measures of performance are well within the measurement uncertainties and the exit profiles of interest agree with the experimental measurements.
Genetic Parallel Programming: design and implementation.
Cheang, Sin Man; Leung, Kwong Sak; Lee, Kin Hong
2006-01-01
This paper presents a novel Genetic Parallel Programming (GPP) paradigm for evolving parallel programs running on a Multi-Arithmetic-Logic-Unit (Multi-ALU) Processor (MAP). The MAP is a Multiple Instruction-streams, Multiple Data-streams (MIMD), general-purpose register machine that can be implemented on modern Very Large-Scale Integrated Circuits (VLSIs) in order to evaluate genetic programs at high speed. For human programmers, writing parallel programs is more difficult than writing sequential programs. However, experimental results show that GPP evolves parallel programs with less computational effort than that of their sequential counterparts. It creates a new approach to evolving a feasible problem solution in parallel program form and then serializes it into a sequential program if required. The effectiveness and efficiency of GPP are investigated using a suite of 14 well-studied benchmark problems. Experimental results show that GPP speeds up evolution substantially.
Thermal energy storage flight experiments
NASA Technical Reports Server (NTRS)
Namkoong, D.
1989-01-01
Consideration is given to the development of an experimental program to study heat transfer, energy storage, fluid movement, and void location under microgravity. Plans for experimental flight packages containing Thermal Energy Storage (TES) material applicable for advanced solar heat receivers are discussed. Candidate materials for TES include fluoride salts, salt eutectics, silicides, and metals. The development of a three-dimensional computer program to describe TES material behavior undergoing melting and freezing under microgravity is also discussed. The TES experiment concept and plans for ground and flight tests are outlined.
Experimental program for real gas flow code validation at NASA Ames Research Center
NASA Technical Reports Server (NTRS)
Deiwert, George S.; Strawa, Anthony W.; Sharma, Surendra P.; Park, Chul
1989-01-01
The experimental program for validating real gas hypersonic flow codes at NASA Ames Rsearch Center is described. Ground-based test facilities used include ballistic ranges, shock tubes and shock tunnels, arc jet facilities and heated-air hypersonic wind tunnels. Also included are large-scale computer systems for kinetic theory simulations and benchmark code solutions. Flight tests consist of the Aeroassist Flight Experiment, the Space Shuttle, Project Fire 2, and planetary probes such as Galileo, Pioneer Venus, and PAET.
Students' Explanations in Complex Learning of Disciplinary Programming
ERIC Educational Resources Information Center
Vieira, Camilo
2016-01-01
Computational Science and Engineering (CSE) has been denominated as the third pillar of science and as a set of important skills to solve the problems of a global society. Along with the theoretical and the experimental approaches, computation offers a third alternative to solve complex problems that require processing large amounts of data, or…
Computer-Aided Engineering Tools | Water Power | NREL
energy converters that will provide a full range of simulation capabilities for single devices and arrays simulation of water power technologies on high-performance computers enables the study of complex systems and experimentation. Such simulation is critical to accelerate progress in energy programs within the U.S. Department
NASA Technical Reports Server (NTRS)
Boyce, Lola; Bast, Callie C.
1992-01-01
The research included ongoing development of methodology that provides probabilistic lifetime strength of aerospace materials via computational simulation. A probabilistic material strength degradation model, in the form of a randomized multifactor interaction equation, is postulated for strength degradation of structural components of aerospace propulsion systems subjected to a number of effects or primative variables. These primative variable may include high temperature, fatigue or creep. In most cases, strength is reduced as a result of the action of a variable. This multifactor interaction strength degradation equation has been randomized and is included in the computer program, PROMISS. Also included in the research is the development of methodology to calibrate the above described constitutive equation using actual experimental materials data together with linear regression of that data, thereby predicting values for the empirical material constraints for each effect or primative variable. This regression methodology is included in the computer program, PROMISC. Actual experimental materials data were obtained from the open literature for materials typically of interest to those studying aerospace propulsion system components. Material data for Inconel 718 was analyzed using the developed methodology.
NASA Technical Reports Server (NTRS)
Smith, Tamara A.; Pavli, Albert J.; Kacynski, Kenneth J.
1987-01-01
The joint Army. Navy, NASA. Air Force (JANNAF) rocket engine peformnace prediction procedure is based on the use of various reference computer programs. One of the reference programs for nozzle analysis is the Two-Dimensional Kinetics (TDK) Program. The purpose of this report is to calibrate the JANNAF procedure incorporated into the December l984 version of the TDK program for the high-area-ratio rocket engine regime. The calibration was accomplished by modeling the performance of a 1030:1 rocket nozzle tested at NASA Lewis Research Center. A detailed description of the experimental test conditions and TDK input parameters is given. The results show that the computer code predicts delivered vacuum specific impulse to within 0.12 to 1.9 percent of the experimental data. Vacuum thrust coefficient predictions were within + or - 1.3 percent of experimental results. Predictions of wall static pressure were within approximately + or - 5 percent of the measured values. An experimental value for inviscid thrust was obtained for the nozzle extension between area ratios of 427.5 and 1030 by using an integration of the measured wall static pressures. Subtracting the measured thrust gain produced by the nozzle between area ratios of 427.5 and 1030 from the inviscid thrust gain yielded experimental drag decrements of 10.85 and 27.00 N (2.44 and 6.07 lb) for mixture ratios of 3.04 and 4.29, respectively. These values correspond to 0.45 and 1.11 percent of the total vacuum thrust. At a mixture ratio of 4.29, the TDK predicted drag decrement was 16.59 N (3.73 lb), or 0.71 percent of the predicted total vacuum thrust.
Computer-aided roll pass design in rolling of airfoil shapes
NASA Technical Reports Server (NTRS)
Akgerman, N.; Lahoti, G. D.; Altan, T.
1980-01-01
This paper describes two computer-aided design (CAD) programs developed for modeling the shape rolling process for airfoil sections. The first program, SHPROL, uses a modular upper-bound method of analysis and predicts the lateral spread, elongation, and roll torque. The second program, ROLPAS, predicts the stresses, roll separating force, the roll torque and the details of metal flow by simulating the rolling process, using the slab method of analysis. ROLPAS is an interactive program; it offers graphic display capabilities and allows the user to interact with the computer via a keyboard, CRT, and a light pen. The accuracy of the computerized models was evaluated by (a) rolling a selected airfoil shape at room temperature from 1018 steel and isothermally at high temperature from Ti-6Al-4V, and (b) comparing the experimental results with computer predictions. The comparisons indicated that the CAD systems, described here, are useful for practical engineering purposes and can be utilized in roll pass design and analysis for airfoil and similar shapes.
Computer programs for eddy-current defect studies
DOE Office of Scientific and Technical Information (OSTI.GOV)
Pate, J. R.; Dodd, C. V.
Several computer programs to aid in the design of eddy-current tests and probes have been written. The programs, written in Fortran, deal in various ways with the response to defects exhibited by four types of probes: the pancake probe, the reflection probe, the circumferential boreside probe, and the circumferential encircling probe. Programs are included which calculate the impedance or voltage change in a coil due to a defect, which calculate and plot the defect sensitivity factor of a coil, and which invert calculated or experimental readings to obtain the size of a defect. The theory upon which the programs aremore » based is the Burrows point defect theory, and thus the calculations of the programs will be more accurate for small defects. 6 refs., 21 figs.« less
Comparison of Novice Programmers' Performances: Blended versus Face-to-Face
ERIC Educational Resources Information Center
Cakiroglu, Unal
2012-01-01
This study investigated the effect of blended learning on novices' understandings of the introductory programming. A quasi-experimental design with participants of preservice computer and instructional technologies teachers, one control group (CG, N =64) and one experimental group (EG, N=61) who received the course 11 weeks. While face-to-face…
User's manual: Subsonic/supersonic advanced panel pilot code
NASA Technical Reports Server (NTRS)
Moran, J.; Tinoco, E. N.; Johnson, F. T.
1978-01-01
Sufficient instructions for running the subsonic/supersonic advanced panel pilot code were developed. This software was developed as a vehicle for numerical experimentation and it should not be construed to represent a finished production program. The pilot code is based on a higher order panel method using linearly varying source and quadratically varying doublet distributions for computing both linearized supersonic and subsonic flow over arbitrary wings and bodies. This user's manual contains complete input and output descriptions. A brief description of the method is given as well as practical instructions for proper configurations modeling. Computed results are also included to demonstrate some of the capabilities of the pilot code. The computer program is written in FORTRAN IV for the SCOPE 3.4.4 operations system of the Ames CDC 7600 computer. The program uses overlay structure and thirteen disk files, and it requires approximately 132000 (Octal) central memory words.
Experimental and analytical investigation of active loads control for aircraft landing gear
NASA Technical Reports Server (NTRS)
Morris, D. L.; Mcgehee, J. R.
1983-01-01
A series hydraulic, active loads control main landing gear from a light, twin-engine civil aircraft was investigated. Tests included landing impact and traversal of simulated runway roughness. It is shown that the active gear is feasible and very effective in reducing the force transmitted to the airframe. Preliminary validation of a multidegree of freedom active gear flexible airframe takeoff and landing analysis computer program, which may be used as a design tool for active gear systems, is accomplished by comparing experimental and computed data for the passive and active gears.
Numerical simulation of NQR/NMR: Applications in quantum computing.
Possa, Denimar; Gaudio, Anderson C; Freitas, Jair C C
2011-04-01
A numerical simulation program able to simulate nuclear quadrupole resonance (NQR) as well as nuclear magnetic resonance (NMR) experiments is presented, written using the Mathematica package, aiming especially applications in quantum computing. The program makes use of the interaction picture to compute the effect of the relevant nuclear spin interactions, without any assumption about the relative size of each interaction. This makes the program flexible and versatile, being useful in a wide range of experimental situations, going from NQR (at zero or under small applied magnetic field) to high-field NMR experiments. Some conditions specifically required for quantum computing applications are implemented in the program, such as the possibility of use of elliptically polarized radiofrequency and the inclusion of first- and second-order terms in the average Hamiltonian expansion. A number of examples dealing with simple NQR and quadrupole-perturbed NMR experiments are presented, along with the proposal of experiments to create quantum pseudopure states and logic gates using NQR. The program and the various application examples are freely available through the link http://www.profanderson.net/files/nmr_nqr.php. Copyright © 2011 Elsevier Inc. All rights reserved.
NASA Technical Reports Server (NTRS)
Sharma, Naveen
1992-01-01
In this paper we briefly describe a combined symbolic and numeric approach for solving mathematical models on parallel computers. An experimental software system, PIER, is being developed in Common Lisp to synthesize computationally intensive and domain formulation dependent phases of finite element analysis (FEA) solution methods. Quantities for domain formulation like shape functions, element stiffness matrices, etc., are automatically derived using symbolic mathematical computations. The problem specific information and derived formulae are then used to generate (parallel) numerical code for FEA solution steps. A constructive approach to specify a numerical program design is taken. The code generator compiles application oriented input specifications into (parallel) FORTRAN77 routines with the help of built-in knowledge of the particular problem, numerical solution methods and the target computer.
Experimental Evaluation and Workload Characterization for High-Performance Computer Architectures
NASA Technical Reports Server (NTRS)
El-Ghazawi, Tarek A.
1995-01-01
This research is conducted in the context of the Joint NSF/NASA Initiative on Evaluation (JNNIE). JNNIE is an inter-agency research program that goes beyond typical.bencbking to provide and in-depth evaluations and understanding of the factors that limit the scalability of high-performance computing systems. Many NSF and NASA centers have participated in the effort. Our research effort was an integral part of implementing JNNIE in the NASA ESS grand challenge applications context. Our research work under this program was composed of three distinct, but related activities. They include the evaluation of NASA ESS high- performance computing testbeds using the wavelet decomposition application; evaluation of NASA ESS testbeds using astrophysical simulation applications; and developing an experimental model for workload characterization for understanding workload requirements. In this report, we provide a summary of findings that covers all three parts, a list of the publications that resulted from this effort, and three appendices with the details of each of the studies using a key publication developed under the respective work.
Graphical User Interface Development for Representing Air Flow Patterns
NASA Technical Reports Server (NTRS)
Chaudhary, Nilika
2004-01-01
In the Turbine Branch, scientists carry out experimental and computational work to advance the efficiency and diminish the noise production of jet engine turbines. One way to do this is by decreasing the heat that the turbine blades receive. Most of the experimental work is carried out by taking a single turbine blade and analyzing the air flow patterns around it, because this data indicates the sections of the turbine blade that are getting too hot. Since the cost of doing turbine blade air flow experiments is very high, researchers try to do computational work that fits the experimental data. The goal of computational fluid dynamics is for scientists to find a numerical way to predict the complex flow patterns around different turbine blades without physically having to perform tests or costly experiments. When visualizing flow patterns, scientists need a way to represent the flow conditions around a turbine blade. A researcher will assign specific zones that surround the turbine blade. In a two-dimensional view, the zones are usually quadrilaterals. The next step is to assign boundary conditions which define how the flow enters or exits one side of a zone. way of setting up computational zones and grids, visualizing flow patterns, and storing all the flow conditions in a file on the computer for future computation. Such a program is necessary because the only method for creating flow pattern graphs is by hand, which is tedious and time-consuming. By using a computer program to create the zones and grids, the graph would be faster to make and easier to edit. Basically, the user would run a program that is an editable graph. The user could click and drag with the mouse to form various zones and grids, then edit the locations of these grids, add flow and boundary conditions, and finally save the graph for future use and analysis. My goal this summer is to create a graphical user interface (GUI) that incorporates all of these elements. I am writing the program in Java, a language that is portable among platforms, because it can run on different operating systems such as Windows and Unix without having to be rewritten. I had no prior experience of programming in Java at the start of my internship; I am continuously learning as I create the program. I have written the part of the program that enables a user to draw several zones, edit them, and store their locations. The next phase of my project is to allow the user to click on the side of a zone and create a boundary condition for it. A previous intern wrote a program that allows the user to input boundary conditions. I can integrate the two programs to create a larger, more usable program. After that, I will develop a way for the user to save the graph for future reference. Another eventual goal is to make the GUI capable of creating three-dimensional zones as well. Researchers such as my mentor, Dr. David Ashpis, need a quick, user-friendly
Analysis and calculation of lightning-induced voltages in aircraft electrical circuits
NASA Technical Reports Server (NTRS)
Plumer, J. A.
1974-01-01
Techniques to calculate the transfer functions relating lightning-induced voltages in aircraft electrical circuits to aircraft physical characteristics and lightning current parameters are discussed. The analytical work was carried out concurrently with an experimental program of measurements of lightning-induced voltages in the electrical circuits of an F89-J aircraft. A computer program, ETCAL, developed earlier to calculate resistive and inductive transfer functions is refined to account for skin effect, providing results more valid over a wider range of lightning waveshapes than formerly possible. A computer program, WING, is derived to calculate the resistive and inductive transfer functions between a basic aircraft wing and a circuit conductor inside it. Good agreement is obtained between transfer inductances calculated by WING and those reduced from measured data by ETCAL. This computer program shows promise of expansion to permit eventual calculation of potential lightning-induced voltages in electrical circuits of complete aircraft in the design stage.
Computer program for plotting and fairing wind-tunnel data
NASA Technical Reports Server (NTRS)
Morgan, H. L., Jr.
1983-01-01
A detailed description of the Langley computer program PLOTWD which plots and fairs experimental wind-tunnel data is presented. The program was written for use primarily on the Langley CDC computer and CALCOMP plotters. The fundamental operating features of the program are that the input data are read and written to a random-access file for use during program execution, that the data for a selected run can be sorted and edited to delete duplicate points, and that the data can be plotted and faired using tension splines, least-squares polynomial, or least-squares cubic-spline curves. The most noteworthy feature of the program is the simplicity of the user-supplied input requirements. Several subroutines are also included that can be used to draw grid lines, zero lines, axis scale values and lables, and legends. A detailed description of the program operational features and each sub-program are presented. The general application of the program is also discussed together with the input and output for two typical plot types. A listing of the program code, user-guide, and output description are presented in appendices. The program has been in use at Langley for several years and has proven to be both easy to use and versatile.
X-38 Experimental Aeroheating at Mach 10
NASA Technical Reports Server (NTRS)
Berry, Scott A.; Horvath, Thomas J.; Weilmuenster, K. James; Alter, Stephan J.; Merski, N. Ronald
2001-01-01
This report provides an update of the hypersonic aerothermodynamic wind tunnel test program conducted at the NASA Langley Research Center in support of the X-38 program. Global surface heat transfer distributions were measured on 0.0177 and 0.0236 scale models of the proposed X-38 configuration at Mach 10 in air. The parametrics that were investigated primarily include freestream unit Reynolds numbers of 0.6 to 2.2 million per foot and body flap deflections of 15, 20, and 25 deg for an angle-of-attack of 40 deg. The model-scale variance was tested to obtain laminar, transitional, and turbulent heating levels on the defected bodyflaps. In addition, a limited investigation of forced boundary layer transition through the use of discrete roughness elements was performed. Comparisons of the present experimental results to computational predictions and previous experimental data were conducted Laminar, transitional, and turbulent heating levels were observed on the deflected body flap, which compared favorably to the computational results and to the predicted heating based on the flight aerothermodynamic database.
Experimental Stage Separation Tool Development in NASA Langley's Aerothermodynamics Laboratory
NASA Technical Reports Server (NTRS)
Murphy, Kelly J.; Scallion, William I.
2005-01-01
As part of the research effort at NASA in support of the stage separation and ascent aerothermodynamics research program, proximity testing of a generic bimese wing-body configuration was conducted in NASA Langley's Aerothermodynamics Laboratory in the 20-Inch Mach 6 Air Tunnel. The objective of this work is the development of experimental tools and testing methodologies to apply to hypersonic stage separation problems for future multi-stage launch vehicle systems. Aerodynamic force and moment proximity data were generated at a nominal Mach number of 6 over a small range of angles of attack. The generic bimese configuration was tested in a belly-to-belly and back-to-belly orientation at 86 relative proximity locations. Over 800 aerodynamic proximity data points were taken to serve as a database for code validation. Longitudinal aerodynamic data generated in this test program show very good agreement with viscous computational predictions. Thus a framework has been established to study separation problems in the hypersonic regime using coordinated experimental and computational tools.
NASA Technical Reports Server (NTRS)
Chen, Shu-cheng S.
2011-01-01
The axial flow turbine off-design computer program AXOD has been upgraded to include the outlet guide vane (OGV) into its acceptable turbine configurations. The mathematical bases and the techniques used for the code implementation are described and discussed in lengths in this paper. This extended capability is verified and validated with two cases of highly loaded fan-drive turbines, designed and tested in the V/STOL Program of NASA. The first case is a 4 1/2-stage turbine with an average stage loading factor of 4.66, designed by Pratt & Whitney Aircraft. The second case is a 3 1/2-stage turbine with an average loading factor of 4.0, designed in-house by the NASA Lewis Research Center (now the NASA Glenn Research Center). Both cases were experimentally tested in the turbine facility located at the Glenn Research Center. The processes conducted in these studies are described in detail in this paper, and the results in comparison with the experimental data are presented and discussed. The comparisons between the AXOD results and the experimental data are in excellent agreement.
X-34 Experimental Aeroheating at Mach 6 and 10
NASA Technical Reports Server (NTRS)
Berry, Scott A.; Horvath, Thomas J.; DiFulvio, Michael; Glass, Christopher; Merski, N. Ronald
1998-01-01
Critical technologies are being developed to support the goals of the NASA Office of Aeronautics and Space Transportation Technology Access to Space initiative for next-generation reusable space transportation systems. From the perspective of aerothermodynamic performance throughout the flight trajectory, the Reusable Launch Vehicle program incorporates conceptual analysis, ground-based testing, and computational fluid dynamics to provide flyable suborbital flight demonstrator vehicles. This report provides an overview of the hypersonic aeroheating wind tunnel test program conducted at the NASA Langley Research Center in support of one of these vehicles, the X-34 small reusable technology demonstrator program. Global surface heat transfer images, surface streamline patterns, and shock shapes were measured on 0.0153- and 0.0183-scale models of proposed X-34 flight vehicles at Mach 6 and 10 in air. The primary parametrics that were investigated include angles-of-attack from 0 to 35 deg. and freestream unit Reynolds numbers from 0.5 to 8 million per foot (which was sufficient to produce laminar, transitional, and turbulent heating data), both with and without control surface deflections. Comparisons of the experimental data to computational predictions are included, along with a discussion of the implications of some of the experimental flow features for the flight vehicle.
Image Segmentation via Fractal Dimension
1987-12-01
Contipany, 1986. 7. Harrington, Steven. Computer Graphics A Programming Aproach. New York: McGrawU-Hill Book Company, 1987. 8. Norev, Moshe. Picture...ITEX TORHS conversion program . My fazily has been a constant source of relaxation and enoouragement, and I wish to thank them for their understanding...Conclusion 29.... ....................... 9........... . -11 III. Experimental Method 3..-.1..................... .. ....... -1 Overviev
TT : a program that implements predictor sort design and analysis
S. P. Verrill; D. W. Green; V. L. Herian
1997-01-01
In studies on wood strength, researchers sometimes replace experimental unit allocation via random sampling with allocation via sorts based on nondestructive measurements of strength predictors such as modulus of elasticity and specific gravity. This report documents TT, a computer program that implements recently published methods to increase the sensitivity of such...
An Experimental Comparison of an Intrinsically Programed Text and a Narrative Text.
ERIC Educational Resources Information Center
Senter, R. J.; And Others
The study compared three methods of instruction in binary and octal arithmetic, i.e., (1) Norman Crowder's branched programed text, "The Arithmetic of Computers," (2) another version of this text modified so that subjects could not see the instructional material while answering "branching" questions, and (3) a narrative text…
Pattern of Plagiarism in Novice Students' Generated Programs: An Experimental Approach
ERIC Educational Resources Information Center
Ahmadzadeh, Marzieh; Mahmoudabadi, Elham; Khodadadi, Farzad
2011-01-01
Anecdotal evidence shows that in computer programming courses plagiarism is a widespread problem. With the growing number of students in such courses, manual plagiarism detection is impractical. This requires instructors to use one of the many available plagiarism detection tools. Prior to choosing one of such tools, a metric that assures the…
New technology in turbine aerodynamics
NASA Technical Reports Server (NTRS)
Glassman, A. J.; Moffitt, T. P.
1972-01-01
A cursory review is presented of some of the recent work that has been done in turbine aerodynamic research at NASA-Lewis Research Center. Topics discussed include the aerodynamic effect of turbine coolant, high work-factor (ratio of stage work to square of blade speed) turbines, and computer methods for turbine design and performance prediction. An extensive bibliography is included. Experimental cooled-turbine aerodynamics programs using two-dimensional cascades, full annular cascades, and cold rotating turbine stage tests are discussed with some typical results presented. Analytically predicted results for cooled blade performance are compared to experimental results. The problems and some of the current programs associated with the use of very high work factors for fan-drive turbines of high-bypass-ratio engines are discussed. Turbines currently being investigated make use of advanced blading concepts designed to maintain high efficiency under conditions of high aerodynamic loading. Computer programs have been developed for turbine design-point performance, off-design performance, supersonic blade profile design, and the calculation of channel velocities for subsonic and transonic flow fields. The use of these programs for the design and analysis of axial and radial turbines is discussed.
Persistence of Learning Gains from Computer Assisted Learning: Experimental Evidence from China
ERIC Educational Resources Information Center
Mo, D.; Zhang, L.; Wang, J.; Huang, W.; Shi, Y.; Boswell, M.; Rozelle, S.
2015-01-01
Computer assisted learning (CAL) programs have been shown to be effective in improving educational outcomes. However, the existing studies on CAL have almost all been conducted over a short period of time. There is very little evidence on how the impact evolves over time. In response, we conducted a clustered randomized experiment involving 2741…
NASA Technical Reports Server (NTRS)
Cebeci, T.; Kaups, K.; Ramsey, J. A.
1977-01-01
The method described utilizes a nonorthogonal coordinate system for boundary-layer calculations. It includes a geometry program that represents the wing analytically, and a velocity program that computes the external velocity components from a given experimental pressure distribution when the external velocity distribution is not computed theoretically. The boundary layer method is general, however, and can also be used for an external velocity distribution computed theoretically. Several test cases were computed by this method and the results were checked with other numerical calculations and with experiments when available. A typical computation time (CPU) on an IBM 370/165 computer for one surface of a wing which roughly consist of 30 spanwise stations and 25 streamwise stations, with 30 points across the boundary layer is less than 30 seconds for an incompressible flow and a little more for a compressible flow.
ARACHNE: A neural-neuroglial network builder with remotely controlled parallel computing
Rusakov, Dmitri A.; Savtchenko, Leonid P.
2017-01-01
Creating and running realistic models of neural networks has hitherto been a task for computing professionals rather than experimental neuroscientists. This is mainly because such networks usually engage substantial computational resources, the handling of which requires specific programing skills. Here we put forward a newly developed simulation environment ARACHNE: it enables an investigator to build and explore cellular networks of arbitrary biophysical and architectural complexity using the logic of NEURON and a simple interface on a local computer or a mobile device. The interface can control, through the internet, an optimized computational kernel installed on a remote computer cluster. ARACHNE can combine neuronal (wired) and astroglial (extracellular volume-transmission driven) network types and adopt realistic cell models from the NEURON library. The program and documentation (current version) are available at GitHub repository https://github.com/LeonidSavtchenko/Arachne under the MIT License (MIT). PMID:28362877
Review of NASA antiskid braking research
NASA Technical Reports Server (NTRS)
Tanner, J. A.
1982-01-01
NASA antiskid braking system research programs are reviewed. These programs include experimental studies of four antiskid systems on the Langley Landing Loads Track, flights tests with a DC-9 airplane, and computer simulation studies. Results from these research efforts include identification of factors contributing to degraded antiskid performance under adverse weather conditions, tire tread temperature measurements during antiskid braking on dry runway surfaces, and an assessment of the accuracy of various brake pressure-torque computer models. This information should lead to the development of better antiskid systems in the future.
NASA Technical Reports Server (NTRS)
Austin, F.; Markowitz, J.; Goldenberg, S.; Zetkov, G. A.
1973-01-01
The formulation of a mathematical model for predicting the dynamic behavior of rotating flexible space station configurations was conducted. The overall objectives of the study were: (1) to develop the theoretical techniques for determining the behavior of a realistically modeled rotating space station, (2) to provide a versatile computer program for the numerical analysis, and (3) to present practical concepts for experimental verification of the analytical results. The mathematical model and its associated computer program are described.
NASA Aeronautics: Research and Technology Program Highlights
NASA Technical Reports Server (NTRS)
1990-01-01
This report contains numerous color illustrations to describe the NASA programs in aeronautics. The basic ideas involved are explained in brief paragraphs. The seven chapters deal with Subsonic aircraft, High-speed transport, High-performance military aircraft, Hypersonic/Transatmospheric vehicles, Critical disciplines, National facilities and Organizations & installations. Some individual aircraft discussed are : the SR-71 aircraft, aerospace planes, the high-speed civil transport (HSCT), the X-29 forward-swept wing research aircraft, and the X-31 aircraft. Critical disciplines discussed are numerical aerodynamic simulation, computational fluid dynamics, computational structural dynamics and new experimental testing techniques.
An Overview of the Semi-Span Super-Sonic Transport (S4T) Wind-Tunnel Model Program
NASA Technical Reports Server (NTRS)
Silva, Walter A.; Perry, Boyd, III; Florance, James R.; Sanetrik, Mark D.; Wieseman, Carol D.; Stevens, William L.; Funk, Christie J.; Christhilf, David M.; Coulson, David A.
2012-01-01
A summary of computational and experimental aeroelastic (AE) and aeroservoelastic (ASE) results for the Semi-Span Super-Sonic Transport (S4T) wind-tunnel model is presented. A broad range of analyses and multiple AE and ASE wind-tunnel tests of the S4T wind-tunnel model have been performed in support of the ASE element in the Supersonics Program, part of the NASA Fundamental Aeronautics Program. This paper is intended to be an overview of multiple papers that comprise a special S4T technical session. Along those lines, a brief description of the design and hardware of the S4T wind-tunnel model will be presented. Computational results presented include linear and nonlinear aeroelastic analyses, and rapid aeroelastic analyses using CFD-based reduced-order models (ROMs). A brief survey of some of the experimental results from two open-loop and two closed-loop wind-tunnel tests performed at the NASA Langley Transonic Dynamics Tunnel (TDT) will be presented as well.
Comparison of LEWICE and GlennICE in the SLD Regime
NASA Technical Reports Server (NTRS)
Wright, William B.; Potapczuk, Mark G.; Levinson, Laurie H.
2008-01-01
A research project is underway at the NASA Glenn Research Center (GRC) to produce computer software that can accurately predict ice growth under any meteorological conditions for any aircraft surface. This report will present results from two different computer programs. The first program, LEWICE version 3.2.2, has been reported on previously. The second program is GlennICE version 0.1. An extensive comparison of the results in a quantifiable manner against the database of ice shapes that have been generated in the GRC Icing Research Tunnel (IRT) has also been performed, including additional data taken to extend the database in the Super-cooled Large Drop (SLD) regime. This paper will show the differences in ice shape between LEWICE 3.2.2, GlennICE, and experimental data. This report will also provide a description of both programs. Comparisons are then made to recent additions to the SLD database and selected previous cases. Quantitative comparisons are shown for horn height, horn angle, icing limit, area, and leading edge thickness. The results show that the predicted results for both programs are within the accuracy limits of the experimental data for the majority of cases.
Optical Activity of Benzil Crystal
NASA Astrophysics Data System (ADS)
Říha, Jan; Vyšín, Ivo
2003-09-01
Optical activity of benzil as an example of optically active matter in the crystalline state only, not in solution, is studied for the wavelengths ranging from 0.320 m to 0.585 m. Previously measured experimental data are approximated by the theoretical set of formulas, which were derived by the use of the three coupled oscillators model. The earlier published formula consisting of six terms differed from the experimental data particularly in the wavelength region (0.380-0.510) m. This formula is replaced by the twelve-term formula which was computed by our specially worked computer program for the interpretation of the experimental data of optical activity based on the Marquardt-Levenberg method of the sum of least squares minimization. The possibility of molecular contribution to the resulting optical activity of benzil is mentioned. The use of Kramers-Kronig transforms for the determination of the circular dichroism curve based on the optical rotatory dispersion result is shown. The theoretically computed circular dichroism is compared with the available experimental data.
Generalized three-dimensional experimental lightning code (G3DXL) user's manual
NASA Technical Reports Server (NTRS)
Kunz, Karl S.
1986-01-01
Information concerning the programming, maintenance and operation of the G3DXL computer program is presented and the theoretical basis for the code is described. The program computes time domain scattering fields and surface currents and charges induced by a driving function on and within a complex scattering object which may be perfectly conducting or a lossy dielectric. This is accomplished by modeling the object with cells within a three-dimensional, rectangular problem space, enforcing the appropriate boundary conditions and differencing Maxwell's equations in time. In the present version of the program, the driving function can be either the field radiated by a lightning strike or a direct lightning strike. The F-106 B aircraft is used as an example scattering object.
ERIC Educational Resources Information Center
School Science Review, 1985
1985-01-01
Presents 23 experiments, activities, field projects and computer programs in the biological and physical sciences. Instructional procedures, experimental designs, materials, and background information are suggested. Topics include fluid mechanics, electricity, crystals, arthropods, limpets, acid neutralization, and software evaluation. (ML)
ERIC Educational Resources Information Center
School Science Review, 1983
1983-01-01
Presents background information, laboratory procedures, classroom materials/activities, and chemistry experiments. Topics include sublimation, electronegativity, electrolysis, experimental aspects of strontianite, halide test, evaluation of present and future computer programs in chemistry, formula building, care of glass/saturated calomel…
Survey of NASA research on crash dynamics
NASA Technical Reports Server (NTRS)
Thomson, R. G.; Carden, H. D.; Hayduk, R. J.
1984-01-01
Ten years of structural crash dynamics research activities conducted on general aviation aircraft by the National Aeronautics and Space Administration (NASA) are described. Thirty-two full-scale crash tests were performed at Langley Research Center, and pertinent data on airframe and seat behavior were obtained. Concurrent with the experimental program, analytical methods were developed to help predict structural behavior during impact. The effects of flight parameters at impact on cabin deceleration pulses at the seat/occupant interface, experimental and analytical correlation of data on load-limiting subfloor and seat configurations, airplane section test results for computer modeling validation, and data from emergency-locator-transmitter (ELT) investigations to determine probable cause of false alarms and nonactivations are assessed. Computer programs which provide designers with analytical methods for predicting accelerations, velocities, and displacements of collapsing structures are also discussed.
WTO — a deterministic approach to 4-fermion physics
NASA Astrophysics Data System (ADS)
Passarino, Giampiero
1996-09-01
The program WTO, which is designed for computing cross sections and other relevant observables in the e+e- annihilation into four fermions, is described. The various quantities are computed over both a completely inclusive experimental set-up and a realistic one, i.e. with cuts on the final state energies, final state angles, scattering angles and final state invariant masses. Initial state QED corrections are included by means of the structure function approach while final state QCD corrections are applicable in their naive formulation. A gauge restoring mechanism is included according to the Fermion-Loop scheme. The program structure is highly modular and particular care has been devoted to computing efficiency and speed.
Three-dimensional transonic potential flow about complex 3-dimensional configurations
NASA Technical Reports Server (NTRS)
Reyhner, T. A.
1984-01-01
An analysis has been developed and a computer code written to predict three-dimensional subsonic or transonic potential flow fields about lifting or nonlifting configurations. Possible condfigurations include inlets, nacelles, nacelles with ground planes, S-ducts, turboprop nacelles, wings, and wing-pylon-nacelle combinations. The solution of the full partial differential equation for compressible potential flow written in terms of a velocity potential is obtained using finite differences, line relaxation, and multigrid. The analysis uses either a cylindrical or Cartesian coordinate system. The computational mesh is not body fitted. The analysis has been programmed in FORTRAN for both the CDC CYBER 203 and the CRAY-1 computers. Comparisons of computed results with experimental measurement are presented. Descriptions of the program input and output formats are included.
NASA Astrophysics Data System (ADS)
Huppert, J.; Michal Lomask, S.; Lazarowitz, R.
2002-08-01
Computer-assisted learning, including simulated experiments, has great potential to address the problem solving process which is a complex activity. It requires a highly structured approach in order to understand the use of simulations as an instructional device. This study is based on a computer simulation program, 'The Growth Curve of Microorganisms', which required tenth grade biology students to use problem solving skills whilst simultaneously manipulating three independent variables in one simulated experiment. The aims were to investigate the computer simulation's impact on students' academic achievement and on their mastery of science process skills in relation to their cognitive stages. The results indicate that the concrete and transition operational students in the experimental group achieved significantly higher academic achievement than their counterparts in the control group. The higher the cognitive operational stage, the higher students' achievement was, except in the control group where students in the concrete and transition operational stages did not differ. Girls achieved equally with the boys in the experimental group. Students' academic achievement may indicate the potential impact a computer simulation program can have, enabling students with low reasoning abilities to cope successfully with learning concepts and principles in science which require high cognitive skills.
Fast ForWord: An Investigation of the Effectiveness of Computer-Assisted Reading Intervention
ERIC Educational Resources Information Center
Soboleski, Penny K.
2011-01-01
The three-fold purpose of this quasi-experimental study was to examine the impact of the computer-based reading program Fast ForWord (FFW) on the reading achievement of second-grade students in an Ohio school district. The sample included 360 students (treatment group, n = 85; control group, n = 275) from four elementary buildings. FFW is an…
ERIC Educational Resources Information Center
Lloret, Miguel; Aguila, Estela; Lloret, Alejandro
2009-01-01
The purpose of this study was to study the effect of a multimedia computing program on the production of activities and self-regulated learning processes in 18 students of the Dentistry postdegree (Celaya, Mexico). A multi-method design (quasi-experimental, pretest-post-test and qualitative: Think aloud protocol) was used. Self-regulated…
Learning Problem-Solving through Making Games at the Game Design and Learning Summer Program
ERIC Educational Resources Information Center
Akcaoglu, Mete
2014-01-01
Today's complex and fast-evolving world necessitates young students to possess design and problem-solving skills more than ever. One alternative method of teaching children problem-solving or thinking skills has been using computer programming, and more recently, game-design tasks. In this pre-experimental study, a group of middle school…
The CGE-PLATO Electronic Laboratory Instructional Programs. (August 1, 1972 Through June 30, 1975).
ERIC Educational Resources Information Center
Neal, J. P.
Twelve PLATO lessons are reproduced in this document to show the status of computer guided experimentation (CGE) instructional programs. The lesson topics include a description of the CGE-PLATO instructional laboratory, an introduction to CGE-PLATO tests and special software routines, router lesson for two electrical engineering courses, and an…
ERIC Educational Resources Information Center
Canelos, James; And Others
This study examined the effects of two cognitive styles--field dependents-independents and reflectivity-impulsivity--on learning from microcomputer-based instruction. In the first of three experimental designs, a programmed instruction text on the human heart was used which contained both visual and verbal information in an instructional display,…
Solid State Cooling with Advanced Oxide Materials
2014-06-03
Department of Materials Science and Engineering , Department of Mechanical Science and Engineering , and Department of Electrical and Computer... Engineering University of Illinois, Urbana-Champaign Program Overview The focus of this program was to probe electro-(magneto-)caloric materials for... engineering systems by developing theoretical and experimental approaches to study thermodynamic properties and effects in thin film systems. Despite
A microcomputer program for analysis of nucleic acid hybridization data
Green, S.; Field, J.K.; Green, C.D.; Beynon, R.J.
1982-01-01
The study of nucleic acid hybridization is facilitated by computer mediated fitting of theoretical models to experimental data. This paper describes a non-linear curve fitting program, using the `Patternsearch' algorithm, written in BASIC for the Apple II microcomputer. The advantages and disadvantages of using a microcomputer for local data processing are discussed. Images PMID:7071017
NASA Technical Reports Server (NTRS)
Blair, M. F.
1991-01-01
A combined experimental and computational program was conducted to examine the heat transfer distribution in a turbine rotor passage geometrically similar to the Space Shuttle Main Engine (SSME) High Pressure Fuel Turbopump (HPFTP). Heat transfer was measured and computed for both the full span suction and pressure surfaces of the rotor airfoil as well as for the hub endwall surface. The objective of the program was to provide a benchmark-quality database for the assessment of rotor heat transfer computational techniques. The experimental portion of the study was conducted in a large scale, ambient temperature, rotating turbine model. The computational portion consisted of the application of a well-posed parabolized Navier-Stokes analysis of the calculation of the three-dimensional viscous flow through ducts simulating a gas turbine package. The results of this assessment indicate that the procedure has the potential to predict the aerodynamics and the heat transfer in a gas turbine passage and can be used to develop detailed three dimensional turbulence models for the prediction of skin friction and heat transfer in complex three dimensional flow passages.
Experimental system for computer network via satellite /CS/. III - Network control processor
NASA Astrophysics Data System (ADS)
Kakinuma, Y.; Ito, A.; Takahashi, H.; Uchida, K.; Matsumoto, K.; Mitsudome, H.
1982-03-01
A network control processor (NCP) has the functions of generating traffics, the control of links and the control of transmitting bursts. The NCP executes protocols, monitors of experiments, gathering and compiling data of measurements, of which programs are loaded on a minicomputer (MELCOM 70/40) with 512KB of memories. The NCP acts as traffic generators, instead of a host computer, in the experiment. For this purpose, 15 fake stations are realized by the software in each user station. This paper describes the configuration of the NCP and the implementation of the protocols for the experimental system.
Maximum likelihood convolutional decoding (MCD) performance due to system losses
NASA Technical Reports Server (NTRS)
Webster, L.
1976-01-01
A model for predicting the computational performance of a maximum likelihood convolutional decoder (MCD) operating in a noisy carrier reference environment is described. This model is used to develop a subroutine that will be utilized by the Telemetry Analysis Program to compute the MCD bit error rate. When this computational model is averaged over noisy reference phase errors using a high-rate interpolation scheme, the results are found to agree quite favorably with experimental measurements.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Dickens, J.K.
1988-04-01
This document provides a discussion of the development of the FORTRAN Monte Carlo program SCINFUL (for scintillator full response), a program designed to provide a calculated full response anticipated for either an NE-213 (liquid) scintillator or an NE-110 (solid) scintillator. The program may also be used to compute angle-integrated spectra of charged particles (p, d, t, /sup 3/He, and ..cap alpha..) following neutron interactions with /sup 12/C. Extensive comparisons with a variety of experimental data are given. There is generally overall good agreement (<10% differences) of results from SCINFUL calculations with measured detector responses, i.e., N(E/sub r/) vs E/sub r/more » where E/sub r/ is the response pulse height, reproduce measured detector responses with an accuracy which, at least partly, depends upon how well the experimental configuration is known. For E/sub n/ < 16 MeV and for E/sub r/ > 15% of the maximum pulse height response, calculated spectra are within +-5% of experiment on the average. For E/sub n/ up to 50 MeV similar good agreement is obtained with experiment for E/sub r/ > 30% of maximum response. For E/sub n/ up to 75 MeV the calculated shape of the response agrees with measurements, but the calculations underpredicts the measured response by up to 30%. 65 refs., 64 figs., 3 tabs.« less
Tempest: GPU-CPU computing for high-throughput database spectral matching.
Milloy, Jeffrey A; Faherty, Brendan K; Gerber, Scott A
2012-07-06
Modern mass spectrometers are now capable of producing hundreds of thousands of tandem (MS/MS) spectra per experiment, making the translation of these fragmentation spectra into peptide matches a common bottleneck in proteomics research. When coupled with experimental designs that enrich for post-translational modifications such as phosphorylation and/or include isotopically labeled amino acids for quantification, additional burdens are placed on this computational infrastructure by shotgun sequencing. To address this issue, we have developed a new database searching program that utilizes the massively parallel compute capabilities of a graphical processing unit (GPU) to produce peptide spectral matches in a very high throughput fashion. Our program, named Tempest, combines efficient database digestion and MS/MS spectral indexing on a CPU with fast similarity scoring on a GPU. In our implementation, the entire similarity score, including the generation of full theoretical peptide candidate fragmentation spectra and its comparison to experimental spectra, is conducted on the GPU. Although Tempest uses the classical SEQUEST XCorr score as a primary metric for evaluating similarity for spectra collected at unit resolution, we have developed a new "Accelerated Score" for MS/MS spectra collected at high resolution that is based on a computationally inexpensive dot product but exhibits scoring accuracy similar to that of the classical XCorr. In our experience, Tempest provides compute-cluster level performance in an affordable desktop computer.
Software for Testing Electroactive Structural Components
NASA Technical Reports Server (NTRS)
Moses, Robert W.; Fox, Robert L.; Dimery, Archie D.; Bryant, Robert G.; Shams, Qamar
2003-01-01
A computer program generates a graphical user interface that, in combination with its other features, facilitates the acquisition and preprocessing of experimental data on the strain response, hysteresis, and power consumption of a multilayer composite-material structural component containing one or more built-in sensor(s) and/or actuator(s) based on piezoelectric materials. This program runs in conjunction with Lab-VIEW software in a computer-controlled instrumentation system. For a test, a specimen is instrumented with appliedvoltage and current sensors and with strain gauges. Once the computational connection to the test setup has been made via the LabVIEW software, this program causes the test instrumentation to step through specified configurations. If the user is satisfied with the test results as displayed by the software, the user activates an icon on a front-panel display, causing the raw current, voltage, and strain data to be digitized and saved. The data are also put into a spreadsheet and can be plotted on a graph. Graphical displays are saved in an image file for future reference. The program also computes and displays the power and the phase angle between voltage and current.
The Shock and Vibration Digest. Volume 7, Number 5, May 1975.
Contents: Dynamic response of fluid-filled shells; News briefs; Short courses; Abstracts from the current literature -- (Analysis and design, computer programs, environments, phenomenology, experimentation, components, systems); Author index ; Literature review; Book reviews.
Membrane proteins structures: A review on computational modeling tools.
Almeida, Jose G; Preto, Antonio J; Koukos, Panagiotis I; Bonvin, Alexandre M J J; Moreira, Irina S
2017-10-01
Membrane proteins (MPs) play diverse and important functions in living organisms. They constitute 20% to 30% of the known bacterial, archaean and eukaryotic organisms' genomes. In humans, their importance is emphasized as they represent 50% of all known drug targets. Nevertheless, experimental determination of their three-dimensional (3D) structure has proven to be both time consuming and rather expensive, which has led to the development of computational algorithms to complement the available experimental methods and provide valuable insights. This review highlights the importance of membrane proteins and how computational methods are capable of overcoming challenges associated with their experimental characterization. It covers various MP structural aspects, such as lipid interactions, allostery, and structure prediction, based on methods such as Molecular Dynamics (MD) and Machine-Learning (ML). Recent developments in algorithms, tools and hybrid approaches, together with the increase in both computational resources and the amount of available data have resulted in increasingly powerful and trustworthy approaches to model MPs. Even though MPs are elementary and important in nature, the determination of their 3D structure has proven to be a challenging endeavor. Computational methods provide a reliable alternative to experimental methods. In this review, we focus on computational techniques to determine the 3D structure of MP and characterize their binding interfaces. We also summarize the most relevant databases and software programs available for the study of MPs. Copyright © 2017 Elsevier B.V. All rights reserved.
Computational structural mechanics for engine structures
NASA Technical Reports Server (NTRS)
Chamis, Christos C.
1988-01-01
The computational structural mechanics (CSM) program at Lewis encompasses the formulation and solution of structural mechanics problems and the development of integrated software systems to computationally simulate the performance, durability, and life of engine structures. It is structured to supplement, complement, and, whenever possible, replace costly experimental efforts. Specific objectives are to investigate unique advantages of parallel and multiprocessing for reformulating and solving structural mechanics and formulating and solving multidisciplinary mechanics and to develop integrated structural system computational simulators for predicting structural performance, evaluating newly developed methods, and identifying and prioritizing improved or missing methods.
Computational structural mechanics for engine structures
NASA Technical Reports Server (NTRS)
Chamis, Christos C.
1989-01-01
The computational structural mechanics (CSM) program at Lewis encompasses the formulation and solution of structural mechanics problems and the development of integrated software systems to computationally simulate the performance, durability, and life of engine structures. It is structured to supplement, complement, and, whenever possible, replace costly experimental efforts. Specific objectives are to investigate unique advantages of parallel and multiprocessing for reformulating and solving structural mechanics and formulating and solving multidisciplinary mechanics and to develop integrated structural system computational simulators for predicting structural performance, evaluating newly developed methods, and identifying and prioritizing improved or missing methods.
NASA Technical Reports Server (NTRS)
1993-01-01
This report summarizes research conducted at the Institute for Computer Applications in Science and Engineering in applied mathematics and computer science during the period April 1, 1993 through September 30, 1993. The major categories of the current ICASE research program are: (1) applied and numerical mathematics, including numerical analysis and algorithm development; (2) theoretical and computational research in fluid mechanics in selected areas of interest to LaRC, including acoustic and combustion; (3) experimental research in transition and turbulence and aerodynamics involving LaRC facilities and scientists; and (4) computer science.
NASA Technical Reports Server (NTRS)
1994-01-01
This report summarizes research conducted at the Institute for Computer Applications in Science and Engineering in applied mathematics, fluid mechanics, and computer science during the period October 1, 1993 through March 31, 1994. The major categories of the current ICASE research program are: (1) applied and numerical mathematics, including numerical analysis and algorithm development; (2) theoretical and computational research in fluid mechanics in selected areas of interest to LaRC, including acoustics and combustion; (3) experimental research in transition and turbulence and aerodynamics involving LaRC facilities and scientists; and (4) computer science.
Cartreine, James Albert; Locke, Steven E; Buckey, Jay C; Sandoval, Luis; Hegel, Mark T
2012-09-25
Computer-automated depression interventions rely heavily on users reading text to receive the intervention. However, text-delivered interventions place a burden on persons with depression and convey only verbal content. The primary aim of this project was to develop a computer-automated treatment for depression that is delivered via interactive media technology. By using branching video and audio, the program simulates the experience of being in therapy with a master clinician who provides six sessions of problem-solving therapy. A secondary objective was to conduct a pilot study of the program's usability, acceptability, and credibility, and to obtain an initial estimate of its efficacy. The program was produced in a professional multimedia production facility and incorporates video, audio, graphics, animation, and text. Failure analyses of patient data are conducted across sessions and across problems to identify ways to help the user improve his or her problem solving. A pilot study was conducted with persons who had minor depression. An experimental group (n = 7) used the program while a waitlist control group (n = 7) was provided with no treatment for 6 weeks. All of the experimental group participants completed the trial, whereas 1 from the control was lost to follow-up. Experimental group participants rated the program high on usability, acceptability, and credibility. The study was not powered to detect clinical improvement, although these pilot data are encouraging. Although the study was not powered to detect treatment effects, participants did find the program highly usable, acceptable, and credible. This suggests that the highly interactive and immersive nature of the program is beneficial. Further clinical trials are warranted. ClinicalTrials.gov NCT00906581; http://clinicaltrials.gov/ct2/show/NCT00906581 (Archived by WebCite at http://www.webcitation.org/6A5Ni5HUp).
Geothermal reservoir engineering research
NASA Technical Reports Server (NTRS)
Ramey, H. J., Jr.; Kruger, P.; Brigham, W. E.; London, A. L.
1974-01-01
The Stanford University research program on the study of stimulation and reservoir engineering of geothermal resources commenced as an interdisciplinary program in September, 1972. The broad objectives of this program have been: (1) the development of experimental and computational data to evaluate the optimum performance of fracture-stimulated geothermal reservoirs; (2) the development of a geothermal reservoir model to evaluate important thermophysical, hydrodynamic, and chemical parameters based on fluid-energy-volume balances as part of standard reservoir engineering practice; and (3) the construction of a laboratory model of an explosion-produced chimney to obtain experimental data on the processes of in-place boiling, moving flash fronts, and two-phase flow in porous and fractured hydrothermal reservoirs.
The grand challenge of managing the petascale facility.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Aiken, R. J.; Mathematics and Computer Science
2007-02-28
This report is the result of a study of networks and how they may need to evolve to support petascale leadership computing and science. As Dr. Ray Orbach, director of the Department of Energy's Office of Science, says in the spring 2006 issue of SciDAC Review, 'One remarkable example of growth in unexpected directions has been in high-end computation'. In the same article Dr. Michael Strayer states, 'Moore's law suggests that before the end of the next cycle of SciDAC, we shall see petaflop computers'. Given the Office of Science's strong leadership and support for petascale computing and facilities, wemore » should expect to see petaflop computers in operation in support of science before the end of the decade, and DOE/SC Advanced Scientific Computing Research programs are focused on making this a reality. This study took its lead from this strong focus on petascale computing and the networks required to support such facilities, but it grew to include almost all aspects of the DOE/SC petascale computational and experimental science facilities, all of which will face daunting challenges in managing and analyzing the voluminous amounts of data expected. In addition, trends indicate the increased coupling of unique experimental facilities with computational facilities, along with the integration of multidisciplinary datasets and high-end computing with data-intensive computing; and we can expect these trends to continue at the petascale level and beyond. Coupled with recent technology trends, they clearly indicate the need for including capability petascale storage, networks, and experiments, as well as collaboration tools and programming environments, as integral components of the Office of Science's petascale capability metafacility. The objective of this report is to recommend a new cross-cutting program to support the management of petascale science and infrastructure. The appendices of the report document current and projected DOE computation facilities, science trends, and technology trends, whose combined impact can affect the manageability and stewardship of DOE's petascale facilities. This report is not meant to be all-inclusive. Rather, the facilities, science projects, and research topics presented are to be considered examples to clarify a point.« less
Modelling parallel programs and multiprocessor architectures with AXE
NASA Technical Reports Server (NTRS)
Yan, Jerry C.; Fineman, Charles E.
1991-01-01
AXE, An Experimental Environment for Parallel Systems, was designed to model and simulate for parallel systems at the process level. It provides an integrated environment for specifying computation models, multiprocessor architectures, data collection, and performance visualization. AXE is being used at NASA-Ames for developing resource management strategies, parallel problem formulation, multiprocessor architectures, and operating system issues related to the High Performance Computing and Communications Program. AXE's simple, structured user-interface enables the user to model parallel programs and machines precisely and efficiently. Its quick turn-around time keeps the user interested and productive. AXE models multicomputers. The user may easily modify various architectural parameters including the number of sites, connection topologies, and overhead for operating system activities. Parallel computations in AXE are represented as collections of autonomous computing objects known as players. Their use and behavior is described. Performance data of the multiprocessor model can be observed on a color screen. These include CPU and message routing bottlenecks, and the dynamic status of the software.
Fortran Program for X-Ray Photoelectron Spectroscopy Data Reformatting
NASA Technical Reports Server (NTRS)
Abel, Phillip B.
1989-01-01
A FORTRAN program has been written for use on an IBM PC/XT or AT or compatible microcomputer (personal computer, PC) that converts a column of ASCII-format numbers into a binary-format file suitable for interactive analysis on a Digital Equipment Corporation (DEC) computer running the VGS-5000 Enhanced Data Processing (EDP) software package. The incompatible floating-point number representations of the two computers were compared, and a subroutine was created to correctly store floating-point numbers on the IBM PC, which can be directly read by the DEC computer. Any file transfer protocol having provision for binary data can be used to transmit the resulting file from the PC to the DEC machine. The data file header required by the EDP programs for an x ray photoelectron spectrum is also written to the file. The user is prompted for the relevant experimental parameters, which are then properly coded into the format used internally by all of the VGS-5000 series EDP packages.
Indexed Retrieval System for Navy Experimental Diving Unit Research and Evaluation Reports.
KWIC computer programs developed by the International Business Machine Corporation (IBM) were so successful in this application that they are now being applied to all of NEDU’s microfilmed research files. (Author)
DOE Office of Scientific and Technical Information (OSTI.GOV)
None
1963-07-01
This second edition is based on data available on March 15, 1961. Sections on constants necessary for the interpretation of experimental data and on digital computer programs for reactor design and reactor physics have been added. 1344 references. (D.C.W.)
Multi-scale computation methods: Their applications in lithium-ion battery research and development
NASA Astrophysics Data System (ADS)
Siqi, Shi; Jian, Gao; Yue, Liu; Yan, Zhao; Qu, Wu; Wangwei, Ju; Chuying, Ouyang; Ruijuan, Xiao
2016-01-01
Based upon advances in theoretical algorithms, modeling and simulations, and computer technologies, the rational design of materials, cells, devices, and packs in the field of lithium-ion batteries is being realized incrementally and will at some point trigger a paradigm revolution by combining calculations and experiments linked by a big shared database, enabling accelerated development of the whole industrial chain. Theory and multi-scale modeling and simulation, as supplements to experimental efforts, can help greatly to close some of the current experimental and technological gaps, as well as predict path-independent properties and help to fundamentally understand path-independent performance in multiple spatial and temporal scales. Project supported by the National Natural Science Foundation of China (Grant Nos. 51372228 and 11234013), the National High Technology Research and Development Program of China (Grant No. 2015AA034201), and Shanghai Pujiang Program, China (Grant No. 14PJ1403900).
NASA Technical Reports Server (NTRS)
Storey, Jedediah M.; Kirk, Daniel; Gutierrez, Hector; Marsell, Brandon; Schallhorn, Paul; Lapilli, Gabriel D.
2015-01-01
Experimental and numerical results are presented from a new cryogenic fluid slosh program at the Florida Institute of Technology (FIT). Water and cryogenic liquid nitrogen are used in various ground-based tests with an approximately 30 cm diameter spherical tank to characterize damping, slosh mode frequencies, and slosh forces. The experimental results are compared to a computational fluid dynamics (CFD) model for validation. An analytical model is constructed from prior work for comparison. Good agreement is seen between experimental, numerical, and analytical results.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Carlsson, Mats; Johansson, Mikael; Larson, Jeffrey
Previous approaches for scheduling a league with round-robin and divisional tournaments involved decomposing the problem into easier subproblems. This approach, used to schedule the top Swedish handball league Elitserien, reduces the problem complexity but can result in suboptimal schedules. This paper presents an integrated constraint programming model that allows to perform the scheduling in a single step. Particular attention is given to identifying implied and symmetry-breaking constraints that reduce the computational complexity significantly. The experimental evaluation of the integrated approach takes considerably less computational effort than the previous approach.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Amadio, G.; et al.
An intensive R&D and programming effort is required to accomplish new challenges posed by future experimental high-energy particle physics (HEP) programs. The GeantV project aims to narrow the gap between the performance of the existing HEP detector simulation software and the ideal performance achievable, exploiting latest advances in computing technology. The project has developed a particle detector simulation prototype capable of transporting in parallel particles in complex geometries exploiting instruction level microparallelism (SIMD and SIMT), task-level parallelism (multithreading) and high-level parallelism (MPI), leveraging both the multi-core and the many-core opportunities. We present preliminary verification results concerning the electromagnetic (EM) physicsmore » models developed for parallel computing architectures within the GeantV project. In order to exploit the potential of vectorization and accelerators and to make the physics model effectively parallelizable, advanced sampling techniques have been implemented and tested. In this paper we introduce a set of automated statistical tests in order to verify the vectorized models by checking their consistency with the corresponding Geant4 models and to validate them against experimental data.« less
Programming Language Software For Graphics Applications
NASA Technical Reports Server (NTRS)
Beckman, Brian C.
1993-01-01
New approach reduces repetitive development of features common to different applications. High-level programming language and interactive environment with access to graphical hardware and software created by adding graphical commands and other constructs to standardized, general-purpose programming language, "Scheme". Designed for use in developing other software incorporating interactive computer-graphics capabilities into application programs. Provides alternative to programming entire applications in C or FORTRAN, specifically ameliorating design and implementation of complex control and data structures typifying applications with interactive graphics. Enables experimental programming and rapid development of prototype software, and yields high-level programs serving as executable versions of software-design documentation.
Computer Simulation Of Cyclic Oxidation
NASA Technical Reports Server (NTRS)
Probst, H. B.; Lowell, C. E.
1990-01-01
Computer model developed to simulate cyclic oxidation of metals. With relatively few input parameters, kinetics of cyclic oxidation simulated for wide variety of temperatures, durations of cycles, and total numbers of cycles. Program written in BASICA and run on any IBM-compatible microcomputer. Used in variety of ways to aid experimental research. In minutes, effects of duration of cycle and/or number of cycles on oxidation kinetics of material surveyed.
Marshal Wrubel and the Electronic Computer as an Astronomical Instrument
NASA Astrophysics Data System (ADS)
Mutschlecner, J. P.; Olsen, K. H.
1998-05-01
In 1960, Marshal H. Wrubel, professor of astrophysics at Indiana University, published an influential review paper under the title, "The Electronic Computer as an Astronomical Instrument." This essay pointed out the enormous potential of the electronic computer as an instrument of observational and theoretical research in astronomy, illustrated programming concepts, and made specific recommendations for the increased use of computers in astronomy. He noted that, with a few scattered exceptions, computer use by the astronomical community had heretofore been "timid and sporadic." This situation was to improve dramatically in the next few years. By the late 1950s, general-purpose, high-speed, "mainframe" computers were just emerging from the experimental, developmental stage, but few were affordable by or available to academic and research institutions not closely associated with large industrial or national defense programs. Yet by 1960 Wrubel had spent a decade actively pioneering and promoting the imaginative application of electronic computation within the astronomical community. Astronomy upper-level undergraduate and graduate students at Indiana were introduced to computing, and Ph.D. candidates who he supervised applied computer techniques to problems in theoretical astrophysics. He wrote an early textbook on programming, taught programming classes, and helped establish and direct the Research Computing Center at Indiana, later named the Wrubel Computing Center in his honor. He and his students created a variety of algorithms and subroutines and exchanged these throughout the astronomical community by distributing the Astronomical Computation News Letter. Nationally as well as internationally, Wrubel actively cooperated with other groups interested in computing applications for theoretical astrophysics, often through his position as secretary of the IAU commission on Stellar Constitution.
Markov Chains For Testing Redundant Software
NASA Technical Reports Server (NTRS)
White, Allan L.; Sjogren, Jon A.
1990-01-01
Preliminary design developed for validation experiment that addresses problems unique to assuring extremely high quality of multiple-version programs in process-control software. Approach takes into account inertia of controlled system in sense it takes more than one failure of control program to cause controlled system to fail. Verification procedure consists of two steps: experimentation (numerical simulation) and computation, with Markov model for each step.
Is a Web Survey as Effective as a Mail Survey? A Field Experiment Among Computer Users
ERIC Educational Resources Information Center
Kiernan, Nancy; Kiernan, Michaela; Oyler, Mary; Gilles, Carolyn
2005-01-01
With the exponential increase in Web access, program evaluators need to understand the methodological benefits and barriers of using the Web to collect survey data from program participants. In this experimental study, the authors examined whether a Web survey can be as effective as the more established mail survey on three measures of survey…
NASA Astrophysics Data System (ADS)
Johnston, Michael A.; Farrell, Damien; Nielsen, Jens Erik
2012-04-01
The exchange of information between experimentalists and theoreticians is crucial to improving the predictive ability of theoretical methods and hence our understanding of the related biology. However many barriers exist which prevent the flow of information between the two disciplines. Enabling effective collaboration requires that experimentalists can easily apply computational tools to their data, share their data with theoreticians, and that both the experimental data and computational results are accessible to the wider community. We present a prototype collaborative environment for developing and validating predictive tools for protein biophysical characteristics. The environment is built on two central components; a new python-based integration module which allows theoreticians to provide and manage remote access to their programs; and PEATDB, a program for storing and sharing experimental data from protein biophysical characterisation studies. We demonstrate our approach by integrating PEATSA, a web-based service for predicting changes in protein biophysical characteristics, into PEATDB. Furthermore, we illustrate how the resulting environment aids method development using the Potapov dataset of experimentally measured ΔΔGfold values, previously employed to validate and train protein stability prediction algorithms.
Designing, programming, and optimizing a (small) quantum computer
NASA Astrophysics Data System (ADS)
Svore, Krysta
In 1982, Richard Feynman proposed to use a computer founded on the laws of quantum physics to simulate physical systems. In the more than thirty years since, quantum computers have shown promise to solve problems in number theory, chemistry, and materials science that would otherwise take longer than the lifetime of the universe to solve on an exascale classical machine. The practical realization of a quantum computer requires understanding and manipulating subtle quantum states while experimentally controlling quantum interference. It also requires an end-to-end software architecture for programming, optimizing, and implementing a quantum algorithm on the quantum device hardware. In this talk, we will introduce recent advances in connecting abstract theory to present-day real-world applications through software. We will highlight recent advancement of quantum algorithms and the challenges in ultimately performing a scalable solution on a quantum device.
Probabilistic lifetime strength of aerospace materials via computational simulation
NASA Technical Reports Server (NTRS)
Boyce, Lola; Keating, Jerome P.; Lovelace, Thomas B.; Bast, Callie C.
1991-01-01
The results of a second year effort of a research program are presented. The research included development of methodology that provides probabilistic lifetime strength of aerospace materials via computational simulation. A probabilistic phenomenological constitutive relationship, in the form of a randomized multifactor interaction equation, is postulated for strength degradation of structural components of aerospace propulsion systems subjected to a number of effects of primitive variables. These primitive variables often originate in the environment and may include stress from loading, temperature, chemical, or radiation attack. This multifactor interaction constitutive equation is included in the computer program, PROMISS. Also included in the research is the development of methodology to calibrate the constitutive equation using actual experimental materials data together with the multiple linear regression of that data.
Concentrator optical characterization using computer mathematical modelling and point source testing
NASA Technical Reports Server (NTRS)
Dennison, E. W.; John, S. L.; Trentelman, G. F.
1984-01-01
The optical characteristics of a paraboloidal solar concentrator are analyzed using the intercept factor curve (a format for image data) to describe the results of a mathematical model and to represent reduced data from experimental testing. This procedure makes it possible not only to test an assembled concentrator, but also to evaluate single optical panels or to conduct non-solar tests of an assembled concentrator. The use of three-dimensional ray tracing computer programs to calculate the mathematical model is described. These ray tracing programs can include any type of optical configuration from simple paraboloids to array of spherical facets and can be adapted to microcomputers or larger computers, which can graphically display real-time comparison of calculated and measured data.
Shi, Yulin; Veidenbaum, Alexander V; Nicolau, Alex; Xu, Xiangmin
2015-01-15
Modern neuroscience research demands computing power. Neural circuit mapping studies such as those using laser scanning photostimulation (LSPS) produce large amounts of data and require intensive computation for post hoc processing and analysis. Here we report on the design and implementation of a cost-effective desktop computer system for accelerated experimental data processing with recent GPU computing technology. A new version of Matlab software with GPU enabled functions is used to develop programs that run on Nvidia GPUs to harness their parallel computing power. We evaluated both the central processing unit (CPU) and GPU-enabled computational performance of our system in benchmark testing and practical applications. The experimental results show that the GPU-CPU co-processing of simulated data and actual LSPS experimental data clearly outperformed the multi-core CPU with up to a 22× speedup, depending on computational tasks. Further, we present a comparison of numerical accuracy between GPU and CPU computation to verify the precision of GPU computation. In addition, we show how GPUs can be effectively adapted to improve the performance of commercial image processing software such as Adobe Photoshop. To our best knowledge, this is the first demonstration of GPU application in neural circuit mapping and electrophysiology-based data processing. Together, GPU enabled computation enhances our ability to process large-scale data sets derived from neural circuit mapping studies, allowing for increased processing speeds while retaining data precision. Copyright © 2014 Elsevier B.V. All rights reserved.
Shi, Yulin; Veidenbaum, Alexander V.; Nicolau, Alex; Xu, Xiangmin
2014-01-01
Background Modern neuroscience research demands computing power. Neural circuit mapping studies such as those using laser scanning photostimulation (LSPS) produce large amounts of data and require intensive computation for post-hoc processing and analysis. New Method Here we report on the design and implementation of a cost-effective desktop computer system for accelerated experimental data processing with recent GPU computing technology. A new version of Matlab software with GPU enabled functions is used to develop programs that run on Nvidia GPUs to harness their parallel computing power. Results We evaluated both the central processing unit (CPU) and GPU-enabled computational performance of our system in benchmark testing and practical applications. The experimental results show that the GPU-CPU co-processing of simulated data and actual LSPS experimental data clearly outperformed the multi-core CPU with up to a 22x speedup, depending on computational tasks. Further, we present a comparison of numerical accuracy between GPU and CPU computation to verify the precision of GPU computation. In addition, we show how GPUs can be effectively adapted to improve the performance of commercial image processing software such as Adobe Photoshop. Comparison with Existing Method(s) To our best knowledge, this is the first demonstration of GPU application in neural circuit mapping and electrophysiology-based data processing. Conclusions Together, GPU enabled computation enhances our ability to process large-scale data sets derived from neural circuit mapping studies, allowing for increased processing speeds while retaining data precision. PMID:25277633
A method for experimental modal separation
NASA Technical Reports Server (NTRS)
Hallauer, W. L., Jr.
1977-01-01
A method is described for the numerical simulation of multiple-shaker modal survey testing using simulated experimental data to optimize the shaker force-amplitude distribution for the purpose of isolating individual modes of vibration. Inertia, damping, stiffness, and model data are stored on magnetic disks, available by direct access to the interactive FORTRAN programs which perform all computations required by this relative force amplitude distribution method.
Simulation of 3-D Nonequilibrium Seeded Air Flow in the NASA-Ames MHD Channel
NASA Technical Reports Server (NTRS)
Gupta, Sumeet; Tannehill, John C.; Mehta, Unmeel B.
2004-01-01
The 3-D nonequilibrium seeded air flow in the NASA-Ames experimental MHD channel has been numerically simulated. The channel contains a nozzle section, a center section, and an accelerator section where magnetic and electric fields can be imposed on the flow. In recent tests, velocity increases of up to 40% have been achieved in the accelerator section. The flow in the channel is numerically computed us ing a 3-D parabolized Navier-Stokes (PNS) algorithm that has been developed to efficiently compute MHD flows in the low magnetic Reynolds number regime: The MHD effects are modeled by introducing source terms into the PNS equations which can then be solved in a very efficient manner. The algorithm has been extended in the present study to account for nonequilibrium seeded air flows. The electrical conductivity of the flow is determined using the program of Park. The new algorithm has been used to compute two test cases that match the experimental conditions. In both cases, magnetic and electric fields are applied to the seeded flow. The computed results are in good agreement with the experimental data.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Stevens, H.E. Jr.
1960-02-16
SYNFAR, the buckling, reflector saviags, flux, and reactivity segments of the pilot code, was assembled, checked out, and placed in production status. A reduction of 50% in the computation time required for SYNFAR was achieved through incorporation of a convergence acceleration technique. Modification of SYNFAR to perrait computation of dynamic flux and reactivity was raade and the option was prepared for checkout. Retails of the APWRC Error and Exit Diagnosis package and the APWRC Library Tape System are given. The latter was checked out except for the simultaneous tape shifting fuaction. Digitalization of basic cross section data was completed formore » fifteen materials. The portion of the Cross Section Data Program which converts the punched card data to magnetic tape form, interpolating as necessary to obtain data at 1001 energy levels, was completed and checked out. The Breit-Wigner Analysis Program, used with the Cross Section Data Program, was checked out. A listing of the Fortran source program, containing definitions of terms used, fiow diagrams, input data forms, and a sample caloulation is contained. The theory and equations developed to compute the scattering parametera, mu and xi , also used by the Croas Section Data Program, were developed. Checkout of the corresponding program, XIMU, was started. Theory and equations for computing an inelastic scattering matrix, for use with the Cross Section Data Program, were developed and a FORTRAN program for evaluating them was started. An aralysis of the results of the experimental program was started using SYNFAR. Multiplication factors for the two cores studied, Nos. 453 and 454, agreed with the experimental value of 1.00 within 0.6%. The experimental program on Core 454 was completed. Experiments performed were determination of temperature coefficient (--8.9 x 10/sup -5/ DELTA k/k per degree centigrade at 35 deg C), per cent fiasions by subcadmium neutrons (18%), intracell thermal flux measurements, and buckling measuremerts. Core 453 was assembled. The cold clean critical mass for this core was 17.5 kg of U/sup 235/ with 134.63 grams of natural boron in the core. A complete series of clean core experiments was performed on this core. Core 452 was also assembled. The critical mass for this core was 14.4 kg of U/sup 235/ with 83.14 grams of natural boron in the core. The critical experiment control rods were calibrated. Material and dimensional specifications of the homogeneous fuel elements were prepared. A number of saruple blocks containing powdered stainless steel and lucite was pressed. Improvements in the process are being made in an attempt to minimize dimensioral variations from block to block. (See also MND-E-2119.) (auth)« less
Virtual Diagnostic Interface: Aerospace Experimentation in the Synthetic Environment
NASA Technical Reports Server (NTRS)
Schwartz, Richard J.; McCrea, Andrew C.
2009-01-01
The Virtual Diagnostics Interface (ViDI) methodology combines two-dimensional image processing and three-dimensional computer modeling to provide comprehensive in-situ visualizations commonly utilized for in-depth planning of wind tunnel and flight testing, real time data visualization of experimental data, and unique merging of experimental and computational data sets in both real-time and post-test analysis. The preparation of such visualizations encompasses the realm of interactive three-dimensional environments, traditional and state of the art image processing techniques, database management and development of toolsets with user friendly graphical user interfaces. ViDI has been under development at the NASA Langley Research Center for over 15 years, and has a long track record of providing unique and insightful solutions to a wide variety of experimental testing techniques and validation of computational simulations. This report will address the various aspects of ViDI and how it has been applied to test programs as varied as NASCAR race car testing in NASA wind tunnels to real-time operations concerning Space Shuttle aerodynamic flight testing. In addition, future trends and applications will be outlined in the paper.
Virtual Diagnostic Interface: Aerospace Experimentation in the Synthetic Environment
NASA Technical Reports Server (NTRS)
Schwartz, Richard J.; McCrea, Andrew C.
2010-01-01
The Virtual Diagnostics Interface (ViDI) methodology combines two-dimensional image processing and three-dimensional computer modeling to provide comprehensive in-situ visualizations commonly utilized for in-depth planning of wind tunnel and flight testing, real time data visualization of experimental data, and unique merging of experimental and computational data sets in both real-time and post-test analysis. The preparation of such visualizations encompasses the realm of interactive three-dimensional environments, traditional and state of the art image processing techniques, database management and development of toolsets with user friendly graphical user interfaces. ViDI has been under development at the NASA Langley Research Center for over 15 years, and has a long track record of providing unique and insightful solutions to a wide variety of experimental testing techniques and validation of computational simulations. This report will address the various aspects of ViDI and how it has been applied to test programs as varied as NASCAR race car testing in NASA wind tunnels to real-time operations concerning Space Shuttle aerodynamic flight testing. In addition, future trends and applications will be outlined in the paper.
Accurate Molecular Dimensions from Stearic Acid Monolayers.
ERIC Educational Resources Information Center
Lane, Charles A.; And Others
1984-01-01
Discusses modifications in the fatty acid monolayer experiment to reduce the inaccurate moleculary data students usually obtain. Copies of the experimental procedure used and a Pascal computer program to work up the data are available from the authors. (JN)
Binary Solid-Liquid Phase Equilibria
ERIC Educational Resources Information Center
Ellison, Herbert R.
1978-01-01
Indicates some of the information that may be obtained from a binary solid-liquid phase equilibria experiment and a method to write a computer program that will plot an ideal phase diagram to which the experimental results may be compared. (Author/CP)
A Computer Program to Evaluate Experimental Therapies for Treating Burned Patients
Flora, Jairus D.; Flora, Sharyl Ann
1980-01-01
Determining the worth of new therapies for burn patients has been difficult because of the rarity of the burn injury and the disparate survival chances associated with different sizes of burns. Recently a burn survival model has been developed that estimates the risk of death from a burn as a function of the patient's age, sex, area of full thickness (third degree) burn, area of partial thickness burn, involvement of the perineum, and time from burn to admission. An alternative risk model uses the total area burned in place of the areas of partial thickness burn and full thickness burn, and is appropriate if the amount of full thickness burn is not determined accurately. This paper describes a program that uses these risk models to correct or standardize for demographic and severity factors, then, after that adjustment, tests whether a group of burn patients who received a new or experimental therapy shows a significantly better survival rate than that predicted by a baseline model. The program is a simple one written in Fortran for easy adaptation to other computer systems.
Configuration Aerodynamics: Past - Present - Future
NASA Technical Reports Server (NTRS)
Wood, Richard M.; Agrawal, Shreekant; Bencze, Daniel P.; Kulfan, Robert M.; Wilson, Douglas L.
1999-01-01
The Configuration Aerodynamics (CA) element of the High Speed Research (HSR) program is managed by a joint NASA and Industry team, referred to as the Technology Integration Development (ITD) team. This team is responsible for the development of a broad range of technologies for improved aerodynamic performance and stability and control characteristics at subsonic to supersonic flight conditions. These objectives are pursued through the aggressive use of advanced experimental test techniques and state of the art computational methods. As the HSR program matures and transitions into the next phase the objectives of the Configuration Aerodynamics ITD are being refined to address the drag reduction needs and stability and control requirements of High Speed Civil Transport (HSCT) aircraft. In addition, the experimental and computational tools are being refined and improved to meet these challenges. The presentation will review the work performed within the Configuration Aerodynamics element in 1994 and 1995 and then discuss the plans for the 1996-1998 time period. The final portion of the presentation will review several observations of the HSR program and the design activity within Configuration Aerodynamics.
Computational/experimental studies of isolated, single component droplet combustion
NASA Technical Reports Server (NTRS)
Dryer, Frederick L.
1993-01-01
Isolated droplet combustion processes have been the subject of extensive experimental and theoretical investigations for nearly 40 years. The gross features of droplet burning are qualitatively embodied by simple theories and are relatively well understood. However, there remain significant aspects of droplet burning, particularly its dynamics, for which additional basic knowledge is needed for thorough interpretations and quantitative explanations of transient phenomena. Spherically-symmetric droplet combustion, which can only be approximated under conditions of both low Reynolds and Grashof numbers, represents the simplest geometrical configuration in which to study the coupled chemical/transport processes inherent within non-premixed flames. The research summarized here, concerns recent results on isolated, single component, droplet combustion under microgravity conditions, a program pursued jointly with F.A. Williams of the University of California, San Diego. The overall program involves developing and applying experimental methods to study the burning of isolated, single component droplets, in various atmospheres, primarily at atmospheric pressure and below, in both drop towers and aboard space-based platforms such as the Space Shuttle or Space Station. Both computational methods and asymptotic methods, the latter pursued mainly at UCSD, are used in developing the experimental test matrix, in analyzing results, and for extending theoretical understanding. Methanol, and the normal alkanes, n-heptane, and n-decane, have been selected as test fuels to study time-dependent droplet burning phenomena. The following sections summarizes the Princeton efforts on this program, describe work in progress, and briefly delineate future research directions.
Capitalizing on Community: the Small College Environment and the Development of Researchers
NASA Astrophysics Data System (ADS)
Stoneking, M. R.
2014-03-01
Liberal arts colleges constitute an important source of and training ground for future scientists. At Lawrence University, we take advantage of our small college environment to prepare physics students for research careers by complementing content acquisition with skill development and project experience distributed throughout the curriculum and with co-curricular elements that are tied to our close-knit supportive physics community. Small classes and frequent contact between physics majors and faculty members offer opportunities for regular and detailed feedback on the development of research relevant skills such as laboratory record-keeping, data analysis, electronic circuit design, computational programming, experimental design and modification, and scientific communication. Part of our approach is to balance collaborative group work on small projects (such as Arduino-based electronics projects and optical design challenges) with independent work (on, for example, advanced laboratory experimental extensions and senior capstone projects). Communal spaces and specialized facilities (experimental and computational) and active on-campus research programs attract eager students to the program, establish a community-based atmosphere, provide unique opportunities for the development of research aptitude, and offer opportunities for genuine contribution to a research program. Recently, we have also been encouraging innovativetendencies in physics majors through intentional efforts to develop personal characteristics, encouraging students to become more tolerant of ambiguity, risk-taking, initiative-seeking, and articulate. Indicators of the success of our approach include the roughly ten physics majors who graduate each year and our program's high ranking among institutions whose graduates go on to receive the Ph.D. in physics. Work supported in part by the National Science Foundation.
A nonproprietary, nonsecret program for calculating Stirling cryocoolers
NASA Technical Reports Server (NTRS)
Martini, W. R.
1985-01-01
A design program for an integrated Stirling cycle cryocooler was written on an IBM-PC computer. The program is easy to use and shows the trends and itemizes the losses. The calculated results were compared with some measured performance values. The program predicts somewhat optimistic performance and needs to be calibrated more with experimental measurements. Adding a multiplier to the friction factor can bring the calculated rsults in line with the limited test results so far available. The program is offered as a good framework on which to build a truly useful design program for all types of cryocoolers.
Comprehensive silicon solar cell computer modeling
NASA Technical Reports Server (NTRS)
Lamorte, M. F.
1984-01-01
The development of an efficient, comprehensive Si solar cell modeling program that has the capability of simulation accuracy of 5 percent or less is examined. A general investigation of computerized simulation is provided. Computer simulation programs are subdivided into a number of major tasks: (1) analytical method used to represent the physical system; (2) phenomena submodels that comprise the simulation of the system; (3) coding of the analysis and the phenomena submodels; (4) coding scheme that results in efficient use of the CPU so that CPU costs are low; and (5) modularized simulation program with respect to structures that may be analyzed, addition and/or modification of phenomena submodels as new experimental data become available, and the addition of other photovoltaic materials.
Applications of computational modeling in ballistics
NASA Technical Reports Server (NTRS)
Sturek, Walter B.
1987-01-01
The development of the technology of ballistics as applied to gun launched Army weapon systems is the main objective of research at the U.S. Army Ballistic Research Laboratory (BRL). The primary research programs at the BRL consist of three major ballistic disciplines: exterior, interior, and terminal. The work done at the BRL in these areas was traditionally highly dependent on experimental testing. A considerable emphasis was placed on the development of computational modeling to augment the experimental testing in the development cycle; however, the impact of the computational modeling to this date is modest. With the availability of supercomputer computational resources recently installed at the BRL, a new emphasis on the application of computational modeling to ballistics technology is taking place. The major application areas are outlined which are receiving considerable attention at the BRL at present and to indicate the modeling approaches involved. An attempt was made to give some information as to the degree of success achieved and indicate the areas of greatest need.
Elevated temperature crack growth
NASA Technical Reports Server (NTRS)
Kim, K. S.; Vanstone, R. H.
1992-01-01
The purpose of this program was to extend the work performed in the base program (CR 182247) into the regime of time-dependent crack growth under isothermal and thermal mechanical fatigue (TMF) loading, where creep deformation also influences the crack growth behavior. The investigation was performed in a two-year, six-task, combined experimental and analytical program. The path-independent integrals for application to time-dependent crack growth were critically reviewed. The crack growth was simulated using a finite element method. The path-independent integrals were computed from the results of finite-element analyses. The ability of these integrals to correlate experimental crack growth data were evaluated under various loading and temperature conditions. The results indicate that some of these integrals are viable parameters for crack growth prediction at elevated temperatures.
Experimental Program to Stimulate Competitive Research (EPSCoR)
NASA Technical Reports Server (NTRS)
Dingerson, Michael R.
1997-01-01
Report includes: (1) CLUSTER: "Studies in Macromolecular Behavior in Microgravity Environment": The Role of Protein Oligomers in Protein Crystallization; Phase Separation Phenomena in Microgravity; Traveling Front Polymerizations; Investigating Mechanisms Affecting Phase Transition Response and Changes in Thermal Transport Properties in ER-Fluids under Normal and Microgravity Conditions. (2) CLUSTER: "Computational/Parallel Processing Studies": Flows in Local Chemical Equilibrium; A Computational Method for Solving Very Large Problems; Modeling of Cavitating Flows.
ERIC Educational Resources Information Center
von Arnim, Albrecht G.; Missra, Anamika
2017-01-01
Leading voices in the biological sciences have called for a transformation in graduate education leading to the PhD degree. One area commonly singled out for growth and innovation is cross-training in computational science. In 1998, the University of Tennessee (UT) founded an intercollegiate graduate program called the UT-ORNL Graduate School of…
ERIC Educational Resources Information Center
Fasting, Rolf B.; Lyster, Solveig-Alma Halaas
2005-01-01
The aim of the present study is to evaluate the effect of MultiFunk, a computer program designed to assist reading, on the reading and spelling proficiency of struggling readers. Fifty-two below-average readers and spellers, in grades 5, 6 and 7, were randomly assigned as experimental and control groups (N = 26 + 26). In addition, 114 classmates,…
X-38 Experimental Controls Laws
NASA Technical Reports Server (NTRS)
Munday, Steve; Estes, Jay; Bordano, Aldo J.
2000-01-01
X-38 Experimental Control Laws X-38 is a NASA JSC/DFRC experimental flight test program developing a series of prototypes for an International Space Station (ISS) Crew Return Vehicle, often called an ISS "lifeboat." X- 38 Vehicle 132 Free Flight 3, currently scheduled for the end of this month, will be the first flight test of a modem FCS architecture called Multi-Application Control-Honeywell (MACH), originally developed by the Honeywell Technology Center. MACH wraps classical P&I outer attitude loops around a modem dynamic inversion attitude rate loop. The dynamic inversion process requires that the flight computer have an onboard aircraft model of expected vehicle dynamics based upon the aerodynamic database. Dynamic inversion is computationally intensive, so some timing modifications were made to implement MACH on the slower flight computers of the subsonic test vehicles. In addition to linear stability margin analyses and high fidelity 6-DOF simulation, hardware-in-the-loop testing is used to verify the implementation of MACH and its robustness to aerodynamic and environmental uncertainties and disturbances.
ERIC Educational Resources Information Center
Burke, Edmund B.
1994-01-01
Outlines basic patent law information that pertains to computer software programs. Topics addressed include protection in other countries; how to obtain patents; kinds of patents; duration; classes of patentable subject matter, including machines and processes; patentability searches; experimental use prior to obtaining a patent; and patent…
μ-PIV measurements of the ensemble flow fields surrounding a migrating semi-infinite bubble.
Yamaguchi, Eiichiro; Smith, Bradford J; Gaver, Donald P
2009-08-01
Microscale particle image velocimetry (μ-PIV) measurements of ensemble flow fields surrounding a steadily-migrating semi-infinite bubble through the novel adaptation of a computer controlled linear motor flow control system. The system was programmed to generate a square wave velocity input in order to produce accurate constant bubble propagation repeatedly and effectively through a fused glass capillary tube. We present a novel technique for re-positioning of the coordinate axis to the bubble tip frame of reference in each instantaneous field through the analysis of the sudden change of standard deviation of centerline velocity profiles across the bubble interface. Ensemble averages were then computed in this bubble tip frame of reference. Combined fluid systems of water/air, glycerol/air, and glycerol/Si-oil were used to investigate flows comparable to computational simulations described in Smith and Gaver (2008) and to past experimental observations of interfacial shape. Fluorescent particle images were also analyzed to measure the residual film thickness trailing behind the bubble. The flow fields and film thickness agree very well with the computational simulations as well as existing experimental and analytical results. Particle accumulation and migration associated with the flow patterns near the bubble tip after long experimental durations are discussed as potential sources of error in the experimental method.
Optimizing Force Deployment and Force Structure for the Rapid Deployment Force
1984-03-01
Analysis . . . . .. .. ... ... 97 Experimental Design . . . . . .. .. .. ... 99 IX. Use of a Flexible Response Surface ........ 10.2 Selection of a...setS . ere designe . arun, programming methodology , where the require: s.stem re..r is input and the model optimizes the num=er. :::pe, cargo. an...to obtain new computer outputs" (Ref 38:23). The methodology can be used with any decision model, linear or nonlinear. Experimental Desion Since the
Teaching braille letters, numerals, punctuation, and contractions to sighted individuals.
Putnam, Brittany C; Tiger, Jeffrey H
2015-01-01
Braille-character recognition is one of the foundational skills required for teachers of braille. Prior research has evaluated computer programming for teaching braille-to-print letter relations (e.g., Scheithauer & Tiger, 2012). In the current study, we developed a program (the Visual Braille Trainer) to teach not only letters but also numerals, punctuation, symbols, and contractions; we evaluated this program with 4 sighted undergraduate participants. Exposure to this program resulted in mastery of all braille-to-print relations for each participant. © Society for the Experimental Analysis of Behavior.
Sector and Sphere: the design and implementation of a high-performance data cloud
Gu, Yunhong; Grossman, Robert L.
2009-01-01
Cloud computing has demonstrated that processing very large datasets over commodity clusters can be done simply, given the right programming model and infrastructure. In this paper, we describe the design and implementation of the Sector storage cloud and the Sphere compute cloud. By contrast with the existing storage and compute clouds, Sector can manage data not only within a data centre, but also across geographically distributed data centres. Similarly, the Sphere compute cloud supports user-defined functions (UDFs) over data both within and across data centres. As a special case, MapReduce-style programming can be implemented in Sphere by using a Map UDF followed by a Reduce UDF. We describe some experimental studies comparing Sector/Sphere and Hadoop using the Terasort benchmark. In these studies, Sector is approximately twice as fast as Hadoop. Sector/Sphere is open source. PMID:19451100
Exploring quantum computing application to satellite data assimilation
NASA Astrophysics Data System (ADS)
Cheung, S.; Zhang, S. Q.
2015-12-01
This is an exploring work on potential application of quantum computing to a scientific data optimization problem. On classical computational platforms, the physical domain of a satellite data assimilation problem is represented by a discrete variable transform, and classical minimization algorithms are employed to find optimal solution of the analysis cost function. The computation becomes intensive and time-consuming when the problem involves large number of variables and data. The new quantum computer opens a very different approach both in conceptual programming and in hardware architecture for solving optimization problem. In order to explore if we can utilize the quantum computing machine architecture, we formulate a satellite data assimilation experimental case in the form of quadratic programming optimization problem. We find a transformation of the problem to map it into Quadratic Unconstrained Binary Optimization (QUBO) framework. Binary Wavelet Transform (BWT) will be applied to the data assimilation variables for its invertible decomposition and all calculations in BWT are performed by Boolean operations. The transformed problem will be experimented as to solve for a solution of QUBO instances defined on Chimera graphs of the quantum computer.
The 1981 NASA ASEE Summer Faculty Fellowship Program, volume 2
NASA Technical Reports Server (NTRS)
Robertson, N. G.; Huang, C. J.
1981-01-01
A collection of papers on miscellaneous subjects in aerospace research is presented. Topics discussed are: (1) Langmuir probe theory and the problem of anisotropic collection; (2) anthropometric program analysis of reach and body movement; (3) analysis of IV characteristics of negatively biased panels in a magnetoplasma; (4) analytic solution to classical two body drag problem; (5) fast variable step size integration algorithm for computer simulations of physiological systems; (6) spectroscopic experimental computer assisted empirical model for the production of energetics of excited oxygen molecules formed by atom recombination shuttle tile surfaces; and (7) capillary priming characteristics of dual passage heat pipe in zero-g.
NASA Technical Reports Server (NTRS)
Roberti, Dino; Ludwig, Reinhold; Looft, Fred J.
1988-01-01
A 3-D computer model of a piston radiator with lenses for focusing and defocusing is presented. To achieve high-resolution imaging, the frequency of the transmitted and received ultrasound must be as high as 10 MHz. Current ultrasonic transducers produce an extremely narrow beam at these high frequencies and thus are not appropriate for imaging schemes such as synthetic-aperture focus techniques (SAFT). Consequently, a numerical analysis program has been developed to determine field intensity patterns that are radiated from ultrasonic transducers with lenses. Lens shapes are described and the field intensities are numerically predicted and compared with experimental results.
Joint University Program for Air Transportation Research, 1986
NASA Technical Reports Server (NTRS)
Morrell, Frederick R. (Compiler)
1988-01-01
The research conducted under the NASA/FAA sponsored Joint University Program for Air Transportation Research is summarized. The Joint University Program is a coordinated set of three grants sponsored by NASA and the FAA, one each with the Mass. Inst. of Tech., Ohio Univ., and Princeton Univ. Completed works, status reports, and bibliographies are presented for research topics, which include computer science, guidance and control theory and practice, aircraft performance, flight dynamics, and applied experimental psychology. An overview of activities is presented.
DOE Office of Scientific and Technical Information (OSTI.GOV)
J.A. Bamberger; L.M. Liljegren; P.S. Lowery
This document presents an analysis of the mechanisms influencing mixing within double-shell slurry tanks. A research program to characterize mixing of slurries within tanks has been proposed. The research program presents a combined experimental and computational approach to produce correlations describing the tank slurry concentration profile (and therefore uniformity) as a function of mixer pump operating conditions. The TEMPEST computer code was used to simulate both a full-scale (prototype) and scaled (model) double-shell waste tank to predict flow patterns resulting from a stationary jet centered in the tank. The simulation results were used to evaluate flow patterns in the tankmore » and to determine whether flow patterns are similar between the full-scale prototype and an existing 1/12-scale model tank. The flow patterns were sufficiently similar to recommend conducting scoping experiments at 1/12-scale. Also, TEMPEST modeled velocity profiles of the near-floor jet were compared to experimental measurements of the near-floor jet with good agreement. Reported values of physical properties of double-shell tank slurries were analyzed to evaluate the range of properties appropriate for conducting scaled experiments. One-twelfth scale scoping experiments are recommended to confirm the prioritization of the dimensionless groups (gravitational settling, Froude, and Reynolds numbers) that affect slurry suspension in the tank. Two of the proposed 1/12-scale test conditions were modeled using the TEMPEST computer code to observe the anticipated flow fields. This information will be used to guide selection of sampling probe locations. Additional computer modeling is being conducted to model a particulate laden, rotating jet centered in the tank. The results of this modeling effort will be compared to the scaled experimental data to quantify the agreement between the code and the 1/12-scale experiment. The scoping experiment results will guide selection of parameters to be varied in the follow-on experiments. Data from the follow-on experiments will be used to develop correlations to describe slurry concentration profile as a function of mixing pump operating conditions. This data will also be used to further evaluate the computer model applications. If the agreement between the experimental data and the code predictions is good, the computer code will be recommended for use to predict slurry uniformity in the tanks under various operating conditions. If the agreement between the code predictions and experimental results is not good, the experimental data correlations will be used to predict slurry uniformity in the tanks within the range of correlation applicability.« less
Visualizing ultrasound through computational modeling
NASA Technical Reports Server (NTRS)
Guo, Theresa W.
2004-01-01
The Doppler Ultrasound Hematocrit Project (DHP) hopes to find non-invasive methods of determining a person s blood characteristics. Because of the limits of microgravity and the space travel environment, it is important to find non-invasive methods of evaluating the health of persons in space. Presently, there is no well developed method of determining blood composition non-invasively. This projects hopes to use ultrasound and Doppler signals to evaluate the characteristic of hematocrit, the percentage by volume of red blood cells within whole blood. These non-invasive techniques may also be developed to be used on earth for trauma patients where invasive measure might be detrimental. Computational modeling is a useful tool for collecting preliminary information and predictions for the laboratory research. We hope to find and develop a computer program that will be able to simulate the ultrasound signals the project will work with. Simulated models of test conditions will more easily show what might be expected from laboratory results thus help the research group make informed decisions before and during experimentation. There are several existing Matlab based computer programs available, designed to interpret and simulate ultrasound signals. These programs will be evaluated to find which is best suited for the project needs. The criteria of evaluation that will be used are 1) the program must be able to specify transducer properties and specify transmitting and receiving signals, 2) the program must be able to simulate ultrasound signals through different attenuating mediums, 3) the program must be able to process moving targets in order to simulate the Doppler effects that are associated with blood flow, 4) the program should be user friendly and adaptable to various models. After a computer program is chosen, two simulation models will be constructed. These models will simulate and interpret an RF data signal and a Doppler signal.
SoftLab: A Soft-Computing Software for Experimental Research with Commercialization Aspects
NASA Technical Reports Server (NTRS)
Akbarzadeh-T, M.-R.; Shaikh, T. S.; Ren, J.; Hubbell, Rob; Kumbla, K. K.; Jamshidi, M
1998-01-01
SoftLab is a software environment for research and development in intelligent modeling/control using soft-computing paradigms such as fuzzy logic, neural networks, genetic algorithms, and genetic programs. SoftLab addresses the inadequacies of the existing soft-computing software by supporting comprehensive multidisciplinary functionalities from management tools to engineering systems. Furthermore, the built-in features help the user process/analyze information more efficiently by a friendly yet powerful interface, and will allow the user to specify user-specific processing modules, hence adding to the standard configuration of the software environment.
Cumulative reports and publications
NASA Technical Reports Server (NTRS)
1993-01-01
A complete list of Institute for Computer Applications in Science and Engineering (ICASE) reports are listed. Since ICASE reports are intended to be preprints of articles that will appear in journals or conference proceedings, the published reference is included when it is available. The major categories of the current ICASE research program are: applied and numerical mathematics, including numerical analysis and algorithm development; theoretical and computational research in fluid mechanics in selected areas of interest to LaRC, including acoustics and combustion; experimental research in transition and turbulence and aerodynamics involving LaRC facilities and scientists; and computer science.
Unsteady thermal blooming of intense laser beams
NASA Astrophysics Data System (ADS)
Ulrich, J. T.; Ulrich, P. B.
1980-01-01
A four dimensional (three space plus time) computer program has been written to compute the nonlinear heating of a gas by an intense laser beam. Unsteady, transient cases are capable of solution and no assumption of a steady state need be made. The transient results are shown to asymptotically approach the steady-state results calculated by the standard three dimensional thermal blooming computer codes. The report discusses the physics of the laser-absorber interaction, the numerical approximation used, and comparisons with experimental data. A flowchart is supplied in the appendix to the report.
NASA Astrophysics Data System (ADS)
Bak, Roman; Matyja, Tomasz
An algorithm and a computer program have been developed for calculating the strength of pressure vessels made of laminated composites. Numerical results for pressure vessels of Kevlar 49 laminates are compared with experimental data in the literature.
The Shock and Vibration Digest. Volume 7, Number 7, July 1975.
Contents: News briefs; Feature article: The application of skeleton curves and limit envelopes to analysis of nonlinear vibration; Abstracts from the current literature--analysis and design, computer programs, environments, phenomenology, experimentation, components, systems; Author index ; Literature review; Book reviews.
Progress Towards a Microgravity CFD Validation Study Using the ISS SPHERES-SLOSH Experiment
NASA Technical Reports Server (NTRS)
Storey, Jedediah M.; Kirk, Daniel; Marsell, Brandon (Editor); Schallhorn, Paul (Editor)
2017-01-01
Understanding, predicting, and controlling fluid slosh dynamics is critical to safety and improving performance of space missions when a significant percentage of the spacecrafts mass is a liquid. Computational fluid dynamics simulations can be used to predict the dynamics of slosh, but these programs require extensive validation. Many CFD programs have been validated by slosh experiments using various fluids in earth gravity, but prior to the ISS SPHERES-Slosh experiment1, little experimental data for long-duration, zero-gravity slosh existed. This paper presents the current status of an ongoing CFD validation study using the ISS SPHERES-Slosh experimental data.
Progress Towards a Microgravity CFD Validation Study Using the ISS SPHERES-SLOSH Experiment
NASA Technical Reports Server (NTRS)
Storey, Jed; Kirk, Daniel (Editor); Marsell, Brandon (Editor); Schallhorn, Paul (Editor)
2017-01-01
Understanding, predicting, and controlling fluid slosh dynamics is critical to safety and improving performance of space missions when a significant percentage of the spacecrafts mass is a liquid. Computational fluid dynamics simulations can be used to predict the dynamics of slosh, but these programs require extensive validation. Many CFD programs have been validated by slosh experiments using various fluids in earth gravity, but prior to the ISS SPHERES-Slosh experiment, little experimental data for long-duration, zero-gravity slosh existed. This paper presents the current status of an ongoing CFD validation study using the ISS SPHERES-Slosh experimental data.
NASA Astrophysics Data System (ADS)
Bateev, A. B.; Filippov, V. P.
2017-01-01
The principle possibility of using computer program Univem MS for Mössbauer spectra fitting as a demonstration material at studying such disciplines as atomic and nuclear physics and numerical methods by students is shown in the article. This program is associated with nuclear-physical parameters such as isomer (or chemical) shift of nuclear energy level, interaction of nuclear quadrupole moment with electric field and of magnetic moment with surrounded magnetic field. The basic processing algorithm in such programs is the Least Square Method. The deviation of values of experimental points on spectra from the value of theoretical dependence is defined on concrete examples. This value is characterized in numerical methods as mean square deviation. The shape of theoretical lines in the program is defined by Gaussian and Lorentzian distributions. The visualization of the studied material on atomic and nuclear physics can be improved by similar programs of the Mössbauer spectroscopy, X-ray Fluorescence Analyzer or X-ray diffraction analysis.
A generalized global alignment algorithm.
Huang, Xiaoqiu; Chao, Kun-Mao
2003-01-22
Homologous sequences are sometimes similar over some regions but different over other regions. Homologous sequences have a much lower global similarity if the different regions are much longer than the similar regions. We present a generalized global alignment algorithm for comparing sequences with intermittent similarities, an ordered list of similar regions separated by different regions. A generalized global alignment model is defined to handle sequences with intermittent similarities. A dynamic programming algorithm is designed to compute an optimal general alignment in time proportional to the product of sequence lengths and in space proportional to the sum of sequence lengths. The algorithm is implemented as a computer program named GAP3 (Global Alignment Program Version 3). The generalized global alignment model is validated by experimental results produced with GAP3 on both DNA and protein sequences. The GAP3 program extends the ability of standard global alignment programs to recognize homologous sequences of lower similarity. The GAP3 program is freely available for academic use at http://bioinformatics.iastate.edu/aat/align/align.html.
Experimental results for a hypersonic nozzle/afterbody flow field
NASA Technical Reports Server (NTRS)
Spaid, Frank W.; Keener, Earl R.; Hui, Frank C. L.
1995-01-01
This study was conducted to experimentally characterize the flow field created by the interaction of a single-expansion ramp-nozzle (SERN) flow with a hypersonic external stream. Data were obtained from a generic nozzle/afterbody model in the 3.5 Foot Hypersonic Wind Tunnel at the NASA Ames Research Center, in a cooperative experimental program involving Ames and McDonnell Douglas Aerospace. The model design and test planning were performed in close cooperation with members of the Ames computational fluid dynamics (CFD) team for the National Aerospace Plane (NASP) program. This paper presents experimental results consisting of oil-flow and shadow graph flow-visualization photographs, afterbody surface-pressure distributions, rake boundary-layer measurements, Preston-tube skin-friction measurements, and flow field surveys with five-hole and thermocouple probes. The probe data consist of impact pressure, flow direction, and total temperature profiles in the interaction flow field.
Aerothermal modeling program, phase 2. Element B: Flow interaction experiment
NASA Technical Reports Server (NTRS)
Nikjooy, M.; Mongia, H. C.; Murthy, S. N. B.; Sullivan, J. P.
1986-01-01
The design process was improved and the efficiency, life, and maintenance costs of the turbine engine hot section was enhanced. Recently, there has been much emphasis on the need for improved numerical codes for the design of efficient combustors. For the development of improved computational codes, there is a need for an experimentally obtained data base to be used at test cases for the accuracy of the computations. The purpose of Element-B is to establish a benchmark quality velocity and scalar measurements of the flow interaction of circular jets with swirling flow typical of that in the dome region of annular combustor. In addition to the detailed experimental effort, extensive computations of the swirling flows are to be compared with the measurements for the purpose of assessing the accuracy of current and advanced turbulence and scalar transport models.
NASA Technical Reports Server (NTRS)
Weller, W. H.
1983-01-01
A program of experimental and analytical research was performed to demonstrate the degree of correlation achieved between measured and computed rotor inplane stability characteristics. The experimental data were obtained from hover and wind tunnel tests of a scaled bearingless main rotor model. Both isolated rotor and free-hub conditions were tested. Test parameters included blade built-in cone and sweep angles; rotor inplane structural stiffness and damping; pitch link stiffness and location; and fuselage damping, inertia, and natural frequency. Analytical results for many test conditions were obtained. In addition, the analytical and experimental results were examined to ascertain the effects of the test parameters on rotor ground and air resonance stability. The results from this program are presented herein in tabular and graphical form.
A comparison of experiment and theory for sound propagation in variable area ducts
NASA Technical Reports Server (NTRS)
Nayfeh, A. H.; Kaiser, J. E.; Marshall, R. L.; Hurst, C. J.
1980-01-01
An experimental and analytical program has been carried out to evaluate sound suppression techniques in ducts that produce refraction effects due to axial velocity gradients. The analytical program employs a computer code based on the method of multiple scales to calculate the influence of axial variations due to slow changes in the cross-sectional area as well as transverse gradients due to the wall boundary layers. Detailed comparisons between the analytical predictions and the experimental measurements have been made. The circumferential variations of pressure amplitudes and phases at several axial positions have been examined in straight and variable area ducts, with hard walls and lined sections, and with and without a mean flow. Reasonable agreement between the theoretical and experimental results has been found.
Swellix: a computational tool to explore RNA conformational space.
Sloat, Nathan; Liu, Jui-Wen; Schroeder, Susan J
2017-11-21
The sequence of nucleotides in an RNA determines the possible base pairs for an RNA fold and thus also determines the overall shape and function of an RNA. The Swellix program presented here combines a helix abstraction with a combinatorial approach to the RNA folding problem in order to compute all possible non-pseudoknotted RNA structures for RNA sequences. The Swellix program builds on the Crumple program and can include experimental constraints on global RNA structures such as the minimum number and lengths of helices from crystallography, cryoelectron microscopy, or in vivo crosslinking and chemical probing methods. The conceptual advance in Swellix is to count helices and generate all possible combinations of helices rather than counting and combining base pairs. Swellix bundles similar helices and includes improvements in memory use and efficient parallelization. Biological applications of Swellix are demonstrated by computing the reduction in conformational space and entropy due to naturally modified nucleotides in tRNA sequences and by motif searches in Human Endogenous Retroviral (HERV) RNA sequences. The Swellix motif search reveals occurrences of protein and drug binding motifs in the HERV RNA ensemble that do not occur in minimum free energy or centroid predicted structures. Swellix presents significant improvements over Crumple in terms of efficiency and memory use. The efficient parallelization of Swellix enables the computation of sequences as long as 418 nucleotides with sufficient experimental constraints. Thus, Swellix provides a practical alternative to free energy minimization tools when multiple structures, kinetically determined structures, or complex RNA-RNA and RNA-protein interactions are present in an RNA folding problem.
NASA Technical Reports Server (NTRS)
Brown, Robert L.; Doyle, Dee; Haines, Richard F.; Slocum, Michael
1989-01-01
As part of the Telescience Testbed Pilot Program, the Universities Space Research Association/ Research Institute for Advanced Computer Science (USRA/RIACS) proposed to support remote communication by providing a network of human/machine interfaces, computer resources, and experimental equipment which allows: remote science, collaboration, technical exchange, and multimedia communication. The telescience workstation is intended to provide a local computing environment for telescience. The purpose of the program are as follows: (1) to provide a suitable environment to integrate existing and new software for a telescience workstation; (2) to provide a suitable environment to develop new software in support of telescience activities; (3) to provide an interoperable environment so that a wide variety of workstations may be used in the telescience program; (4) to provide a supportive infrastructure and a common software base; and (5) to advance, apply, and evaluate the telescience technolgy base. A prototype telescience computing environment designed to bring practicing scientists in domains other than their computer science into a modern style of doing their computing was created and deployed. This environment, the Telescience Windowing Environment, Phase 1 (TeleWEn-1), met some, but not all of the goals stated above. The TeleWEn-1 provided a window-based workstation environment and a set of tools for text editing, document preparation, electronic mail, multimedia mail, raster manipulation, and system management.
NASA Technical Reports Server (NTRS)
Goldman, L. J.; Seasholtz, R. G.
1982-01-01
Experimental measurements of the velocity components in the blade to blade (axial tangential) plane were obtained with an axial flow turbine stator passage and were compared with calculations from three turbomachinery computer programs. The theoretical results were calculated from a quasi three dimensional inviscid code, a three dimensional inviscid code, and a three dimensional viscous code. Parameter estimation techniques and a particle dynamics calculation were used to assess the accuracy of the laser measurements, which allow a rational basis for comparison of the experimenal and theoretical results. The general agreement of the experimental data with the results from the two inviscid computer codes indicates the usefulness of these calculation procedures for turbomachinery blading. The comparison with the viscous code, while generally reasonable, was not as good as for the inviscid codes.
NASA Technical Reports Server (NTRS)
Coe, H. H.; Huller, F. T.
1980-01-01
Bearing inner and outer race temperatures and the amount of heat transferred to the lubricant were calculated by using the computer program CYBEAN. The results obtained were compared with previously reported experimental data for a 118 mm bore roller bearing that operated at shaft speeds to 25,500 rpm, radial loads to 8,900 N (2000 lb), and total lubricant flow rates to 0.0102 cu m/min (2.7 gal/min). The calculated results compared well with the experimental data.
Sczyrba, Alexander; Hofmann, Peter; Belmann, Peter; Koslicki, David; Janssen, Stefan; Dröge, Johannes; Gregor, Ivan; Majda, Stephan; Fiedler, Jessika; Dahms, Eik; Bremges, Andreas; Fritz, Adrian; Garrido-Oter, Ruben; Jørgensen, Tue Sparholt; Shapiro, Nicole; Blood, Philip D.; Gurevich, Alexey; Bai, Yang; Turaev, Dmitrij; DeMaere, Matthew Z.; Chikhi, Rayan; Nagarajan, Niranjan; Quince, Christopher; Meyer, Fernando; Balvočiūtė, Monika; Hansen, Lars Hestbjerg; Sørensen, Søren J.; Chia, Burton K. H.; Denis, Bertrand; Froula, Jeff L.; Wang, Zhong; Egan, Robert; Kang, Dongwan Don; Cook, Jeffrey J.; Deltel, Charles; Beckstette, Michael; Lemaitre, Claire; Peterlongo, Pierre; Rizk, Guillaume; Lavenier, Dominique; Wu, Yu-Wei; Singer, Steven W.; Jain, Chirag; Strous, Marc; Klingenberg, Heiner; Meinicke, Peter; Barton, Michael; Lingner, Thomas; Lin, Hsin-Hung; Liao, Yu-Chieh; Silva, Genivaldo Gueiros Z.; Cuevas, Daniel A.; Edwards, Robert A.; Saha, Surya; Piro, Vitor C.; Renard, Bernhard Y.; Pop, Mihai; Klenk, Hans-Peter; Göker, Markus; Kyrpides, Nikos C.; Woyke, Tanja; Vorholt, Julia A.; Schulze-Lefert, Paul; Rubin, Edward M.; Darling, Aaron E.; Rattei, Thomas; McHardy, Alice C.
2018-01-01
In metagenome analysis, computational methods for assembly, taxonomic profiling and binning are key components facilitating downstream biological data interpretation. However, a lack of consensus about benchmarking datasets and evaluation metrics complicates proper performance assessment. The Critical Assessment of Metagenome Interpretation (CAMI) challenge has engaged the global developer community to benchmark their programs on datasets of unprecedented complexity and realism. Benchmark metagenomes were generated from ~700 newly sequenced microorganisms and ~600 novel viruses and plasmids, including genomes with varying degrees of relatedness to each other and to publicly available ones and representing common experimental setups. Across all datasets, assembly and genome binning programs performed well for species represented by individual genomes, while performance was substantially affected by the presence of related strains. Taxonomic profiling and binning programs were proficient at high taxonomic ranks, with a notable performance decrease below the family level. Parameter settings substantially impacted performances, underscoring the importance of program reproducibility. While highlighting current challenges in computational metagenomics, the CAMI results provide a roadmap for software selection to answer specific research questions. PMID:28967888
Computer programming for generating visual stimuli.
Bukhari, Farhan; Kurylo, Daniel D
2008-02-01
Critical to vision research is the generation of visual displays with precise control over stimulus metrics. Generating stimuli often requires adapting commercial software or developing specialized software for specific research applications. In order to facilitate this process, we give here an overview that allows nonexpert users to generate and customize stimuli for vision research. We first give a review of relevant hardware and software considerations, to allow the selection of display hardware, operating system, programming language, and graphics packages most appropriate for specific research applications. We then describe the framework of a generic computer program that can be adapted for use with a broad range of experimental applications. Stimuli are generated in the context of trial events, allowing the display of text messages, the monitoring of subject responses and reaction times, and the inclusion of contingency algorithms. This approach allows direct control and management of computer-generated visual stimuli while utilizing the full capabilities of modern hardware and software systems. The flowchart and source code for the stimulus-generating program may be downloaded from www.psychonomic.org/archive.
NASA Technical Reports Server (NTRS)
Anderson, E. C.; Lewis, C. H.
1971-01-01
Turbulent boundary layer flows of non-reacting gases are predicted for both interal (nozzle) and external flows. Effects of favorable pressure gradients on two eddy viscosity models were studied in rocket and hypervelocity wind tunnel flows. Nozzle flows of equilibrium air with stagnation temperatures up to 10,000 K were computed. Predictions of equilibrium nitrogen flows through hypervelocity nozzles were compared with experimental data. A slender spherically blunted cone was studied at 70,000 ft altitude and 19,000 ft/sec. in the earth's atmosphere. Comparisons with available experimental data showed good agreement. A computer program was developed and fully documented during this investigation for use by interested individuals.
Instrumentation and control of harmonic oscillators via a single-board microprocessor-FPGA device.
Picone, Rico A R; Davis, Solomon; Devine, Cameron; Garbini, Joseph L; Sidles, John A
2017-04-01
We report the development of an instrumentation and control system instantiated on a microprocessor-field programmable gate array (FPGA) device for a harmonic oscillator comprising a portion of a magnetic resonance force microscope. The specific advantages of the system are that it minimizes computation, increases maintainability, and reduces the technical barrier required to enter the experimental field of magnetic resonance force microscopy. Heterodyne digital control and measurement yields computational advantages. A single microprocessor-FPGA device improves system maintainability by using a single programming language. The system presented requires significantly less technical expertise to instantiate than the instrumentation of previous systems, yet integrity of performance is retained and demonstrated with experimental data.
Instrumentation and control of harmonic oscillators via a single-board microprocessor-FPGA device
NASA Astrophysics Data System (ADS)
Picone, Rico A. R.; Davis, Solomon; Devine, Cameron; Garbini, Joseph L.; Sidles, John A.
2017-04-01
We report the development of an instrumentation and control system instantiated on a microprocessor-field programmable gate array (FPGA) device for a harmonic oscillator comprising a portion of a magnetic resonance force microscope. The specific advantages of the system are that it minimizes computation, increases maintainability, and reduces the technical barrier required to enter the experimental field of magnetic resonance force microscopy. Heterodyne digital control and measurement yields computational advantages. A single microprocessor-FPGA device improves system maintainability by using a single programming language. The system presented requires significantly less technical expertise to instantiate than the instrumentation of previous systems, yet integrity of performance is retained and demonstrated with experimental data.
Computational Fluid Dynamics Program at NASA Ames Research Center
NASA Technical Reports Server (NTRS)
Holst, Terry L.
1989-01-01
The Computational Fluid Dynamics (CFD) Program at NASA Ames Research Center is reviewed and discussed. The technical elements of the CFD Program are listed and briefly discussed. These elements include algorithm research, research and pilot code development, scientific visualization, advanced surface representation, volume grid generation, and numerical optimization. Next, the discipline of CFD is briefly discussed and related to other areas of research at NASA Ames including experimental fluid dynamics, computer science research, computational chemistry, and numerical aerodynamic simulation. These areas combine with CFD to form a larger area of research, which might collectively be called computational technology. The ultimate goal of computational technology research at NASA Ames is to increase the physical understanding of the world in which we live, solve problems of national importance, and increase the technical capabilities of the aerospace community. Next, the major programs at NASA Ames that either use CFD technology or perform research in CFD are listed and discussed. Briefly, this list includes turbulent/transition physics and modeling, high-speed real gas flows, interdisciplinary research, turbomachinery demonstration computations, complete aircraft aerodynamics, rotorcraft applications, powered lift flows, high alpha flows, multiple body aerodynamics, and incompressible flow applications. Some of the individual problems actively being worked in each of these areas is listed to help define the breadth or extent of CFD involvement in each of these major programs. State-of-the-art examples of various CFD applications are presented to highlight most of these areas. The main emphasis of this portion of the presentation is on examples which will not otherwise be treated at this conference by the individual presentations. Finally, a list of principal current limitations and expected future directions is given.
United States Air Force Research Initiation Program. 1984 Research Reports. Volume 2.
1986-05-01
105th Winter Annual Meeting, Symposium on Experimental Measurements and Techniques In 11. Chiu, H.H., Dynamics of Vortex Shedding and Quasi -large...conservation of quantities such as mass, momentum and energy over any group of control volumes and therefore, over the whole computational domain. The...phenomena. Peliaole and accirate experimental data ’or flowfielcs with high levels of turnulence are neginnirg to appear in literatuream. The advances in non
NASA Technical Reports Server (NTRS)
Giles, G. L.; Wallas, M.
1981-01-01
User documentation is presented for a computer program which considers the nonlinear properties of the strain isolator pad (SIP) in the static stress analysis of the shuttle thermal protection system. This program is generalized to handle an arbitrary SIP footprint including cutouts for instrumentation and filler bar. Multiple SIP surfaces are defined to model tiles in unique locations such as leading edges, intersections, and penetrations. The nonlinearity of the SIP is characterized by experimental stress displacement data for both normal and shear behavior. Stresses in the SIP are calculated using a Newton iteration procedure to determine the six rigid body displacements of the tile which develop reaction forces in the SIP to equilibrate the externally applied loads. This user documentation gives an overview of the analysis capabilities, a detailed description of required input data and an example to illustrate use of the program.
ERIC Educational Resources Information Center
West, Rose
1988-01-01
Describes a science activity done with sixth graders during a unit on outer space. Uses the "Discovery Lab" software program to introduce controlled and experimental variables to the children. Discusses the coordination of library research, computer time, and creative drawing to study planetary environments by designing representative aliens. (CW)
Relative Importance of Nitrite Oxidation by Hypochlorous Acid under Chloramination Conditions
The importance of nitrite’s oxidation by tree chlorine under chloramination conditions was evaluated using batch kinetic experiments and a chloramine model implemented into the computer program AWUASIM. The experimental data was best represented with the inclusion of a reaction b...
The Construction of an Embankment with Frozen Soil.
1980-05-01
Idewlrty by block numlbe) Inls paper presents the construction procedure, data and analysis from an experimental field program to determine the rippability ...finite element computer solution. Also the frozen chunks, produced in the ripping operation, were analyzed to determine the rippability of the frozen
McKay, E
2000-01-01
An innovative research program was devised to investigate the interactive effect of instructional strategies enhanced with text-plus-textual metaphors or text-plus-graphical metaphors, and cognitive style on the acquisition of programming concepts. The Cognitive Styles Analysis (CSA) program (Riding,1991) was used to establish the participants' cognitive style. The QUEST Interactive Test Analysis System (Adams and Khoo,1996) provided the cognitive performance measuring tool, which ensured an absence of error measurement in the programming knowledge testing instruments. Therefore, reliability of the instrumentation was assured through the calibration techniques utilized by the QUEST estimate; providing predictability of the research design. A means analysis of the QUEST data, using the Cohen (1977) approach to size effect and statistical power further quantified the significance of the findings. The experimental methodology adopted for this research links the disciplines of instructional science, cognitive psychology, and objective measurement to provide reliable mechanisms for beneficial use in the evaluation of cognitive performance by the education, training and development sectors. Furthermore, the research outcomes will be of interest to educators, cognitive psychologists, communications engineers, and computer scientists specializing in computer-human interactions.
Accommodation of practical constraints by a linear programming jet select. [for Space Shuttle
NASA Technical Reports Server (NTRS)
Bergmann, E.; Weiler, P.
1983-01-01
An experimental spacecraft control system will be incorporated into the Space Shuttle flight software and exercised during a forthcoming mission to evaluate its performance and handling qualities. The control system incorporates a 'phase space' control law to generate rate change requests and a linear programming jet select to compute jet firings. Posed as a linear programming problem, jet selection must represent the rate change request as a linear combination of jet acceleration vectors where the coefficients are the jet firing times, while minimizing the fuel expended in satisfying that request. This problem is solved in real time using a revised Simplex algorithm. In order to implement the jet selection algorithm in the Shuttle flight control computer, it was modified to accommodate certain practical features of the Shuttle such as limited computer throughput, lengthy firing times, and a large number of control jets. To the authors' knowledge, this is the first such application of linear programming. It was made possible by careful consideration of the jet selection problem in terms of the properties of linear programming and the Simplex algorithm. These modifications to the jet select algorithm may by useful for the design of reaction controlled spacecraft.
NASA high performance computing and communications program
NASA Technical Reports Server (NTRS)
Holcomb, Lee; Smith, Paul; Hunter, Paul
1993-01-01
The National Aeronautics and Space Administration's HPCC program is part of a new Presidential initiative aimed at producing a 1000-fold increase in supercomputing speed and a 100-fold improvement in available communications capability by 1997. As more advanced technologies are developed under the HPCC program, they will be used to solve NASA's 'Grand Challenge' problems, which include improving the design and simulation of advanced aerospace vehicles, allowing people at remote locations to communicate more effectively and share information, increasing scientist's abilities to model the Earth's climate and forecast global environmental trends, and improving the development of advanced spacecraft. NASA's HPCC program is organized into three projects which are unique to the agency's mission: the Computational Aerosciences (CAS) project, the Earth and Space Sciences (ESS) project, and the Remote Exploration and Experimentation (REE) project. An additional project, the Basic Research and Human Resources (BRHR) project exists to promote long term research in computer science and engineering and to increase the pool of trained personnel in a variety of scientific disciplines. This document presents an overview of the objectives and organization of these projects as well as summaries of individual research and development programs within each project.
Incinerator ash dissolution model for the system: Plutonium, nitric acid and hydrofluoric acid
DOE Office of Scientific and Technical Information (OSTI.GOV)
Brown, E V
1988-06-01
This research accomplished two goals. The first was to develop a computer program to simulate a cascade dissolver system. This program would be used to predict the bulk rate of dissolution in incinerator ash. The other goal was to verify the model in a single-stage dissolver system using Dy/sub 2/O/sub 3/. PuO/sub 2/ (and all of the species in the incinerator ash) was assumed to exist as spherical particles. A model was used to calculate the bulk rate of plutonium oxide dissolution using fluoride as a catalyst. Once the bulk rate of PuO/sub 2/ dissolution and the dissolution rate ofmore » all soluble species were calculated, mass and energy balances were written. A computer program simulating the cascade dissolver system was then developed. Tests were conducted on a single-stage dissolver. A simulated incinerator ash mixture was made and added to the dissolver. CaF/sub 2/ was added to the mixture as a catalyst. A 9M HNO/sub 3/ solution was pumped into the dissolver system. Samples of the dissolver effluent were analyzed for dissolved and F concentrations. The computer program proved satisfactory in predicting the F concentrations in the dissolver effluent. The experimental sparge air flow rate was predicted to within 5.5%. The experimental percentage of solids dissolved (51.34%) compared favorably to the percentage of incinerator ash dissolved (47%) in previous work. No general conclusions on model verification could be reached. 56 refs., 11 figs., 24 tabs.« less
The Computer as a Tool for Learning
Starkweather, John A.
1986-01-01
Experimenters from the beginning recognized the advantages computers might offer in medical education. Several medical schools have gained experience in such programs in automated instruction. Television images and graphic display combined with computer control and user interaction are effective for teaching problem solving. The National Board of Medical Examiners has developed patient-case simulation for examining clinical skills, and the National Library of Medicine has experimented with combining media. Advances from the field of artificial intelligence and the availability of increasingly powerful microcomputers at lower cost will aid further development. Computers will likely affect existing educational methods, adding new capabilities to laboratory exercises, to self-assessment and to continuing education. PMID:3544511
Sobie, Eric A
2011-09-13
This two-part lecture introduces students to the scientific computing language MATLAB. Prior computer programming experience is not required. The lectures present basic concepts of computer programming logic that tend to cause difficulties for beginners in addition to concepts that relate specifically to the MATLAB language syntax. The lectures begin with a discussion of vectors, matrices, and arrays. Because many types of biological data, such as fluorescence images and DNA microarrays, are stored as two-dimensional objects, processing these data is a form of array manipulation, and MATLAB is especially adept at handling such array objects. The students are introduced to basic commands in MATLAB, as well as built-in functions that provide useful shortcuts. The second lecture focuses on the differences between MATLAB scripts and MATLAB functions and describes when one method of programming organization might be preferable to the other. The principles are illustrated through the analysis of experimental data, specifically measurements of intracellular calcium concentration in live cells obtained using confocal microscopy.
Sobie, Eric A.
2014-01-01
This two-part lecture introduces students to the scientific computing language MATLAB. Prior computer programming experience is not required. The lectures present basic concepts of computer programming logic that tend to cause difficulties for beginners in addition to concepts that relate specifically to the MATLAB language syntax. The lectures begin with a discussion of vectors, matrices, and arrays. Because many types of biological data, such as fluorescence images and DNA microarrays, are stored as two-dimensional objects, processing these data is a form of array manipulation, and MATLAB is especially adept at handling such array objects. The students are introduced to basic commands in MATLAB, as well as built-in functions that provide useful shortcuts. The second lecture focuses on the differences between MATLAB scripts and MATLAB functions and describes when one method of programming organization might be preferable to the other. The principles are illustrated through the analysis of experimental data, specifically measurements of intracellular calcium concentration in live cells obtained using confocal microscopy. PMID:21934110
Ad Hoc modeling, expert problem solving, and R&T program evaluation
NASA Technical Reports Server (NTRS)
Silverman, B. G.; Liebowitz, J.; Moustakis, V. S.
1983-01-01
A simplified cost and time (SCAT) analysis program utilizing personal-computer technology is presented and demonstrated in the case of the NASA-Goddard end-to-end data system. The difficulties encountered in implementing complex program-selection and evaluation models in the research and technology field are outlined. The prototype SCAT system described here is designed to allow user-friendly ad hoc modeling in real time and at low cost. A worksheet constructed on the computer screen displays the critical parameters and shows how each is affected when one is altered experimentally. In the NASA case, satellite data-output and control requirements, ground-facility data-handling capabilities, and project priorities are intricately interrelated. Scenario studies of the effects of spacecraft phaseout or new spacecraft on throughput and delay parameters are shown. The use of a network of personal computers for higher-level coordination of decision-making processes is suggested, as a complement or alternative to complex large-scale modeling.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Woods, Brian; Gutowska, Izabela; Chiger, Howard
Computer simulations of nuclear reactor thermal-hydraulic phenomena are often used in the design and licensing of nuclear reactor systems. In order to assess the accuracy of these computer simulations, computer codes and methods are often validated against experimental data. This experimental data must be of sufficiently high quality in order to conduct a robust validation exercise. In addition, this experimental data is generally collected at experimental facilities that are of a smaller scale than the reactor systems that are being simulated due to cost considerations. Therefore, smaller scale test facilities must be designed and constructed in such a fashion tomore » ensure that the prototypical behavior of a particular nuclear reactor system is preserved. The work completed through this project has resulted in scaling analyses and conceptual design development for a test facility capable of collecting code validation data for the following high temperature gas reactor systems and events— 1. Passive natural circulation core cooling system, 2. pebble bed gas reactor concept, 3. General Atomics Energy Multiplier Module reactor, and 4. prismatic block design steam-water ingress event. In the event that code validation data for these systems or events is needed in the future, significant progress in the design of an appropriate integral-type test facility has already been completed as a result of this project. Where applicable, the next step would be to begin the detailed design development and material procurement. As part of this project applicable scaling analyses were completed and test facility design requirements developed. Conceptual designs were developed for the implementation of these design requirements at the Oregon State University (OSU) High Temperature Test Facility (HTTF). The original HTTF is based on a ¼-scale model of a high temperature gas reactor concept with the capability for both forced and natural circulation flow through a prismatic core with an electrical heat source. The peak core region temperature capability is 1400°C. As part of this project, an inventory of test facilities that could be used for these experimental programs was completed. Several of these facilities showed some promise, however, upon further investigation it became clear that only the OSU HTTF had the power and/or peak temperature limits that would allow for the experimental programs envisioned herein. Thus the conceptual design and feasibility study development focused on examining the feasibility of configuring the current HTTF to collect validation data for these experimental programs. In addition to the scaling analyses and conceptual design development, a test plan was developed for the envisioned modified test facility. This test plan included a discussion on an appropriate shakedown test program as well as the specific matrix tests. Finally, a feasibility study was completed to determine the cost and schedule considerations that would be important to any test program developed to investigate these designs and events.« less
Performance Evaluation in Network-Based Parallel Computing
NASA Technical Reports Server (NTRS)
Dezhgosha, Kamyar
1996-01-01
Network-based parallel computing is emerging as a cost-effective alternative for solving many problems which require use of supercomputers or massively parallel computers. The primary objective of this project has been to conduct experimental research on performance evaluation for clustered parallel computing. First, a testbed was established by augmenting our existing SUNSPARCs' network with PVM (Parallel Virtual Machine) which is a software system for linking clusters of machines. Second, a set of three basic applications were selected. The applications consist of a parallel search, a parallel sort, a parallel matrix multiplication. These application programs were implemented in C programming language under PVM. Third, we conducted performance evaluation under various configurations and problem sizes. Alternative parallel computing models and workload allocations for application programs were explored. The performance metric was limited to elapsed time or response time which in the context of parallel computing can be expressed in terms of speedup. The results reveal that the overhead of communication latency between processes in many cases is the restricting factor to performance. That is, coarse-grain parallelism which requires less frequent communication between processes will result in higher performance in network-based computing. Finally, we are in the final stages of installing an Asynchronous Transfer Mode (ATM) switch and four ATM interfaces (each 155 Mbps) which will allow us to extend our study to newer applications, performance metrics, and configurations.
Abstractions for DNA circuit design.
Lakin, Matthew R; Youssef, Simon; Cardelli, Luca; Phillips, Andrew
2012-03-07
DNA strand displacement techniques have been used to implement a broad range of information processing devices, from logic gates, to chemical reaction networks, to architectures for universal computation. Strand displacement techniques enable computational devices to be implemented in DNA without the need for additional components, allowing computation to be programmed solely in terms of nucleotide sequences. A major challenge in the design of strand displacement devices has been to enable rapid analysis of high-level designs while also supporting detailed simulations that include known forms of interference. Another challenge has been to design devices capable of sustaining precise reaction kinetics over long periods, without relying on complex experimental equipment to continually replenish depleted species over time. In this paper, we present a programming language for designing DNA strand displacement devices, which supports progressively increasing levels of molecular detail. The language allows device designs to be programmed using a common syntax and then analysed at varying levels of detail, with or without interference, without needing to modify the program. This allows a trade-off to be made between the level of molecular detail and the computational cost of analysis. We use the language to design a buffered architecture for DNA devices, capable of maintaining precise reaction kinetics for a potentially unbounded period. We test the effectiveness of buffered gates to support long-running computation by designing a DNA strand displacement system capable of sustained oscillations.
NACA0012 benchmark model experimental flutter results with unsteady pressure distributions
NASA Technical Reports Server (NTRS)
Rivera, Jose A., Jr.; Dansberry, Bryan E.; Bennett, Robert M.; Durham, Michael H.; Silva, Walter A.
1992-01-01
The Structural Dynamics Division at NASA Langley Research Center has started a wind tunnel activity referred to as the Benchmark Models Program. The primary objective of this program is to acquire measured dynamic instability and corresponding pressure data that will be useful for developing and evaluating aeroelastic type computational fluid dynamics codes currently in use or under development. The program is a multi-year activity that will involve testing of several different models to investigate various aeroelastic phenomena. This paper describes results obtained from a second wind tunnel test of the first model in the Benchmark Models Program. This first model consisted of a rigid semispan wing having a rectangular planform and a NACA 0012 airfoil shape which was mounted on a flexible two degree of freedom mount system. Experimental flutter boundaries and corresponding unsteady pressure distribution data acquired over two model chords located at the 60 and 95 percent span stations are presented.
BioBlocks: Programming Protocols in Biology Made Easier.
Gupta, Vishal; Irimia, Jesús; Pau, Iván; Rodríguez-Patón, Alfonso
2017-07-21
The methods to execute biological experiments are evolving. Affordable fluid handling robots and on-demand biology enterprises are making automating entire experiments a reality. Automation offers the benefit of high-throughput experimentation, rapid prototyping, and improved reproducibility of results. However, learning to automate and codify experiments is a difficult task as it requires programming expertise. Here, we present a web-based visual development environment called BioBlocks for describing experimental protocols in biology. It is based on Google's Blockly and Scratch, and requires little or no experience in computer programming to automate the execution of experiments. The experiments can be specified, saved, modified, and shared between multiple users in an easy manner. BioBlocks is open-source and can be customized to execute protocols on local robotic platforms or remotely, that is, in the cloud. It aims to serve as a de facto open standard for programming protocols in Biology.
NASA Astrophysics Data System (ADS)
Leuca, Maxim
CFD (Computational Fluid Dynamics) is a computational tool for studying flow in science and technology. The Aerospace Industry uses increasingly the CFD modeling and design phase of the aircraft, so the precision with which phenomena are simulated boundary layer is very important. The research efforts are focused on optimizing the aerodynamic performance of airfoils to predict the drag and delay the laminar-turbulent transition. CFD codes must be fast and efficient to model complex geometries for aerodynamic flows. The resolution of the boundary layer equations requires a large amount of computing resources for viscous flows. CFD codes are commonly used to simulate aerodynamic flows, require normal meshes to the wall, extremely fine, and, by consequence, the calculations are very expensive. . This thesis proposes a new approach to solve the equations of boundary layer for laminar and turbulent flows using an approach based on the finite difference method. Integrated into a code of panels, this concept allows to solve airfoils avoiding the use of iterative algorithms, usually computing time and often involving convergence problems. The main advantages of panels methods are their simplicity and ability to obtain, with minimal computational effort, solutions in complex flow conditions for relatively complicated configurations. To verify and validate the developed program, experimental data are used as references when available. Xfoil code is used to obtain data as a pseudo references. Pseudo-reference, as in the absence of experimental data, we cannot really compare two software together. Xfoil is a program that has proven to be accurate and inexpensive computing resources. Developed by Drela (1985), this program uses the method with two integral to design and analyze profiles of wings at low speed (Drela et Youngren, 2014), (Drela, 2003). NACA 0012, NACA 4412, and ATR-42 airfoils have been used for this study. For the airfoils NACA 0012 and NACA 4412 the calculations are made using the Mach number M =0.17 and Reynolds number Re = 6x10 6 conditions for which we have experimental results. For the airfoil ATR-42 the calculations are made using the Mach number M =0.1 and Reynolds number Re=536450 as it was analysed in LARCASE's Price-Paidoussis wind tunnel. Keywords: boundary layer, direct method, displacement thickness, finite differences, Xfoil code.
Simulation of Transcritical CO2 Refrigeration System with Booster Hot Gas Bypass in Tropical Climate
NASA Astrophysics Data System (ADS)
Santosa, I. D. M. C.; Sudirman; Waisnawa, IGNS; Sunu, PW; Temaja, IW
2018-01-01
A Simulation computer becomes significant important for performance analysis since there is high cost and time allocation to build an experimental rig, especially for CO2 refrigeration system. Besides, to modify the rig also need additional cos and time. One of computer program simulation that is very eligible to refrigeration system is Engineering Equation System (EES). In term of CO2 refrigeration system, environmental issues becomes priority on the refrigeration system development since the Carbon dioxide (CO2) is natural and clean refrigerant. This study aims is to analysis the EES simulation effectiveness to perform CO2 transcritical refrigeration system with booster hot gas bypass in high outdoor temperature. The research was carried out by theoretical study and numerical analysis of the refrigeration system using the EES program. Data input and simulation validation were obtained from experimental and secondary data. The result showed that the coefficient of performance (COP) decreased gradually with the outdoor temperature variation increasing. The results show the program can calculate the performance of the refrigeration system with quick running time and accurate. So, it will be significant important for the preliminary reference to improve the CO2 refrigeration system design for the hot climate temperature.
AAFE RADSCAT data reduction programs user's guide
NASA Technical Reports Server (NTRS)
Claassen, J. P.
1976-01-01
Theory, design and operation of the computer programs which automate the reduction of joint radiometer and scatterometer observations are presented. The programs reduce scatterometer measurements to the normalized scattering coefficient; whereas the radiometer measurements are converted into antenna temperatures. The programs are both investigator and user oriented. Supplementary parameters are provided to aid in the interpretation of the observations. A hierarchy of diagnostics is available to evaluate the operation of the instrument, the conduct of the experiments and the quality of the records. General descriptions of the programs and their data products are also presented. This document therefore serves as a user's guide to the programs and is therefore intended to serve both the experimenter and the program operator.
NASA Astrophysics Data System (ADS)
Ouyang, Lizhi
A systematic improvement and extension of the orthogonalized linear combinations of atomic orbitals method was carried out using a combined computational and theoretical approach. For high performance parallel computing, a Beowulf class personal computer cluster was constructed. It also served as a parallel program development platform that helped us to port the programs of the method to the national supercomputer facilities. The program, received a language upgrade from Fortran 77 to Fortran 90, and a dynamic memory allocation feature. A preliminary parallel High Performance Fortran version of the program has been developed as well. To be of more benefit though, scalability improvements are needed. In order to circumvent the difficulties of the analytical force calculation in the method, we developed a geometry optimization scheme using the finite difference approximation based on the total energy calculation. The implementation of this scheme was facilitated by the powerful general utility lattice program, which offers many desired features such as multiple optimization schemes and usage of space group symmetry. So far, many ceramic oxides have been tested with the geometry optimization program. Their optimized geometries were in excellent agreement with the experimental data. For nine ceramic oxide crystals, the optimized cell parameters differ from the experimental ones within 0.5%. Moreover, the geometry optimization was recently used to predict a new phase of TiNx. The method has also been used to investigate a complex Vitamin B12-derivative, the OHCbl crystals. In order to overcome the prohibitive disk I/O demand, an on-demand version of the method was developed. Based on the electronic structure calculation of the OHCbl crystal, a partial density of states analysis and a bond order analysis were carried out. The calculated bonding of the corrin ring of OHCbl model was coincident with the big open-ring pi bond. One interesting find of the calculation was that the Co-OH bond was weak. This, together with the ongoing projects studying different Vitamin B12 derivatives, might help us to answer questions about the Co-C cleavage of the B12 coenzyme, which is involved in many important B12 enzymatic reactions.
Effect of computer game intervention on the attention capacity of mentally retarded children.
Rezaiyan, Akram; Mohammadi, Eesa; Fallah, Parviz A
2007-10-01
The low rate of understanding and the limited domain of attention are two important and distinctive characteristics of mental retardation. One of the important strategies to provide effective education for these subjects is to lessen the backgrounds and impacts of these affective factors. The aim of this research was to find the effect of computer games program on the amount of mentally retarded persons' attention. Sixty educable male mentally retarded subjects were selected from two 24-h care centres in Tehran. The Toulouse-Pieron Scale was used to determinate the subjects' attention at pre-post test. Members of the experimental group were subjected to 35 sets of computer games. After use of the sets of computer games, the attention scores of the subjects were assessed immediately after the intervention and 5 weeks later. The results showed that exactly after the intervention the average attention scores of the experimental group were significantly higher than those of the control group. But, 5 weeks after the intervention, there was actually no significant difference.
DOE Office of Scientific and Technical Information (OSTI.GOV)
HOLM,ELIZABETH A.; BATTAILE,CORBETT C.; BUCHHEIT,THOMAS E.
2000-04-01
Computational materials simulations have traditionally focused on individual phenomena: grain growth, crack propagation, plastic flow, etc. However, real materials behavior results from a complex interplay between phenomena. In this project, the authors explored methods for coupling mesoscale simulations of microstructural evolution and micromechanical response. In one case, massively parallel (MP) simulations for grain evolution and microcracking in alumina stronglink materials were dynamically coupled. In the other, codes for domain coarsening and plastic deformation in CuSi braze alloys were iteratively linked. this program provided the first comparison of two promising ways to integrate mesoscale computer codes. Coupled microstructural/micromechanical codes were appliedmore » to experimentally observed microstructures for the first time. In addition to the coupled codes, this project developed a suite of new computational capabilities (PARGRAIN, GLAD, OOF, MPM, polycrystal plasticity, front tracking). The problem of plasticity length scale in continuum calculations was recognized and a solution strategy was developed. The simulations were experimentally validated on stockpile materials.« less
ALCF Data Science Program: Productive Data-centric Supercomputing
NASA Astrophysics Data System (ADS)
Romero, Nichols; Vishwanath, Venkatram
The ALCF Data Science Program (ADSP) is targeted at big data science problems that require leadership computing resources. The goal of the program is to explore and improve a variety of computational methods that will enable data-driven discoveries across all scientific disciplines. The projects will focus on data science techniques covering a wide area of discovery including but not limited to uncertainty quantification, statistics, machine learning, deep learning, databases, pattern recognition, image processing, graph analytics, data mining, real-time data analysis, and complex and interactive workflows. Project teams will be among the first to access Theta, ALCFs forthcoming 8.5 petaflops Intel/Cray system. The program will transition to the 200 petaflop/s Aurora supercomputing system when it becomes available. In 2016, four projects have been selected to kick off the ADSP. The selected projects span experimental and computational sciences and range from modeling the brain to discovering new materials for solar-powered windows to simulating collision events at the Large Hadron Collider (LHC). The program will have a regular call for proposals with the next call expected in Spring 2017.http://www.alcf.anl.gov/alcf-data-science-program This research used resources of the ALCF, which is a DOE Office of Science User Facility supported under Contract DE-AC02-06CH11357.
NASA Technical Reports Server (NTRS)
Raphael, B.; Fikes, R.; Waldinger, R.
1973-01-01
The results are summarised of a project aimed at the design and implementation of computer languages to aid in expressing problem solving procedures in several areas of artificial intelligence including automatic programming, theorem proving, and robot planning. The principal results of the project were the design and implementation of two complete systems, QA4 and QLISP, and their preliminary experimental use. The various applications of both QA4 and QLISP are given.
Computation of airfoil buffet boundaries
NASA Technical Reports Server (NTRS)
Levy, L. L., Jr.; Bailey, H. E.
1981-01-01
The ILLIAC IV computer has been programmed with an implicit, finite-difference code for solving the thin layer compressible Navier-Stokes equation. Results presented for the case of the buffet boundaries of a conventional and a supercritical airfoil section at high Reynolds numbers are found to be in agreement with experimentally determined buffet boundaries, especially at the higher freestream Mach numbers and lower lift coefficients where the onset of unsteady flows is associated with shock wave-induced boundary layer separation.
NASA Technical Reports Server (NTRS)
Trosset, Michael W.
1999-01-01
Comprehensive computational experiments to assess the performance of algorithms for numerical optimization require (among other things) a practical procedure for generating pseudorandom nonlinear objective functions. We propose a procedure that is based on the convenient fiction that objective functions are realizations of stochastic processes. This report details the calculations necessary to implement our procedure for the case of certain stationary Gaussian processes and presents a specific implementation in the statistical programming language S-PLUS.
Dapor, Maurizio
2018-03-29
Quantum information theory deals with quantum noise in order to protect physical quantum bits (qubits) from its effects. A single electron is an emblematic example of a qubit, and today it is possible to experimentally produce polarized ensembles of electrons. In this paper, the theory of the polarization of electron beams elastically scattered by atoms is briefly summarized. Then the POLARe program suite, a set of computer programs aimed at the calculation of the spin-polarization parameters of electron beams elastically interacting with atomic targets, is described. Selected results of the program concerning Ar, Kr, and Xe atoms are presented together with the comparison with experimental data about the Sherman function for low kinetic energy of the incident electrons (1.5eV-350eV). It is demonstrated that the quantum-relativistic theory of the polarization of electron beams elastically scattered by atoms is in good agreement with experimental data down to energies smaller than a few eV.
USSAERO computer program development, versions B and C
NASA Technical Reports Server (NTRS)
Woodward, F. A.
1980-01-01
Versions B and C of the unified subsonic and supersonic aerodynamic analysis program, USSAERO, are described. Version B incorporates a new symmetrical singularity method to provide improved surface pressure distributions on wings in subsonic flow. Version C extends the range of application of the program to include the analysis of multiple engine nacelles or finned external stores. In addition, nonlinear compressibility effects in high subsonic and supersonic flows are approximated using a correction based on the local Mach number at panel control points. Several examples are presented comparing the results of these programs with other panel methods and experimental data.
NASA Technical Reports Server (NTRS)
Ludewig, M.; Omori, S.; Rao, G. L.
1974-01-01
Tests were conducted to determine the experimental pressure drop and velocity data for water flowing through woven screens. The types of materials used are dutch twill and square weave fabrics. Pressure drop measures were made at four locations in a rectangular channel. The data are presented as change in pressure compared with the average entry velocity and the numerical relationship is determined by dividing the volumetric flow rate by the screen area open to flow. The equations of continuity and momentum are presented. A computer program listing an extension of a theoretical model and data from that computer program are included.
Video Games: A Human Factors Guide to Visual Display Design and Instructional System Design
1984-04-01
Electronic video games have many of the same technological and psychological characteristics that are found in military computer-based systems. For...both of which employ video games as experimental stimuli, are presented here. The first research program seeks to identify and exploit the...characteristics of video games in the design of game-based training devices. The second program is designed to explore the effects of electronic video display
DOE Office of Scientific and Technical Information (OSTI.GOV)
Syring, R.P.; Grubb, R.L.
1979-09-30
This document reports on the following: (1) experimental determination of the response of 16 basic structural elements and 7 B-52 components to simulated nuclear overpressure environments (utilizing Sandia Corporation's Thunderpipe Shock Tube), (2) analysis of these test specimens utilizing the NOVA-2 computer program, and (3) correlation of test and analysis results.
A structurally oriented simulation system
NASA Technical Reports Server (NTRS)
Aran, Z.
1973-01-01
The computer program SOSS (Structurally Oriented Simulation System) is designed to be used as an experimental aid in the study of reliable systems. Basically, SOSS can simulate the structure and behavior of a discrete-time, finite-state, time-invariant system at various levels of structural definition. A general description of the program is given along with its modes of operation, command language of the basic system, future features to be incorporated in SOSS, and an example of usage.
NASA Technical Reports Server (NTRS)
Murphy, Kelly J.; Bunning, Pieter G.; Pamadi, Bandu N.; Scallion, William I.; Jones, Kenneth M.
2004-01-01
An overview of research efforts at NASA in support of the stage separation and ascent aerothermodynamics research program is presented. The objective of this work is to develop a synergistic suite of experimental, computational, and engineering tools and methods to apply to vehicle separation across the transonic to hypersonic speed regimes. Proximity testing of a generic bimese wing-body configuration is on-going in the transonic (Mach numbers 0.6, 1.05, and 1.1), supersonic (Mach numbers 2.3, 3.0, and 4.5) and hypersonic (Mach numbers 6 and 10) speed regimes in four wind tunnel facilities at the NASA Langley Research Center. An overset grid, Navier-Stokes flow solver has been enhanced and demonstrated on a matrix of proximity cases and on a dynamic separation simulation of the bimese configuration. Steady-state predictions with this solver were in excellent agreement with wind tunnel data at Mach 3 as were predictions via a Cartesian-grid Euler solver. Experimental and computational data have been used to evaluate multi-body enhancements to the widely-used Aerodynamic Preliminary Analysis System, an engineering methodology, and to develop a new software package, SepSim, for the simulation and visualization of vehicle motions in a stage separation scenario. Web-based software will be used for archiving information generated from this research program into a database accessible to the user community. Thus, a framework has been established to study stage separation problems using coordinated experimental, computational, and engineering tools.
Experimental and analytical study of close-coupled ventral nozzles for ASTOVL aircraft
NASA Technical Reports Server (NTRS)
Mcardle, Jack G.; Smith, C. Frederic
1990-01-01
Flow in a generic ventral nozzle system was studied experimentally and analytically with a block version of the PARC3D computational fluid dynamics program (a full Navier-Stokes equation solver) in order to evaluate the program's ability to predict system performance and internal flow patterns. For the experimental work a one-third-size model tailpipe with a single large rectangular ventral nozzle mounted normal to the tailpipe axis was tested with unheated air at steady-state pressure ratios up to 4.0. The end of the tailpipe was closed to simulate a blocked exhaust nozzle. Measurements showed about 5 1/2 percent flow-turning loss, reasonable nozzle performance coefficients, and a significant aftward axial component of thrust due to flow turning loss, reasonable nozzle performance coefficients, and a significant aftward axial component of thrust due to flow turning more than 90 deg. Flow behavior into and through the ventral duct is discussed and illustrated with paint streak flow visualization photographs. For the analytical work the same ventral system configuration was modeled with two computational grids to evaluate the effect of grid density. Both grids gave good results. The finer-grid solution produced more detailed flow patterns and predicted performance parameters, such as thrust and discharge coefficient, within 1 percent of the measured values. PARC3D flow visualization images are shown for comparison with the paint streak photographs. Modeling and computational issues encountered in the analytical work are discussed.
Computation of wind tunnel model deflections. [for transport type solid wing
NASA Technical Reports Server (NTRS)
Mehrotra, S. C.; Gloss, B. B.
1981-01-01
The experimental deflections for a transport type solid wing model were measured for several single point load conditions. These deflections were compared with those obtained by structural modeling of the wing by using plate and solid elements of Structural Performance Analysis and Redesign (SPAR) program. The solid element representation of the wing showed better agreement with the experimental deflections than the plate representation. The difference between the measured and calculated deflections is about 5 percent.
Study of Residual Stresses and Distortion in Structural Weldments in High-Strength Steels.
1981-11-30
diagram of the material under consideration which is either known experimentally or can be derived from the TTT diagram. Using the CCT diagram and the...strains are incorporated into the finite element program ADINA replacing the conventional thermal strains, T TH e i" I Computed CCT Diagram . Experimentally...derived CCT diagrams exist today for several steel alloys, enabling one to predict the microstructure [history during the cooling stage of the
Analytical and scale model research aimed at improved hangglider design
NASA Technical Reports Server (NTRS)
Kroo, I.; Chang, L. S.
1979-01-01
Research consisted of a theoretical analysis which attempts to predict aerodynamic characteristics using lifting surface theory and finite-element structural analysis as well as an experimental investigation using 1/5 scale elastically similar models in the NASA Ames 2m x 3m (7' x 10') wind tunnel. Experimental data were compared with theoretical results in the development of a computer program which may be used in the design and evaluation of ultralight gliders.
NASA Astrophysics Data System (ADS)
Hartmann, Jürgen; Nawroth, Thomas; Dose, Klaus
1984-12-01
Carbodiimide-mediated peptide synthesis in aqueous solution has been studied with respect to self-ordering of amino acids. The copolymerisation of amino acids in the presence of glutamic acid or pyroglutamic acid leads to short pyroglutamyl peptides. Without pyroglutamic acid the formation of higher polymers is favoured. The interactions of the amino acids and the peptides, however, are very complex. Therefore, the experimental results are rather difficult to explain. Some of the experimental results, however, can be explained with the aid of computer simulation programs. Regarding only the tripeptide fraction the copolymerisation of pyroGlu, Ala and Leu, as well as the simulated copolymerisation lead to pyroGlu-Ala-Leu as the main reaction product. The amino acid composition of the insoluble peptides formed during the copolymerisation of Ser, Gly, Ala, Val, Phe, Leu and Ile corresponds in part to the computer-simulated copolymerisation data.
Early phase drug discovery: cheminformatics and computational techniques in identifying lead series.
Duffy, Bryan C; Zhu, Lei; Decornez, Hélène; Kitchen, Douglas B
2012-09-15
Early drug discovery processes rely on hit finding procedures followed by extensive experimental confirmation in order to select high priority hit series which then undergo further scrutiny in hit-to-lead studies. The experimental cost and the risk associated with poor selection of lead series can be greatly reduced by the use of many different computational and cheminformatic techniques to sort and prioritize compounds. We describe the steps in typical hit identification and hit-to-lead programs and then describe how cheminformatic analysis assists this process. In particular, scaffold analysis, clustering and property calculations assist in the design of high-throughput screening libraries, the early analysis of hits and then organizing compounds into series for their progression from hits to leads. Additionally, these computational tools can be used in virtual screening to design hit-finding libraries and as procedures to help with early SAR exploration. Copyright © 2012 Elsevier Ltd. All rights reserved.
Heterogeneous concurrent computing with exportable services
NASA Technical Reports Server (NTRS)
Sunderam, Vaidy
1995-01-01
Heterogeneous concurrent computing, based on the traditional process-oriented model, is approaching its functionality and performance limits. An alternative paradigm, based on the concept of services, supporting data driven computation, and built on a lightweight process infrastructure, is proposed to enhance the functional capabilities and the operational efficiency of heterogeneous network-based concurrent computing. TPVM is an experimental prototype system supporting exportable services, thread-based computation, and remote memory operations that is built as an extension of and an enhancement to the PVM concurrent computing system. TPVM offers a significantly different computing paradigm for network-based computing, while maintaining a close resemblance to the conventional PVM model in the interest of compatibility and ease of transition Preliminary experiences have demonstrated that the TPVM framework presents a natural yet powerful concurrent programming interface, while being capable of delivering performance improvements of upto thirty percent.
NASA Technical Reports Server (NTRS)
Rivera, Jose A., Jr.; Dansberry, Bryan E.; Farmer, Moses G.; Eckstrom, Clinton V.; Seidel, David A.; Bennett, Robert M.
1991-01-01
The Structural Dynamics Div. at NASA-Langley has started a wind tunnel activity referred to as the Benchmark Models Program. The objective is to acquire test data that will be useful for developing and evaluating aeroelastic type Computational Fluid Dynamics codes currently in use or under development. The progress is described which was achieved in testing the first model in the Benchmark Models Program. Experimental flutter boundaries are presented for a rigid semispan model (NACA 0012 airfoil section) mounted on a flexible mount system. Also, steady and unsteady pressure measurements taken at the flutter condition are presented. The pressure data were acquired over the entire model chord located at the 60 pct. span station.
NASA Technical Reports Server (NTRS)
Roozeboom, Nettie H.; Lee, Henry C.; Simurda, Laura J.; Zilliac, Gregory G.; Pulliam, Thomas H.
2016-01-01
Wing-body juncture flow fields on commercial aircraft configurations are challenging to compute accurately. The NASA Advanced Air Vehicle Program's juncture flow committee is designing an experiment to provide data to improve Computational Fluid Dynamics (CFD) modeling in the juncture flow region. Preliminary design of the model was done using CFD, yet CFD tends to over-predict the separation in the juncture flow region. Risk reduction wind tunnel tests were requisitioned by the committee to obtain a better understanding of the flow characteristics of the designed models. NASA Ames Research Center's Fluid Mechanics Lab performed one of the risk reduction tests. The results of one case, accompanied by CFD simulations, are presented in this paper. Experimental results suggest the wall mounted wind tunnel model produces a thicker boundary layer on the fuselage than the CFD predictions, resulting in a larger wing horseshoe vortex suppressing the side of body separation in the juncture flow region. Compared to experimental results, CFD predicts a thinner boundary layer on the fuselage generates a weaker wing horseshoe vortex resulting in a larger side of body separation.
NASA Technical Reports Server (NTRS)
Bailey, F. R.; Kutler, Paul
1988-01-01
Discussed are the capabilities of NASA's Numerical Aerodynamic Simulation (NAS) Program and its application as an advanced supercomputing system for computational fluid dynamics (CFD) research. First, the paper describes the NAS computational system, called the NAS Processing System Network, and the advanced computational capabilities it offers as a consequence of carrying out the NAS pathfinder objective. Second, it presents examples of pioneering CFD research accomplished during NAS's first operational year. Examples are included which illustrate CFD applications for predicting fluid phenomena, complementing and supplementing experimentation, and aiding in design. Finally, pacing elements and future directions for CFD and NAS are discussed.
Joint University Program for Air Transportation Research, 1987
NASA Technical Reports Server (NTRS)
Morrell, Frederick R. (Compiler)
1989-01-01
The research conducted during 1987 under the NASA/FAA sponsored Joint University Program for Air Transportation Research is summarized. The Joint University Program is a coordinated set of 3 grants sponsored by NASA-Langley and the FAA, one each with the MIT, Ohio Univ., and Princeton Univ. Completed works, status reports, and annotated bibliographies are presented for research topics, which include computer science, guidance and control theory and practice, aircraft performance, flight dynamics, and applied experimental psychology. An overview of the year's activities for each university is also presented.
A Computational Model of Reasoning from the Clinical Literature
Rennels, Glenn D.
1986-01-01
This paper explores the premise that a formalized representation of empirical studies can play a central role in computer-based decision support. The specific motivations underlying this research include the following propositions: 1. Reasoning from experimental evidence contained in the clinical literature is central to the decisions physicians make in patient care. 2. A computational model, based upon a declarative representation for published reports of clinical studies, can drive a computer program that selectively tailors knowledge of the clinical literature as it is applied to a particular case. 3. The development of such a computational model is an important first step toward filling a void in computer-based decision support systems. Furthermore, the model may help us better understand the general principles of reasoning from experimental evidence both in medicine and other domains. Roundsman is a developmental computer system which draws upon structured representations of the clinical literature in order to critique plans for the management of primary breast cancer. Roundsman is able to produce patient-specific analyses of breast cancer management options based on the 24 clinical studies currently encoded in its knowledge base. The Roundsman system is a first step in exploring how the computer can help to bring a critical analysis of the relevant literature to the physician, structured around a particular patient and treatment decision.
Alabama NASA EPSCoR Preparation Grant Program: Grant No. NCC5-391
NASA Technical Reports Server (NTRS)
Gregory, John C.
2003-01-01
The funded research projects under the Experimental Program to Stimulate Cooperative Research (EPSCoR) grant program and the student fellowship awards are summarized in this report. The projects include: 1) Crystallization of Dehydratase/DcoH: A Target in Lung Disease; 2) Measuring Velocity Profiles in Liquid Metals using an Ultrasonic Doppler Velocimeter; 3) Synthesis, Structure, and Properties of New Thermoelectric Materials; 4) Computational Determination of Structures and Reactivity of Phenol-Formaldehyde Resins; 5) Synthesis of Microbial Polyesters in the NASA Bioreactor; 6) Visualization of Flow-Fields in Magnetocombustion; 7) Synthesis of Fluorescent Saccharide Derivatives. The student fellowship awards include: 1) Distributed Fusion of Satellite Images; 2) Study of the Relationship between Urban Development, Local Climate, and Water Quality for the Atlanta, Georgia Metrop; 3) Computer Simulation of the Effectiveness of a Spring-Loaded Exercise Device.
An Introduction to Programming for Bioscientists: A Python-Based Primer
Mura, Cameron
2016-01-01
Computing has revolutionized the biological sciences over the past several decades, such that virtually all contemporary research in molecular biology, biochemistry, and other biosciences utilizes computer programs. The computational advances have come on many fronts, spurred by fundamental developments in hardware, software, and algorithms. These advances have influenced, and even engendered, a phenomenal array of bioscience fields, including molecular evolution and bioinformatics; genome-, proteome-, transcriptome- and metabolome-wide experimental studies; structural genomics; and atomistic simulations of cellular-scale molecular assemblies as large as ribosomes and intact viruses. In short, much of post-genomic biology is increasingly becoming a form of computational biology. The ability to design and write computer programs is among the most indispensable skills that a modern researcher can cultivate. Python has become a popular programming language in the biosciences, largely because (i) its straightforward semantics and clean syntax make it a readily accessible first language; (ii) it is expressive and well-suited to object-oriented programming, as well as other modern paradigms; and (iii) the many available libraries and third-party toolkits extend the functionality of the core language into virtually every biological domain (sequence and structure analyses, phylogenomics, workflow management systems, etc.). This primer offers a basic introduction to coding, via Python, and it includes concrete examples and exercises to illustrate the language’s usage and capabilities; the main text culminates with a final project in structural bioinformatics. A suite of Supplemental Chapters is also provided. Starting with basic concepts, such as that of a “variable,” the Chapters methodically advance the reader to the point of writing a graphical user interface to compute the Hamming distance between two DNA sequences. PMID:27271528
An Introduction to Programming for Bioscientists: A Python-Based Primer.
Ekmekci, Berk; McAnany, Charles E; Mura, Cameron
2016-06-01
Computing has revolutionized the biological sciences over the past several decades, such that virtually all contemporary research in molecular biology, biochemistry, and other biosciences utilizes computer programs. The computational advances have come on many fronts, spurred by fundamental developments in hardware, software, and algorithms. These advances have influenced, and even engendered, a phenomenal array of bioscience fields, including molecular evolution and bioinformatics; genome-, proteome-, transcriptome- and metabolome-wide experimental studies; structural genomics; and atomistic simulations of cellular-scale molecular assemblies as large as ribosomes and intact viruses. In short, much of post-genomic biology is increasingly becoming a form of computational biology. The ability to design and write computer programs is among the most indispensable skills that a modern researcher can cultivate. Python has become a popular programming language in the biosciences, largely because (i) its straightforward semantics and clean syntax make it a readily accessible first language; (ii) it is expressive and well-suited to object-oriented programming, as well as other modern paradigms; and (iii) the many available libraries and third-party toolkits extend the functionality of the core language into virtually every biological domain (sequence and structure analyses, phylogenomics, workflow management systems, etc.). This primer offers a basic introduction to coding, via Python, and it includes concrete examples and exercises to illustrate the language's usage and capabilities; the main text culminates with a final project in structural bioinformatics. A suite of Supplemental Chapters is also provided. Starting with basic concepts, such as that of a "variable," the Chapters methodically advance the reader to the point of writing a graphical user interface to compute the Hamming distance between two DNA sequences.
Formal versus Grass-Roots Training: Women, Work, and Computers.
ERIC Educational Resources Information Center
Brunet, Jean; Proulx, Serge
1989-01-01
Examines programs in Montreal, Canada, that offer microcomputing training--traditional private courses as well as an experimental, neighborhood-oriented "popular laboratory." Finds that both are used by men to advance their careers but that women use them to catch up and survive economically in a transformed workplace. (SR)
The Teaching and Learning Environment SAIDA: Some Features and Lessons.
ERIC Educational Resources Information Center
Grandbastien, Monique; Morinet-Lambert, Josette
Written in ADA language, SAIDA, a Help System for Data Implementation, is an experimental teaching and learning environment which uses artificial intelligence techniques to teach a computer science course on abstract data representations. The application domain is teaching advanced programming concepts which have not received much attention from…
NASA Technical Reports Server (NTRS)
Andrews, E. H., Jr.; Mackley, E. A.
1976-01-01
Computer program performance results of a Mach 6 hypersonic research engine during supersonic and subsonic combustion modes were presented. The combustion mode transition was successfully performed, exit surveys made, and effects of altitude, angle of attack, and inlet spike position were determined during these tests.
Production Techniques for Computer-Based Learning Material.
ERIC Educational Resources Information Center
Moonen, Jef; Schoenmaker, Jan
Experiences in the development of educational software in the Netherlands have included the use of individual and team approaches, the determination of software content and how it should be presented, and the organization of the entire development process, from experimental programs to prototype to final product. Because educational software is a…
Management Education: An Experimental Course.
ERIC Educational Resources Information Center
Gutelius, Paul Payne
The thesis describes the design, implementation, and evaluation of a course in the theory and practice of management. It gives an appraisal of programmed learning techniques and compares three methods of teaching management--by readings, by cases, and by computer gaming. Additionally, it relates student reactions to the opportunity to select one…
Positron Annihilation Gamma Ray Lineshape Studies of Defects in Solids.
1980-06-24
of R. Waki in the development of anneal probably polygonized the zinc more completely the computer programs and for other experimental than did the...positron lifetime measurements. The assistance of R. Waki was greatly appreciated. References /1/ B.D. BOGGS and J.G. BYRNE, Metallurg.Trans. 4, 2153
Systems and Software Producibility Collaboration and Experimental Environment (SPRUCE)
2009-04-23
Research Manhattan Project Like Research – Transition timeframe needed • Current generation programs – DoD acquisitions over next 1-5 years • Next...Specific Computing Plant B a s i c Transformational Research Manhattan Project Like Research B a s i c 16 • Sponsored by Lockheed Martin
Validation of NASA Thermal Ice Protection Computer Codes. Part 3; The Validation of Antice
NASA Technical Reports Server (NTRS)
Al-Khalil, Kamel M.; Horvath, Charles; Miller, Dean R.; Wright, William B.
2001-01-01
An experimental program was generated by the Icing Technology Branch at NASA Glenn Research Center to validate two ice protection simulation codes: (1) LEWICE/Thermal for transient electrothermal de-icing and anti-icing simulations, and (2) ANTICE for steady state hot gas and electrothermal anti-icing simulations. An electrothermal ice protection system was designed and constructed integral to a 36 inch chord NACA0012 airfoil. The model was fully instrumented with thermo-couples, RTD'S, and heat flux gages. Tests were conducted at several icing environmental conditions during a two week period at the NASA Glenn Icing Research Tunnel. Experimental results of running-wet and evaporative cases were compared to the ANTICE computer code predictions and are presented in this paper.
Numerical simulation of the helium gas spin-up channel performance of the relativity gyroscope
NASA Technical Reports Server (NTRS)
Karr, Gerald R.; Edgell, Josephine; Zhang, Burt X.
1991-01-01
The dependence of the spin-up system efficiency on each geometrical parameter of the spin-up channel and the exhaust passage of the Gravity Probe-B (GPB) is individually investigated. The spin-up model is coded into a computer program which simulates the spin-up process. Numerical results reveal optimal combinations of the geometrical parameters for the ultimate spin-up performance. Comparisons are also made between the numerical results and experimental data. The experimental leakage rate can only be reached when the gap between the channel lip and the rotor surface increases beyond physical limit. The computed rotating frequency is roughly twice as high as the measured ones although the spin-up torques fairly match.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Hameka, H.F.; Jensen, J.O.
1993-05-01
This report presents the computed optimized geometry and vibrational IR and Raman frequencies of the V-agent VX. The computations are performed with the Gaussian 90 Program Package using 6-31G* basis sets. We assign the vibrational frequencies and correct each frequency by multiplying it with a previously derived 6-31G* correction factor. The result is a computer-generated prediction of the IR and Raman spectra of VX. This study was intended as a blind test of the utility of IR spectral prediction. Therefore, we intentionally did not look at experimental data on the IR and Raman spectra of VX.... IR Spectra, VX, Ramanmore » spectra, Computer predictions.« less
Tang, Qi-Yi; Zhang, Chuan-Xi
2013-04-01
A comprehensive but simple-to-use software package called DPS (Data Processing System) has been developed to execute a range of standard numerical analyses and operations used in experimental design, statistics and data mining. This program runs on standard Windows computers. Many of the functions are specific to entomological and other biological research and are not found in standard statistical software. This paper presents applications of DPS to experimental design, statistical analysis and data mining in entomology. © 2012 The Authors Insect Science © 2012 Institute of Zoology, Chinese Academy of Sciences.
Prediction of Scour below Flip Bucket using Soft Computing Techniques
NASA Astrophysics Data System (ADS)
Azamathulla, H. Md.; Ab Ghani, Aminuddin; Azazi Zakaria, Nor
2010-05-01
The accurate prediction of the depth of scour around hydraulic structure (trajectory spillways) has been based on the experimental studies and the equations developed are mainly empirical in nature. This paper evaluates the performance of the soft computing (intelligence) techiques, Adaptive Neuro-Fuzzy System (ANFIS) and Genetic expression Programming (GEP) approach, in prediction of scour below a flip bucket spillway. The results are very promising, which support the use of these intelligent techniques in prediction of highly non-linear scour parameters.
NASA Technical Reports Server (NTRS)
Ahrens, Thomas J.
2001-01-01
This research is computational /theoretical and complements the Caltech experimental program. We have developed an understanding of the basic physical processes and produced computational models and implemented these into Eulerian and Lagrangian finite element codes. The key issues we have addressed include the conditions required for: faulting (strain localization), elastic moduli weakening, dynamic weakening (layering elastic instabilities and fluidization), bulking (creation of porosity at zero pressure) and compaction of pores, frictional melting (creation of pseudotachylytes), partial and selective devolatilization of materials (e.g. CaCO3, water/ice mixtures), and debris flows.
Estimating Noise Levels In An Enclosed Space
NASA Technical Reports Server (NTRS)
Azzi, Elias
1995-01-01
GEGS Acoustic Analysis Program (GAAP) developed to compute composite profile of noise in Spacelab module on basis of data on noise produced by equipment, data on locations of equipment, and equipment-operating schedules. Impetus for development of GAAP provided by noise that generated in Spacelab Module during SLS-1 mission because of concurrent operation of many pieces of experimental and subsystem equipment. Although originally intended specifically to help compute noise in Spacelab, also applicable to any region with multiple sources of noise. Written in FORTRAN 77.
An Experimental Investigation of the Boundary Layer under Pack Ice
1975-01-01
current-meter interface ( CMIF ) consists of a very stable, 20-Kllz crystal oscillator and counter, a master memory-address buffer, and a buffer for each...data channel to a specific location in the computer’s memory, The CMIF also generates computer interrupts at a rate determined by the program (12.8... CMIF can handle up to 128 channels and is designed so that even if all channels have simultaneous dipulses, the processing delay is less than .05 msec
High-freezing-point fuel studies
NASA Technical Reports Server (NTRS)
Tolle, F. F.
1980-01-01
Considerable progress in developing the experimental and analytical techniques needed to design airplanes to accommodate fuels with less stringent low temperature specifications is reported. A computer technique for calculating fuel temperature profiles in full tanks was developed. The computer program is being extended to include the case of partially empty tanks. Ultimately, the completed package is to be incorporated into an aircraft fuel tank thermal analyser code to permit the designer to fly various thermal exposure patterns, study fuel temperatures versus time, and determine holdup.
Modeling of transient heat pipe operation
NASA Technical Reports Server (NTRS)
Colwell, Gene T.
1989-01-01
Mathematical models and an associated computer program for heat pipe startup from the frozen state have been developed. Finite element formulations of the governing equations are written for each heat pipe region for each operating condition during startup from the frozen state. The various models were checked against analytical and experimental data available in the literature for three specific types of operation. Computations using the methods developed were made for a space shuttle reentry mission where a heat pipe cooled leading edge was used on the wing.
The Physician's Workstation: Recording a Physical Examination Using a Controlled Vocabulary
Cimino, James J.; Barnett, G. Octo
1987-01-01
A system has been developed which runs on MS-DOS personal computers and serves as an experimental model of a physician's workstation. The program provides an interface to a controlled vocabulary which allows rapid selection of appropriate terms and modifiers for entry of clinical information. Because it captures patient descriptions, it has the ability to serve as an intermediary between the physician and computer-based medical knowledge resources. At present, the vocabulary permits rapid, reliable representation of cardiac physical examination findings.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Syring, R.P.; Grubb, R.L.
1979-09-30
This document reports on the following: (1) experimental determination of the response of 16 basic structural elements and 7 B-52 components to simulated nuclear overpressure environments (utilizing Sandia Corporation's Thunderpipe Shock Tube), (2) analysis of these test specimens utilizing the NOVA-2 computer program, and (3) correlation of test and analysis results.
ASTROP2 users manual: A program for aeroelastic stability analysis of propfans
NASA Technical Reports Server (NTRS)
Narayanan, G. V.; Kaza, K. R. V.
1991-01-01
A user's manual is presented for the aeroelastic stability and response of propulsion systems computer program called ASTROP2. The ASTROP2 code preforms aeroelastic stability analysis of rotating propfan blades. This analysis uses a two-dimensional, unsteady cascade aerodynamics model and a three-dimensional, normal-mode structural model. Analytical stability results from this code are compared with published experimental results of a rotating composite advanced turboprop model and of nonrotating metallic wing model.
Predicting oligonucleotide affinity to nucleic acid targets.
Mathews, D H; Burkard, M E; Freier, S M; Wyatt, J R; Turner, D H
1999-01-01
A computer program, OligoWalk, is reported that predicts the equilibrium affinity of complementary DNA or RNA oligonucleotides to an RNA target. This program considers the predicted stability of the oligonucleotide-target helix and the competition with predicted secondary structure of both the target and the oligonucleotide. Both unimolecular and bimolecular oligonucleotide self structure are considered with a user-defined concentration. The application of OligoWalk is illustrated with three comparisons to experimental results drawn from the literature. PMID:10580474
Support for Implications of Compressive Sensing Concepts to Imaging Systems
2015-08-02
34pretty picture" is not only not needed, but is not ALLOWED due to privacy concerns. Remember the huge controversy caused by mmW imagers seeing people...in 2003, for experimental studies of quantum degenerate atomic gases. From 2004-2006 he was a postdoctoral researcher in the Electrical and...Computer Engineering at the University of Arizona. He was recently also a program manager at DARPA/DSO where he started programs on quantum information
Computations of Internal and External Axisymmetric Nozzle Aerodynamics at Transonic Speeds
NASA Technical Reports Server (NTRS)
Dalbello, Teryn; Georgiadis, Nicholas; Yoder, Dennis; Keith, Theo
2003-01-01
Computational Fluid Dynamics (CFD) analyses of axisymmetric circular-arc boattail nozzles have been completed in support of NASA's Next Generation Launch Technology Program to investigate the effects of high-speed nozzle geometries on the nozzle internal flow and the surrounding boattail regions. These computations span the very difficult transonic flight regime, with shock-induced separations and strong adverse pressure gradients. External afterbody and internal nozzle pressure distributions computed with the Wind code are compared with experimental data. A range of turbulence models were examined in Wind, including an Explicit Algebraic Stress model (EASM). Computations on two nozzle geometries have been completed at freestream Mach numbers ranging from 0.6 to 0.9, driven by nozzle pressure ratios (NPR) ranging from 2.9 to 5. Results obtained on converging-only geometry indicate reasonable agreement to experimental data, with the EASM and Shear Stress Transport (SST) turbulence models providing the best agreement. Calculations completed on a converging-diverging geometry involving large-scale internal flow separation did not converge to a true steady-state solution when run with variable timestepping (steady-state). Calculations obtained using constant timestepping (time-accurate) indicate less variations in flow properties compared with steady-state solutions. This failure to converge to a steady-state solution was found to be the result of difficulties in using variable time-stepping with large-scale separations present in the flow. Nevertheless, time-averaged boattail surface pressure coefficient and internal nozzle pressures show fairly good agreement with experimental data. The SST turbulence model demonstrates the best over-all agreement with experimental data.
NASA Technical Reports Server (NTRS)
Marshall, F. J.; Deffenbaugh, F. D.
1974-01-01
A method is developed to determine the flow field of a body of revolution in separated flow. The computer was used to integrate various solutions and solution properties of the sub-flow fields which made up the entire flow field without resorting to a finite difference solution to the complete Navier-Stokes equations. The technique entails the use of the unsteady cross flow analogy and a new solution to the two-dimensional unsteady separated flow problem based upon an unsteady, discrete-vorticity wake. Data for the forces and moments on aerodynamic bodies at low speeds and high angle of attack (outside the range of linear inviscid theories) such that the flow is substantially separated are produced which compare well with experimental data. In addition, three dimensional steady separated regions and wake vortex patterns are determined. The computer program developed to perform the numerical calculations is described.
Theoretical and experimental studies in support of the geophysical fluid flow experiment
NASA Technical Reports Server (NTRS)
Hart, J.; Toomre, J.; Gilman, P.
1984-01-01
Computer programming was completed for digital acquisition of temperature and velocity data generated by the Geophysical Fluid Flow Cell (GFFC) during the upcoming Spacelab 3 mission. A set of scenarios was developed which covers basic electro-hydrodynamic instability, highly supercritical convection with isothermal boundaries, convection with imposed thermal forcing, and some stably stratified runs to look at large-scale thermohaline ocean circulations. The extent to which the GFFC experimental results apply to more complicated circumstances within the Sun or giant planets was assessed.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Yamankaradeniz, R.; Horuz, I.
In this study, the characteristics of solar assisted heat pump are investigated theoretically and experimentally for clear days during the seven months of the winter season in Istanbul/Turkey. A theoretical model was developed and a computer program was written on this basis. The characteristics such as: daily average collector efficiency and solar radiation, monthly average heat transfer at the condenser, monthly average cooling capacity, the mean COP and the mean COP for total system were examined. The theoretical results were found to be in good agreement with the experimental values.
Report on twisted nematic and supertwisted nematic device characterization program
NASA Technical Reports Server (NTRS)
1995-01-01
In this study we measured the optical characteristics of normally white twisted nematic (NWTN) and super twisted nematic (STN ) cells. Though no dynamic computer model was available, the static observations were compared with computer simulated behavior. The measurements were taken as a function of both viewing angle and applied voltage and included in the static case not only luminance but also contrast ratio and chromaticity . We employed the computer model Twist Cell Optics, developed at Kent State in conjunction with this study, and whose optical modeling foundation, Iike the ViDEOS program, is the 4 x 4 matrix method of Berreman. In order to resolve discrepancies between the experimental and modeled data the optical parameters of the individual cell components, where not known, were determined using refractometry, profilometry, and various forms of ellipsometry. The resulting agreement between experiment and model is quite good due primarily to a better understanding of the structure and optics of dichroic sheet polarizers. A description of the model and test cells employed are given in section 2. Section 3 contains the experimental data gathered and section 4 gives examples of the fit between model and experiment. Also included with this report are a pair of papers which resulted from the research and which detail the polarizer properties and some of the cell characterization methods.
Trace contaminant control simulation computer program, version 8.1
NASA Technical Reports Server (NTRS)
Perry, J. L.
1994-01-01
The Trace Contaminant Control Simulation computer program is a tool for assessing the performance of various process technologies for removing trace chemical contamination from a spacecraft cabin atmosphere. Included in the simulation are chemical and physical adsorption by activated charcoal, chemical adsorption by lithium hydroxide, absorption by humidity condensate, and low- and high-temperature catalytic oxidation. Means are provided for simulating regenerable as well as nonregenerable systems. The program provides an overall mass balance of chemical contaminants in a spacecraft cabin given specified generation rates. Removal rates are based on device flow rates specified by the user and calculated removal efficiencies based on cabin concentration and removal technology experimental data. Versions 1.0 through 8.0 are documented in NASA TM-108409. TM-108409 also contains a source file listing for version 8.0. Changes to version 8.0 are documented in this technical memorandum and a source file listing for the modified version, version 8.1, is provided. Detailed descriptions for the computer program subprograms are extracted from TM-108409 and modified as necessary to reflect version 8.1. Version 8.1 supersedes version 8.0. Information on a separate user's guide is available from the author.
Comparison of optimization algorithms in intensity-modulated radiation therapy planning
NASA Astrophysics Data System (ADS)
Kendrick, Rachel
Intensity-modulated radiation therapy is used to better conform the radiation dose to the target, which includes avoiding healthy tissue. Planning programs employ optimization methods to search for the best fluence of each photon beam, and therefore to create the best treatment plan. The Computational Environment for Radiotherapy Research (CERR), a program written in MATLAB, was used to examine some commonly-used algorithms for one 5-beam plan. Algorithms include the genetic algorithm, quadratic programming, pattern search, constrained nonlinear optimization, simulated annealing, the optimization method used in Varian EclipseTM, and some hybrids of these. Quadratic programing, simulated annealing, and a quadratic/simulated annealing hybrid were also separately compared using different prescription doses. The results of each dose-volume histogram as well as the visual dose color wash were used to compare the plans. CERR's built-in quadratic programming provided the best overall plan, but avoidance of the organ-at-risk was rivaled by other programs. Hybrids of quadratic programming with some of these algorithms seems to suggest the possibility of better planning programs, as shown by the improved quadratic/simulated annealing plan when compared to the simulated annealing algorithm alone. Further experimentation will be done to improve cost functions and computational time.
The FTS atomic spectrum tool (FAST) for rapid analysis of line spectra
NASA Astrophysics Data System (ADS)
Ruffoni, M. P.
2013-07-01
The FTS Atomic Spectrum Tool (FAST) is an interactive graphical program designed to simplify the analysis of atomic emission line spectra obtained from Fourier transform spectrometers. Calculated, predicted and/or known experimental line parameters are loaded alongside experimentally observed spectral line profiles for easy comparison between new experimental data and existing results. Many such line profiles, which could span numerous spectra, may be viewed simultaneously to help the user detect problems from line blending or self-absorption. Once the user has determined that their experimental line profile fits are good, a key feature of FAST is the ability to calculate atomic branching fractions, transition probabilities, and oscillator strengths-and their uncertainties-which is not provided by existing analysis packages. Program SummaryProgram title: FAST: The FTS Atomic Spectrum Tool Catalogue identifier: AEOW_v1_0 Program summary URL: http://cpc.cs.qub.ac.uk/summaries/AEOW_v1_0.html Program obtainable from: CPC Program Library, Queen's University, Belfast, N. Ireland Licensing provisions: GNU General Public License version 3 No. of lines in distributed program, including test data, etc.: 293058 No. of bytes in distributed program, including test data, etc.: 13809509 Distribution format: tar.gz Programming language: C++. Computer: Intel x86-based systems. Operating system: Linux/Unix/Windows. RAM: 8 MB minimum. About 50-200 MB for a typical analysis. Classification: 2.2, 2.3, 21.2. Nature of problem: Visualisation of atomic line spectra including the comparison of theoretical line parameters with experimental atomic line profiles. Accurate intensity calibration of experimental spectra, and the determination of observed relative line intensities that are needed for calculating atomic branching fractions and oscillator strengths. Solution method: FAST is centred around a graphical interface, where a user may view sets of experimental line profiles and compare them to calculated data (such as from the Kurucz database [1]), predicted line parameters, and/or previously known experimental results. With additional information on the spectral response of the spectrometer, obtained from a calibrated standard light source, FT spectra may be intensity calibrated. In turn, this permits the user to calculate atomic branching fractions and oscillator strengths, and their respective uncertainties. Running time: Open ended. Defined by the user. References: [1] R.L. Kurucz (2007). URL http://kurucz.harvard.edu/atoms/.
NASA Astrophysics Data System (ADS)
Karakaya, Mustafa; Kürekçi, Mehmet; Eskiyurt, Buse; Sert, Yusuf; Çırak, Çağrı
2015-01-01
In present study, the experimental and theoretical harmonic vibrational frequencies of gliclazide molecule have been investigated. The experimental FT-IR (400-4000 cm-1) and Laser-Raman spectra (100-4000 cm-1) of the molecule in the solid phase were recorded. Theoretical vibrational frequencies and geometric parameters (bond lengths and bond angles) have been calculated using ab initio Hartree Fock (HF), density functional theory (B3LYP hybrid function) methods with 6-311++G(d,p) and 6-31G(d,p) basis sets by Gaussian 09W program. The assignments of the vibrational frequencies were performed by potential energy distribution (PED) analysis by using VEDA 4 program. Theoretical optimized geometric parameters and vibrational frequencies have been compared with the corresponding experimental data, and they have been shown to be in a good agreement with each other. Also, the highest occupied molecular orbital (HOMO) and lowest unoccupied molecular orbital (LUMO) energies have been found.
NASA Astrophysics Data System (ADS)
Çırak, Çağrı; Sert, Yusuf; Ucun, Fatih
In the present study, the experimental and theoretical vibrational spectra of 5-bromo-2'-deoxyuridine were investigated. The experimental FT-IR (400-4000 cm-1) and μ-Raman spectra (100-4000 cm-1) of the molecule in the solid phase were recorded. Theoretical vibrational frequencies and geometric parameters (bond lengths and bond angles) were calculated using ab initio Hartree Fock (HF) and density functional B3LYP method with 6-31G(d), 6-31G(d,p), 6-311++G(d) and 6-311++G(d,p) basis sets by Gaussian program, for the first time. The assignments of vibrational frequencies were performed by potential energy distribution by using VEDA 4 program. The optimized geometric parameters and theoretical vibrational frequencies are compared with the corresponding experimental data and they were seen to be in a good agreement with the each other. Also, the highest occupied molecular orbital (HOMO) and lowest unoccupied molecular orbital (LUMO) energies were found.
Karakaya, Mustafa; Kürekçi, Mehmet; Eskiyurt, Buse; Sert, Yusuf; Çırak, Çağrı
2015-01-25
In present study, the experimental and theoretical harmonic vibrational frequencies of gliclazide molecule have been investigated. The experimental FT-IR (400-4000 cm(-1)) and Laser-Raman spectra (100-4000 cm(-1)) of the molecule in the solid phase were recorded. Theoretical vibrational frequencies and geometric parameters (bond lengths and bond angles) have been calculated using ab initio Hartree Fock (HF), density functional theory (B3LYP hybrid function) methods with 6-311++G(d,p) and 6-31G(d,p) basis sets by Gaussian 09W program. The assignments of the vibrational frequencies were performed by potential energy distribution (PED) analysis by using VEDA 4 program. Theoretical optimized geometric parameters and vibrational frequencies have been compared with the corresponding experimental data, and they have been shown to be in a good agreement with each other. Also, the highest occupied molecular orbital (HOMO) and lowest unoccupied molecular orbital (LUMO) energies have been found. Copyright © 2014 Elsevier B.V. All rights reserved.
NASA High Performance Computing and Communications program
NASA Technical Reports Server (NTRS)
Holcomb, Lee; Smith, Paul; Hunter, Paul
1994-01-01
The National Aeronautics and Space Administration's HPCC program is part of a new Presidential initiative aimed at producing a 1000-fold increase in supercomputing speed and a 1(X)-fold improvement in available communications capability by 1997. As more advanced technologies are developed under the HPCC program, they will be used to solve NASA's 'Grand Challenge' problems, which include improving the design and simulation of advanced aerospace vehicles, allowing people at remote locations to communicate more effectively and share information, increasing scientists' abilities to model the Earth's climate and forecast global environmental trends, and improving the development of advanced spacecraft. NASA's HPCC program is organized into three projects which are unique to the agency's mission: the Computational Aerosciences (CAS) project, the Earth and Space Sciences (ESS) project, and the Remote Exploration and Experimentation (REE) project. An additional project, the Basic Research and Human Resources (BRHR) project, exists to promote long term research in computer science and engineering and to increase the pool of trained personnel in a variety of scientific disciplines. This document presents an overview of the objectives and organization of these projects, as well as summaries of early accomplishments and the significance, status, and plans for individual research and development programs within each project. Areas of emphasis include benchmarking, testbeds, software and simulation methods.
NASA Astrophysics Data System (ADS)
Yim, Keun Soo
This dissertation summarizes experimental validation and co-design studies conducted to optimize the fault detection capabilities and overheads in hybrid computer systems (e.g., using CPUs and Graphics Processing Units, or GPUs), and consequently to improve the scalability of parallel computer systems using computational accelerators. The experimental validation studies were conducted to help us understand the failure characteristics of CPU-GPU hybrid computer systems under various types of hardware faults. The main characterization targets were faults that are difficult to detect and/or recover from, e.g., faults that cause long latency failures (Ch. 3), faults in dynamically allocated resources (Ch. 4), faults in GPUs (Ch. 5), faults in MPI programs (Ch. 6), and microarchitecture-level faults with specific timing features (Ch. 7). The co-design studies were based on the characterization results. One of the co-designed systems has a set of source-to-source translators that customize and strategically place error detectors in the source code of target GPU programs (Ch. 5). Another co-designed system uses an extension card to learn the normal behavioral and semantic execution patterns of message-passing processes executing on CPUs, and to detect abnormal behaviors of those parallel processes (Ch. 6). The third co-designed system is a co-processor that has a set of new instructions in order to support software-implemented fault detection techniques (Ch. 7). The work described in this dissertation gains more importance because heterogeneous processors have become an essential component of state-of-the-art supercomputers. GPUs were used in three of the five fastest supercomputers that were operating in 2011. Our work included comprehensive fault characterization studies in CPU-GPU hybrid computers. In CPUs, we monitored the target systems for a long period of time after injecting faults (a temporally comprehensive experiment), and injected faults into various types of program states that included dynamically allocated memory (to be spatially comprehensive). In GPUs, we used fault injection studies to demonstrate the importance of detecting silent data corruption (SDC) errors that are mainly due to the lack of fine-grained protections and the massive use of fault-insensitive data. This dissertation also presents transparent fault tolerance frameworks and techniques that are directly applicable to hybrid computers built using only commercial off-the-shelf hardware components. This dissertation shows that by developing understanding of the failure characteristics and error propagation paths of target programs, we were able to create fault tolerance frameworks and techniques that can quickly detect and recover from hardware faults with low performance and hardware overheads.
Computational Materials Program for Alloy Design
NASA Technical Reports Server (NTRS)
Bozzolo, Guillermo
2005-01-01
The research program sponsored by this grant, "Computational Materials Program for Alloy Design", covers a period of time of enormous change in the emerging field of computational materials science. The computational materials program started with the development of the BFS method for alloys, a quantum approximate method for atomistic analysis of alloys specifically tailored to effectively deal with the current challenges in the area of atomistic modeling and to support modern experimental programs. During the grant period, the program benefited from steady growth which, as detailed below, far exceeds its original set of goals and objectives. Not surprisingly, by the end of this grant, the methodology and the computational materials program became an established force in the materials communitiy, with substantial impact in several areas. Major achievements during the duration of the grant include the completion of a Level 1 Milestone for the HITEMP program at NASA Glenn, consisting of the planning, development and organization of an international conference held at the Ohio Aerospace Institute in August of 2002, finalizing a period of rapid insertion of the methodology in the research community worlwide. The conference, attended by citizens of 17 countries representing various fields of the research community, resulted in a special issue of the leading journal in the area of applied surface science. Another element of the Level 1 Milestone was the presentation of the first version of the Alloy Design Workbench software package, currently known as "adwTools". This software package constitutes the first PC-based piece of software for atomistic simulations for both solid alloys and surfaces in the market.Dissemination of results and insertion in the materials community worldwide was a primary focus during this period. As a result, the P.I. was responsible for presenting 37 contributed talks, 19 invited talks, and publishing 71 articles in peer-reviewed journals, as detailed later in this Report.
NASA Technical Reports Server (NTRS)
Walls, Laurie K.; Kirk, Daniel; deLuis, Kavier; Haberbusch, Mark S.
2011-01-01
As space programs increasingly investigate various options for long duration space missions the accurate prediction of propellant behavior over long periods of time in microgravity environment has become increasingly imperative. This has driven the development of a detailed, physics-based understanding of slosh behavior of cryogenic propellants over a range of conditions and environments that are relevant for rocket and space storage applications. Recent advancements in computational fluid dynamics (CFD) models and hardware capabilities have enabled the modeling of complex fluid behavior in microgravity environment. Historically, launch vehicles with moderate duration upper stage coast periods have contained very limited instrumentation to quantify propellant stratification and boil-off in these environments, thus the ability to benchmark these complex computational models is of great consequence. To benchmark enhanced CFD models, recent work focuses on establishing an extensive experimental database of liquid slosh under a wide range of relevant conditions. In addition, a mass gauging system specifically designed to provide high fidelity measurements for both liquid stratification and liquid/ullage position in a micro-gravity environment has been developed. This pUblication will summarize the various experimental programs established to produce this comprehensive database and unique flight measurement techniques.
Toward computer simulation of high-LET in vitro survival curves.
Heuskin, A-C; Michiels, C; Lucas, S
2013-09-21
We developed a Monte Carlo based computer program called MCSC (Monte Carlo Survival Curve) able to predict the survival fraction of cells irradiated in vitro with a broad beam of high linear energy transfer particles. Three types of cell responses are studied: the usual high dose response, the bystander effect and the low-dose hypersensitivity (HRS). The program models the broad beam irradiation and double strand break distribution following Poisson statistics. The progression of cells through the cell cycle is taken into account while the repair takes place. Input parameters are experimentally determined for A549 lung carcinoma cells irradiated with 10 and 20 keV µm(-1) protons, 115 keV µm(-1) alpha particles and for EAhy926 endothelial cells exposed to 115 keV µm(-1) alpha particles. Results of simulations are presented and compared with experimental survival curves obtained for A549 and EAhy296 cells. Results are in good agreement with experimental data for both cell lines and all irradiation protocols. The benefits of MCSC are several: the gain of time that would have been spent performing time-consuming clonogenic assays, the capacity to estimate survival fraction of cell lines not forming colonies and possibly the evaluation of radiosensitivity parameters of given individuals.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Jones, Eric D.
1999-06-17
In the world of computer-based data acquisition and control, the graphical interface program LabVIEW from National Instruments is so ubiquitous that in many ways it has almost become the laboratory standard. To date, there have been approximately fifteen books concerning LabVIEW, but Professor Essick's treatise takes on a completely different tack than all of the previous discussions. In the more standard treatments of the ways and wherefores of LabVIEW such as LabVIEW Graphical Programming: Practical Applications in Instrumentation and Control by Gary W. Johnson (McGraw Hill, NY 1997), the emphasis has been instructing the reader how to program LabVIEW tomore » create a Virtual Instrument (VI) on the computer for interfacing to a particular instruments. LabVIEW is written in G a graphical programming language developed by National Instruments. In the past the emphasis has been on training the experimenter to learn G . Without going into details here, G incorporates the usual loops, arithmetic expressions, etc., found in many programming languages, but in an icon (graphical) environment. The net result being that LabVIEW contains all of the standard methods needed for interfacing to instruments, data acquisition, data analysis, graphics, and also methodology to incorporate programs written in other languages into LabVIEW. Historically, according to Professor Essick, he developed a series of experiments for an upper division laboratory course for computer-based instrumentation. His observation was that while many students had the necessary background in computer programming languages, there were students who had virtually no concept about writing a computer program let alone a computer- based interfacing program. Thus the beginnings of a concept for not only teaching computer- based instrumentation techniques, but aiso a method for the beginner to experience writing a com- puter program. Professor Essick saw LabVIEW as the perfect environment in which to teach computer-based research skills. With this goal in mind, he has succeeded admirably. Advanced LabVIEW Labs presents a series of chapters devoted to not only introducing the reader to LabVIEW, but also to the concepts necessary for writing a successful computer pro- gram. Each chapter is an assignment for the student and is suitable for a ten week course. The first topic introduces the while loop and waveform chart VI'S. After learning how to launch LabVIEW, the student then leans how to use LabVIEW functions such as sine and cosine. The beauty of thk and subsequent chapters, the student is introduced immediately to computer-based instruction by learning how to display the results in graph form on the screen. At each point along the way, the student is not only introduced to another LabVIEW operation, but also to such subjects as spread sheets for data storage, numerical integration, Fourier transformations', curve fitting algorithms, etc. The last few chapters conclude with the purpose of the learning module, and that is, com- puter-based instrumentation. Computer-based laboratory projects such as analog-to-digital con- version, digitizing oscilloscopes treated. Advanced Lab VIEW Labs finishes with a treatment on GPIB interfacing and finally, the student is asked to create an operating VI for temperature con- trol. This is an excellent text, not only as an treatise on LabVIEW but also as an introduction to computer programming logic. All programmers, who are struggling to not only learning how interface computers to instruments, but also trying understand top down programming and other programming language techniques, should add Advanced Lab-VIEW Labs to their computer library.« less
DOE Office of Scientific and Technical Information (OSTI.GOV)
Jones, Eric D.
1999-06-17
In the world of computer-based data acquisition and control, the graphical interface program LabVIEW from National Instruments is so ubiquitous that in many ways it has almost become the laboratory standard. To date, there have been approximately fifteen books concerning LabVIEW, but Professor Essick's treatise takes on a completely different tack than all of the previous discussions. In the more standard treatments of the ways and wherefores of LabVIEW such as LabVIEW Graphical Programming: Practical Applications in Instrumentation and Control by Gary W. Johnson (McGraw Hill, NY 1997), the emphasis has been instructing the reader how to program LabVIEW tomore » create a Virtual Instrument (VI) on the computer for interfacing to a particular instruments. LabVIEW is written in "G" a graphical programming language developed by National Instruments. In the past the emphasis has been on training the experimenter to learn "G". Without going into details here, "G" incorporates the usual loops, arithmetic expressions, etc., found in many programming languages, but in an icon (graphical) environment. The net result being that LabVIEW contains all of the standard methods needed for interfacing to instruments, data acquisition, data analysis, graphics, and also methodology to incorporate programs written in other languages into LabVIEW. Historically, according to Professor Essick, he developed a series of experiments for an upper division laboratory course for computer-based instrumentation. His observation was that while many students had the necessary background in computer programming languages, there were students who had virtually no concept about writing a computer program let alone a computer- based interfacing program. Thus the beginnings of a concept for not only teaching computer- based instrumentation techniques, but aiso a method for the beginner to experience writing a com- puter program. Professor Essick saw LabVIEW as the "perfect environment in which to teach computer-based research skills." With this goal in mind, he has succeeded admirably. Advanced LabVIEW Labs presents a series of chapters devoted to not only introducing the reader to LabVIEW, but also to the concepts necessary for writing a successful computer pro- gram. Each chapter is an assignment for the student and is suitable for a ten week course. The first topic introduces the while loop and waveform chart VI'S. After learning how to launch LabVIEW, the student then leans how to use LabVIEW functions such as sine and cosine. The beauty of thk and subsequent chapters, the student is introduced immediately to computer-based instruction by learning how to display the results in graph form on the screen. At each point along the way, the student is not only introduced to another LabVIEW operation, but also to such subjects as spread sheets for data storage, numerical integration, Fourier transformations', curve fitting algorithms, etc. The last few chapters conclude with the purpose of the learning module, and that is, com- puter-based instrumentation. Computer-based laboratory projects such as analog-to-digital con- version, digitizing oscilloscopes treated. Advanced Lab VIEW Labs finishes with a treatment on GPIB interfacing and finally, the student is asked to create an operating VI for temperature con- trol. This is an excellent text, not only as an treatise on LabVIEW but also as an introduction to computer programming logic. All programmers, who are struggling to not only learning how interface computers to instruments, but also trying understand top down programming and other programming language techniques, should add Advanced Lab-VIEW Labs to their computer library.« less
Double degree master program: Optical Design
NASA Astrophysics Data System (ADS)
Bakholdin, Alexey; Kujawinska, Malgorzata; Livshits, Irina; Styk, Adam; Voznesenskaya, Anna; Ezhova, Kseniia; Ermolayeva, Elena; Ivanova, Tatiana; Romanova, Galina; Tolstoba, Nadezhda
2015-10-01
Modern tendencies of higher education require development of master programs providing achievement of learning outcomes corresponding to quickly variable job market needs. ITMO University represented by Applied and Computer Optics Department and Optical Design and Testing Laboratory jointly with Warsaw University of Technology represented by the Institute of Micromechanics and Photonics at The Faculty of Mechatronics have developed a novel international master double-degree program "Optical Design" accumulating the expertise of both universities including experienced teaching staff, educational technologies, and experimental resources. The program presents studies targeting research and professional activities in high-tech fields connected with optical and optoelectronics devices, optical engineering, numerical methods and computer technologies. This master program deals with the design of optical systems of various types, assemblies and layouts using computer modeling means; investigation of light distribution phenomena; image modeling and formation; development of optical methods for image analysis and optical metrology including optical testing, materials characterization, NDT and industrial control and monitoring. The goal of this program is training a graduate capable to solve a wide range of research and engineering tasks in optical design and metrology leading to modern manufacturing and innovation. Variability of the program structure provides its flexibility and adoption according to current job market demands and personal learning paths for each student. In addition considerable proportion of internship and research expands practical skills. Some special features of the "Optical Design" program which implements the best practices of both Universities, the challenges and lessons learnt during its realization are presented in the paper.
Meta-T: TetrisⓇ as an experimental paradigm for cognitive skills research.
Lindstedt, John K; Gray, Wayne D
2015-12-01
Studies of human performance in complex tasks using video games are an attractive prospect, but many existing games lack a comprehensive way to modify the game and track performance beyond basic levels of analysis. Meta-T provides experimenters a tool to study behavior in a dynamic task environment with time-stressed decision-making and strong perceptual-motor elements, offering a host of experimental manipulations with a robust and detailed logging system for all user events, system events, and screen objects. Its experimenter-friendly interface provides control over detailed parameters of the task environment without need for programming expertise. Support for eye-tracking and computational cognitive modeling extend the paradigm's scope.
A DNA network as an information processing system.
Santini, Cristina Costa; Bath, Jonathan; Turberfield, Andrew J; Tyrrell, Andy M
2012-01-01
Biomolecular systems that can process information are sought for computational applications, because of their potential for parallelism and miniaturization and because their biocompatibility also makes them suitable for future biomedical applications. DNA has been used to design machines, motors, finite automata, logic gates, reaction networks and logic programs, amongst many other structures and dynamic behaviours. Here we design and program a synthetic DNA network to implement computational paradigms abstracted from cellular regulatory networks. These show information processing properties that are desirable in artificial, engineered molecular systems, including robustness of the output in relation to different sources of variation. We show the results of numerical simulations of the dynamic behaviour of the network and preliminary experimental analysis of its main components.
NASA Technical Reports Server (NTRS)
Dow, J. W.
1972-01-01
A numerical solution of the turbulent mass transport equation utilizing the concept of eddy diffusivity is presented as an efficient method of investigating turbulent mass transport in boundary layer type flows. A FORTRAN computer program is used to study the two-dimensional diffusion of ammonia, from a line source on the surface, into a turbulent boundary layer over a flat plate. The results of the numerical solution are compared with experimental data to verify the results of the solution. Several other solutions to diffusion problems are presented to illustrate the versatility of the computer program and to provide some insight into the problem of mass diffusion as a whole.
Einstein coefficients and oscillator strengths for low lying state of CO molecules
NASA Astrophysics Data System (ADS)
Swer, S.; Syiemiong, A.; Ram, M.; Jha, A. K.; Saxena, A.
2018-04-01
Einstein Coefficients and Oscillator Strengths for different state of CO molecule have been calculated using LEROY'S LEVEL program and MOLCAS ab initio code. Using the wave function derived from Morse potential and transition dipole moment obtained from ab initio calculation, The potential energy functions were computed for these states using the spectroscopic constants. The Morse potential of these states and electronic transition dipole moment of the transition calculated in a recent ab initio study have been used in LEVEL program to produce transition dipole matrix element for a large number of bands. Einstein Coefficients have also been used to compute the radiative lifetimes of several vibrational levels and the calculated values are compared with other theoretical results and experimental values.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Deline, C.
Computer modeling is able to predict the performance of distributed power electronics (microinverters, power optimizers) in PV systems. However, details about partial shade and other mismatch must be known in order to give the model accurate information to go on. This talk will describe recent updates in NREL’s System Advisor Model program to model partial shading losses with and without distributed power electronics, along with experimental validation results. Computer modeling is able to predict the performance of distributed power electronics (microinverters, power optimizers) in PV systems. However, details about partial shade and other mismatch must be known in order tomore » give the model accurate information to go on. This talk will describe recent updates in NREL’s System Advisor Model program to model partial shading losses.« less
The use of experimental structures to model protein dynamics.
Katebi, Ataur R; Sankar, Kannan; Jia, Kejue; Jernigan, Robert L
2015-01-01
The number of solved protein structures submitted in the Protein Data Bank (PDB) has increased dramatically in recent years. For some specific proteins, this number is very high-for example, there are over 550 solved structures for HIV-1 protease, one protein that is essential for the life cycle of human immunodeficiency virus (HIV) which causes acquired immunodeficiency syndrome (AIDS) in humans. The large number of structures for the same protein and its variants include a sample of different conformational states of the protein. A rich set of structures solved experimentally for the same protein has information buried within the dataset that can explain the functional dynamics and structural mechanism of the protein. To extract the dynamics information and functional mechanism from the experimental structures, this chapter focuses on two methods-Principal Component Analysis (PCA) and Elastic Network Models (ENM). PCA is a widely used statistical dimensionality reduction technique to classify and visualize high-dimensional data. On the other hand, ENMs are well-established simple biophysical method for modeling the functionally important global motions of proteins. This chapter covers the basics of these two. Moreover, an improved ENM version that utilizes the variations found within a given set of structures for a protein is described. As a practical example, we have extracted the functional dynamics and mechanism of HIV-1 protease dimeric structure by using a set of 329 PDB structures of this protein. We have described, step by step, how to select a set of protein structures, how to extract the needed information from the PDB files for PCA, how to extract the dynamics information using PCA, how to calculate ENM modes, how to measure the congruency between the dynamics computed from the principal components (PCs) and the ENM modes, and how to compute entropies using the PCs. We provide the computer programs or references to software tools to accomplish each step and show how to use these programs and tools. We also include computer programs to generate movies based on PCs and ENM modes and describe how to visualize them.
The Use of Experimental Structures to Model Protein Dynamics
Katebi, Ataur R.; Sankar, Kannan; Jia, Kejue; Jernigan, Robert L.
2014-01-01
Summary The number of solved protein structures submitted in the Protein Data Bank (PDB) has increased dramatically in recent years. For some specific proteins, this number is very high – for example, there are over 550 solved structures for HIV-1 protease, one protein that is essential for the life cycle of human immunodeficiency virus (HIV) which causes acquired immunodeficiency syndrome (AIDS) in humans. The large number of structures for the same protein and its variants include a sample of different conformational states of the protein. A rich set of structures solved experimentally for the same protein has information buried within the dataset that can explain the functional dynamics and structural mechanism of the protein. To extract the dynamics information and functional mechanism from the experimental structures, this chapter focuses on two methods – Principal Component Analysis (PCA) and Elastic Network Models (ENM). PCA is a widely used statistical dimensionality reduction technique to classify and visualize high-dimensional data. On the other hand, ENMs are well-established simple biophysical method for modeling the functionally important global motions of proteins. This chapter covers the basics of these two. Moreover, an improved ENM version that utilizes the variations found within a given set of structures for a protein is described. As a practical example, we have extracted the functional dynamics and mechanism of HIV-1 protease dimeric structure by using a set of 329 PDB structures of this protein. We have described, step by step, how to select a set of protein structures, how to extract the needed information from the PDB files for PCA, how to extract the dynamics information using PCA, how to calculate ENM modes, how to measure the congruency between the dynamics computed from the principal components (PCs) and the ENM modes, and how to compute entropies using the PCs. We provide the computer programs or references to software tools to accomplish each step and show how to use these programs and tools. We also include computer programs to generate movies based on PCs and ENM modes and describe how to visualize them. PMID:25330965
Analysis of helium-ion scattering with a desktop computer
NASA Astrophysics Data System (ADS)
Butler, J. W.
1986-04-01
This paper describes a program written in an enhanced BASIC language for a desktop computer, for simulating the energy spectra of high-energy helium ions scattered into two concurrent detectors (backward and glancing). The program is designed for 512-channel spectra from samples containing up to 8 elements and 55 user-defined layers. The program is intended to meet the needs of analyses in materials sciences, such as metallurgy, where more than a few elements may be present, where several elements may be near each other in the periodic table, and where relatively deep structure may be important. These conditions preclude the use of completely automatic procedures for obtaining the sample composition directly from the scattered ion spectrum. Therefore, efficient methods are needed for entering and editing large amounts of composition data, with many iterations and with much feedback of information from the computer to the user. The internal video screen is used exclusively for verbal and numeric communications between user and computer. The composition matrix is edited on screen with a two-dimension forms-fill-in text editor and with many automatic procedures, such as doubling the number of layers with appropriate interpolations and extrapolations. The control center of the program is a bank of 10 keys that initiate on-event branching of program flow. The experimental and calculated spectra, including those of individual elements if desired, are displayed on an external color monitor, with an optional inset plot of the depth concentration profiles of the elements in the sample.
Incorporating computational resources in a cancer research program
Woods, Nicholas T.; Jhuraney, Ankita; Monteiro, Alvaro N.A.
2015-01-01
Recent technological advances have transformed cancer genetics research. These advances have served as the basis for the generation of a number of richly annotated datasets relevant to the cancer geneticist. In addition, many of these technologies are now within reach of smaller laboratories to answer specific biological questions. Thus, one of the most pressing issues facing an experimental cancer biology research program in genetics is incorporating data from multiple sources to annotate, visualize, and analyze the system under study. Fortunately, there are several computational resources to aid in this process. However, a significant effort is required to adapt a molecular biology-based research program to take advantage of these datasets. Here, we discuss the lessons learned in our laboratory and share several recommendations to make this transition effectively. This article is not meant to be a comprehensive evaluation of all the available resources, but rather highlight those that we have incorporated into our laboratory and how to choose the most appropriate ones for your research program. PMID:25324189
NASA Technical Reports Server (NTRS)
Carlson, Harry W.
1985-01-01
The purpose here is to show how two linearized theory computer programs in combination may be used for the design of low speed wing flap systems capable of high levels of aerodynamic efficiency. A fundamental premise of the study is that high levels of aerodynamic performance for flap systems can be achieved only if the flow about the wing remains predominantly attached. Based on this premise, a wing design program is used to provide idealized attached flow camber surfaces from which candidate flap systems may be derived, and, in a following step, a wing evaluation program is used to provide estimates of the aerodynamic performance of the candidate systems. Design strategies and techniques that may be employed are illustrated through a series of examples. Applicability of the numerical methods to the analysis of a representative flap system (although not a system designed by the process described here) is demonstrated in a comparison with experimental data.
NASA Technical Reports Server (NTRS)
Gallardo, V. C.; Storace, A. S.; Gaffney, E. F.; Bach, L. J.; Stallone, M. J.
1981-01-01
The component element method was used to develop a transient dynamic analysis computer program which is essentially based on modal synthesis combined with a central, finite difference, numerical integration scheme. The methodology leads to a modular or building-block technique that is amenable to computer programming. To verify the analytical method, turbine engine transient response analysis (TETRA), was applied to two blade-out test vehicles that had been previously instrumented and tested. Comparison of the time dependent test data with those predicted by TETRA led to recommendations for refinement or extension of the analytical method to improve its accuracy and overcome its shortcomings. The development of working equations, their discretization, numerical solution scheme, the modular concept of engine modelling, the program logical structure and some illustrated results are discussed. The blade-loss test vehicles (rig full engine), the type of measured data, and the engine structural model are described.
BLISS: a computer program for the protection of blood donors. Technical report
DOE Office of Scientific and Technical Information (OSTI.GOV)
Catsimpoolas, N.; Cooke, C.; Valeri, C.R.
1982-06-28
A BASIC program has been developed for the Hewlett-Packard Model 9845 desk-top computer which allows the creation of blood donor files for subsequent retrieval, update, and correction. A similar modified version was developed for hte HP 9835 Model. This software system has been called BLISS which stands for Blood Information and Security System. In addition to its function as a file management system, BLISS provides warnings before a donation is performed to protect the donor from excessive exposure to radioactivity and DMSO levels, from too frequent of donations of blood, and from adverse reactions. The program can also be usedmore » to select donors who have participated in specific studies and to list the experimental details which have been stored in the file. The BLISS system has been actively utilized at the Naval Blood Research Laboratory in Boston and contains the files of over 750 donors.« less
Testing by artificial intelligence: computational alternatives to the determination of mutagenicity.
Klopman, G; Rosenkranz, H S
1992-08-01
In order to develop methods for evaluating the predictive performance of computer-driven structure-activity methods (SAR) as well as to determine the limits of predictivity, we investigated the behavior of two Salmonella mutagenicity data bases: (a) a subset from the Genetox Program and (b) one from the U.S. National Toxicology Program (NTP). For molecules common to the two data bases, the experimental concordance was 76% when "marginals" were included and 81% when they were excluded. Three SAR methods were evaluated: CASE, MULTICASE and CASE/Graph Indices (CASE/GI). The programs "learned" the Genetox data base and used it to predict NTP molecules that were not present in the Genetox compilation. The concordances were 72, 80 and 47% respectively. Obviously, the MULTICASE version is superior and approaches the 85% interlaboratory variability observed for the Salmonella mutagenicity assays when the latter was carried out under carefully controlled conditions.
Durability evaluation of ceramic components using CARES/LIFE
NASA Technical Reports Server (NTRS)
Nemeth, Noel N.; Powers, Lynn M.; Janosik, Lesley A.; Gyekenyesi, John P.
1994-01-01
The computer program CARES/LIFE calculates the time-dependent reliability of monolithic ceramic components subjected to thermomechanical and/or proof test loading. This program is an extension of the CARES (Ceramics Analysis and Reliability Evaluation of Structures) computer program. CARES/LIFE accounts for the phenomenon of subcritical crack growth (SCG) by utilizing the power law, Paris law, or Walker equation. The two-parameter Weibull cumulative distribution function is used to characterize the variation in component strength. The effects of multiaxial stresses are modeled using either the principle of independent action (PIA), the Weibull normal stress averaging method (NSA), or the Batdorf theory. Inert strength and fatigue parameters are estimated from rupture strength data of naturally flawed specimens loaded in static, dynamic, or cyclic fatigue. Application of this design methodology is demonstrated using experimental data from alumina bar and disk flexure specimens which exhibit SCG when exposed to water.
Durability evaluation of ceramic components using CARES/LIFE
DOE Office of Scientific and Technical Information (OSTI.GOV)
Nemeth, N.N.; Janosik, L.A.; Gyekenyesi, J.P.
1996-01-01
The computer program CARES/LIFE calculates the time-dependent reliability of monolithic ceramic components subjected to thermomechanical and/or proof test loading. This program is an extension of the CARES (Ceramics Analysis and Reliability Evaluation of Structures) computer program. CARES/LIFE accounts for the phenomenon of subcritical crack growth (SCG) by utilizing the power law, Paris law, or Walker equation. The two-parameter Weibull cumulative distribution function is used to characterize the variation in component strength. The effects of multiaxial stresses are modeled using either the principle of independent action (PIA), the Weibull normal stress averaging method (NSA), or the Batdorf theory. Inert strength andmore » fatigue parameters are estimated from rupture strength data of naturally flawed specimens loaded in static, dynamic, or cyclic fatigue. Application of this design methodology is demonstrated using experimental data from alumina bar and disk flexure specimens, which exhibit SCG when exposed to water.« less
Optimization Techniques for Analysis of Biological and Social Networks
2012-03-28
analyzing a new metaheuristic technique, variable objective search. 3. Experimentation and application: Implement the proposed algorithms , test and fine...alternative mathematical programming formulations, their theoretical analysis, the development of exact algorithms , and heuristics. Originally, clusters...systematic fashion under a unifying theoretical and algorithmic framework. Optimization, Complex Networks, Social Network Analysis, Computational
The Effects of a Computerized Study Program on the Acquisition of Science Vocabulary
ERIC Educational Resources Information Center
Rollins, Karen F.
2012-01-01
The following study examined the difference in science vocabulary acquisition comparing computer-assisted learning and a traditional study review sheet. Fourth and fifth grade students from a suburban school in central Texas were randomly selected and randomly assigned to either experimental group or control group. Both groups were given a…
Pasadena City College SIGI Project Research Design. Pilot Study.
ERIC Educational Resources Information Center
Risser, John J.; Tulley, John E.
A pilot study evaluation of SIGI (System of Interactive Guidance and Information) at Pasadena City College in 1974-75 tested the effectiveness of an experimental research design for an expanded field test of the system the following year. (SIGI is a computer based career guidance program designed by Educational Testing Service to assist community…
Noise and Vibration Risk Prevention Virtual Web for Ubiquitous Training
ERIC Educational Resources Information Center
Redel-Macías, María Dolores; Cubero-Atienza, Antonio J.; Martínez-Valle, José Miguel; Pedrós-Pérez, Gerardo; del Pilar Martínez-Jiménez, María
2015-01-01
This paper describes a new Web portal offering experimental labs for ubiquitous training of university engineering students in work-related risk prevention. The Web-accessible computer program simulates the noise and machine vibrations met in the work environment, in a series of virtual laboratories that mimic an actual laboratory and provide the…
The Use of a Computer Simulation to Promote Conceptual Change: A Quasi-Experimental Study
ERIC Educational Resources Information Center
Trundle, Kathy Cabe; Bell, Randy L.
2010-01-01
This mixed-methods investigation compared the effectiveness of three instructional approaches in achieving desired conceptual change among early childhood preservice teachers (n = 157). Each of the three treatments employed inquiry-based instruction on moon phases using data collected from: (1) the planetarium software program, Starry Night[TM],…
NASIC at MIT. Final Report, 1 March 1974 through 28 February 1975.
ERIC Educational Resources Information Center
Benenfeld, Alan R.; And Others
Computer-based reference search services were provided to users on a fee-for-service basis at the Massachusetts Institute of Technology as the first, and experimental, note in the development of the Northeast Academic Science Information Center (NASIC). Development of a training program for information specialists and training materials is…
NASA Technical Reports Server (NTRS)
Staveland, Lowell
1994-01-01
This is the experimental and software detailed design report for the prototype task loading model (TLM) developed as part of the man-machine integration design and analysis system (MIDAS), as implemented and tested in phase 6 of the Army-NASA Aircrew/Aircraft Integration (A3I) Program. The A3I program is an exploratory development effort to advance the capabilities and use of computational representations of human performance and behavior in the design, synthesis, and analysis of manned systems. The MIDAS TLM computationally models the demands designs impose on operators to aide engineers in the conceptual design of aircraft crewstations. This report describes TLM and the results of a series of experiments which were run this phase to test its capabilities as a predictive task demand modeling tool. Specifically, it includes discussions of: the inputs and outputs of TLM, the theories underlying it, the results of the test experiments, the use of the TLM as both stand alone tool and part of a complete human operator simulation, and a brief introduction to the TLM software design.
Fernandez, Elizabeth; Bergado Rosado, Jorge A.; Rodriguez Perez, Daymi; Salazar Santana, Sonia; Torres Aguilar, Maydane; Bringas, Maria Luisa
2017-01-01
Many training programs have been designed using modern software to restore the impaired cognitive functions in patients with acquired brain damage (ABD). The objective of this study was to evaluate the effectiveness of a computer-based training program of attention and memory in patients with ABD, using a two-armed parallel group design, where the experimental group (n = 50) received cognitive stimulation using RehaCom software, and the control group (n = 30) received the standard cognitive stimulation (non-computerized) for eight weeks. In order to assess the possible cognitive changes after the treatment, a post-pre experimental design was employed using the following neuropsychological tests: Wechsler Memory Scale (WMS) and Trail Making test A and B. The effectiveness of the training procedure was statistically significant (p < 0.05) when it established the comparison between the performance in these scales, before and after the training period, in each patient and between the two groups. The training group had statistically significant (p < 0.001) changes in focused attention (Trail A), two subtests (digit span and logical memory), and the overall score of WMS. Finally, we discuss the advantages of computerized training rehabilitation and further directions of this line of work. PMID:29301194
Arora, Sanjeevani; Huwe, Peter J.; Sikder, Rahmat; Shah, Manali; Browne, Amanda J.; Lesh, Randy; Nicolas, Emmanuelle; Deshpande, Sanat; Hall, Michael J.; Dunbrack, Roland L.; Golemis, Erica A.
2017-01-01
ABSTRACT The cancer-predisposing Lynch Syndrome (LS) arises from germline mutations in DNA mismatch repair (MMR) genes, predominantly MLH1, MSH2, MSH6, and PMS2. A major challenge for clinical diagnosis of LS is the frequent identification of variants of uncertain significance (VUS) in these genes, as it is often difficult to determine variant pathogenicity, particularly for missense variants. Generic programs such as SIFT and PolyPhen-2, and MMR gene-specific programs such as PON-MMR and MAPP-MMR, are often used to predict deleterious or neutral effects of VUS in MMR genes. We evaluated the performance of multiple predictive programs in the context of functional biologic data for 15 VUS in MLH1, MSH2, and PMS2. Using cell line models, we characterized VUS predicted to range from neutral to pathogenic on mRNA and protein expression, basal cellular viability, viability following treatment with a panel of DNA-damaging agents, and functionality in DNA damage response (DDR) signaling, benchmarking to wild-type MMR proteins. Our results suggest that the MMR gene-specific classifiers do not always align with the experimental phenotypes related to DDR. Our study highlights the importance of complementary experimental and computational assessment to develop future predictors for the assessment of VUS. PMID:28494185
Validation of NASA Thermal Ice Protection Computer Codes. Part 1; Program Overview
NASA Technical Reports Server (NTRS)
Miller, Dean; Bond, Thomas; Sheldon, David; Wright, William; Langhals, Tammy; Al-Khalil, Kamel; Broughton, Howard
1996-01-01
The Icing Technology Branch at NASA Lewis has been involved in an effort to validate two thermal ice protection codes developed at the NASA Lewis Research Center. LEWICE/Thermal (electrothermal deicing & anti-icing), and ANTICE (hot-gas & electrothermal anti-icing). The Thermal Code Validation effort was designated as a priority during a 1994 'peer review' of the NASA Lewis Icing program, and was implemented as a cooperative effort with industry. During April 1996, the first of a series of experimental validation tests was conducted in the NASA Lewis Icing Research Tunnel(IRT). The purpose of the April 96 test was to validate the electrothermal predictive capabilities of both LEWICE/Thermal, and ANTICE. A heavily instrumented test article was designed and fabricated for this test, with the capability of simulating electrothermal de-icing and anti-icing modes of operation. Thermal measurements were then obtained over a range of test conditions, for comparison with analytical predictions. This paper will present an overview of the test, including a detailed description of: (1) the validation process; (2) test article design; (3) test matrix development; and (4) test procedures. Selected experimental results will be presented for de-icing and anti-icing modes of operation. Finally, the status of the validation effort at this point will be summarized. Detailed comparisons between analytical predictions and experimental results are contained in the following two papers: 'Validation of NASA Thermal Ice Protection Computer Codes: Part 2- The Validation of LEWICE/Thermal' and 'Validation of NASA Thermal Ice Protection Computer Codes: Part 3-The Validation of ANTICE'
Development of a prototype automatic controller for liquid cooling garment inlet temperature
NASA Technical Reports Server (NTRS)
Weaver, C. S.; Webbon, B. W.; Montgomery, L. D.
1982-01-01
The development of a computer control of a liquid cooled garment (LCG) inlet temperature is descirbed. An adaptive model of the LCG is used to predict the heat-removal rates for various inlet temperatures. An experimental system that contains a microcomputer was constructed. The LCG inlet and outlet temperatures and the heat exchanger outlet temperature form the inputs to the computer. The adaptive model prediction method of control is successful during tests where the inlet temperature is automatically chosen by the computer. It is concluded that the program can be implemented in a microprocessor of a size that is practical for a life support back-pack.
Internal fluid mechanics research on supercomputers for aerospace propulsion systems
NASA Technical Reports Server (NTRS)
Miller, Brent A.; Anderson, Bernhard H.; Szuch, John R.
1988-01-01
The Internal Fluid Mechanics Division of the NASA Lewis Research Center is combining the key elements of computational fluid dynamics, aerothermodynamic experiments, and advanced computational technology to bring internal computational fluid mechanics (ICFM) to a state of practical application for aerospace propulsion systems. The strategies used to achieve this goal are to: (1) pursue an understanding of flow physics, surface heat transfer, and combustion via analysis and fundamental experiments, (2) incorporate improved understanding of these phenomena into verified 3-D CFD codes, and (3) utilize state-of-the-art computational technology to enhance experimental and CFD research. Presented is an overview of the ICFM program in high-speed propulsion, including work in inlets, turbomachinery, and chemical reacting flows. Ongoing efforts to integrate new computer technologies, such as parallel computing and artificial intelligence, into high-speed aeropropulsion research are described.
Computing with dynamical systems based on insulator-metal-transition oscillators
NASA Astrophysics Data System (ADS)
Parihar, Abhinav; Shukla, Nikhil; Jerry, Matthew; Datta, Suman; Raychowdhury, Arijit
2017-04-01
In this paper, we review recent work on novel computing paradigms using coupled oscillatory dynamical systems. We explore systems of relaxation oscillators based on linear state transitioning devices, which switch between two discrete states with hysteresis. By harnessing the dynamics of complex, connected systems, we embrace the philosophy of "let physics do the computing" and demonstrate how complex phase and frequency dynamics of such systems can be controlled, programmed, and observed to solve computationally hard problems. Although our discussion in this paper is limited to insulator-to-metallic state transition devices, the general philosophy of such computing paradigms can be translated to other mediums including optical systems. We present the necessary mathematical treatments necessary to understand the time evolution of these systems and demonstrate through recent experimental results the potential of such computational primitives.
Effect of Automatic Processing on Specification of Problem Solutions for Computer Programs.
1981-03-01
Number 7 ± 2" item limitaion on human short-term memory capability (Miller, 1956) should be a guiding principle in program design. Yourdon and...input either a single example solution or multiple exam’- le solutions in sequence. If a participant’s P1 has a low value - near 0 - it may be concluded... Principles in Experimental Design, Winer ,1971). 55 Table 12 ANOVA Resultt, For Performance Measure 2 Sb DF MS F Source of Variation Between Subjects
1978-07-01
were input into the computer program. The program was numerically intergrated with time by using a fourth-order Runge-Kutta integration algorithm with...equations of motion are numerically intergrated to provide time histories of the aircraft spinning motion. A.2 EQUATIONS DEFINING THE FORCE AND MOMENT...by Cy or Cn. 50 AE DC-TR-77-126 A . 4 where EQUATIONS FOR TRANSFERRING AERODYNAMIC DATA INPUTS TO THE PROPER HORIZONTAL CENTER OF GRAVITY
NASA Technical Reports Server (NTRS)
Andrews, E. H., Jr.; Mackley, E. A.
1976-01-01
The NASA Hypersonic Research Engine (HRE) Project was initiated for the purpose of advancing the technology of airbreathing propulsion for hypersonic flight. A large component (inlet, combustor, and nozzle) and structures development program was encompassed by the project. The tests of a full-scale (18 in. diameter cowl and 87 in. long) HRE concept, designated the Aerothermodynamic Integration Model (AIM), at Mach numbers of 5, 6, and 7. Computer program results for Mach 6 component integration tests are presented.
Mathematical modeling of high and low temperature heat pipes
NASA Technical Reports Server (NTRS)
Chi, S. W.
1971-01-01
Mathematical models are developed for calculating heat-transfer limitations of high-temperature heat pipes and heat-transfer limitations and temperature gradient of low temperature heat pipes. Calculated results are compared with the available experimental data from various sources to increase confidence in the present math models. Complete listings of two computer programs for high- and low-temperature heat pipes respectively are appended. These programs enable the performance of heat pipes with wrapped-screen, rectangular-groove or screen-covered rectangular-groove wick to be predicted.
ERIC Educational Resources Information Center
Saktanli, S. Cem
2011-01-01
This experimental study was done to see if using computer supported notation and vocalization program for teaching songs instead of using block flute accompanied song teaching has any significant effect on students' singing behavior. The study group is composed of the 5th, 6th and 7th graders of 2008-2009 educational term in T.O.K.I. Yahya Kemal…
Summary of nozzle-exhaust plume flowfield analyses related to space shuttle applications
NASA Technical Reports Server (NTRS)
Penny, M. M.
1975-01-01
Exhaust plume shape simulation is studied, with the major effort directed toward computer program development and analytical support of various plume related problems associated with the space shuttle. Program development centered on (1) two-phase nozzle-exhaust plume flows, (2) plume impingement, and (3) support of exhaust plume simulation studies. Several studies were also conducted to provide full-scale data for defining exhaust plume simulation criteria. Model nozzles used in launch vehicle test were analyzed and compared to experimental calibration data.
Light aircraft crash safety program
NASA Technical Reports Server (NTRS)
Thomson, R. G.; Hayduk, R. J.
1974-01-01
NASA is embarked upon research and development tasks aimed at providing the general aviation industry with a reliable crashworthy airframe design technology. The goals of the NASA program are: reliable analytical techniques for predicting the nonlinear behavior of structures; significant design improvements of airframes; and simulated full-scale crash test data. The analytical tools will include both simplified procedures for estimating energy absorption characteristics and more complex computer programs for analysis of general airframe structures under crash loading conditions. The analytical techniques being developed both in-house and under contract are described, and a comparison of some analytical predictions with experimental results is shown.
Application of a neural network to simulate analysis in an optimization process
NASA Technical Reports Server (NTRS)
Rogers, James L.; Lamarsh, William J., II
1992-01-01
A new experimental software package called NETS/PROSSS aimed at reducing the computing time required to solve a complex design problem is described. The software combines a neural network for simulating the analysis program with an optimization program. The neural network is applied to approximate results of a finite element analysis program to quickly obtain a near-optimal solution. Results of the NETS/PROSSS optimization process can also be used as an initial design in a normal optimization process and make it possible to converge to an optimum solution with significantly fewer iterations.
Joint University Program for Air Transportation Research, 1988-1989
NASA Technical Reports Server (NTRS)
Morrell, Frederick R. (Compiler)
1990-01-01
The research conducted during 1988 to 1989 under the NASA/FAA-sponsored Joint University Program for Air Transportation Research is summarized. The Joint University Program is a coordinated set of three grants sponsored by NASA Langley Research Center and the Federal Aviation Administration, one each with the Massachusetts Institute of Technology, Ohio University, and Princeton University. Completed works, status reports, and annotated bibliographies are presented for research topics, which include computer science, guidance and control theory and practice, aircraft performance, flight dynamics, and applied experimental psychology. An overview of the year's activities for each university is also presented.
Microgravity sciences application visiting scientist program
NASA Technical Reports Server (NTRS)
1994-01-01
Contract NAS8-38785, Microgravity Experimental and Theoretical Research, is a project involving a large number of individual research programs related to: determination of the structure of human serum albumin and other biomedically important proteins; analysis of thermodynamic properties of various proteins and models of protein nucleation; development of experimental techniques for the growth of protein crystals in space; study of the physics of electrical double layers in the mechanics of liquid interfaces; computational analysis of vapor crystal growth processes in microgravity; analysis of the influence of magnetic fields in damping residual flows in directional solidification processes; crystal growth and characterization of II-VI semiconductor alloys; and production of thin films for nonlinear optics. It is not intended that the programs will be necessarily limited to this set at any one time. The visiting scientists accomplishing these programs shall serve on-site at MSFC to take advantage of existing laboratory facilities and the daily opportunities for technical communications with various senior scientists.
A simulation exercise of a cavity-type solar receiver using the HEAP program
NASA Technical Reports Server (NTRS)
Lansing, F. L.
1979-01-01
A computer program has been developed at JPL to support the advanced studies of solar receivers in high concentration solar-thermal-electric power plants. This work presents briefly the program methodology, input data required, expected output results, capabilities and limitations. The program was used to simulate an existing 5 kwt experimental receiver of a cavity type. The receiver is located at the focus of a paraboloid dish and is connected to a Stirling engine. Both steady state and transient performance simulation were given. Details about the receiver modeling were also presented to illustrate the procedure followed. Simulated temperature patterns were found in good agreement with test data obtained by high temperature thermocouples. The simulated receiver performance was extrapolated to various operating conditions not attained experimentally. The results of the parameterization study were fitted to a general performance expression to determine the receiver characteristic constraints. The latter were used to optimize the receiver operating conditions to obtain the highest overall conversion efficiency.
NASA Technical Reports Server (NTRS)
Blatt, M. H.; Bradshaw, R. D.; Risberg, J. A.
1980-01-01
Technology areas critical to the development of cryogenic capillary devices were studied. Passive cooling of capillary devices was investigated with an analytical and experimental study of wicking flow. Capillary device refilling with settled fluid was studied using an analytical and experimental program that resulted in successful correlation of a versatile computer program with test data. The program was used to predict Centaur D-1S LO2 and LH2 start basket refilling. Comparisons were made between the baseline Centaur D-1S propellant feed system and feed system alternatives including systems using capillary devices. The preferred concepts from the Centaur D-1S study were examined for APOTV and POTV vehicles for delivery and round trip transfer of payloads between LEO and GEO. Mission profiles were determined to provide propellant usage timelines and the payload partials were defined.
Romanian MRE Rocket Engines Program - An Early Endeavor
NASA Astrophysics Data System (ADS)
Rugescu, R. E.
2002-01-01
(MRE) was initiated in the years '60 of the past century at the Chair of Aerospace Sciences "Elie Carafoli" from the "Politehnica" University in Bucharest (PUB). Consisting of theoretical and experimental investigations in the form of computational methods and technological solutions for small size MRE-s and the concept of the test stand for these engines, the program ended in the construction of the first Romanian liquid rocket motors. Hermann Oberth and Dorin Pavel, were known from 1923, no experimental practice was yet tempted, at the time level of 1960. It was the intention of the developers at PUB to cover this gap and initiate a feasible, low-cost, demonstrative program of designing and testing experimental models of MRE. The research program was oriented towards future development of small size space carrier vehicles for scientific applications only, as an independent program with no connection to other defense programs imagined by the authorities in Bucharest, at that time. Consequently the entire financial support was assured by "Politehnica" university. computerized methods in the thermochemistry of heterogeneous combustion, for both steady and unsteady flows with chemical reactions and two phase flows. The research was gradually extended to the production of a professional CAD program for steady-state heat transfer simulations and the loading capacity analyses of the double wall, cooled thrust chamber. The resulting computer codes were run on a 360-30 IMB machine, beginning in 1968. Some of the computational methods were first exposed at the 9th International Conference on Applied Mechanics, held in Bucharest between June 23-27, 1969. hot testing of a series of storable propellant, variable thrust, variable geometry, liquid rocket motors, with a maximal thrust of 200N. A remotely controlled, portable test bad, actuated either automatically or manually and consisting of a 6-modules construction was built for this motor series, with a simple 8 analog-channel and 5 digital-channel data measuring and recording system. The first hot test firing of the MRE-1B motor took place successfully on April 9th, 1969 in Bucharest, at the "Elie Carafoli" Chair of UPB. The research program continued with the development of a series of solid, double base propellant rocket and ram-rocket motors, with emphasize on the optimization of the gasdynamic contour of the engine, in order to increase the flight performances. Increments of up to 8% in specific thrust were measured on the test stand, with mass savings and no extra costs. The test firing of the first Romanian, air-breathing ram-rocket engine took place successfully in august 1987 at the Chemical Works in Fagaras, Romania. Astronautics", founded in Bucharest. The principles and history of the "MRE" research program are presented in the proposed paper.
Solution of the lossy nonlinear Tricomi equation with application to sonic boom focusing
NASA Astrophysics Data System (ADS)
Salamone, Joseph A., III
Sonic boom focusing theory has been augmented with new terms that account for mean flow effects in the direction of propagation and also for atmospheric absorption/dispersion due to molecular relaxation due to oxygen and nitrogen. The newly derived model equation was numerically implemented using a computer code. The computer code was numerically validated using a spectral solution for nonlinear propagation of a sinusoid through a lossy homogeneous medium. An additional numerical check was performed to verify the linear diffraction component of the code calculations. The computer code was experimentally validated using measured sonic boom focusing data from the NASA sponsored Superboom Caustic and Analysis Measurement Program (SCAMP) flight test. The computer code was in good agreement with both the numerical and experimental validation. The newly developed code was applied to examine the focusing of a NASA low-boom demonstration vehicle concept. The resulting pressure field was calculated for several supersonic climb profiles. The shaping efforts designed into the signatures were still somewhat evident despite the effects of sonic boom focusing.
NASA Technical Reports Server (NTRS)
Bucinell, Ronald B.
1997-01-01
The objective of this project was to model the 5-3/4 inch pressure vessels used on the NASA RTOP program in an attempt to learn more about how impact damage forms and what are the residual effects of the resulting damage. A global-local finite element model was developed for the bottle and the states of stress in the bottles were determined down to the constituent level. The experimental data that was generated on the NASA RTOP program was not in a form that enabled the model developed under this grant to be correlated with the experimental data. As a result of this exercise it is recommended that an experimental program be designed using statistical design of experiment techniques to generate data that can be used to isolate the phenomenon that control the formation of impact damage. This data should include residual property determinations so that models for post impact structural integrity can be developed. It is also recommended that the global-local methodology be integrated directly into the finite element code. This will require considerable code development.
Environment of Space Interactions with Space Systems
NASA Technical Reports Server (NTRS)
2004-01-01
The primary product of this research project was a computer program named SAVANT. This program uses the Displacement Damage Dose (DDD) method of calculating radiation damage to solar cells. This calculation method was developed at the Naval Research Laboratory, and uses fundamental physical properties of the solar cell materials to predict radiation damage to the solar cells. This means that fewer experimental measurements are required to characterize the radiation damage to the cells, which results in a substantial cost savings to qualify solar cells for orbital missions. In addition, the DDD method makes it easier to characterize cells that are already being used, but have not been fully tested using the older technique of characterizing radiation damage. The computer program combines an orbit generator with NASA's AP-8 and AE-8 models of trapped protons and electrons. This allows the user to specify an orbit, and the program will calculate how the spacecraft moves during the mission, and the radiation environment that it encounters. With the spectrum of the particles, the program calculates how they would slow down while traversing the coverglass, and provides a slowed-down spectrum.
SHAREv2: fluctuations and a comprehensive treatment of decay feed-down
NASA Astrophysics Data System (ADS)
Torrieri, G.; Jeon, S.; Letessier, J.; Rafelski, J.
2006-11-01
This the user's manual for SHARE version 2. SHARE [G. Torrieri, S. Steinke, W. Broniowski, W. Florkowski, J. Letessier, J. Rafelski, Comput. Phys. Comm. 167 (2005) 229] (Statistical Hadronization with Resonances) is a collection of programs designed for the statistical analysis of particle production in relativistic heavy-ion collisions. While the structure of the program remains similar to v1.x, v2 provides several new features such as evaluation of statistical fluctuations of particle yields, and a greater versatility, in particular regarding decay feed-down and input/output structure. This article describes all the new features, with emphasis on statistical fluctuations. Program summaryTitle of program:SHAREv2 Catalogue identifier:ADVD_v2_0 Program summary URL:http://cpc.cs.qub.ac.uk/summaries/ADVD_v2_0 Program obtainable from: CPC Program Library, Queen's University of Belfast, N. Ireland Computer:PC, Pentium III, 512 MB RAM not hardware dependent Operating system:Linux: RedHat 6.1, 7.2, FEDORA, etc. not system dependent Programming language:FORTRAN77 Size of the package:167 KB directory, without libraries (see http://wwwasdoc.web.cern.ch/wwwasdoc/minuit/minmain.html, http://wwwasd.web.cern.ch/wwwasd/cernlib.html for details on library requirements) Number of lines in distributed program, including test data, etc.:26 101 Number of bytes in distributed program, including test data, etc.:170 346 Distribution format:tar.gzip file Computer:Any computer with an f77 compiler Nature of the physical problem:Event-by-event fluctuations have been recognized to be the physical observable capable to constrain particle production models. Therefore, consideration of event-by-event fluctuations is required for a decisive falsification or constraining of (variants of) particle production models based on (grand-, micro-) canonical statistical mechanics phase space, the so called statistical hadronization models (SHM). As in the case of particle yields, to properly compare model calculations to data it is necessary to consistently take into account resonance decays. However, event-by-event fluctuations are more sensitive than particle yields to experimental acceptance issues, and a range of techniques needs to be implemented to extract 'physical' fluctuations from an experimental event-by-event measurement. Method of solving the problem:The techniques used within the SHARE suite of programs [G. Torrieri, S. Steinke, W. Broniowski, W. Florkowski, J. Letessier, J. Rafelski, Comput. Phys. Comm. 167 (2005) 229; SHAREv1] are updated and extended to fluctuations. A full particle data-table, decay tree, and set of experimental feed-down coefficients are provided. Unlike SHAREv1.x, experimental acceptance feed-down coefficients can be entered for any resonance decay. SHAREv2 can calculate yields, fluctuations, and bulk properties of the fireball from provided thermal parameters; alternatively, parameters can be obtained from fits to experimental data, via the MINUIT fitting algorithm [F. James, M. Roos, Comput. Phys. Comm. 10 (1975) 343]. Fits can also be analyzed for significance, parameter and data point sensitivity. Averages and fluctuations at freeze-out of both the stable particles and the hadronic resonances are set according to a statistical prescription, calculated via a series of Bessel functions, using CERN library programs. We also have the option of including finite particle widths of the resonances. A χ minimization algorithm, also from the CERN library programs, is used to perform and analyze the fit. Please see SHAREv1 for more details on these. Purpose:The vast amount of high quality soft hadron production data, from experiments running at the SPS, RHIC, in past at the AGS, and in the near future at the LHC, offers the opportunity for statistical particle production model falsification. This task has turned out to be difficult when considering solely particle yields addressed in the context of SHAREv1.x. For this reason physical conditions at freeze-out remain contested. Inclusion in the analysis of event-by-event fluctuations appears to resolve this issue. Similarly, a thorough analysis including both fluctuations and average multiplicities gives a way to explore the presence and strength of interactions following hadronization (when hadrons form), ending with thermal freeze-out (when all interactions cease). SHAREv2 with fluctuations will also help determine which statistical ensemble (if any), e.g., canonical or grand-canonical, is more physically appropriate for analyzing a given system. Together with resonances, fluctuations can also be used for a direct estimate of the extent the system re-interacts between chemical and thermal freeze-out. We hope and expect that SHAREv2 will contribute to decide if any of the statistical hadronization model variants has a genuine physical connection to hadron particle production. Computation time survey:We encounter, in the FORTRAN version computation, times up to seconds for evaluation of particle yields. These rise by up to a factor of 300 in the process of minimization and a further factor of a few when χ/N profiles and contours with chemical non-equilibrium are requested. Summary of new features (w.r.t. SHAREv1.x)Fluctuations:In addition to particle yields, ratios and bulk quantities SHAREv2 can calculate, fit and analyze statistical fluctuations of particles and particle ratios Decays:SHAREv2 has the flexibility to account for any experimental method of allowing for decay feed-downs to the particle yields Charm flavor:Charmed particles have been added to the decay tree, allowing as an option study of statistical hadronization of J/ψ, χ, D, etc. Quark chemistry:Chemical non-equilibrium yields for both u and d flavors, as opposed to generically light quarks q, are considered; η- η mixing, etc., are properly dealt with, and chemical non-equilibrium can be studied for each flavor separately Misc:Many new commands and features have been introduced and added to the basic user interface. For example, it is possible to study combinations of particles and their ratios. It is also possible to combine all the input files into one file. SHARE compatibility and manual:This write-up is an update and extension of SHAREv1. The user should consult SHAREv1 regarding the principles of user interface and for all particle yield related physics and program instructions, other than the parameter additions and minor changes described here. SHAREv2 is downward compatible for the changes of the user interface, offering the user of SHAREv1 a computer generated revised input files compatible with SHAREv2.
NASA Technical Reports Server (NTRS)
Shanks, Robert E.
1958-01-01
A theoretical method is presented for predicting the dynamic lateral stability characteristics of an airplane towed in tandem by a much larger airplane. Values of period and time to damp to one-half amplitude and rolling motions calculated by an analog computer have been correlated with results of two experimental investigations conducted in the Langley free-flight tunnel which were part of a U.S. Air Force program (Project FICON) to develop a satisfactory arrangement by which a bomber could tow a parasite fighter. In general, the theoretical results agree with the experimental results.
Secondary Structure Predictions for Long RNA Sequences Based on Inversion Excursions and MapReduce.
Yehdego, Daniel T; Zhang, Boyu; Kodimala, Vikram K R; Johnson, Kyle L; Taufer, Michela; Leung, Ming-Ying
2013-05-01
Secondary structures of ribonucleic acid (RNA) molecules play important roles in many biological processes including gene expression and regulation. Experimental observations and computing limitations suggest that we can approach the secondary structure prediction problem for long RNA sequences by segmenting them into shorter chunks, predicting the secondary structures of each chunk individually using existing prediction programs, and then assembling the results to give the structure of the original sequence. The selection of cutting points is a crucial component of the segmenting step. Noting that stem-loops and pseudoknots always contain an inversion, i.e., a stretch of nucleotides followed closely by its inverse complementary sequence, we developed two cutting methods for segmenting long RNA sequences based on inversion excursions: the centered and optimized method. Each step of searching for inversions, chunking, and predictions can be performed in parallel. In this paper we use a MapReduce framework, i.e., Hadoop, to extensively explore meaningful inversion stem lengths and gap sizes for the segmentation and identify correlations between chunking methods and prediction accuracy. We show that for a set of long RNA sequences in the RFAM database, whose secondary structures are known to contain pseudoknots, our approach predicts secondary structures more accurately than methods that do not segment the sequence, when the latter predictions are possible computationally. We also show that, as sequences exceed certain lengths, some programs cannot computationally predict pseudoknots while our chunking methods can. Overall, our predicted structures still retain the accuracy level of the original prediction programs when compared with known experimental secondary structure.
Zhmurov, A; Dima, R I; Kholodov, Y; Barsegov, V
2010-11-01
Theoretical exploration of fundamental biological processes involving the forced unraveling of multimeric proteins, the sliding motion in protein fibers and the mechanical deformation of biomolecular assemblies under physiological force loads is challenging even for distributed computing systems. Using a C(α)-based coarse-grained self organized polymer (SOP) model, we implemented the Langevin simulations of proteins on graphics processing units (SOP-GPU program). We assessed the computational performance of an end-to-end application of the program, where all the steps of the algorithm are running on a GPU, by profiling the simulation time and memory usage for a number of test systems. The ∼90-fold computational speedup on a GPU, compared with an optimized central processing unit program, enabled us to follow the dynamics in the centisecond timescale, and to obtain the force-extension profiles using experimental pulling speeds (v(f) = 1-10 μm/s) employed in atomic force microscopy and in optical tweezers-based dynamic force spectroscopy. We found that the mechanical molecular response critically depends on the conditions of force application and that the kinetics and pathways for unfolding change drastically even upon a modest 10-fold increase in v(f). This implies that, to resolve accurately the free energy landscape and to relate the results of single-molecule experiments in vitro and in silico, molecular simulations should be carried out under the experimentally relevant force loads. This can be accomplished in reasonable wall-clock time for biomolecules of size as large as 10(5) residues using the SOP-GPU package. © 2010 Wiley-Liss, Inc.
Composition of web services using Markov decision processes and dynamic programming.
Uc-Cetina, Víctor; Moo-Mena, Francisco; Hernandez-Ucan, Rafael
2015-01-01
We propose a Markov decision process model for solving the Web service composition (WSC) problem. Iterative policy evaluation, value iteration, and policy iteration algorithms are used to experimentally validate our approach, with artificial and real data. The experimental results show the reliability of the model and the methods employed, with policy iteration being the best one in terms of the minimum number of iterations needed to estimate an optimal policy, with the highest Quality of Service attributes. Our experimental work shows how the solution of a WSC problem involving a set of 100,000 individual Web services and where a valid composition requiring the selection of 1,000 services from the available set can be computed in the worst case in less than 200 seconds, using an Intel Core i5 computer with 6 GB RAM. Moreover, a real WSC problem involving only 7 individual Web services requires less than 0.08 seconds, using the same computational power. Finally, a comparison with two popular reinforcement learning algorithms, sarsa and Q-learning, shows that these algorithms require one or two orders of magnitude and more time than policy iteration, iterative policy evaluation, and value iteration to handle WSC problems of the same complexity.
Generation of an Aerothermal Data Base for the X33 Spacecraft
NASA Technical Reports Server (NTRS)
Roberts, Cathy; Huynh, Loc
1998-01-01
The X-33 experimental program is a cooperative program between industry and NASA, managed by Lockheed-Martin Skunk Works to develop an experimental vehicle to demonstrate new technologies for a single-stage-to-orbit, fully reusable launch vehicle (RLV). One of the new technologies to be demonstrated is an advanced Thermal Protection System (TPS) being designed by BF Goodrich (formerly Rohr, Inc.) with support from NASA. The calculation of an aerothermal database is crucial to identifying the critical design environment data for the TPS. The NASA Ames X-33 team has generated such a database using Computational Fluid Dynamics (CFD) analyses, engineering analysis methods and various programs to compare and interpolate the results from the CFD and the engineering analyses. This database, along with a program used to query the database, is used extensively by several X-33 team members to help them in designing the X-33. This paper will describe the methods used to generate this database, the program used to query the database, and will show some of the aerothermal analysis results for the X-33 aircraft.
Experiments in cooperative-arm object manipulation with a two-armed free-flying robot. Ph.D. Thesis
NASA Technical Reports Server (NTRS)
Koningstein, Ross
1990-01-01
Developing computed-torque controllers for complex manipulator systems using current techniques and tools is difficult because they address the issues pertinent to simulation, as opposed to control. A new formulation of computed-torque (CT) control that leads to an automated computer-torque robot controller program is presented. This automated tool is used for simulations and experimental demonstrations of endpoint and object control from a free-flying robot. A new computed-torque formulation states the multibody control problem in an elegant, homogeneous, and practical form. A recursive dynamics algorithm is presented that numerically evaluates kinematics and dynamics terms for multibody systems given a topological description. Manipulators may be free-flying, and may have closed-chain constraints. With the exception of object squeeze-force control, the algorithm does not deal with actuator redundancy. The algorithm is used to implement an automated 2D computed-torque dynamics and control package that allows joint, endpoint, orientation, momentum, and object squeeze-force control. This package obviates the need for hand-derivation of kinematics and dynamics, and is used for both simulation and experimental control. Endpoint control experiments are performed on a laboratory robot that has two arms to manipulate payloads, and uses an air bearing to achieve very-low drag characteristics. Simulations and experimental data for endpoint and object controllers are presented for the experimental robot - a complex dynamic system. There is a certain rather wide set of conditions under which CT endpoint controllers can neglect robot base accelerations (but not motions) and achieve comparable performance including base accelerations in the model. The regime over which this simplification holds is explored by simulation and experiment.
Thermal and fluid-dynamics behavior of circulating systems in the case of pressure relief
NASA Astrophysics Data System (ADS)
Moeller, L.
Aspects of safety in the case of large-scale installations with operational high-pressure conditions must be an important consideration already during the design of such installations, taking into account all conceivable disturbances. Within an analysis of such disturbances, studies related to pressure relief processes will have to occupy a central position. For such studies, it is convenient to combine experiments involving small-scale models of the actual installation with suitable computational programs. The experiments can be carried out at lower pressures and temperatures if the actual fluid is replaced by another medium, such as, for instance, a refrigerant. This approach has been used in the present investigation. The obtained experimental data are employed as a basis for a verification of the results provided by the computational model 'Frelap-UK' which has been expressly developed for the analysis of system behavior in the case of pressure relief. It is found that the computer fluid-dynamics characteristics agree with the experimental results.
Frequency-Domain Identification Of Aeroelastic Modes
NASA Technical Reports Server (NTRS)
Acree, C. W., Jr.; Tischler, Mark B.
1991-01-01
Report describes flight measurements and frequency-domain analyses of aeroelastic vibrational modes of wings of XV-15 tilt-rotor aircraft. Begins with description of flight-test methods. Followed by brief discussion of methods of analysis, which include Fourier-transform computations using chirp z transformers, use of coherence and other spectral functions, and methods and computer programs to obtain frequencies and damping coefficients from measurements. Includes brief description of results of flight tests and comparisions among various experimental and theoretical results. Ends with section on conclusions and recommended improvements in techniques.
The prediction of three-dimensional liquid-propellant rocket nozzle admittances
NASA Technical Reports Server (NTRS)
Bell, W. A.; Zinn, B. T.
1973-01-01
Crocco's three-dimensional nozzle admittance theory is extended to be applicable when the amplitudes of the combustor and nozzle oscillations increase or decrease with time. An analytical procedure and a computer program for determining nozzle admittance values from the extended theory are presented and used to compute the admittances of a family of liquid-propellant rocket nozzles. The calculated results indicate that the nozzle geometry entrance Mach number and temporal decay coefficient significantly affect the nozzle admittance values. The theoretical predictions are shown to be in good agreement with available experimental data.
Contextual Fraction as a Measure of Contextuality.
Abramsky, Samson; Barbosa, Rui Soares; Mansfield, Shane
2017-08-04
We consider the contextual fraction as a quantitative measure of contextuality of empirical models, i.e., tables of probabilities of measurement outcomes in an experimental scenario. It provides a general way to compare the degree of contextuality across measurement scenarios; it bears a precise relationship to violations of Bell inequalities; its value, and a witnessing inequality, can be computed using linear programing; it is monotonic with respect to the "free" operations of a resource theory for contextuality; and it measures quantifiable advantages in informatic tasks, such as games and a form of measurement-based quantum computing.
Contextual Fraction as a Measure of Contextuality
NASA Astrophysics Data System (ADS)
Abramsky, Samson; Barbosa, Rui Soares; Mansfield, Shane
2017-08-01
We consider the contextual fraction as a quantitative measure of contextuality of empirical models, i.e., tables of probabilities of measurement outcomes in an experimental scenario. It provides a general way to compare the degree of contextuality across measurement scenarios; it bears a precise relationship to violations of Bell inequalities; its value, and a witnessing inequality, can be computed using linear programing; it is monotonic with respect to the "free" operations of a resource theory for contextuality; and it measures quantifiable advantages in informatic tasks, such as games and a form of measurement-based quantum computing.
A study of the electromagnetic interaction between planetary bodies and the solar wind
NASA Technical Reports Server (NTRS)
Schwartz, K.
1971-01-01
Theoretical and computational techniques were developed for calculating the time dependent electromagnetic response of a radially inhomogeneous moon. The techniques were used to analyze the experimental data from the LSM (lunar surface magnetometer) thus providing an in-depth diagnostic of the Lunar interior. The theory was also incorporated into an existing computer code designed to calculate the thermal evolution of planetary bodies. The program will provide a tool for examining the effect of heating from the TE mode (poloidal magnetic field) as well as the TM mode (toroidal magnetic field).
NASA Astrophysics Data System (ADS)
Domanskyi, Sergii; Schilling, Joshua E.; Gorshkov, Vyacheslav; Libert, Sergiy; Privman, Vladimir
2016-09-01
We develop a theoretical approach that uses physiochemical kinetics modelling to describe cell population dynamics upon progression of viral infection in cell culture, which results in cell apoptosis (programmed cell death) and necrosis (direct cell death). Several model parameters necessary for computer simulation were determined by reviewing and analyzing available published experimental data. By comparing experimental data to computer modelling results, we identify the parameters that are the most sensitive to the measured system properties and allow for the best data fitting. Our model allows extraction of parameters from experimental data and also has predictive power. Using the model we describe interesting time-dependent quantities that were not directly measured in the experiment and identify correlations among the fitted parameter values. Numerical simulation of viral infection progression is done by a rate-equation approach resulting in a system of "stiff" equations, which are solved by using a novel variant of the stochastic ensemble modelling approach. The latter was originally developed for coupled chemical reactions.
NASA Astrophysics Data System (ADS)
Domanskyi, Sergii; Schilling, Joshua; Gorshkov, Vyacheslav; Libert, Sergiy; Privman, Vladimir
We develop a theoretical approach that uses physiochemical kinetics modelling to describe cell population dynamics upon progression of viral infection in cell culture, which results in cell apoptosis (programmed cell death) and necrosis (direct cell death). Several model parameters necessary for computer simulation were determined by reviewing and analyzing available published experimental data. By comparing experimental data to computer modelling results, we identify the parameters that are the most sensitive to the measured system properties and allow for the best data fitting. Our model allows extraction of parameters from experimental data and also has predictive power. Using the model we describe interesting time-dependent quantities that were not directly measured in the experiment and identify correlations among the fitted parameter values. Numerical simulation of viral infection progression is done by a rate-equation approach resulting in a system of ``stiff'' equations, which are solved by using a novel variant of the stochastic ensemble modelling approach. The latter was originally developed for coupled chemical reactions.
NASA Astrophysics Data System (ADS)
Maślak, Mariusz; Pazdanowski, Michał; Woźniczka, Piotr
2018-01-01
Validation of fire resistance for the same steel frame bearing structure is performed here using three different numerical models, i.e. a bar one prepared in the SAFIR environment, and two 3D models developed within the framework of Autodesk Simulation Mechanical (ASM) and an alternative one developed in the environment of the Abaqus code. The results of the computer simulations performed are compared with the experimental results obtained previously, in a laboratory fire test, on a structure having the same characteristics and subjected to the same heating regimen. Comparison of the experimental and numerically determined displacement evolution paths for selected nodes of the considered frame during the simulated fire exposure constitutes the basic criterion applied to evaluate the validity of the numerical results obtained. The experimental and numerically determined estimates of critical temperature specific to the considered frame and related to the limit state of bearing capacity in fire have been verified as well.
Influence of rubbing on rotor dynamics, part 1
NASA Technical Reports Server (NTRS)
Muszynska, Agnes; Bently, Donald E.; Franklin, Wesley D.; Hayashida, Robert D.; Kingsley, Lori M.; Curry, Arthur E.
1989-01-01
The results of analytical and experimental research on rotor-to-stationary element rubbing in rotating machines are presented. A characterization of physical phenomena associated with rubbing, as well as a literature survey on the subject of rub is given. The experimental results were obtained from two rubbing rotor rigs: one, which dynamically simulates the space shuttle main engine high pressure fuel turbopump (HPFTP), and the second one, much simpler, a two-mode rotor rig, designed for more generic studies on rotor-to-stator rubbing. Two areas were studied: generic rotor-to-stator rub-related dynamic phenomena affecting rotating machine behavior and applications to the space shuttle HPFTP. An outline of application of dynamic stiffness methodology for identification of rotor/bearing system modal parameters is given. The mathematical model of rotor/bearing/seal system under rub condition is given. The computer program was developed to calculate rotor responses. Compared with experimental results the computed results prove an adequacy of the model.
NASA Technical Reports Server (NTRS)
DeChant, Lawrence Justin
1998-01-01
In spite of rapid advances in both scalar and parallel computational tools, the large number of variables involved in both design and inverse problems make the use of sophisticated fluid flow models impractical, With this restriction, it is concluded that an important family of methods for mathematical/computational development are reduced or approximate fluid flow models. In this study a combined perturbation/numerical modeling methodology is developed which provides a rigorously derived family of solutions. The mathematical model is computationally more efficient than classical boundary layer but provides important two-dimensional information not available using quasi-1-d approaches. An additional strength of the current methodology is its ability to locally predict static pressure fields in a manner analogous to more sophisticated parabolized Navier Stokes (PNS) formulations. To resolve singular behavior, the model utilizes classical analytical solution techniques. Hence, analytical methods have been combined with efficient numerical methods to yield an efficient hybrid fluid flow model. In particular, the main objective of this research has been to develop a system of analytical and numerical ejector/mixer nozzle models, which require minimal empirical input. A computer code, DREA Differential Reduced Ejector/mixer Analysis has been developed with the ability to run sufficiently fast so that it may be used either as a subroutine or called by an design optimization routine. Models are of direct use to the High Speed Civil Transport Program (a joint government/industry project seeking to develop an economically.viable U.S. commercial supersonic transport vehicle) and are currently being adopted by both NASA and industry. Experimental validation of these models is provided by comparison to results obtained from open literature and Limited Exclusive Right Distribution (LERD) sources, as well as dedicated experiments performed at Texas A&M. These experiments have been performed using a hydraulic/gas flow analog. Results of comparisons of DREA computations with experimental data, which include entrainment, thrust, and local profile information, are overall good. Computational time studies indicate that DREA provides considerably more information at a lower computational cost than contemporary ejector nozzle design models. Finally. physical limitations of the method, deviations from experimental data, potential improvements and alternative formulations are described. This report represents closure to the NASA Graduate Researchers Program. Versions of the DREA code and a user's guide may be obtained from the NASA Lewis Research Center.
Enlist micros: Training science teachers to use microcomputers
NASA Astrophysics Data System (ADS)
Baird, William E.; Ellis, James D.; Kuerbis, Paul J.
A National Science Foundation grant to the Biological Sciences Curriculum Study (BSCS) at The Colorado College supported the design and production of training materials to encourage literacy of science teachers in the use of microcomputers. ENLIST Micros is based on results of a national needs assessment that identified 22 compentencies needed by K-12 science teachers to use microcomputers for instruction. A writing team developed the 16-hour training program in the summer of 1985, and field-test coordinators tested it with 18 preservice or in-service groups during the 1985-86 academic year at 15 sites within the United States. The training materials consist of video programs, interactive computer disks for the Apple II series microcomputer, a training manual for participants, and a guide for the group leader. The experimental materials address major areas of educational computing: awareness, applications, implementation, evaluation, and resources. Each chapter contains activities developed for this program, such as viewing video segments of science teachers who are using computers effectively and running commercial science and training courseware. Role playing and small-group interaction help the teachers overcome their reluctance to use computers and plan for effective implementation of microcomputers in the school. This study examines the implementation of educational computing among 47 science teachers who completed the ENLIST Micros training at a southern university. We present results of formative evaluation for that site. Results indicate that both elementary and secondary teachers benefit from the training program and demonstrate gains in attitudes toward computer use. Participating teachers said that the program met its stated objectives and helped them obtain needed skills. Only 33 percent of these teachers, however, reported using computers one year after the training. In June 1986, the BSCS initiated a follow up to the ENLIST Micros curriculum to develop, evaluate, and disseminate a complete model of teacher enhancement for educational computing in the sciences. In that project, we use the ENLIST Micros curriculum as the first step in a training process. The project includes seminars that introduce additional skills: It contains provisions for sharing among participants, monitors use of computers in participants' classrooms, provides structured coaching of participants' use of computers in their classrooms, and offers planned observations of peers using computers in their science teaching.
1981-03-31
logic testing element and a concomitant testability criterion ideally suited to dynamic circuit applications and appro- priate for automatic computer...making connections automatically . PF is an experimental feature which provides users with only four different chip sizes (full, half, quarter, and eighth...initial solution is found constructively which is improved by pair-wise swapping. Results show, however, that the constructive initial sorter , which
A Comparison Study between a Traditional and Experimental Program.
ERIC Educational Resources Information Center
Dogan, Hamide
This paper is part of a dissertation defended in January 2001 as part of the author's Ph.D. requirement. The study investigated the effects of use of Mathematica, a computer algebra system, in learning basic linear algebra concepts, It was done by means of comparing two first year linear algebra classes, one traditional and one Mathematica…
Center of Excellence for Hypersonics Research
2012-01-25
detailed simulations of actual combustor configurations, and ultimately for the optimization of hypersonic air - breathing propulsion system flow paths... vehicle development programs. The Center engaged leading experts in experimental and computational analysis of hypersonic flows to provide research...advanced hypersonic vehicles and space access systems will require significant advances in the design methods and ground testing techniques to ensure
ERIC Educational Resources Information Center
Austin Coll., Sherman, TX.
The Birkman Method is proprietary, and consists of a battery of psychological instruments, an occupational interest survey, and associated reports, a self-awareness seminar based on the questionnaire, and the supporting computer software and data banks. The questions and occupations are designed to explore basic areas of a person's values,…
The Mind Research Network - Mental Illness Neuroscience Discovery Grant
DOE Office of Scientific and Technical Information (OSTI.GOV)
Roberts, J.; Calhoun, V.
The scientific and technological programs of the Mind Research Network (MRN), reflect DOE missions in basic science and associated instrumentation, computational modeling, and experimental techniques. MRN's technical goals over the course of this project have been to develop and apply integrated, multi-modality functional imaging techniques derived from a decade of DOE-support research and technology development.
Evaluating Intention to Use Remote Robotics Experimentation in Programming Courses
ERIC Educational Resources Information Center
Cheng, Pericles L.
2017-01-01
The Digital Agenda for Europe (2015) states that there will be 825,000 unfilled vacancies for Information and Communications Technology by 2020. This lack of IT professionals stems from the small number of students graduating in computer science. To retain more students in the field, teachers can use remote robotic experiments to explain difficult…
NASA Technical Reports Server (NTRS)
Graham, A. B.
1977-01-01
Small- and large-scale models of supersonic cruise fighter vehicles were used to determine the effectiveness of airframe/propulsion integration concepts for improved low-speed performance and stability and control characteristics. Computer programs were used for engine/airframe sizing studies to yield optimum vehicle performance.
ERIC Educational Resources Information Center
Vollands, Stacy R.; And Others
A study evaluated the effect software for self-assessment and management of reading practice had on reading achievement and motivation in two primary schools in Aberdeen, Scotland. The program utilized was The Accelerated Reader (AR) which was designed to enable curriculum based assessment of reading comprehension within the classroom. Students…
NASA Astrophysics Data System (ADS)
Green, Sheldon
1993-09-01
A computer program error led to erroneous results in the titled paper. Corrected generalized IOS cross sections are significantly changed, especially at lower collision energies. These changes tend to cancel in predicted Raman linewidths; there is a systematic increase of 10-15 %, changing quantitative, but not qualitative, comparisons with experimental data.
Computational Performance of Group IV Personnel in Vocational Training Programs. Final Report.
ERIC Educational Resources Information Center
Main, Ray E.; Harrigan, Robert J.
The document evaluates Navy Group Four personnel gains in basic arithmetic skills after taking experimental courses in linear measurement and recipe conversion. Categorized as Mental Group Four by receiving scores from the 10th to the 30th percentile of the Armed Forces Qualification Test, trainees received instruction tailored to the level of…
Applications Development for a Parallel COTS Spaceborne Computer
NASA Technical Reports Server (NTRS)
Katz, Daniel S.; Springer, Paul L.; Granat, Robert; Turmon, Michael
2000-01-01
This presentation reviews the Remote Exploration and Experimentation Project (REE) program for utilization of scalable supercomputing technology in space. The implementation of REE will be the use of COTS hardware and software to the maximum extent possible, keeping overhead low. Since COTS systems will be used, with little or no special modification, there will be significant cost reduction.
Computational toxicology using the OpenTox application programming interface and Bioclipse
2011-01-01
Background Toxicity is a complex phenomenon involving the potential adverse effect on a range of biological functions. Predicting toxicity involves using a combination of experimental data (endpoints) and computational methods to generate a set of predictive models. Such models rely strongly on being able to integrate information from many sources. The required integration of biological and chemical information sources requires, however, a common language to express our knowledge ontologically, and interoperating services to build reliable predictive toxicology applications. Findings This article describes progress in extending the integrative bio- and cheminformatics platform Bioclipse to interoperate with OpenTox, a semantic web framework which supports open data exchange and toxicology model building. The Bioclipse workbench environment enables functionality from OpenTox web services and easy access to OpenTox resources for evaluating toxicity properties of query molecules. Relevant cases and interfaces based on ten neurotoxins are described to demonstrate the capabilities provided to the user. The integration takes advantage of semantic web technologies, thereby providing an open and simplifying communication standard. Additionally, the use of ontologies ensures proper interoperation and reliable integration of toxicity information from both experimental and computational sources. Conclusions A novel computational toxicity assessment platform was generated from integration of two open science platforms related to toxicology: Bioclipse, that combines a rich scriptable and graphical workbench environment for integration of diverse sets of information sources, and OpenTox, a platform for interoperable toxicology data and computational services. The combination provides improved reliability and operability for handling large data sets by the use of the Open Standards from the OpenTox Application Programming Interface. This enables simultaneous access to a variety of distributed predictive toxicology databases, and algorithm and model resources, taking advantage of the Bioclipse workbench handling the technical layers. PMID:22075173
Punzalan, Florencio Rusty; Kunieda, Yoshitoshi; Amano, Akira
2015-01-01
Clinical and experimental studies involving human hearts can have certain limitations. Methods such as computer simulations can be an important alternative or supplemental tool. Physiological simulation at the tissue or organ level typically involves the handling of partial differential equations (PDEs). Boundary conditions and distributed parameters, such as those used in pharmacokinetics simulation, add to the complexity of the PDE solution. These factors can tailor PDE solutions and their corresponding program code to specific problems. Boundary condition and parameter changes in the customized code are usually prone to errors and time-consuming. We propose a general approach for handling PDEs and boundary conditions in computational models using a replacement scheme for discretization. This study is an extension of a program generator that we introduced in a previous publication. The program generator can generate code for multi-cell simulations of cardiac electrophysiology. Improvements to the system allow it to handle simultaneous equations in the biological function model as well as implicit PDE numerical schemes. The replacement scheme involves substituting all partial differential terms with numerical solution equations. Once the model and boundary equations are discretized with the numerical solution scheme, instances of the equations are generated to undergo dependency analysis. The result of the dependency analysis is then used to generate the program code. The resulting program code are in Java or C programming language. To validate the automatic handling of boundary conditions in the program code generator, we generated simulation code using the FHN, Luo-Rudy 1, and Hund-Rudy cell models and run cell-to-cell coupling and action potential propagation simulations. One of the simulations is based on a published experiment and simulation results are compared with the experimental data. We conclude that the proposed program code generator can be used to generate code for physiological simulations and provides a tool for studying cardiac electrophysiology. PMID:26356082
Computer-based cognitive retraining for adults with chronic acquired brain injury: a pilot study.
Li, Kitsum; Robertson, Julie; Ramos, Joshua; Gella, Stephanie
2013-10-01
This study evaluated the effectiveness of a computer-based cognitive retraining (CBCR) program on improving memory and attention deficits in individuals with a chronic acquired brain injury (ABI). Twelve adults with a chronic ABI demonstrating deficits in memory and attention were recruited from a convenience sample from the community. Using a quasi-experimental one-group pretest-posttest design, a significant improvement was found in both memory and attention scores postintervention using the cognitive screening tool. This study supported the effectiveness of CBCR programs in improving cognitive deficits in memory and attention in individuals with chronic ABI. Further research is recommended to validate these findings with a larger ABI population and to investigate transfer to improvement in occupational performance that supports daily living skills.
Predicting the Coupling Properties of Axially-Textured Materials.
Fuentes-Cobas, Luis E; Muñoz-Romero, Alejandro; Montero-Cabrera, María E; Fuentes-Montero, Luis; Fuentes-Montero, María E
2013-10-30
A description of methods and computer programs for the prediction of "coupling properties" in axially-textured polycrystals is presented. Starting data are the single-crystal properties, texture and stereography. The validity and proper protocols for applying the Voigt, Reuss and Hill approximations to estimate coupling properties effective values is analyzed. Working algorithms for predicting mentioned averages are given. Bunge's symmetrized spherical harmonics expansion of orientation distribution functions, inverse pole figures and (single and polycrystals) physical properties is applied in all stages of the proposed methodology. The established mathematical route has been systematized in a working computer program. The discussion of piezoelectricity in a representative textured ferro-piezoelectric ceramic illustrates the application of the proposed methodology. Polycrystal coupling properties, predicted by the suggested route, are fairly close to experimentally measured ones.
Predicting the Coupling Properties of Axially-Textured Materials
Fuentes-Cobas, Luis E.; Muñoz-Romero, Alejandro; Montero-Cabrera, María E.; Fuentes-Montero, Luis; Fuentes-Montero, María E.
2013-01-01
A description of methods and computer programs for the prediction of “coupling properties” in axially-textured polycrystals is presented. Starting data are the single-crystal properties, texture and stereography. The validity and proper protocols for applying the Voigt, Reuss and Hill approximations to estimate coupling properties effective values is analyzed. Working algorithms for predicting mentioned averages are given. Bunge’s symmetrized spherical harmonics expansion of orientation distribution functions, inverse pole figures and (single and polycrystals) physical properties is applied in all stages of the proposed methodology. The established mathematical route has been systematized in a working computer program. The discussion of piezoelectricity in a representative textured ferro-piezoelectric ceramic illustrates the application of the proposed methodology. Polycrystal coupling properties, predicted by the suggested route, are fairly close to experimentally measured ones. PMID:28788370
NASA Technical Reports Server (NTRS)
Wright, R. M.; Hwang, K. C.
1973-01-01
The sorbent behavior of solid amine resin IR-45 with regard to potential use in regenerative CO2-removal systems for manned spacecraft is considered. Measurements of equilibrium sorption capacity of IR-45 for water and for CO2 are reported, and the dynamic mass transfer behavior of IR-45 beds is studied under conditions representative of those expected in a manned spacecraft. A digital computer program was written for the transient performance prediction of CO2 removal systems comprised of solid amine beds. Also evaluated are systems employing inorganic molecular-sieve sorbents. Tests show that there is definitely an effect of water loading on the absorption rate.
Computational Aeroelastic Analysis of the Semi-Span Super-Sonic Transport (S4T) Wind-Tunnel Model
NASA Technical Reports Server (NTRS)
Sanetrik, Mark D.; Silva, Walter A.; Hur, Jiyoung
2012-01-01
A summary of the computational aeroelastic analysis for the Semi-Span Super-Sonic Transport (S4T) wind-tunnel model is presented. A broad range of analysis techniques, including linear, nonlinear and Reduced Order Models (ROMs) were employed in support of a series of aeroelastic (AE) and aeroservoelastic (ASE) wind-tunnel tests conducted in the Transonic Dynamics Tunnel (TDT) at NASA Langley Research Center. This research was performed in support of the ASE element in the Supersonics Program, part of NASA's Fundamental Aeronautics Program. The analysis concentrated on open-loop flutter predictions, which were in good agreement with experimental results. This paper is one in a series that comprise a special S4T technical session, which summarizes the S4T project.
[Computer simulation of thyroid regulatory mechanisms in health and malignancy].
Abduvaliev, A A; Gil'dieva, M S; Khidirov, B N; Saĭdalieva, M; Saatov, T S
2010-07-01
The paper describes a computer model for regulation of the number of thyroid follicular cells in health and malignancy. The authors'computer program for mathematical simulation of the regulatory mechanisms of a thyroid follicular cellular community cannot be now referred to as good commercial products. For commercialization of this product, it is necessary to draw up a direct relation of the introduced corrected values from the actually existing normal values, such as the peripheral blood concentrations of thyroid hormones or the mean values of endocrine tissue mitotic activity. However, the described computer program has been also used in researches by our scientific group in the study of thyroid cancer. The available biological experimental data and theoretical provisions on thyroid structural and functional organization at the cellular level allow one to construct mathematical models for quantitative analysis of the regulation of the size of a cellular community of a thyroid follicle in health and abnormalities, by using the method for simulation of the regulatory mechanisms of living systems and the equations of cellular community regulatory communities.
A maximum likelihood convolutional decoder model vs experimental data comparison
NASA Technical Reports Server (NTRS)
Chen, R. Y.
1979-01-01
This article describes the comparison of a maximum likelihood convolutional decoder (MCD) prediction model and the actual performance of the MCD at the Madrid Deep Space Station. The MCD prediction model is used to develop a subroutine that has been utilized by the Telemetry Analysis Program (TAP) to compute the MCD bit error rate for a given signal-to-noise ratio. The results indicate that that the TAP can predict quite well compared to the experimental measurements. An optimal modulation index also can be found through TAP.
Experimental studies of the Eppler 61 airfoil at low Reynolds numbers
NASA Technical Reports Server (NTRS)
Burns, T. F.; Mueller, T. J.
1982-01-01
The results of an experimental study to document the effects of separation and transition on the performance of an airfoil designed for low Reynolds number operation are presented. Lift, drag and flow visualization data were obtained for the Eppler 61 airfoil section for chord Reynolds numbers from about 30,000 to over 200,000. Smoke flow visualization was employed to document the boundary layer behavior and was correlated with the Eppler airfoil design and analysis computer program. Laminar separation, transition and turbulent reattachment had significant effects on the performance of this airfoil.
Buoyancy-Driven Instabilities in Single-Bubble Sonoluminescence
NASA Technical Reports Server (NTRS)
Matula, Thomas J.
2003-01-01
The principal objectives of this study are to determine how gravity affects the emission of light from single-bubble sonoluminescence (SBSL), and whether or not the bubble extinction is directly related to gravity. Our experimental task involves designing glass or quartz spherical levitation cells that generate very stable SL bubbles. The cells must have minimized vibration, and some temperature control. The experimental system will reside in a light-tight enclosure. Aside from acceleration, the frequency, pressure amplitude, and light intensity must be measured. A computer program will be constructed to perform all aspects of the experiment.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Rice, R.E.
Results are presented of studies conducted by Aerojet Nuclear Company (ANC) in FY 1975 to support the Nuclear Regulatory Commission (NRC) on the boiling water reactor blowdown heat transfer (BWR-BDHT) program. The support provided by ANC is that of an independent assessor of the program to ensure that the data obtained are adequate for verification of analytical models used for predicting reactor response to a postulated loss-of-coolant accident. The support included reviews of program plans, objectives, measurements, and actual data. Additional activity included analysis of experimental system performance and evaluation of the RELAP4 computer code as applied to the experiments.
The benchmark aeroelastic models program: Description and highlights of initial results
NASA Technical Reports Server (NTRS)
Bennett, Robert M.; Eckstrom, Clinton V.; Rivera, Jose A., Jr.; Dansberry, Bryan E.; Farmer, Moses G.; Durham, Michael H.
1991-01-01
An experimental effort was implemented in aeroelasticity called the Benchmark Models Program. The primary purpose of this program is to provide the necessary data to evaluate computational fluid dynamic codes for aeroelastic analysis. It also focuses on increasing the understanding of the physics of unsteady flows and providing data for empirical design. An overview is given of this program and some results obtained in the initial tests are highlighted. The tests that were completed include measurement of unsteady pressures during flutter of rigid wing with a NACA 0012 airfoil section and dynamic response measurements of a flexible rectangular wing with a thick circular arc airfoil undergoing shock boundary layer oscillations.
Ecologic study of children's use of a computer nutrition education program.
Matheson, D; Achterberg, C
2001-01-01
The purpose of this research was to describe the context created by students as they worked in groups on a nutrition computer-assisted instruction (CAI) program. Students worked on the program in groups of three. Observational methods were used to collect data from students in two sixth-grade classrooms that were part of an experimental program designed to restructure the educational process. Thirty-two students, from 12 groups, were observed as they completed the program. The groups were assigned by the teachers according to standard principles of cooperative learning. Students completed "Ship to Shore," a program designed specifically for this research. The program required three to five 50-minute classroom periods to complete. The objectives of the program were to change children's knowledge structure of basic nutrition concepts and to increase children's critical thinking skills related to nutrition concepts. We collected observational data focused on three domains: (1) student-computer interaction, (2) student-student interaction, and (3) students' thinking and learning skills. Grounded theory methods were used to analyze the data. Specifically, the constant-comparative method was used to develop open coding categories, defined by properties and described by dimensions. The open coding categories were in turn used in axial coding to differentiate students' learning styles. Five styles of student interaction were defined. These included (1) dominant directors (n = 6; 19%), (2) passive actors (n = 5; 16%), (3) action-oriented students (n = 7; 22%), (4) content-oriented students (n = 8; 25%), and (5) problem solvers (n = 5; 16%). The "student style" groups were somewhat gender specific. The dominant directors and passive actors were girls and the action-oriented and content-oriented students were boys. The problem solvers group was mixed gender. Children's responses to computer-based nutrition education are highly variable. Based on the results of this research, nutrition educators may recommend that nutrition CAI programs be implemented in mixed gender groups.
Analysis of impact of general-purpose graphics processor units in supersonic flow modeling
NASA Astrophysics Data System (ADS)
Emelyanov, V. N.; Karpenko, A. G.; Kozelkov, A. S.; Teterina, I. V.; Volkov, K. N.; Yalozo, A. V.
2017-06-01
Computational methods are widely used in prediction of complex flowfields associated with off-normal situations in aerospace engineering. Modern graphics processing units (GPU) provide architectures and new programming models that enable to harness their large processing power and to design computational fluid dynamics (CFD) simulations at both high performance and low cost. Possibilities of the use of GPUs for the simulation of external and internal flows on unstructured meshes are discussed. The finite volume method is applied to solve three-dimensional unsteady compressible Euler and Navier-Stokes equations on unstructured meshes with high resolution numerical schemes. CUDA technology is used for programming implementation of parallel computational algorithms. Solutions of some benchmark test cases on GPUs are reported, and the results computed are compared with experimental and computational data. Approaches to optimization of the CFD code related to the use of different types of memory are considered. Speedup of solution on GPUs with respect to the solution on central processor unit (CPU) is compared. Performance measurements show that numerical schemes developed achieve 20-50 speedup on GPU hardware compared to CPU reference implementation. The results obtained provide promising perspective for designing a GPU-based software framework for applications in CFD.
After-effects of human-computer interaction indicated by P300 of the event-related brain potential.
Trimmel, M; Huber, R
1998-05-01
After-effects of human-computer interaction (HCI) were investigated by using the P300 component of the event-related brain potential (ERP). Forty-nine subjects (naive non-users, beginners, experienced users, programmers) completed three paper/pencil tasks (text editing, solving intelligence test items, filling out a questionnaire on sensation seeking) and three HCI tasks (text editing, executing a tutor program or programming, playing Tetris). The sequence of 7-min tasks was randomized between subjects and balanced between groups. After each experimental condition ERPs were recorded during an acoustic discrimination task at F3, F4, Cz, P3 and P4. Data indicate that: (1) mental after-effects of HCI can be detected by P300 of the ERP; (2) HCI showed in general a reduced amplitude; (3) P300 amplitude varied also with type of task, mainly at F4 where it was smaller after cognitive tasks (intelligence test/programming) and larger after emotion-based tasks (sensation seeking/Tetris); (4) cognitive tasks showed shorter latencies; (5) latencies were widely location-independent (within the range of 356-358 ms at F3, F4, P3 and P4) after executing the tutor program or programming; and (6) all observed after-effects were independent of the user's experience in operating computers and may therefore reflect short-term after-effects only and no structural changes of information processing caused by HCI.
Identification of human microRNA targets from isolated argonaute protein complexes.
Beitzinger, Michaela; Peters, Lasse; Zhu, Jia Yun; Kremmer, Elisabeth; Meister, Gunter
2007-06-01
MicroRNAs (miRNAs) constitute a class of small non-coding RNAs that regulate gene expression on the level of translation and/or mRNA stability. Mammalian miRNAs associate with members of the Argonaute (Ago) protein family and bind to partially complementary sequences in the 3' untranslated region (UTR) of specific target mRNAs. Computer algorithms based on factors such as free binding energy or sequence conservation have been used to predict miRNA target mRNAs. Based on such predictions, up to one third of all mammalian mRNAs seem to be under miRNA regulation. However, due to the low degree of complementarity between the miRNA and its target, such computer programs are often imprecise and therefore not very reliable. Here we report the first biochemical identification approach of miRNA targets from human cells. Using highly specific monoclonal antibodies against members of the Ago protein family, we co-immunoprecipitate Ago-bound mRNAs and identify them by cloning. Interestingly, most of the identified targets are also predicted by different computer programs. Moreover, we randomly analyzed six different target candidates and were able to experimentally validate five as miRNA targets. Our data clearly indicate that miRNA targets can be experimentally identified from Ago complexes and therefore provide a new tool to directly analyze miRNA function.
An interactive multimedia program to prevent HIV transmission in men with intellectual disability.
Wells, Jennifer; Clark, Khaya; Sarno, Karen
2014-05-01
The efficacy of a computer-based interactive multimedia HIV/AIDS prevention program for men with intellectual disability (ID) was examined using a quasi-experimental within-subjects design. Thirty-seven men with mild to moderate intellectual disability evaluated the program. The pretest and posttest instruments assessed HIV/AIDS knowledge (high-risk fluids, HIV transmission, and condom facts) and condom application skills. All outcome measures showed statistically significant gains from pretest to posttest, with medium to large effect sizes. In addition, a second study was conducted with twelve service providers who work with men with ID. Service providers reviewed the HIV/AIDS prevention program, completed a demographics questionnaire, and a program satisfaction survey. Overall, service providers rated the program highly on several outcome measures (stimulation, relevance, and usability).
Center for Center for Technology for Advanced Scientific Component Software (TASCS)
DOE Office of Scientific and Technical Information (OSTI.GOV)
Kostadin, Damevski
A resounding success of the Scientific Discovery through Advanced Computing (SciDAC) program is that high-performance computational science is now universally recognized as a critical aspect of scientific discovery [71], complementing both theoretical and experimental research. As scientific communities prepare to exploit unprecedented computing capabilities of emerging leadership-class machines for multi-model simulations at the extreme scale [72], it is more important than ever to address the technical and social challenges of geographically distributed teams that combine expertise in domain science, applied mathematics, and computer science to build robust and flexible codes that can incorporate changes over time. The Center for Technologymore » for Advanced Scientific Component Software (TASCS)1 tackles these these issues by exploiting component-based software development to facilitate collaborative high-performance scientific computing.« less
DOE Office of Scientific and Technical Information (OSTI.GOV)
Latanision, R.M.
1990-12-01
Electrochemical corrosion is pervasive in virtually all engineering systems and in virtually all industrial circumstances. Although engineers now understand how to design systems to minimize corrosion in many instances, many fundamental questions remain poorly understood and, therefore, the development of corrosion control strategies is based more on empiricism than on a deep understanding of the processes by which metals corrode in electrolytes. Fluctuations in potential, or current, in electrochemical systems have been observed for many years. To date, all investigations of this phenomenon have utilized non-deterministic analyses. In this work it is proposed to study electrochemical noise from a deterministicmore » viewpoint by comparison of experimental parameters, such as first and second order moments (non-deterministic), with computer simulation of corrosion at metal surfaces. In this way it is proposed to analyze the origins of these fluctuations and to elucidate the relationship between these fluctuations and kinetic parameters associated with metal dissolution and cathodic reduction reactions. This research program addresses in essence two areas of interest: (a) computer modeling of corrosion processes in order to study the electrochemical processes on an atomistic scale, and (b) experimental investigations of fluctuations in electrochemical systems and correlation of experimental results with computer modeling. In effect, the noise generated by mathematical modeling will be analyzed and compared to experimental noise in electrochemical systems. 1 fig.« less
NASA Astrophysics Data System (ADS)
Mariappan, G.; Sundaraganesan, N.
2014-04-01
Vibrational assignments for the 7-amino-2-methylchromone (abbreviated as 7A2MC) molecule using a combination of experimental vibrational spectroscopic measurements and ab initio computational methods are reported. The optimized geometry, intermolecular hydrogen bonding, first order hyperpolarizability and harmonic vibrational wavenumbers of 7A2MC have been investigated with the help of B3LYP density functional theory method. The calculated molecular geometry parameters, the theoretically computed vibrational frequencies for monomer and dimer and relative peak intensities were compared with experimental data. DFT calculations using the B3LYP method and 6-31 + G(d,p) basis set were found to yield results that are very comparable to experimental IR and Raman spectra. Detailed vibrational assignments were performed with DFT calculations and the potential energy distribution (PED) obtained from the Vibrational Energy Distribution Analysis (VEDA) program. Natural Bond Orbital (NBO) study revealed the characteristics of the electronic delocalization of the molecular structure. 13C and 1H NMR spectra have been recorded and 13C and 1H nuclear magnetic resonance chemical shifts of the molecule have been calculated using the gauge independent atomic orbital (GIAO) method. Furthermore, All the possible calculated values are analyzed using correlation coefficients linear fitting equation and are shown strong correlation with the experimental data.
Development of an Experiment High Performance Nozzle Research Program
NASA Technical Reports Server (NTRS)
2004-01-01
As proposed in the above OAI/NASA Glenn Research Center (GRC) Co-Operative Agreement the objective of the work was to provide consultation and assistance to the NASA GRC GTX Rocket Based Combined Cycle (RBCC) Program Team in planning and developing requirements, scale model concepts, and plans for an experimental nozzle research program. The GTX was one of the launch vehicle concepts being studied as a possible future replacement for the aging NASA Space Shuttle, and was one RBCC element in the ongoing NASA Access to Space R&D Program (Reference 1). The ultimate program objective was the development of an appropriate experimental research program to evaluate and validate proposed nozzle concepts, and thereby result in the optimization of a high performance nozzle for the GTX launch vehicle. Included in this task were the identification of appropriate existing test facilities, development of requirements for new non-existent test rigs and fixtures, develop scale nozzle model concepts, and propose corresponding test plans. Also included were the evaluation of originally proposed and alternate nozzle designs (in-house and contractor), evaluation of Computational Fluid Dynamics (CFD) study results, and make recommendations for geometric changes to result in improved nozzle thrust coefficient performance (Cfg).
NASA Technical Reports Server (NTRS)
Vadyak, J.; Hoffman, J. D.
1978-01-01
The influence of molecular transport is included in the computation by treating viscous and thermal diffusion terms in the governing partial differential equations as correction terms in the method of characteristics scheme. The development of a production type computer program is reported which is capable of calculating the flow field in a variety of axisymmetric mixed-compression aircraft inlets. The results agreed well with those produced by the two-dimensional method characteristics when axisymmetric flow fields are computed. For three-dimensional flow fields, the results agree well with experimental data except in regions of high viscous interaction and boundary layer removal.
Continuum Absorption Coefficient of Atoms and Ions
NASA Technical Reports Server (NTRS)
Armaly, B. F.
1979-01-01
The rate of heat transfer to the heat shield of a Jupiter probe has been estimated to be one order of magnitude higher than any previously experienced in an outer space exploration program. More than one-third of this heat load is due to an emission of continuum radiation from atoms and ions. The existing computer code for calculating the continuum contribution to the total load utilizes a modified version of Biberman's approximate method. The continuum radiation absorption cross sections of a C - H - O - N ablation system were examined in detail. The present computer code was evaluated and updated by being compared with available exact and approximate calculations and correlations of experimental data. A detailed calculation procedure, which can be applied to other atomic species, is presented. The approximate correlations can be made to agree with the available exact and experimental data.
A small Unix-based data acquisition system
DOE Office of Scientific and Technical Information (OSTI.GOV)
Engberg, D.; Glanzman, T.
1994-02-01
The proposed SLAC B Factory detector plans to use Unix-based machines for all aspects of computing, including real-time data acquisition and experimental control. A R and D program has been established to investigate the use of Unix in the various aspects of experimental computation. Earlier R and D work investigated the basic real-time aspects of the IBM RS/6000 workstation running AIX, which claims to be a real-time operating system. The next step in this R and D is the construction of a prototype data acquisition system which attempts to exercise many of the features needed in the final on-line systemmore » in a realistic situation. For this project, the authors have combined efforts with a team studying the use of novel cell designs and gas mixtures in a new prototype drift chamber.« less
Extreme Scale Computing to Secure the Nation
DOE Office of Scientific and Technical Information (OSTI.GOV)
Brown, D L; McGraw, J R; Johnson, J R
2009-11-10
Since the dawn of modern electronic computing in the mid 1940's, U.S. national security programs have been dominant users of every new generation of high-performance computer. Indeed, the first general-purpose electronic computer, ENIAC (the Electronic Numerical Integrator and Computer), was used to calculate the expected explosive yield of early thermonuclear weapons designs. Even the U. S. numerical weather prediction program, another early application for high-performance computing, was initially funded jointly by sponsors that included the U.S. Air Force and Navy, agencies interested in accurate weather predictions to support U.S. military operations. For the decades of the cold war, national securitymore » requirements continued to drive the development of high performance computing (HPC), including advancement of the computing hardware and development of sophisticated simulation codes to support weapons and military aircraft design, numerical weather prediction as well as data-intensive applications such as cryptography and cybersecurity U.S. national security concerns continue to drive the development of high-performance computers and software in the U.S. and in fact, events following the end of the cold war have driven an increase in the growth rate of computer performance at the high-end of the market. This mainly derives from our nation's observance of a moratorium on underground nuclear testing beginning in 1992, followed by our voluntary adherence to the Comprehensive Test Ban Treaty (CTBT) beginning in 1995. The CTBT prohibits further underground nuclear tests, which in the past had been a key component of the nation's science-based program for assuring the reliability, performance and safety of U.S. nuclear weapons. In response to this change, the U.S. Department of Energy (DOE) initiated the Science-Based Stockpile Stewardship (SBSS) program in response to the Fiscal Year 1994 National Defense Authorization Act, which requires, 'in the absence of nuclear testing, a progam to: (1) Support a focused, multifaceted program to increase the understanding of the enduring stockpile; (2) Predict, detect, and evaluate potential problems of the aging of the stockpile; (3) Refurbish and re-manufacture weapons and components, as required; and (4) Maintain the science and engineering institutions needed to support the nation's nuclear deterrent, now and in the future'. This program continues to fulfill its national security mission by adding significant new capabilities for producing scientific results through large-scale computational simulation coupled with careful experimentation, including sub-critical nuclear experiments permitted under the CTBT. To develop the computational science and the computational horsepower needed to support its mission, SBSS initiated the Accelerated Strategic Computing Initiative, later renamed the Advanced Simulation & Computing (ASC) program (sidebar: 'History of ASC Computing Program Computing Capability'). The modern 3D computational simulation capability of the ASC program supports the assessment and certification of the current nuclear stockpile through calibration with past underground test (UGT) data. While an impressive accomplishment, continued evolution of national security mission requirements will demand computing resources at a significantly greater scale than we have today. In particular, continued observance and potential Senate confirmation of the Comprehensive Test Ban Treaty (CTBT) together with the U.S administration's promise for a significant reduction in the size of the stockpile and the inexorable aging and consequent refurbishment of the stockpile all demand increasing refinement of our computational simulation capabilities. Assessment of the present and future stockpile with increased confidence of the safety and reliability without reliance upon calibration with past or future test data is a long-term goal of the ASC program. This will be accomplished through significant increases in the scientific bases that underlie the computational tools. Computer codes must be developed that replace phenomenology with increased levels of scientific understanding together with an accompanying quantification of uncertainty. These advanced codes will place significantly higher demands on the computing infrastructure than do the current 3D ASC codes. This article discusses not only the need for a future computing capability at the exascale for the SBSS program, but also considers high performance computing requirements for broader national security questions. For example, the increasing concern over potential nuclear terrorist threats demands a capability to assess threats and potential disablement technologies as well as a rapid forensic capability for determining a nuclear weapons design from post-detonation evidence (nuclear counterterrorism).« less
The VERRUN and VERNAL software systems for steady-state visual evoked response experimentation
NASA Technical Reports Server (NTRS)
Levison, W. H.; Zacharias, G. L.
1984-01-01
Two digital computer programs were developed for use in experiments involving steady-state visual evoked response (VER): VERRUN, whose primary functions are to generate a sum-of-sines (SOS) stimulus and to digitize and store electro-cortical response; and VERNAL, which provides both time- and frequency-domain metrics of the evoked response. These programs were coded in FORTRAN for operation on the PDP-11/34, using the RSX-11 Operating System, and the PDP-11/23, using the RT-11 Operating System. Users' and programmers' guides to these programs are provided, and guidelines for model analysis of VER data are suggested.
OJPOT: online judge & practice oriented teaching idea in programming courses
NASA Astrophysics Data System (ADS)
Wang, Gui Ping; Chen, Shu Yu; Yang, Xin; Feng, Rui
2016-05-01
Practical abilities are important for students from majors including Computer Science and Engineering, and Electrical Engineering. Along with the popularity of ACM International Collegiate Programming Contest (ACM/ICPC) and other programming contests, online judge (OJ) websites achieve rapid development, thus providing a new kind of programming practice, i.e. online practice. Due to fair and timely feedback results from OJ websites, online practice outperforms traditional programming practice. In order to promote students' practical abilities in programming and algorithm designing, this article presents a novel teaching idea, online judge & practice oriented teaching (OJPOT). OJPOT is applied to Programming Foundation course. OJPOT cultivates students' practical abilities through various kinds of programming practice, such as programming contests, online practice and course project. To verify the effectiveness of this novel teaching idea, this study conducts empirical research. The experimental results show that OJPOT works effectively in enhancing students' practical abilities compared with the traditional teaching idea.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Bess, John D.
2014-03-01
PROTEUS is a zero-power research reactor based on a cylindrical graphite annulus with a central cylindrical cavity. The graphite annulus remains basically the same for all experimental programs, but the contents of the central cavity are changed according to the type of reactor being investigated. Through most of its service history, PROTEUS has represented light-water reactors, but from 1992 to 1996 PROTEUS was configured as a pebble-bed reactor (PBR) critical facility and designated as HTR-PROTEUS. The nomenclature was used to indicate that this series consisted of High Temperature Reactor experiments performed in the PROTEUS assembly. During this period, seventeen criticalmore » configurations were assembled and various reactor physics experiments were conducted. These experiments included measurements of criticality, differential and integral control rod and safety rod worths, kinetics, reaction rates, water ingress effects, and small sample reactivity effects (Ref. 3). HTR-PROTEUS was constructed, and the experimental program was conducted, for the purpose of providing experimental benchmark data for assessment of reactor physics computer codes. Considerable effort was devoted to benchmark calculations as a part of the HTR-PROTEUS program. References 1 and 2 provide detailed data for use in constructing models for codes to be assessed. Reference 3 is a comprehensive summary of the HTR-PROTEUS experiments and the associated benchmark program. This document draws freely from these references. Only Cores 9 and 10 are evaluated in this benchmark report due to similarities in their construction. The other core configurations of the HTR-PROTEUS program are evaluated in their respective reports as outlined in Section 1.0. Cores 9 and 10 were evaluated and determined to be acceptable benchmark experiments.« less
DOE Office of Scientific and Technical Information (OSTI.GOV)
John D. Bess
2013-03-01
PROTEUS is a zero-power research reactor based on a cylindrical graphite annulus with a central cylindrical cavity. The graphite annulus remains basically the same for all experimental programs, but the contents of the central cavity are changed according to the type of reactor being investigated. Through most of its service history, PROTEUS has represented light-water reactors, but from 1992 to 1996 PROTEUS was configured as a pebble-bed reactor (PBR) critical facility and designated as HTR-PROTEUS. The nomenclature was used to indicate that this series consisted of High Temperature Reactor experiments performed in the PROTEUS assembly. During this period, seventeen criticalmore » configurations were assembled and various reactor physics experiments were conducted. These experiments included measurements of criticality, differential and integral control rod and safety rod worths, kinetics, reaction rates, water ingress effects, and small sample reactivity effects (Ref. 3). HTR-PROTEUS was constructed, and the experimental program was conducted, for the purpose of providing experimental benchmark data for assessment of reactor physics computer codes. Considerable effort was devoted to benchmark calculations as a part of the HTR-PROTEUS program. References 1 and 2 provide detailed data for use in constructing models for codes to be assessed. Reference 3 is a comprehensive summary of the HTR-PROTEUS experiments and the associated benchmark program. This document draws freely from these references. Only Cores 9 and 10 are evaluated in this benchmark report due to similarities in their construction. The other core configurations of the HTR-PROTEUS program are evaluated in their respective reports as outlined in Section 1.0. Cores 9 and 10 were evaluated and determined to be acceptable benchmark experiments.« less
DOE Office of Scientific and Technical Information (OSTI.GOV)
John D. Bess
2013-03-01
PROTEUS is a zero-power research reactor based on a cylindrical graphite annulus with a central cylindrical cavity. The graphite annulus remains basically the same for all experimental programs, but the contents of the central cavity are changed according to the type of reactor being investigated. Through most of its service history, PROTEUS has represented light-water reactors, but from 1992 to 1996 PROTEUS was configured as a pebble-bed reactor (PBR) critical facility and designated as HTR-PROTEUS. The nomenclature was used to indicate that this series consisted of High Temperature Reactor experiments performed in the PROTEUS assembly. During this period, seventeen criticalmore » configurations were assembled and various reactor physics experiments were conducted. These experiments included measurements of criticality, differential and integral control rod and safety rod worths, kinetics, reaction rates, water ingress effects, and small sample reactivity effects (Ref. 3). HTR-PROTEUS was constructed, and the experimental program was conducted, for the purpose of providing experimental benchmark data for assessment of reactor physics computer codes. Considerable effort was devoted to benchmark calculations as a part of the HTR-PROTEUS program. References 1 and 2 provide detailed data for use in constructing models for codes to be assessed. Reference 3 is a comprehensive summary of the HTR-PROTEUS experiments and the associated benchmark program. This document draws freely from these references. Only Cores 9 and 10 are evaluated in this benchmark report due to similarities in their construction. The other core configurations of the HTR-PROTEUS program are evaluated in their respective reports as outlined in Section 1.0. Cores 9 and 10 were evaluated and determined to be acceptable benchmark experiments.« less
Experimental Criticality Benchmarks for SNAP 10A/2 Reactor Cores
DOE Office of Scientific and Technical Information (OSTI.GOV)
Krass, A.W.
2005-12-19
This report describes computational benchmark models for nuclear criticality derived from descriptions of the Systems for Nuclear Auxiliary Power (SNAP) Critical Assembly (SCA)-4B experimental criticality program conducted by Atomics International during the early 1960's. The selected experimental configurations consist of fueled SNAP 10A/2-type reactor cores subject to varied conditions of water immersion and reflection under experimental control to measure neutron multiplication. SNAP 10A/2-type reactor cores are compact volumes fueled and moderated with the hydride of highly enriched uranium-zirconium alloy. Specifications for the materials and geometry needed to describe a given experimental configuration for a model using MCNP5 are provided. Themore » material and geometry specifications are adequate to permit user development of input for alternative nuclear safety codes, such as KENO. A total of 73 distinct experimental configurations are described.« less
An inlet analysis for the NASA hypersonic research engine aerothermodynamic integration model
NASA Technical Reports Server (NTRS)
Andrews, E. H., Jr.; Russell, J. W.; Mackley, E. A.; Simmonds, A. L.
1974-01-01
A theoretical analysis for the inlet of the NASA Hypersonic Research Engine (HRE) Aerothermodynamic Integration Model (AIM) has been undertaken by use of a method-of-characteristics computer program. The purpose of the analysis was to obtain pretest information on the full-scale HRE inlet in support of the experimental AIM program (completed May 1974). Mass-flow-ratio and additive-drag-coefficient schedules were obtained that well defined the range effected in the AIM tests. Mass-weighted average inlet total-pressure recovery, kinetic energy efficiency, and throat Mach numbers were obtained.
Suppression of combustion oscillations with mechanical damping devices
NASA Technical Reports Server (NTRS)
1971-01-01
Nonarray absorbing devices were investigated for use in rocket thrust chambers as instability suppressors. A theory for designing absorbing devices suitable for rocket application is derived, and a nonarray computer program is developed. The experimental program used to verify the theory is discussed. It is concluded that individual acoustical devices can be designed for maximum energy absorption, and it is recommended that single resonators be designed so that the ratio of the aperture diameter to the product of the quarter-wave length and cavity backing depth is less than one.
An improved multiple linear regression and data analysis computer program package
NASA Technical Reports Server (NTRS)
Sidik, S. M.
1972-01-01
NEWRAP, an improved version of a previous multiple linear regression program called RAPIER, CREDUC, and CRSPLT, allows for a complete regression analysis including cross plots of the independent and dependent variables, correlation coefficients, regression coefficients, analysis of variance tables, t-statistics and their probability levels, rejection of independent variables, plots of residuals against the independent and dependent variables, and a canonical reduction of quadratic response functions useful in optimum seeking experimentation. A major improvement over RAPIER is that all regression calculations are done in double precision arithmetic.
The use of numerical programs in research and academic institutions
NASA Astrophysics Data System (ADS)
Scupi, A. A.
2016-08-01
This paper is conceived on the idea that numerical programs using computer models of physical processes can be used both for scientific research and academic teaching to study different phenomena. Computational Fluid Dynamics (CFD) is used today on a large scale in research and academic institutions. CFD development is not limited to computer simulations of fluid flow phenomena. Analytical solutions for most fluid dynamics problems are already available for ideal or simplified situations for different situations. CFD is based on the Navier- Stokes (N-S) equations characterizing the flow of a single phase of any liquid. For multiphase flows the integrated N-S equations are complemented with equations of the Volume of Fluid Model (VOF) and with energy equations. Different turbulent models were used in the paper, each one of them with practical engineering applications: the flow around aerodynamic surfaces used as unconventional propulsion system, multiphase flows in a settling chamber and pneumatic transport systems, heat transfer in a heat exchanger etc. Some of them numerical results were validated by experimental results. Numerical programs are also used in academic institutions where certain aspects of various phenomena are presented to students (Bachelor, Master and PhD) for a better understanding of the phenomenon itself.
SPLASH program for three dimensional fluid dynamics with free surface boundaries
NASA Astrophysics Data System (ADS)
Yamaguchi, A.
1996-05-01
This paper describes a three dimensional computer program SPLASH that solves Navier-Stokes equations based on the Arbitrary Lagrangian Eulerian (ALE) finite element method. SPLASH has been developed for application to the fluid dynamics problems including the moving boundary of a liquid metal cooled Fast Breeder Reactor (FBR). To apply SPLASH code to the free surface behavior analysis, a capillary model using a cubic Spline function has been developed. Several sample problems, e.g., free surface oscillation, vortex shedding development, and capillary tube phenomena, are solved to verify the computer program. In the analyses, the numerical results are in good agreement with the theoretical value or experimental observance. Also SPLASH code has been applied to an analysis of a free surface sloshing experiment coupled with forced circulation flow in a rectangular tank. This is a simplified situation of the flow field in a reactor vessel of the FBR. The computational simulation well predicts the general behavior of the fluid flow inside and the free surface behavior. Analytical capability of the SPLASH code has been verified in this study and the application to more practical problems such as FBR design and safety analysis is under way.
Autonomous mobile robot research using the HERMIES-III robot
DOE Office of Scientific and Technical Information (OSTI.GOV)
Pin, F.G.; Beckerman, M.; Spelt, P.F.
1989-01-01
This paper reports on the status and future directions in the research, development and experimental validation of intelligent control techniques for autonomous mobile robots using the HERMIES-III robot at the Center for Engineering Systems Advanced research (CESAR) at Oak Ridge National Laboratory (ORNL). HERMIES-III is the fourth robot in a series of increasingly more sophisticated and capable experimental test beds developed at CESAR. HERMIES-III is comprised of a battery powered, onmi-directional wheeled platform with a seven degree-of-freedom manipulator arm, video cameras, sonar range sensors, laser imaging scanner and a dual computer system containing up to 128 NCUBE nodes in hypercubemore » configuration. All electronics, sensors, computers, and communication equipment required for autonomous operation of HERMIES-III are located on board along with sufficient battery power for three to four hours of operation. The paper first provides a more detailed description of the HERMIES-III characteristics, focussing on the new areas of research and demonstration now possible at CESAR with this new test-bed. The initial experimental program is then described with emphasis placed on autonomous performance of human-scale tasks (e.g., valve manipulation, use of tools), integration of a dexterous manipulator and platform motion in geometrically complex environments, and effective use of multiple cooperating robots (HERMIES-IIB and HERMIES- III). The paper concludes with a discussion of the integration problems and safety considerations necessarily arising from the set-up of an experimental program involving human-scale, multi-autonomous mobile robots performance. 10 refs., 3 figs.« less
A comparison of fitness-case sampling methods for genetic programming
NASA Astrophysics Data System (ADS)
Martínez, Yuliana; Naredo, Enrique; Trujillo, Leonardo; Legrand, Pierrick; López, Uriel
2017-11-01
Genetic programming (GP) is an evolutionary computation paradigm for automatic program induction. GP has produced impressive results but it still needs to overcome some practical limitations, particularly its high computational cost, overfitting and excessive code growth. Recently, many researchers have proposed fitness-case sampling methods to overcome some of these problems, with mixed results in several limited tests. This paper presents an extensive comparative study of four fitness-case sampling methods, namely: Interleaved Sampling, Random Interleaved Sampling, Lexicase Selection and Keep-Worst Interleaved Sampling. The algorithms are compared on 11 symbolic regression problems and 11 supervised classification problems, using 10 synthetic benchmarks and 12 real-world data-sets. They are evaluated based on test performance, overfitting and average program size, comparing them with a standard GP search. Comparisons are carried out using non-parametric multigroup tests and post hoc pairwise statistical tests. The experimental results suggest that fitness-case sampling methods are particularly useful for difficult real-world symbolic regression problems, improving performance, reducing overfitting and limiting code growth. On the other hand, it seems that fitness-case sampling cannot improve upon GP performance when considering supervised binary classification.
ERIC Educational Resources Information Center
Kirby, Frederick C.; Castagna, Paul A.
The purpose of this study is to estimate costs and benefits and to compute alternative benefit-cost ratios for both the individuals and the Federal Government as a result of investing time and resources in the Training and Technology (TAT) Project. TAT is a continuing experimental program in training skilled workers for private industry. The five…
ERIC Educational Resources Information Center
Spaniol, Mayra Muller; Shalev, Lilach; Kossyvaki, Lila; Mevorach, Carmel
2018-01-01
This study assessed the effectiveness of an attention intervention program (Computerized Progressive Attentional Training; CPAT) in improving academic performance of children with ASD. Fifteen 6-10 year olds with ASD attending a mainstream and a special school were assigned to an experimental (CPAT; n = 8) and active control (computer games; n =…
ERIC Educational Resources Information Center
Burns, Barbara; Hagerman, Alison
1989-01-01
Discussion of achievement motivation and children's ideas about themselves as learners focuses on a study of third graders that examined the effects of LOGO programing on performance. Incremental and entity theories of intelligence are explained, and treatment of the experimental group and the control group are described. (26 references) (LRW)
ERIC Educational Resources Information Center
Korkmaz, Özgen
2016-01-01
The purpose of the present research is to designate the effects of Scratch-based game activities on students' attitudes towards learning computer programming, self-efficacy beliefs and levels of academic achievement. The research was conducted through a pre-test-post-test control group quasi-experimental study. The study group consists of 49…
Hierarchy of simulation models for a turbofan gas engine
NASA Technical Reports Server (NTRS)
Longenbaker, W. E.; Leake, R. J.
1977-01-01
Steady-state and transient performance of an F-100-like turbofan gas engine are modeled by a computer program, DYNGEN, developed by NASA. The model employs block data maps and includes about 25 states. Low-order nonlinear analytical and linear techniques are described in terms of their application to the model. Experimental comparisons illustrating the accuracy of each model are presented.
Students' explanations in complex learning of disciplinary programming
NASA Astrophysics Data System (ADS)
Vieira, Camilo
Computational Science and Engineering (CSE) has been denominated as the third pillar of science and as a set of important skills to solve the problems of a global society. Along with the theoretical and the experimental approaches, computation offers a third alternative to solve complex problems that require processing large amounts of data, or representing complex phenomena that are not easy to experiment with. Despite the relevance of CSE, current professionals and scientists are not well prepared to take advantage of this set of tools and methods. Computation is usually taught in an isolated way from engineering disciplines, and therefore, engineers do not know how to exploit CSE affordances. This dissertation intends to introduce computational tools and methods contextualized within the Materials Science and Engineering curriculum. Considering that learning how to program is a complex task, the dissertation explores effective pedagogical practices that can support student disciplinary and computational learning. Two case studies will be evaluated to identify the characteristics of effective worked examples in the context of CSE. Specifically, this dissertation explores students explanations of these worked examples in two engineering courses with different levels of transparency: a programming course in materials science and engineering glass box and a thermodynamics course involving computational representations black box. Results from this study suggest that students benefit in different ways from writing in-code comments. These benefits include but are not limited to: connecting xv individual lines of code to the overall problem, getting familiar with the syntax, learning effective algorithm design strategies, and connecting computation with their discipline. Students in the glass box context generate higher quality explanations than students in the black box context. These explanations are related to students prior experiences. Specifically, students with low ability to do programming engage in a more thorough explanation process than students with high ability. This dissertation concludes proposing an adaptation to the instructional principles of worked-examples for the context of CSE education.
A memory module for experimental data handling
NASA Astrophysics Data System (ADS)
De Blois, J.
1985-02-01
A compact CAMAC memory module for experimental data handling was developed to eliminate the need of direct memory access in computer controlled measurements. When using autonomous controllers it also makes measurements more independent of the program and enlarges the available space for programs in the memory of the micro-computer. The memory module has three modes of operation: an increment-, a list- and a fifo mode. This is achieved by connecting the main parts, being: the memory (MEM), the fifo buffer (FIFO), the address buffer (BUF), two counters (AUX and ADDR) and a readout register (ROR), by an internal 24-bit databus. The time needed for databus operations is 1 μs, for measuring cycles as well as for CAMAC cycles. The FIFO provides temporary data storage during CAMAC cycles and separates the memory part from the application part. The memory is variable from 1 to 64K (24 bits) by using different types of memory chips. The application part, which forms 1/3 of the module, will be specially designed for each application and is added to the memory chian internal connector. The memory unit will be used in Mössbauer experiments and in thermal neutron scattering experiments.