Tsou, Christina; Haynes, Emma; Warner, Wayne D; Gray, Gordon; Thompson, Sandra C
2015-04-23
The need for better partnerships between Aboriginal organisations and mainstream agencies demands attention on process and relational elements of these partnerships, and improving partnership functioning through transformative or iterative evaluation procedures. This paper presents the findings of a literature review which examines the usefulness of existing partnership tools to the Australian Aboriginal-mainstream partnership (AMP) context. Three sets of best practice principles for successful AMP were selected based on authors' knowledge and experience. Items in each set of principles were separated into process and relational elements and used to guide the analysis of partnership assessment tools. The review and analysis of partnership assessment tools were conducted in three distinct but related parts. Part 1- identify and select reviews of partnership tools; part 2 - identify and select partnership self-assessment tool; part 3 - analysis of selected tools using AMP principles. The focus on relational and process elements in the partnership tools reviewed is consistent with the focus of Australian AMP principles by reconciliation advocates; however, historical context, lived experience, cultural context and approaches of Australian Aboriginal people represent key deficiencies in the tools reviewed. The overall assessment indicated that the New York Partnership Self-Assessment Tool and the VicHealth Partnership Analysis Tools reflect the greatest number of AMP principles followed by the Nuffield Partnership Assessment Tool. The New York PSAT has the strongest alignment with the relational elements while VicHealth and Nuffield tools showed greatest alignment with the process elements in the chosen AMP principles. Partnership tools offer opportunities for providing evidence based support to partnership development. The multiplicity of tools in existence and the reported uniqueness of each partnership, mean the development of a generic partnership analysis for AMP may not be a viable option for future effort.
Stochastic Simulation Tool for Aerospace Structural Analysis
NASA Technical Reports Server (NTRS)
Knight, Norman F.; Moore, David F.
2006-01-01
Stochastic simulation refers to incorporating the effects of design tolerances and uncertainties into the design analysis model and then determining their influence on the design. A high-level evaluation of one such stochastic simulation tool, the MSC.Robust Design tool by MSC.Software Corporation, has been conducted. This stochastic simulation tool provides structural analysts with a tool to interrogate their structural design based on their mathematical description of the design problem using finite element analysis methods. This tool leverages the analyst's prior investment in finite element model development of a particular design. The original finite element model is treated as the baseline structural analysis model for the stochastic simulations that are to be performed. A Monte Carlo approach is used by MSC.Robust Design to determine the effects of scatter in design input variables on response output parameters. The tool was not designed to provide a probabilistic assessment, but to assist engineers in understanding cause and effect. It is driven by a graphical-user interface and retains the engineer-in-the-loop strategy for design evaluation and improvement. The application problem for the evaluation is chosen to be a two-dimensional shell finite element model of a Space Shuttle wing leading-edge panel under re-entry aerodynamic loading. MSC.Robust Design adds value to the analysis effort by rapidly being able to identify design input variables whose variability causes the most influence in response output parameters.
Rapid SAW Sensor Development Tools
NASA Technical Reports Server (NTRS)
Wilson, William C.; Atkinson, Gary M.
2007-01-01
The lack of integrated design tools for Surface Acoustic Wave (SAW) devices has led us to develop tools for the design, modeling, analysis, and automatic layout generation of SAW devices. These tools enable rapid development of wireless SAW sensors. The tools developed have been designed to integrate into existing Electronic Design Automation (EDA) tools to take advantage of existing 3D modeling, and Finite Element Analysis (FEA). This paper presents the SAW design, modeling, analysis, and automated layout generation tools.
The dynamic analysis of drum roll lathe for machining of rollers
NASA Astrophysics Data System (ADS)
Qiao, Zheng; Wu, Dongxu; Wang, Bo; Li, Guo; Wang, Huiming; Ding, Fei
2014-08-01
An ultra-precision machine tool for machining of the roller has been designed and assembled, and due to the obvious impact which dynamic characteristic of machine tool has on the quality of microstructures on the roller surface, the dynamic characteristic of the existing machine tool is analyzed in this paper, so is the influence of circumstance that a large scale and slender roller is fixed in the machine on dynamic characteristic of the machine tool. At first, finite element model of the machine tool is built and simplified, and based on that, the paper carries on with the finite element mode analysis and gets the natural frequency and shaking type of four steps of the machine tool. According to the above model analysis results, the weak stiffness systems of machine tool can be further improved and the reasonable bandwidth of control system of the machine tool can be designed. In the end, considering the shock which is caused by Z axis as a result of fast positioning frequently to feeding system and cutting tool, transient analysis is conducted by means of ANSYS analysis in this paper. Based on the results of transient analysis, the vibration regularity of key components of machine tool and its impact on cutting process are explored respectively.
The Applications of Finite Element Analysis in Proximal Humeral Fractures.
Ye, Yongyu; You, Wei; Zhu, Weimin; Cui, Jiaming; Chen, Kang; Wang, Daping
2017-01-01
Proximal humeral fractures are common and most challenging, due to the complexity of the glenohumeral joint, especially in the geriatric population with impacted fractures, that the development of implants continues because currently the problems with their fixation are not solved. Pre-, intra-, and postoperative assessments are crucial in management of those patients. Finite element analysis, as one of the valuable tools, has been implemented as an effective and noninvasive method to analyze proximal humeral fractures, providing solid evidence for management of troublesome patients. However, no review article about the applications and effects of finite element analysis in assessing proximal humeral fractures has been reported yet. This review article summarized the applications, contribution, and clinical significance of finite element analysis in assessing proximal humeral fractures. Furthermore, the limitations of finite element analysis, the difficulties of more realistic simulation, and the validation and also the creation of validated FE models were discussed. We concluded that although some advancements in proximal humeral fractures researches have been made by using finite element analysis, utility of this powerful tool for routine clinical management and adequate simulation requires more state-of-the-art studies to provide evidence and bases.
RSAT: regulatory sequence analysis tools.
Thomas-Chollier, Morgane; Sand, Olivier; Turatsinze, Jean-Valéry; Janky, Rekin's; Defrance, Matthieu; Vervisch, Eric; Brohée, Sylvain; van Helden, Jacques
2008-07-01
The regulatory sequence analysis tools (RSAT, http://rsat.ulb.ac.be/rsat/) is a software suite that integrates a wide collection of modular tools for the detection of cis-regulatory elements in genome sequences. The suite includes programs for sequence retrieval, pattern discovery, phylogenetic footprint detection, pattern matching, genome scanning and feature map drawing. Random controls can be performed with random gene selections or by generating random sequences according to a variety of background models (Bernoulli, Markov). Beyond the original word-based pattern-discovery tools (oligo-analysis and dyad-analysis), we recently added a battery of tools for matrix-based detection of cis-acting elements, with some original features (adaptive background models, Markov-chain estimation of P-values) that do not exist in other matrix-based scanning tools. The web server offers an intuitive interface, where each program can be accessed either separately or connected to the other tools. In addition, the tools are now available as web services, enabling their integration in programmatic workflows. Genomes are regularly updated from various genome repositories (NCBI and EnsEMBL) and 682 organisms are currently supported. Since 1998, the tools have been used by several hundreds of researchers from all over the world. Several predictions made with RSAT were validated experimentally and published.
[Short interspersed repetitive sequences (SINEs) and their use as a phylogenetic tool].
Kramerov, D A; Vasetskiĭ, N S
2009-01-01
The data on one of the most common repetitive elements of eukaryotic genomes, short interspersed elements (SINEs), are reviewed. Their structure, origin, and functioning in the genome are discussed. The variation and abundance of these neutral genomic markers makes them a convenient and reliable tool for phylogenetic analysis. The main methods of such analysis are presented, and the potential and limitations of this approach are discussed using specific examples.
Status of the Combustion Devices Injector Technology Program at the NASA MSFC
NASA Technical Reports Server (NTRS)
Jones, Gregg; Protz, Christopher; Trinh, Huu; Tucker, Kevin; Nesman, Tomas; Hulka, James
2005-01-01
To support the NASA Space Exploration Mission, an in-house program called Combustion Devices Injector Technology (CDIT) is being conducted at the NASA Marshall Space Flight Center (MSFC) for the fiscal year 2005. CDIT is focused on developing combustor technology and analysis tools to improve reliability and durability of upper-stage and in-space liquid propellant rocket engines. The three areas of focus include injector/chamber thermal compatibility, ignition, and combustion stability. In the compatibility and ignition areas, small-scale single- and multi-element hardware experiments will be conducted to demonstrate advanced technological concepts as well as to provide experimental data for validation of computational analysis tools. In addition, advanced analysis tools will be developed to eventually include 3-dimensional and multi- element effects and improve capability and validity to analyze heat transfer and ignition in large, multi-element injectors.
NASA Technical Reports Server (NTRS)
Hairr, John W.; Dorris, William J.; Ingram, J. Edward; Shah, Bharat M.
1993-01-01
Interactive Stiffened Panel Analysis (ISPAN) modules, written in FORTRAN, were developed to provide an easy to use tool for creating finite element models of composite material stiffened panels. The modules allow the user to interactively construct, solve and post-process finite element models of four general types of structural panel configurations using only the panel dimensions and properties as input data. Linear, buckling and post-buckling solution capability is provided. This interactive input allows rapid model generation and solution by non finite element users. The results of a parametric study of a blade stiffened panel are presented to demonstrate the usefulness of the ISPAN modules. Also, a non-linear analysis of a test panel was conducted and the results compared to measured data and previous correlation analysis.
Elasto-Plastic Analysis of Tee Joints Using HOT-SMAC
NASA Technical Reports Server (NTRS)
Arnold, Steve M. (Technical Monitor); Bednarcyk, Brett A.; Yarrington, Phillip W.
2004-01-01
The Higher Order Theory - Structural/Micro Analysis Code (HOT-SMAC) software package is applied to analyze the linearly elastic and elasto-plastic response of adhesively bonded tee joints. Joints of this type are finding an increasing number of applications with the increased use of composite materials within advanced aerospace vehicles, and improved tools for the design and analysis of these joints are needed. The linearly elastic results of the code are validated vs. finite element analysis results from the literature under different loading and boundary conditions, and new results are generated to investigate the inelastic behavior of the tee joint. The comparison with the finite element results indicates that HOT-SMAC is an efficient and accurate alternative to the finite element method and has a great deal of potential as an analysis tool for a wide range of bonded joints.
Finite-element analysis of NiTi wire deflection during orthodontic levelling treatment
NASA Astrophysics Data System (ADS)
Razali, M. F.; Mahmud, A. S.; Mokhtar, N.; Abdullah, J.
2016-02-01
Finite-element analysis is an important product development tool in medical devices industry for design and failure analysis of devices. This tool helps device designers to quickly explore various design options, optimizing specific designs and providing a deeper insight how a device is actually performing. In this study, three-dimensional finite-element models of superelastic nickel-titanium arch wire engaged in a three brackets system were developed. The aim was to measure the effect of binding friction developed on wire-bracket interaction towards the remaining recovery force available for tooth movement. Uniaxial and three brackets bending test were modelled and validated against experimental works. The prediction made by the three brackets bending models shows good agreement with the experimental results.
Multiple heteroatom substitution to graphene nanoribbon
Meyer, Ernst
2018-01-01
Substituting heteroatoms into nanostructured graphene elements, such as graphene nanoribbons, offers the possibility for atomic engineering of electronic properties. To characterize these substitutions, functionalized atomic force microscopy (AFM)—a tool to directly resolve chemical structures—is one of the most promising tools, yet the chemical analysis of heteroatoms has been rarely performed. We synthesized multiple heteroatom-substituted graphene nanoribbons and showed that AFM can directly resolve elemental differences and can be correlated to the van der Waals radii, as well as the modulated local electron density caused by the substitution. This elemental-sensitive measurement takes an important step in the analysis of functionalized two-dimensional carbon materials. PMID:29662955
Toward automatic finite element analysis
NASA Technical Reports Server (NTRS)
Kela, Ajay; Perucchio, Renato; Voelcker, Herbert
1987-01-01
Two problems must be solved if the finite element method is to become a reliable and affordable blackbox engineering tool. Finite element meshes must be generated automatically from computer aided design databases and mesh analysis must be made self-adaptive. The experimental system described solves both problems in 2-D through spatial and analytical substructuring techniques that are now being extended into 3-D.
Ablative Thermal Response Analysis Using the Finite Element Method
NASA Technical Reports Server (NTRS)
Dec John A.; Braun, Robert D.
2009-01-01
A review of the classic techniques used to solve ablative thermal response problems is presented. The advantages and disadvantages of both the finite element and finite difference methods are described. As a first step in developing a three dimensional finite element based ablative thermal response capability, a one dimensional computer tool has been developed. The finite element method is used to discretize the governing differential equations and Galerkin's method of weighted residuals is used to derive the element equations. A code to code comparison between the current 1-D tool and the 1-D Fully Implicit Ablation and Thermal Response Program (FIAT) has been performed.
Automation Tools for Finite Element Analysis of Adhesively Bonded Joints
NASA Technical Reports Server (NTRS)
Tahmasebi, Farhad; Brodeur, Stephen J. (Technical Monitor)
2002-01-01
This article presents two new automation creation tools that obtain stresses and strains (Shear and peel) in adhesively bonded joints. For a given adhesively bonded joint Finite Element model, in which the adhesive is characterised using springs, these automation tools read the corresponding input and output files, use the spring forces and deformations to obtain the adhesive stresses and strains, sort the stresses and strains in descending order, and generate plot files for 3D visualisation of the stress and strain fields. Grids (nodes) and elements can be numbered in any order that is convenient for the user. Using the automation tools, trade-off studies, which are needed for design of adhesively bonded joints, can be performed very quickly.
eShadow: A tool for comparing closely related sequences
DOE Office of Scientific and Technical Information (OSTI.GOV)
Ovcharenko, Ivan; Boffelli, Dario; Loots, Gabriela G.
2004-01-15
Primate sequence comparisons are difficult to interpret due to the high degree of sequence similarity shared between such closely related species. Recently, a novel method, phylogenetic shadowing, has been pioneered for predicting functional elements in the human genome through the analysis of multiple primate sequence alignments. We have expanded this theoretical approach to create a computational tool, eShadow, for the identification of elements under selective pressure in multiple sequence alignments of closely related genomes, such as in comparisons of human to primate or mouse to rat DNA. This tool integrates two different statistical methods and allows for the dynamic visualizationmore » of the resulting conservation profile. eShadow also includes a versatile optimization module capable of training the underlying Hidden Markov Model to differentially predict functional sequences. This module grants the tool high flexibility in the analysis of multiple sequence alignments and in comparing sequences with different divergence rates. Here, we describe the eShadow comparative tool and its potential uses for analyzing both multiple nucleotide and protein alignments to predict putative functional elements. The eShadow tool is publicly available at http://eshadow.dcode.org/« less
Birch, Ivan; Vernon, Wesley; Walker, Jeremy; Saxelby, Jai
2013-10-01
Gait analysis from closed circuit camera footage is now commonly used as evidence in criminal trials. The biomechanical analysis of human gait is a well established science in both clinical and laboratory settings. However, closed circuit camera footage is rarely of the quality of that taken in the more controlled clinical and laboratory environments. The less than ideal quality of much of this footage for use in gait analysis is associated with a range of issues, the combination of which can often render the footage unsuitable for use in gait analysis. The aim of this piece of work was to develop a tool for assessing the suitability of closed circuit camera footage for the purpose of forensic gait analysis. A Delphi technique was employed with a small sample of expert forensic gait analysis practitioners, to identify key quality elements of CCTV footage used in legal proceedings. Five elements of the footage were identified and then subdivided into 15 contributing sub-elements, each of which was scored using a 5-point Likert scale. A Microsoft Excel worksheet was developed to calculate automatically an overall score from the fifteen sub-element scores. Five expert witnesses experienced in using CCTV footage for gait analysis then trialled the prototype tool on current case footage. A repeatability study was also undertaken using standardized CCTV footage. The results showed the tool to be a simple and repeatable means of assessing the suitability of closed circuit camera footage for use in forensic gait analysis. The inappropriate use of poor quality footage could lead to challenges to the practice of forensic gait analysis. All parties involved in criminal proceedings must therefore understand the fitness for purpose of any footage used. The development of this tool could offer a method of achieving this goal, and help to assure the continued role of forensic gait analysis as an aid to the identification process. Copyright © 2013 Elsevier Ltd and Faculty of Forensic and Legal Medicine. All rights reserved.
[Progress in the application of laser ablation ICP-MS to surface microanalysis in material science].
Zhang, Yong; Jia, Yun-hai; Chen, Ji-wen; Shen, Xue-jing; Liu, Ying; Zhao, Leiz; Li, Dong-ling; Hang, Peng-cheng; Zhao, Zhen; Fan, Wan-lun; Wang, Hai-zhou
2014-08-01
In the present paper, apparatus and theory of surface analysis is introduced, and the progress in the application of laser ablation ICP-MS to microanalysis in ferrous, nonferrous and semiconductor field is reviewed in detail. Compared with traditional surface analytical tools, such as SEM/EDS (scanning electron microscopy/energy dispersive spectrum), EPMA (electron probe microanalysis analysis), AES (auger energy spectrum), etc. the advantage is little or no sample preparation, adjustable spatial resolution according to analytical demand, multi-element analysis and high sensitivity. It is now a powerful complementary method to traditional surface analytical tool. With the development of LA-ICP-MS technology maturing, more and more analytical workers will use this powerful tool in the future, and LA-ICP-MS will be a super star in elemental analysis field just like LIBS (Laser-induced breakdown spectroscopy).
Integrated Sensitivity Analysis Workflow
DOE Office of Scientific and Technical Information (OSTI.GOV)
Friedman-Hill, Ernest J.; Hoffman, Edward L.; Gibson, Marcus J.
2014-08-01
Sensitivity analysis is a crucial element of rigorous engineering analysis, but performing such an analysis on a complex model is difficult and time consuming. The mission of the DART Workbench team at Sandia National Laboratories is to lower the barriers to adoption of advanced analysis tools through software integration. The integrated environment guides the engineer in the use of these integrated tools and greatly reduces the cycle time for engineering analysis.
Nonlinear transient analysis via energy minimization
NASA Technical Reports Server (NTRS)
Kamat, M. P.; Knight, N. F., Jr.
1978-01-01
The formulation basis for nonlinear transient analysis of finite element models of structures using energy minimization is provided. Geometric and material nonlinearities are included. The development is restricted to simple one and two dimensional finite elements which are regarded as being the basic elements for modeling full aircraft-like structures under crash conditions. The results indicate the effectiveness of the technique as a viable tool for this purpose.
NASA Astrophysics Data System (ADS)
Ee, K. C.; Dillon, O. W.; Jawahir, I. S.
2004-06-01
This paper discusses the influence of major chip-groove parameters of a cutting tool on the chip formation process in orthogonal machining using finite element (FE) methods. In the FE formulation, a thermal elastic-viscoplastic material model is used together with a modified Johnson-Cook material law for the flow stress. The chip back-flow angle and the chip up-curl radius are calculated for a range of cutting conditions by varying the chip-groove parameters. The analysis provides greater understanding of the effectiveness of chip-groove configurations and points a way to correlate cutting conditions with tool-wear when machining with a grooved cutting tool.
Finite Element Analysis of Surface Residual Stress in Functionally Gradient Cemented Carbide Tool
NASA Astrophysics Data System (ADS)
Su, Chuangnan; Liu, Deshun; Tang, Siwen; Li, Pengnan; Qiu, Xinyi
2018-03-01
A component distribution model is proposed for three-component functionally gradient cemented carbide (FGCC) based on electron probe microanalysis results obtained for gradient layer thickness, microstructure, and elemental distribution. The residual surface stress of FGCC-T5 tools occurring during the fabrication process is analyzed using an ANSYS-implemented finite element method (FEM) and X-ray diffraction. A comparison of the experimental and calculated values verifies the feasibility of using FEM to analyze the residual surface stress in FGCC-T5 tools. The effects of the distribution index, geometrical shape, substrate thickness, gradient layer thickness, and position of the cobalt-rich layer on residual surface stress are studied in detail.
Mangado, Nerea; Piella, Gemma; Noailly, Jérôme; Pons-Prats, Jordi; Ballester, Miguel Ángel González
2016-01-01
Computational modeling has become a powerful tool in biomedical engineering thanks to its potential to simulate coupled systems. However, real parameters are usually not accurately known, and variability is inherent in living organisms. To cope with this, probabilistic tools, statistical analysis and stochastic approaches have been used. This article aims to review the analysis of uncertainty and variability in the context of finite element modeling in biomedical engineering. Characterization techniques and propagation methods are presented, as well as examples of their applications in biomedical finite element simulations. Uncertainty propagation methods, both non-intrusive and intrusive, are described. Finally, pros and cons of the different approaches and their use in the scientific community are presented. This leads us to identify future directions for research and methodological development of uncertainty modeling in biomedical engineering. PMID:27872840
Mangado, Nerea; Piella, Gemma; Noailly, Jérôme; Pons-Prats, Jordi; Ballester, Miguel Ángel González
2016-01-01
Computational modeling has become a powerful tool in biomedical engineering thanks to its potential to simulate coupled systems. However, real parameters are usually not accurately known, and variability is inherent in living organisms. To cope with this, probabilistic tools, statistical analysis and stochastic approaches have been used. This article aims to review the analysis of uncertainty and variability in the context of finite element modeling in biomedical engineering. Characterization techniques and propagation methods are presented, as well as examples of their applications in biomedical finite element simulations. Uncertainty propagation methods, both non-intrusive and intrusive, are described. Finally, pros and cons of the different approaches and their use in the scientific community are presented. This leads us to identify future directions for research and methodological development of uncertainty modeling in biomedical engineering.
NASA Astrophysics Data System (ADS)
Scaini, C.; Felpeto, A.; Martí, J.; Carniel, R.
2014-05-01
This paper presents a GIS-based methodology to estimate damages produced by volcanic eruptions. The methodology is constituted by four parts: definition and simulation of eruptive scenarios, exposure analysis, vulnerability assessment and estimation of expected damages. Multi-hazard eruptive scenarios are defined for the Teide-Pico Viejo active volcanic complex, and simulated through the VORIS tool. The exposure analysis identifies the elements exposed to the hazard at stake and focuses on the relevant assets for the study area. The vulnerability analysis is based on previous studies on the built environment and complemented with the analysis of transportation and urban infrastructures. Damage assessment is performed associating a qualitative damage rating to each combination of hazard and vulnerability. This operation consists in a GIS-based overlap, performed for each hazardous phenomenon considered and for each element. The methodology is then automated into a GIS-based tool using an ArcGIS® program. Given the eruptive scenarios and the characteristics of the exposed elements, the tool produces expected damage maps. The tool is applied to the Icod Valley (North of Tenerife Island) which is likely to be affected by volcanic phenomena in case of eruption from both the Teide-Pico Viejo volcanic complex and North-West basaltic rift. Results are thematic maps of vulnerability and damage that can be displayed at different levels of detail, depending on the user preferences. The aim of the tool is to facilitate territorial planning and risk management in active volcanic areas.
Laser induced breakdown spectroscopy (LIBS) as a rapid tool for material analysis
NASA Astrophysics Data System (ADS)
Hussain, T.; Gondal, M. A.
2013-06-01
Laser induced breakdown spectroscopy (LIBS) is a novel technique for elemental analysis based on laser-generated plasma. In this technique, laser pulses are applied for ablation of the sample, resulting in the vaporization and ionization of sample in hot plasma which is finally analyzed by the spectrometer. The elements are identified by their unique spectral signatures. LIBS system was developed for elemental analysis of solid and liquid samples. The developed system was applied for qualitative as well as quantitative measurement of elemental concentration present in iron slag and open pit ore samples. The plasma was generated by focusing a pulsed Nd:YAG laser at 1064 nm on test samples to study the capabilities of LIBS as a rapid tool for material analysis. The concentrations of various elements of environmental significance such as cadmium, calcium, magnesium, chromium, manganese, titanium, barium, phosphorus, copper, iron, zinc etc., in these samples were determined. Optimal experimental conditions were evaluated for improving the sensitivity of developed LIBS system through parametric dependence study. The laser-induced breakdown spectroscopy (LIBS) results were compared with the results obtained using standard analytical technique such as inductively couple plasma emission spectroscopy (ICP). Limit of detection (LOD) of our LIBS system were also estimated for the above mentioned elements. This study demonstrates that LIBS could be highly appropriate for rapid online analysis of iron slag and open pit waste.
Simultaneous Aerodynamic and Structural Design Optimization (SASDO) for a 3-D Wing
NASA Technical Reports Server (NTRS)
Gumbert, Clyde R.; Hou, Gene J.-W.; Newman, Perry A.
2001-01-01
The formulation and implementation of an optimization method called Simultaneous Aerodynamic and Structural Design Optimization (SASDO) is shown as an extension of the Simultaneous Aerodynamic Analysis and Design Optimization (SAADO) method. It is extended by the inclusion of structure element sizing parameters as design variables and Finite Element Method (FEM) analysis responses as constraints. The method aims to reduce the computational expense. incurred in performing shape and sizing optimization using state-of-the-art Computational Fluid Dynamics (CFD) flow analysis, FEM structural analysis and sensitivity analysis tools. SASDO is applied to a simple. isolated, 3-D wing in inviscid flow. Results show that the method finds the saine local optimum as a conventional optimization method with some reduction in the computational cost and without significant modifications; to the analysis tools.
Localized Overheating Phenomena and Optimization of Spark-Plasma Sintering Tooling Design
Giuntini, Diletta; Olevsky, Eugene A.; Garcia-Cardona, Cristina; Maximenko, Andrey L.; Yurlova, Maria S.; Haines, Christopher D.; Martin, Darold G.; Kapoor, Deepak
2013-01-01
The present paper shows the application of a three-dimensional coupled electrical, thermal, mechanical finite element macro-scale modeling framework of Spark Plasma Sintering (SPS) to an actual problem of SPS tooling overheating, encountered during SPS experimentation. The overheating phenomenon is analyzed by varying the geometry of the tooling that exhibits the problem, namely by modeling various tooling configurations involving sequences of disk-shape spacers with step-wise increasing radii. The analysis is conducted by means of finite element simulations, intended to obtain temperature spatial distributions in the graphite press-forms, including punches, dies, and spacers; to identify the temperature peaks and their respective timing, and to propose a more suitable SPS tooling configuration with the avoidance of the overheating as a final aim. Electric currents-based Joule heating, heat transfer, mechanical conditions, and densification are imbedded in the model, utilizing the finite-element software COMSOL™, which possesses a distinguishing ability of coupling multiple physics. Thereby the implementation of a finite element method applicable to a broad range of SPS procedures is carried out, together with the more specific optimization of the SPS tooling design when dealing with excessive heating phenomena. PMID:28811398
Patient-specific finite element modeling of bones.
Poelert, Sander; Valstar, Edward; Weinans, Harrie; Zadpoor, Amir A
2013-04-01
Finite element modeling is an engineering tool for structural analysis that has been used for many years to assess the relationship between load transfer and bone morphology and to optimize the design and fixation of orthopedic implants. Due to recent developments in finite element model generation, for example, improved computed tomography imaging quality, improved segmentation algorithms, and faster computers, the accuracy of finite element modeling has increased vastly and finite element models simulating the anatomy and properties of an individual patient can be constructed. Such so-called patient-specific finite element models are potentially valuable tools for orthopedic surgeons in fracture risk assessment or pre- and intraoperative planning of implant placement. The aim of this article is to provide a critical overview of current themes in patient-specific finite element modeling of bones. In addition, the state-of-the-art in patient-specific modeling of bones is compared with the requirements for a clinically applicable patient-specific finite element method, and judgment is passed on the feasibility of application of patient-specific finite element modeling as a part of clinical orthopedic routine. It is concluded that further development in certain aspects of patient-specific finite element modeling are needed before finite element modeling can be used as a routine clinical tool.
CFD Methods and Tools for Multi-Element Airfoil Analysis
NASA Technical Reports Server (NTRS)
Rogers, Stuart E.; George, Michael W. (Technical Monitor)
1995-01-01
This lecture will discuss the computational tools currently available for high-lift multi-element airfoil analysis. It will present an overview of a number of different numerical approaches, their current capabilities, short-comings, and computational costs. The lecture will be limited to viscous methods, including inviscid/boundary layer coupling methods, and incompressible and compressible Reynolds-averaged Navier-Stokes methods. Both structured and unstructured grid generation approaches will be presented. Two different structured grid procedures are outlined, one which uses multi-block patched grids, the other uses overset chimera grids. Turbulence and transition modeling will be discussed.
ERIC Educational Resources Information Center
Delgadillo, Elizabeth Montoya; Vivier, Laurent
2016-01-01
Mathematical working space (MWS) is a model that is used in research in mathematics education, particularly in the field of geometry. Some MWS elements are independent of the field while other elements must be adapted to the field in question. In this paper, we develop the MWS model for the field of analysis with an identification of paradigms. We…
NASA Technical Reports Server (NTRS)
Ferragut, N. J.
1982-01-01
The Shuttle Pointed Autonomous Research Tool for Astronomy (SPARTAN) family of spacecraft are intended to operate with minimum interfaces with the U.S. Space Shuttle in order to increase flight opportunities. The SPARTAN I Spacecraft was designed to enhance structural capabilities and increase reliability. The approach followed results from work experience which evolved from sounding rocket projects. Structural models were developed to do the analyses necessary to satisfy safety requirements for Shuttle hardware. A loads analysis must also be performed. Stress analysis calculations will be performed on the main structural elements and subcomponents. Attention is given to design considerations and program definition, the schematic representation of a finite element model used for SPARTAN I spacecraft, details of loads analysis, the stress analysis, and fracture mechanics plan implications.
An Analysis of the Elements of Collaboration Associated with Top Collaborative Tools
2010-03-01
lets you access your e-mail, calendar, and files from any web browser anywhere in the world. Web based www.hotoffice.com Noodle Vialect’s (parent...www.taroby.org Yuuguu Yuuguu is an instant screen sharing, web conferencing, remote support, desktop remote control and messaging tool. Client...Office, Noodle , Novlet, Revizr, Taroby, and Yuuguu) received all seven NS ratings (see Table 20 below). The overall ratings for the major elements
An Integrated Tool for System Analysis of Sample Return Vehicles
NASA Technical Reports Server (NTRS)
Samareh, Jamshid A.; Maddock, Robert W.; Winski, Richard G.
2012-01-01
The next important step in space exploration is the return of sample materials from extraterrestrial locations to Earth for analysis. Most mission concepts that return sample material to Earth share one common element: an Earth entry vehicle. The analysis and design of entry vehicles is multidisciplinary in nature, requiring the application of mass sizing, flight mechanics, aerodynamics, aerothermodynamics, thermal analysis, structural analysis, and impact analysis tools. Integration of a multidisciplinary problem is a challenging task; the execution process and data transfer among disciplines should be automated and consistent. This paper describes an integrated analysis tool for the design and sizing of an Earth entry vehicle. The current tool includes the following disciplines: mass sizing, flight mechanics, aerodynamics, aerothermodynamics, and impact analysis tools. Python and Java languages are used for integration. Results are presented and compared with the results from previous studies.
Some Observations on the Current Status of Performing Finite Element Analyses
NASA Technical Reports Server (NTRS)
Raju, Ivatury S.; Knight, Norman F., Jr; Shivakumar, Kunigal N.
2015-01-01
Aerospace structures are complex high-performance structures. Advances in reliable and efficient computing and modeling tools are enabling analysts to consider complex configurations, build complex finite element models, and perform analysis rapidly. Many of the early career engineers of today are very proficient in the usage of modern computers, computing engines, complex software systems, and visualization tools. These young engineers are becoming increasingly efficient in building complex 3D models of complicated aerospace components. However, the current trends demonstrate blind acceptance of the results of the finite element analysis results. This paper is aimed at raising an awareness of this situation. Examples of the common encounters are presented. To overcome the current trends, some guidelines and suggestions for analysts, senior engineers, and educators are offered.
Canizo, Brenda V; Escudero, Leticia B; Pérez, María B; Pellerano, Roberto G; Wuilloud, Rodolfo G
2018-03-01
The feasibility of the application of chemometric techniques associated with multi-element analysis for the classification of grape seeds according to their provenance vineyard soil was investigated. Grape seed samples from different localities of Mendoza province (Argentina) were evaluated. Inductively coupled plasma mass spectrometry (ICP-MS) was used for the determination of twenty-nine elements (Ag, As, Ce, Co, Cs, Cu, Eu, Fe, Ga, Gd, La, Lu, Mn, Mo, Nb, Nd, Ni, Pr, Rb, Sm, Te, Ti, Tl, Tm, U, V, Y, Zn and Zr). Once the analytical data were collected, supervised pattern recognition techniques such as linear discriminant analysis (LDA), partial least square discriminant analysis (PLS-DA), k-nearest neighbors (k-NN), support vector machine (SVM) and Random Forest (RF) were applied to construct classification/discrimination rules. The results indicated that nonlinear methods, RF and SVM, perform best with up to 98% and 93% accuracy rate, respectively, and therefore are excellent tools for classification of grapes. Copyright © 2017 Elsevier Ltd. All rights reserved.
Jaswal, Brij Bir S; Kumar, Vinay; Sharma, Jitendra; Rai, Pradeep K; Gondal, Mohammed A; Gondal, Bilal; Singh, Vivek K
2016-04-01
Laser-induced breakdown spectroscopy (LIBS) is an emerging analytical technique with numerous advantages such as rapidity, multi-elemental analysis, no specific sample preparation requirements, non-destructiveness, and versatility. It has been proven to be a robust elemental analysis tool attracting interest because of being applied to a wide range of materials including biomaterials. In this paper, we have performed spectroscopic studies on gallstones which are heterogeneous in nature using LIBS and wavelength dispersive X-ray fluorescence (WD-XRF) techniques. It has been observed that the presence and relative concentrations of trace elements in different kind of gallstones (cholesterol and pigment gallstones) can easily be determined using LIBS technique. From the experiments carried out on gallstones for trace elemental mapping and detection, it was found that LIBS is a robust tool for such biomedical applications. The stone samples studied in the present paper were classified using the Fourier transform infrared (FTIR) spectroscopy. WD-XRF spectroscopy has been applied for the qualitative and quantitative analysis of major and trace elements present in the gallstone which was compared with the LIBS data. The results obtained in the present paper show interesting prospects for LIBS and WD-XRF to study cholelithiasis better.
Modelling of tunnelling processes and rock cutting tool wear with the particle finite element method
NASA Astrophysics Data System (ADS)
Carbonell, Josep Maria; Oñate, Eugenio; Suárez, Benjamín
2013-09-01
Underground construction involves all sort of challenges in analysis, design, project and execution phases. The dimension of tunnels and their structural requirements are growing, and so safety and security demands do. New engineering tools are needed to perform a safer planning and design. This work presents the advances in the particle finite element method (PFEM) for the modelling and the analysis of tunneling processes including the wear of the cutting tools. The PFEM has its foundation on the Lagrangian description of the motion of a continuum built from a set of particles with known physical properties. The method uses a remeshing process combined with the alpha-shape technique to detect the contacting surfaces and a finite element method for the mechanical computations. A contact procedure has been developed for the PFEM which is combined with a constitutive model for predicting the excavation front and the wear of cutting tools. The material parameters govern the coupling of frictional contact and wear between the interacting domains at the excavation front. The PFEM allows predicting several parameters which are relevant for estimating the performance of a tunnelling boring machine such as wear in the cutting tools, the pressure distribution on the face of the boring machine and the vibrations produced in the machinery and the adjacent soil/rock. The final aim is to help in the design of the excavating tools and in the planning of the tunnelling operations. The applications presented show that the PFEM is a promising technique for the analysis of tunnelling problems.
NASA Astrophysics Data System (ADS)
Dahing, Lahasen@Normanshah; Yahya, Redzuan; Yahya, Roslan; Hassan, Hearie
2014-09-01
In this study, principle of prompt gamma neutron activation analysis has been used as a technique to determine the elements in the sample. The system consists of collimated isotopic neutron source, Cf-252 with HPGe detector and Multichannel Analysis (MCA). Concrete with size of 10×10×10 cm3 and 15×15×15 cm3 were analysed as sample. When neutrons enter and interact with elements in the concrete, the neutron capture reaction will occur and produce characteristic prompt gamma ray of the elements. The preliminary result of this study demonstrate the major element in the concrete was determined such as Si, Mg, Ca, Al, Fe and H as well as others element, such as Cl by analysis the gamma ray lines respectively. The results obtained were compared with NAA and XRF techniques as a part of reference and validation. The potential and the capability of neutron induced prompt gamma as tool for multi elemental analysis qualitatively to identify the elements present in the concrete sample discussed.
Large Angle Transient Dynamics (LATDYN) user's manual
NASA Technical Reports Server (NTRS)
Abrahamson, A. Louis; Chang, Che-Wei; Powell, Michael G.; Wu, Shih-Chin; Bingel, Bradford D.; Theophilos, Paula M.
1991-01-01
A computer code for modeling the large angle transient dynamics (LATDYN) of structures was developed to investigate techniques for analyzing flexible deformation and control/structure interaction problems associated with large angular motions of spacecraft. This type of analysis is beyond the routine capability of conventional analytical tools without simplifying assumptions. In some instances, the motion may be sufficiently slow and the spacecraft (or component) sufficiently rigid to simplify analyses of dynamics and controls by making pseudo-static and/or rigid body assumptions. The LATDYN introduces a new approach to the problem by combining finite element structural analysis, multi-body dynamics, and control system analysis in a single tool. It includes a type of finite element that can deform and rotate through large angles at the same time, and which can be connected to other finite elements either rigidly or through mechanical joints. The LATDYN also provides symbolic capabilities for modeling control systems which are interfaced directly with the finite element structural model. Thus, the nonlinear equations representing the structural model are integrated along with the equations representing sensors, processing, and controls as a coupled system.
NASA Technical Reports Server (NTRS)
Cirillo, William M.; Earle, Kevin D.; Goodliff, Kandyce E.; Reeves, J. D.; Stromgren, Chel; Andraschko, Mark R.; Merrill, R. Gabe
2008-01-01
NASA s Constellation Program employs a strategic analysis methodology in providing an integrated analysis capability of Lunar exploration scenarios and to support strategic decision-making regarding those scenarios. The strategic analysis methodology integrates the assessment of the major contributors to strategic objective satisfaction performance, affordability, and risk and captures the linkages and feedbacks between all three components. Strategic analysis supports strategic decision making by senior management through comparable analysis of alternative strategies, provision of a consistent set of high level value metrics, and the enabling of cost-benefit analysis. The tools developed to implement the strategic analysis methodology are not element design and sizing tools. Rather, these models evaluate strategic performance using predefined elements, imported into a library from expert-driven design/sizing tools or expert analysis. Specific components of the strategic analysis tool set include scenario definition, requirements generation, mission manifesting, scenario lifecycle costing, crew time analysis, objective satisfaction benefit, risk analysis, and probabilistic evaluation. Results from all components of strategic analysis are evaluated a set of pre-defined figures of merit (FOMs). These FOMs capture the high-level strategic characteristics of all scenarios and facilitate direct comparison of options. The strategic analysis methodology that is described in this paper has previously been applied to the Space Shuttle and International Space Station Programs and is now being used to support the development of the baseline Constellation Program lunar architecture. This paper will present an overview of the strategic analysis methodology and will present sample results from the application of the strategic analysis methodology to the Constellation Program lunar architecture.
Analysis of the electromagnetic wave resistivity tool in deviated well drilling
NASA Astrophysics Data System (ADS)
Zhang, Yumei; Xu, Lijun; Cao, Zhang
2014-04-01
Electromagnetic wave resistivity (EWR) tools are used to provide real-time measurements of resistivity in the formation around the tool in Logging While Drilling (LWD). In this paper, the acquired resistivity information in the formation is analyzed to extract more information, including dipping angle and azimuth direction of the drill. A finite element (FM) model of EWR tool working in layered earth formations is established. Numerical analysis and FM simulations are employed to analyze the amplitude ratio and phase difference between the voltages measured at the two receivers of the EWR tool in deviated well drilling.
Analysis of Brick Masonry Wall using Applied Element Method
NASA Astrophysics Data System (ADS)
Lincy Christy, D.; Madhavan Pillai, T. M.; Nagarajan, Praveen
2018-03-01
The Applied Element Method (AEM) is a versatile tool for structural analysis. Analysis is done by discretising the structure as in the case of Finite Element Method (FEM). In AEM, elements are connected by a set of normal and shear springs instead of nodes. AEM is extensively used for the analysis of brittle materials. Brick masonry wall can be effectively analyzed in the frame of AEM. The composite nature of masonry wall can be easily modelled using springs. The brick springs and mortar springs are assumed to be connected in series. The brick masonry wall is analyzed and failure load is determined for different loading cases. The results were used to find the best aspect ratio of brick to strengthen brick masonry wall.
Integrated Data Visualization and Virtual Reality Tool
NASA Technical Reports Server (NTRS)
Dryer, David A.
1998-01-01
The Integrated Data Visualization and Virtual Reality Tool (IDVVRT) Phase II effort was for the design and development of an innovative Data Visualization Environment Tool (DVET) for NASA engineers and scientists, enabling them to visualize complex multidimensional and multivariate data in a virtual environment. The objectives of the project were to: (1) demonstrate the transfer and manipulation of standard engineering data in a virtual world; (2) demonstrate the effects of design and changes using finite element analysis tools; and (3) determine the training and engineering design and analysis effectiveness of the visualization system.
Probabilistic finite elements for fatigue and fracture analysis
NASA Astrophysics Data System (ADS)
Belytschko, Ted; Liu, Wing Kam
Attenuation is focused on the development of Probabilistic Finite Element Method (PFEM), which combines the finite element method with statistics and reliability methods, and its application to linear, nonlinear structural mechanics problems and fracture mechanics problems. The computational tool based on the Stochastic Boundary Element Method is also given for the reliability analysis of a curvilinear fatigue crack growth. The existing PFEM's have been applied to solve for two types of problems: (1) determination of the response uncertainty in terms of the means, variance and correlation coefficients; and (2) determination the probability of failure associated with prescribed limit states.
Probabilistic finite elements for fatigue and fracture analysis
NASA Technical Reports Server (NTRS)
Belytschko, Ted; Liu, Wing Kam
1992-01-01
Attenuation is focused on the development of Probabilistic Finite Element Method (PFEM), which combines the finite element method with statistics and reliability methods, and its application to linear, nonlinear structural mechanics problems and fracture mechanics problems. The computational tool based on the Stochastic Boundary Element Method is also given for the reliability analysis of a curvilinear fatigue crack growth. The existing PFEM's have been applied to solve for two types of problems: (1) determination of the response uncertainty in terms of the means, variance and correlation coefficients; and (2) determination the probability of failure associated with prescribed limit states.
3-d finite element model development for biomechanics: a software demonstration
DOE Office of Scientific and Technical Information (OSTI.GOV)
Hollerbach, K.; Hollister, A.M.; Ashby, E.
1997-03-01
Finite element analysis is becoming an increasingly important part of biomechanics and orthopedic research, as computational resources become more powerful, and data handling algorithms become more sophisticated. Until recently, tools with sufficient power did not exist or were not accessible to adequately model complicated, three-dimensional, nonlinear biomechanical systems. In the past, finite element analyses in biomechanics have often been limited to two-dimensional approaches, linear analyses, or simulations of single tissue types. Today, we have the resources to model fully three-dimensional, nonlinear, multi-tissue, and even multi-joint systems. The authors will present the process of developing these kinds of finite element models,more » using human hand and knee examples, and will demonstrate their software tools.« less
User-defined Material Model for Thermo-mechanical Progressive Failure Analysis
NASA Technical Reports Server (NTRS)
Knight, Norman F., Jr.
2008-01-01
Previously a user-defined material model for orthotropic bimodulus materials was developed for linear and nonlinear stress analysis of composite structures using either shell or solid finite elements within a nonlinear finite element analysis tool. Extensions of this user-defined material model to thermo-mechanical progressive failure analysis are described, and the required input data are documented. The extensions include providing for temperature-dependent material properties, archival of the elastic strains, and a thermal strain calculation for materials exhibiting a stress-free temperature.
NASA Astrophysics Data System (ADS)
Murrill, Steven R.; Franck, Charmaine C.; Espinola, Richard L.; Petkie, Douglas T.; De Lucia, Frank C.; Jacobs, Eddie L.
2011-11-01
The U.S. Army Research Laboratory (ARL) and the U.S. Army Night Vision and Electronic Sensors Directorate (NVESD) have developed a terahertz-band imaging system performance model/tool for detection and identification of concealed weaponry. The details of the MATLAB-based model which accounts for the effects of all critical sensor and display components, and for the effects of atmospheric attenuation, concealment material attenuation, and active illumination, were reported on at the 2005 SPIE Europe Security & Defence Symposium (Brugge). An advanced version of the base model that accounts for both the dramatic impact that target and background orientation can have on target observability as related to specular and Lambertian reflections captured by an active-illumination-based imaging system, and for the impact of target and background thermal emission, was reported on at the 2007 SPIE Defense and Security Symposium (Orlando). This paper will provide a comprehensive review of an enhanced, user-friendly, Windows-executable, terahertz-band imaging system performance analysis and design tool that now includes additional features such as a MODTRAN-based atmospheric attenuation calculator and advanced system architecture configuration inputs that allow for straightforward performance analysis of active or passive systems based on scanning (single- or line-array detector element(s)) or staring (focal-plane-array detector elements) imaging architectures. This newly enhanced THz imaging system design tool is an extension of the advanced THz imaging system performance model that was developed under the Defense Advanced Research Project Agency's (DARPA) Terahertz Imaging Focal-Plane Technology (TIFT) program. This paper will also provide example system component (active-illumination source and detector) trade-study analyses using the new features of this user-friendly THz imaging system performance analysis and design tool.
NASA Technical Reports Server (NTRS)
Grillenbeck, Anton M.; Dillinger, Stephan A.; Elliott, Kenny B.
1998-01-01
Theoretical and experimental studies have been performed to investigate the potential and limitations of the modal characterization of a typical spacecraft bus structure by means of active structure elements. The aim of these studies has been test and advance tools for performing an accurate on-orbit modal identification which may be characterized by the availability of a generally very limited test instrumentation, autonomous excitation capabilities by active structure elements and a zero-g environment. The NASA LARC CSI Evolutionary Testbed provided an excellent object for the experimental part of this study program. The main subjects of investigation were: (1) the selection of optimum excitation and measurement to unambiguously identify modes of interest; (2) the applicability of different types of excitation means with focus on active structure elements; and (3) the assessment of the modal identification potential of different types of excitation functions and modal analysis tools. Conventional as well as dedicated modal analysis tools were applied to determine modal parameters and mode shapes. The results will be presented and discussed based on orthogonality checks as well as on suitable indicators for the quality of the acquired modes with respect to modal purity. In particular, the suitability for modal analysis of the acquired frequency response functions as obtained by excitation with active structure elements will be demonstrated with the help of reciprocity checks. Finally, the results will be summarized in a procedure to perform an on-orbit modal identification, including an indication of limitation to be observed.
DOT National Transportation Integrated Search
2009-09-01
More and more, transportation system operators are seeing the benefits of strengthening links between planning and operations. A critical element in improving transportation decision-making and the effectiveness of transportation systems related to o...
Analysis and Design of Rotors at Ultra-Low Reynolds Numbers
NASA Technical Reports Server (NTRS)
Kunz, Peter J.; Strawn, Roger C.
2003-01-01
Design tools have been developed for ultra-low Reynolds number rotors, combining enhanced actuator-ring / blade-element theory with airfoil section data based on two-dimensional Navier-Stokes calculations. This performance prediction method is coupled with an optimizer for both design and analysis applications. Performance predictions from these tools have been compared with three-dimensional Navier Stokes analyses and experimental data for a 2.5 cm diameter rotor with chord Reynolds numbers below 10,000. Comparisons among the analyses and experimental data show reasonable agreement both in the global thrust and power required, but the spanwise distributions of these quantities exhibit significant deviations. The study also reveals that three-dimensional and rotational effects significantly change local airfoil section performance. The magnitude of this issue, unique to this operating regime, may limit the applicability of blade-element type methods for detailed rotor design at ultra-low Reynolds numbers, but these methods are still useful for evaluating concept feasibility and rapidly generating initial designs for further analysis and optimization using more advanced tools.
Inferring transposons activity chronology by TRANScendence - TEs database and de-novo mining tool.
Startek, Michał Piotr; Nogły, Jakub; Gromadka, Agnieszka; Grzebelus, Dariusz; Gambin, Anna
2017-10-16
The constant progress in sequencing technology leads to ever increasing amounts of genomic data. In the light of current evidence transposable elements (TEs for short) are becoming useful tools for learning about the evolution of host genome. Therefore the software for genome-wide detection and analysis of TEs is of great interest. Here we describe the computational tool for mining, classifying and storing TEs from newly sequenced genomes. This is an online, web-based, user-friendly service, enabling users to upload their own genomic data, and perform de-novo searches for TEs. The detected TEs are automatically analyzed, compared to reference databases, annotated, clustered into families, and stored in TEs repository. Also, the genome-wide nesting structure of found elements are detected and analyzed by new method for inferring evolutionary history of TEs. We illustrate the functionality of our tool by performing a full-scale analyses of TE landscape in Medicago truncatula genome. TRANScendence is an effective tool for the de-novo annotation and classification of transposable elements in newly-acquired genomes. Its streamlined interface makes it well-suited for evolutionary studies.
Probabilistic finite elements for fatigue and fracture analysis
NASA Astrophysics Data System (ADS)
Belytschko, Ted; Liu, Wing Kam
1993-04-01
An overview of the probabilistic finite element method (PFEM) developed by the authors and their colleagues in recent years is presented. The primary focus is placed on the development of PFEM for both structural mechanics problems and fracture mechanics problems. The perturbation techniques are used as major tools for the analytical derivation. The following topics are covered: (1) representation and discretization of random fields; (2) development of PFEM for the general linear transient problem and nonlinear elasticity using Hu-Washizu variational principle; (3) computational aspects; (4) discussions of the application of PFEM to the reliability analysis of both brittle fracture and fatigue; and (5) a stochastic computational tool based on stochastic boundary element (SBEM). Results are obtained for the reliability index and corresponding probability of failure for: (1) fatigue crack growth; (2) defect geometry; (3) fatigue parameters; and (4) applied loads. These results show that initial defect is a critical parameter.
Probabilistic finite elements for fatigue and fracture analysis
NASA Technical Reports Server (NTRS)
Belytschko, Ted; Liu, Wing Kam
1993-01-01
An overview of the probabilistic finite element method (PFEM) developed by the authors and their colleagues in recent years is presented. The primary focus is placed on the development of PFEM for both structural mechanics problems and fracture mechanics problems. The perturbation techniques are used as major tools for the analytical derivation. The following topics are covered: (1) representation and discretization of random fields; (2) development of PFEM for the general linear transient problem and nonlinear elasticity using Hu-Washizu variational principle; (3) computational aspects; (4) discussions of the application of PFEM to the reliability analysis of both brittle fracture and fatigue; and (5) a stochastic computational tool based on stochastic boundary element (SBEM). Results are obtained for the reliability index and corresponding probability of failure for: (1) fatigue crack growth; (2) defect geometry; (3) fatigue parameters; and (4) applied loads. These results show that initial defect is a critical parameter.
ERIC Educational Resources Information Center
Hope, Kempe Ronald, Sr.
2013-01-01
The purpose of this article is to provide an assessment and analysis of public sector performance contracting as a performance management tool in Kenya. It aims to demonstrate that performance contracting remains a viable and important tool for improving public sector performance as a key element of the on-going public sector transformation…
A Probabilistic Approach to Model Update
NASA Technical Reports Server (NTRS)
Horta, Lucas G.; Reaves, Mercedes C.; Voracek, David F.
2001-01-01
Finite element models are often developed for load validation, structural certification, response predictions, and to study alternate design concepts. In rare occasions, models developed with a nominal set of parameters agree with experimental data without the need to update parameter values. Today, model updating is generally heuristic and often performed by a skilled analyst with in-depth understanding of the model assumptions. Parameter uncertainties play a key role in understanding the model update problem and therefore probabilistic analysis tools, developed for reliability and risk analysis, may be used to incorporate uncertainty in the analysis. In this work, probability analysis (PA) tools are used to aid the parameter update task using experimental data and some basic knowledge of potential error sources. Discussed here is the first application of PA tools to update parameters of a finite element model for a composite wing structure. Static deflection data at six locations are used to update five parameters. It is shown that while prediction of individual response values may not be matched identically, the system response is significantly improved with moderate changes in parameter values.
Competencies in Organizational E-Learning: Concepts and Tools
ERIC Educational Resources Information Center
Sicilia, Miguel-Angel, Ed.
2007-01-01
"Competencies in Organizational E-Learning: Concepts and Tools" provides a comprehensive view of the way competencies can be used to drive organizational e-learning, including the main conceptual elements, competency gap analysis, advanced related computing topics, the application of semantic Web technologies, and the integration of competencies…
Kuhn-Tucker optimization based reliability analysis for probabilistic finite elements
NASA Technical Reports Server (NTRS)
Liu, W. K.; Besterfield, G.; Lawrence, M.; Belytschko, T.
1988-01-01
The fusion of probability finite element method (PFEM) and reliability analysis for fracture mechanics is considered. Reliability analysis with specific application to fracture mechanics is presented, and computational procedures are discussed. Explicit expressions for the optimization procedure with regard to fracture mechanics are given. The results show the PFEM is a very powerful tool in determining the second-moment statistics. The method can determine the probability of failure or fracture subject to randomness in load, material properties and crack length, orientation, and location.
Process Improvement Through Tool Integration in Aero-Mechanical Design
NASA Technical Reports Server (NTRS)
Briggs, Clark
2010-01-01
Emerging capabilities in commercial design tools promise to significantly improve the multi-disciplinary and inter-disciplinary design and analysis coverage for aerospace mechanical engineers. This paper explores the analysis process for two example problems of a wing and flap mechanical drive system and an aircraft landing gear door panel. The examples begin with the design solid models and include various analysis disciplines such as structural stress and aerodynamic loads. Analytical methods include CFD, multi-body dynamics with flexible bodies and structural analysis. Elements of analysis data management, data visualization and collaboration are also included.
Solernou, Albert; Hanson, Benjamin S; Richardson, Robin A; Welch, Robert; Read, Daniel J; Harlen, Oliver G; Harris, Sarah A
2018-03-01
Fluctuating Finite Element Analysis (FFEA) is a software package designed to perform continuum mechanics simulations of proteins and other globular macromolecules. It combines conventional finite element methods with stochastic thermal noise, and is appropriate for simulations of large proteins and protein complexes at the mesoscale (length-scales in the range of 5 nm to 1 μm), where there is currently a paucity of modelling tools. It requires 3D volumetric information as input, which can be low resolution structural information such as cryo-electron tomography (cryo-ET) maps or much higher resolution atomistic co-ordinates from which volumetric information can be extracted. In this article we introduce our open source software package for performing FFEA simulations which we have released under a GPLv3 license. The software package includes a C ++ implementation of FFEA, together with tools to assist the user to set up the system from Electron Microscopy Data Bank (EMDB) or Protein Data Bank (PDB) data files. We also provide a PyMOL plugin to perform basic visualisation and additional Python tools for the analysis of FFEA simulation trajectories. This manuscript provides a basic background to the FFEA method, describing the implementation of the core mechanical model and how intermolecular interactions and the solvent environment are included within this framework. We provide prospective FFEA users with a practical overview of how to set up an FFEA simulation with reference to our publicly available online tutorials and manuals that accompany this first release of the package.
A Web-Based Course Assessment Tool with Direct Mapping to Student Outcomes
ERIC Educational Resources Information Center
Ibrahim, Walid; Atif, Yacine; Shuaib, Khaled; Sampson, Demetrios
2015-01-01
The assessment of curriculum outcomes is an essential element for continuous academic improvement. However, the collection, aggregation and analysis of assessment data are notoriously complex and time-consuming processes. At the same time, only few developments of supporting electronic processes and tools for continuous academic program assessment…
Scalable Implementation of Finite Elements by NASA _ Implicit (ScIFEi)
NASA Technical Reports Server (NTRS)
Warner, James E.; Bomarito, Geoffrey F.; Heber, Gerd; Hochhalter, Jacob D.
2016-01-01
Scalable Implementation of Finite Elements by NASA (ScIFEN) is a parallel finite element analysis code written in C++. ScIFEN is designed to provide scalable solutions to computational mechanics problems. It supports a variety of finite element types, nonlinear material models, and boundary conditions. This report provides an overview of ScIFEi (\\Sci-Fi"), the implicit solid mechanics driver within ScIFEN. A description of ScIFEi's capabilities is provided, including an overview of the tools and features that accompany the software as well as a description of the input and output le formats. Results from several problems are included, demonstrating the efficiency and scalability of ScIFEi by comparing to finite element analysis using a commercial code.
Janamian, Tina; Upham, Susan J; Crossland, Lisa; Jackson, Claire L
2016-04-18
To conduct a systematic review of the literature to identify existing online primary care quality improvement tools and resources to support organisational improvement related to the seven elements in the Primary Care Practice Improvement Tool (PC-PIT), with the identified tools and resources to progress to a Delphi study for further assessment of relevance and utility. Systematic review of the international published and grey literature. CINAHL, Embase and PubMed databases were searched in March 2014 for articles published between January 2004 and December 2013. GreyNet International and other relevant websites and repositories were also searched in March-April 2014 for documents dated between 1992 and 2012. All citations were imported into a bibliographic database. Published and unpublished tools and resources were included in the review if they were in English, related to primary care quality improvement and addressed any of the seven PC-PIT elements of a high-performing practice. Tools and resources that met the eligibility criteria were then evaluated for their accessibility, relevance, utility and comprehensiveness using a four-criteria appraisal framework. We used a data extraction template to systematically extract information from eligible tools and resources. A content analysis approach was used to explore the tools and resources and collate relevant information: name of the tool or resource, year and country of development, author, name of the organisation that provided access and its URL, accessibility information or problems, overview of each tool or resource and the quality improvement element(s) it addresses. If available, a copy of the tool or resource was downloaded into the bibliographic database, along with supporting evidence (published or unpublished) on its use in primary care. This systematic review identified 53 tools and resources that can potentially be provided as part of a suite of tools and resources to support primary care practices in improving the quality of their practice, to achieve improved health outcomes.
Transducer Analysis and ATILA++ Model Development
2016-10-10
the ATILA finite element software package. This will greatly enhance the state-of-the-art in transducer performance prediction and provide a tool...refereed publication. 15 IMPACT/APPLICATIONS This work is helping to enable the expansion of the functionality of the A TILA ++ finite element ...Sb. GRANT NUMBER N00014-13-1-0196 Sc. PROGRAM ELEMENT NUMBER 6. AUTHOR(S) Sd. PROJECT NUMBER Richard J. Meyer, Jr. 20675 Douglas C. Markley Se
Ramírez Hernández, Javier; Bonete Pérez, María José; Martínez Espinosa, Rosa María
2014-12-17
1) to propose a new classification of the trace elements based on a study of the recently reported research; 2) to offer detailed and actualized information about trace elements. the analysis of the research results recently reported reveals that the advances of the molecular analysis techniques point out the importance of certain trace elements in human health. A detailed analysis of the catalytic function related to several elements not considered essential o probably essentials up to now is also offered. To perform the integral analysis of the enzymes containing trace elements informatics tools have been used. Actualized information about physiological role, kinetics, metabolism, dietetic sources and factors promoting trace elements scarcity or toxicity is also presented. Oligotherapy uses catalytic active trace elements with therapeutic proposals. The new trace element classification here presented will be of high interest for different professional sectors: doctors and other professions related to medicine; nutritionist, pharmaceutics, etc. Using this new classification and approaches, new therapeutic strategies could be designed to mitigate symptomatology related to several pathologies, particularly carential and metabolic diseases. Copyright AULA MEDICA EDICIONES 2014. Published by AULA MEDICA. All rights reserved.
Numerical analysis of thermal drilling technique on titanium sheet metal
NASA Astrophysics Data System (ADS)
Kumar, R.; Hynes, N. Rajesh Jesudoss
2018-05-01
Thermal drilling is a technique used in drilling of sheet metal for various applications. It involves rotating conical tool with high speed in order to drill the sheet metal and formed a hole with bush below the surface of sheet metal. This article investigates the finite element analysis of thermal drilling on Ti6Al4Valloy sheet metal. This analysis was carried out by means of DEFORM-3D simulation software to simulate the performance characteristics of thermal drilling technique. Due to the contribution of high temperature deformation in this technique, the output performances which are difficult to measure by the experimental approach, can be successfully achieved by finite element method. Therefore, the modeling and simulation of thermal drilling is an essential tool to predict the strain rate, stress distribution and temperature of the workpiece.
Dcode.org anthology of comparative genomic tools.
Loots, Gabriela G; Ovcharenko, Ivan
2005-07-01
Comparative genomics provides the means to demarcate functional regions in anonymous DNA sequences. The successful application of this method to identifying novel genes is currently shifting to deciphering the non-coding encryption of gene regulation across genomes. To facilitate the practical application of comparative sequence analysis to genetics and genomics, we have developed several analytical and visualization tools for the analysis of arbitrary sequences and whole genomes. These tools include two alignment tools, zPicture and Mulan; a phylogenetic shadowing tool, eShadow for identifying lineage- and species-specific functional elements; two evolutionary conserved transcription factor analysis tools, rVista and multiTF; a tool for extracting cis-regulatory modules governing the expression of co-regulated genes, Creme 2.0; and a dynamic portal to multiple vertebrate and invertebrate genome alignments, the ECR Browser. Here, we briefly describe each one of these tools and provide specific examples on their practical applications. All the tools are publicly available at the http://www.dcode.org/ website.
RNA-Seq Analysis to Measure the Expression of SINE Retroelements.
Román, Ángel Carlos; Morales-Hernández, Antonio; Fernández-Salguero, Pedro M
2016-01-01
The intrinsic features of retroelements, like their repetitive nature and disseminated presence in their host genomes, demand the use of advanced methodologies for their bioinformatic and functional study. The short length of SINE (short interspersed elements) retrotransposons makes such analyses even more complex. Next-generation sequencing (NGS) technologies are currently one of the most widely used tools to characterize the whole repertoire of gene expression in a specific tissue. In this chapter, we will review the molecular and computational methods needed to perform NGS analyses on SINE elements. We will also describe new methods of potential interest for researchers studying repetitive elements. We intend to outline the general ideas behind the computational analyses of NGS data obtained from SINE elements, and to stimulate other scientists to expand our current knowledge on SINE biology using RNA-seq and other NGS tools.
Otolith edge fingerprints as approach for stock identification of Genidens barbus
NASA Astrophysics Data System (ADS)
Avigliano, Esteban; Maichak de Carvalho, Barbara; Leisen, Mathieu; Romero, Rurik; Velasco, Gonzalo; Vianna, Marcelo; Barra, Fernando; Volpedo, Alejandra Vanina
2017-07-01
The purpose of this paper is to assess the use of multi-elemental otolith fingerprints as a tool to delimit catfish Genidens barbus fish stocks in four estuaries from the southwestern Atlantic Ocean. Barium:Calcium (Ca), Magnesium:Ca, Manganese:Ca, Sodium:Ca and Strontium:Ca ratios in the otolith edge were determined by LA-ICPMS. PERMANOVA analysis reveal significant differences in the multi-element signatures among estuaries (p = 0.0001-0.002). Reclassification rates of quadratic discriminant analysis are high, averaging 89.9% (78-100%). The new data presented here show that the otolith chemistry is a potential tool for stock identification, and indicates the presence of at least four stocks which should probably be handled independently.
Key elements of high-quality practice organisation in primary health care: a systematic review.
Crossland, Lisa; Janamian, Tina; Jackson, Claire L
2014-08-04
To identify elements that are integral to high-quality practice and determine considerations relating to high-quality practice organisation in primary care. A narrative systematic review of published and grey literature. Electronic databases (PubMed, CINAHL, the Cochrane Library, Embase, Emerald Insight, PsycInfo, the Primary Health Care Research and Information Service website, Google Scholar) were searched in November 2013 and used to identify articles published in English from 2002 to 2013. Reference lists of included articles were searched for relevant unpublished articles and reports. Data were configured at the study level to allow for the inclusion of findings from a broad range of study types. Ten elements were most often included in the existing organisational assessment tools. A further three elements were identified from an inductive thematic analysis of descriptive articles, and were noted as important considerations in effective quality improvement in primary care settings. Although there are some validated tools available to primary care that identify and build quality, most are single-strategy approaches developed outside health care settings. There are currently no validated organisational improvement tools, designed specifically for primary health care, which combine all elements of practice improvement and whose use does not require extensive external facilitation.
Biomedical applications of laser-induced breakdown spectroscopy (LIBS)
NASA Astrophysics Data System (ADS)
Unnikrishnan, V. K.; Nayak, Rajesh; Bhat, Sujatha; Mathew, Stanley; Kartha, V. B.; Santhosh, C.
2015-03-01
LIBS has been proven to be a robust elemental analysis tool attracting interest because of the wide applications. LIBS can be used for analysis of any type of samples i.e. environmental/physiological, regardless of its state of matter. Conventional spectroscopy techniques are good in analytical performance, but their sample preparation method is mostly destructive and time consuming. Also, almost all these methods are incapable of analysing multi elements simaltaneously. On the other hand, LIBS has many potential advantages such as simplicity in the experimental setup, less sample preparation, less destructive analysis of sample etc. In this paper, we report some of the biomedical applications of LIBS. From the experiments carried out on clinical samples (calcified tissues or teeth and gall stones) for trace elemental mapping and detection, it was found that LIBS is a robust tool for such applications. It is seen that the presence and relative concentrations of major elements (calcium, phosphorus and magnesium) in human calcified tissue (tooth) can be easily determined using LIBS technique. The importance of this study comes in anthropology where tooth and bone are main samples from which reliable data can be easily retrieved. Similarly, elemental composition of bile juice and gall stone collected from the same subject using LIBS was found to be similar. The results show interesting prospects for LIBS to study cholelithiasis (the presence of stones in the gall bladder, is a common disease of the gastrointestinal tract) better.
Automation of a DXA-based finite element tool for clinical assessment of hip fracture risk.
Luo, Yunhua; Ahmed, Sharif; Leslie, William D
2018-03-01
Finite element analysis of medical images is a promising tool for assessing hip fracture risk. Although a number of finite element models have been developed for this purpose, none of them have been routinely used in clinic. The main reason is that the computer programs that implement the finite element models have not been completely automated, and heavy training is required before clinicians can effectively use them. By using information embedded in clinical dual energy X-ray absorptiometry (DXA), we completely automated a DXA-based finite element (FE) model that we previously developed for predicting hip fracture risk. The automated FE tool can be run as a standalone computer program with the subject's raw hip DXA image as input. The automated FE tool had greatly improved short-term precision compared with the semi-automated version. To validate the automated FE tool, a clinical cohort consisting of 100 prior hip fracture cases and 300 matched controls was obtained from a local community clinical center. Both the automated FE tool and femoral bone mineral density (BMD) were applied to discriminate the fracture cases from the controls. Femoral BMD is the gold standard reference recommended by the World Health Organization for screening osteoporosis and for assessing hip fracture risk. The accuracy was measured by the area under ROC curve (AUC) and odds ratio (OR). Compared with femoral BMD (AUC = 0.71, OR = 2.07), the automated FE tool had a considerably improved accuracy (AUC = 0.78, OR = 2.61 at the trochanter). This work made a large step toward applying our DXA-based FE model as a routine clinical tool for the assessment of hip fracture risk. Furthermore, the automated computer program can be embedded into a web-site as an internet application. Copyright © 2017 Elsevier B.V. All rights reserved.
NASA Astrophysics Data System (ADS)
Murrill, Steven R.; Jacobs, Eddie L.; Franck, Charmaine C.; Petkie, Douglas T.; De Lucia, Frank C.
2015-10-01
The U.S. Army Research Laboratory (ARL) has continued to develop and enhance a millimeter-wave (MMW) and submillimeter- wave (SMMW)/terahertz (THz)-band imaging system performance prediction and analysis tool for both the detection and identification of concealed weaponry, and for pilotage obstacle avoidance. The details of the MATLAB-based model which accounts for the effects of all critical sensor and display components, for the effects of atmospheric attenuation, concealment material attenuation, and active illumination, were reported on at the 2005 SPIE Europe Security and Defence Symposium (Brugge). An advanced version of the base model that accounts for both the dramatic impact that target and background orientation can have on target observability as related to specular and Lambertian reflections captured by an active-illumination-based imaging system, and for the impact of target and background thermal emission, was reported on at the 2007 SPIE Defense and Security Symposium (Orlando). Further development of this tool that includes a MODTRAN-based atmospheric attenuation calculator and advanced system architecture configuration inputs that allow for straightforward performance analysis of active or passive systems based on scanning (single- or line-array detector element(s)) or staring (focal-plane-array detector elements) imaging architectures was reported on at the 2011 SPIE Europe Security and Defence Symposium (Prague). This paper provides a comprehensive review of a newly enhanced MMW and SMMW/THz imaging system analysis and design tool that now includes an improved noise sub-model for more accurate and reliable performance predictions, the capability to account for postcapture image contrast enhancement, and the capability to account for concealment material backscatter with active-illumination- based systems. Present plans for additional expansion of the model's predictive capabilities are also outlined.
An analysis of pilot error-related aircraft accidents
NASA Technical Reports Server (NTRS)
Kowalsky, N. B.; Masters, R. L.; Stone, R. B.; Babcock, G. L.; Rypka, E. W.
1974-01-01
A multidisciplinary team approach to pilot error-related U.S. air carrier jet aircraft accident investigation records successfully reclaimed hidden human error information not shown in statistical studies. New analytic techniques were developed and applied to the data to discover and identify multiple elements of commonality and shared characteristics within this group of accidents. Three techniques of analysis were used: Critical element analysis, which demonstrated the importance of a subjective qualitative approach to raw accident data and surfaced information heretofore unavailable. Cluster analysis, which was an exploratory research tool that will lead to increased understanding and improved organization of facts, the discovery of new meaning in large data sets, and the generation of explanatory hypotheses. Pattern recognition, by which accidents can be categorized by pattern conformity after critical element identification by cluster analysis.
An Array of Qualitative Data Analysis Tools: A Call for Data Analysis Triangulation
ERIC Educational Resources Information Center
Leech, Nancy L.; Onwuegbuzie, Anthony J.
2007-01-01
One of the most important steps in the qualitative research process is analysis of data. The purpose of this article is to provide elements for understanding multiple types of qualitative data analysis techniques available and the importance of utilizing more than one type of analysis, thus utilizing data analysis triangulation, in order to…
A three-dimensional inverse finite element analysis of the heel pad.
Chokhandre, Snehal; Halloran, Jason P; van den Bogert, Antonie J; Erdemir, Ahmet
2012-03-01
Quantification of plantar tissue behavior of the heel pad is essential in developing computational models for predictive analysis of preventive treatment options such as footwear for patients with diabetes. Simulation based studies in the past have generally adopted heel pad properties from the literature, in return using heel-specific geometry with material properties of a different heel. In exceptional cases, patient-specific material characterization was performed with simplified two-dimensional models, without further evaluation of a heel-specific response under different loading conditions. The aim of this study was to conduct an inverse finite element analysis of the heel in order to calculate heel-specific material properties in situ. Multidimensional experimental data available from a previous cadaver study by Erdemir et al. ("An Elaborate Data Set Characterizing the Mechanical Response of the Foot," ASME J. Biomech. Eng., 131(9), pp. 094502) was used for model development, optimization, and evaluation of material properties. A specimen-specific three-dimensional finite element representation was developed. Heel pad material properties were determined using inverse finite element analysis by fitting the model behavior to the experimental data. Compression dominant loading, applied using a spherical indenter, was used for optimization of the material properties. The optimized material properties were evaluated through simulations representative of a combined loading scenario (compression and anterior-posterior shear) with a spherical indenter and also of a compression dominant loading applied using an elevated platform. Optimized heel pad material coefficients were 0.001084 MPa (μ), 9.780 (α) (with an effective Poisson's ratio (ν) of 0.475), for a first-order nearly incompressible Ogden material model. The model predicted structural response of the heel pad was in good agreement for both the optimization (<1.05% maximum tool force, 0.9% maximum tool displacement) and validation cases (6.5% maximum tool force, 15% maximum tool displacement). The inverse analysis successfully predicted the material properties for the given specimen-specific heel pad using the experimental data for the specimen. The modeling framework and results can be used for accurate predictions of the three-dimensional interaction of the heel pad with its surroundings.
LoRTE: Detecting transposon-induced genomic variants using low coverage PacBio long read sequences.
Disdero, Eric; Filée, Jonathan
2017-01-01
Population genomic analysis of transposable elements has greatly benefited from recent advances of sequencing technologies. However, the short size of the reads and the propensity of transposable elements to nest in highly repeated regions of genomes limits the efficiency of bioinformatic tools when Illumina or 454 technologies are used. Fortunately, long read sequencing technologies generating read length that may span the entire length of full transposons are now available. However, existing TE population genomic softwares were not designed to handle long reads and the development of new dedicated tools is needed. LoRTE is the first tool able to use PacBio long read sequences to identify transposon deletions and insertions between a reference genome and genomes of different strains or populations. Tested against simulated and genuine Drosophila melanogaster PacBio datasets, LoRTE appears to be a reliable and broadly applicable tool to study the dynamic and evolutionary impact of transposable elements using low coverage, long read sequences. LoRTE is an efficient and accurate tool to identify structural genomic variants caused by TE insertion or deletion. LoRTE is available for download at http://www.egce.cnrs-gif.fr/?p=6422.
Hair: A Diagnostic Tool to Complement Blood Serum and Urine.
ERIC Educational Resources Information Center
Maugh, Thomas H., II
1978-01-01
Trace elements and some drugs can be identified in hair and it seems likely that other organic chemicals will be identifiable in the future. Since hair is so easily collected, stored, and analyzed it promises to be an ideal complement to serum and urine analysis as a diagnostic tool. (BB)
Parameter identification and optimization of slide guide joint of CNC machine tools
NASA Astrophysics Data System (ADS)
Zhou, S.; Sun, B. B.
2017-11-01
The joint surface has an important influence on the performance of CNC machine tools. In order to identify the dynamic parameters of slide guide joint, the parametric finite element model of the joint is established and optimum design method is used based on the finite element simulation and modal test. Then the mode that has the most influence on the dynamics of slip joint is found through harmonic response analysis. Take the frequency of this mode as objective, the sensitivity analysis of the stiffness of each joint surface is carried out using Latin Hypercube Sampling and Monte Carlo Simulation. The result shows that the vertical stiffness of slip joint surface constituted by the bed and the slide plate has the most obvious influence on the structure. Therefore, this stiffness is taken as the optimization variable and the optimal value is obtained through studying the relationship between structural dynamic performance and stiffness. Take the stiffness values before and after optimization into the FEM of machine tool, and it is found that the dynamic performance of the machine tool is improved.
NASA Astrophysics Data System (ADS)
Abass, K. I.
2016-11-01
Single Point Incremental Forming process (SPIF) is a forming technique of sheet material based on layered manufacturing principles. The edges of sheet material are clamped while the forming tool is moved along the tool path. The CNC milling machine is used to manufacturing the product. SPIF involves extensive plastic deformation and the description of the process is more complicated by highly nonlinear boundary conditions, namely contact and frictional effects have been accomplished. However, due to the complex nature of these models, numerical approaches dominated by Finite Element Analysis (FEA) are now in widespread use. The paper presents the data and main results of a study on effect of using preforming blank in SPIF through FEA. The considered SPIF has been studied under certain process conditions referring to the test work piece, tool, etc., applying ANSYS 11. The results show that the simulation model can predict an ideal profile of processing track, the behaviour of contact tool-workpiece, the product accuracy by evaluation its thickness, surface strain and the stress distribution along the deformed blank section during the deformation stages.
NASA Technical Reports Server (NTRS)
Panontin, Tina; Carvalho, Robert; Keller, Richard
2004-01-01
Contents include the folloving:Overview of the Application; Input Data; Analytical Process; Tool's Output; and Application of the Results of the Analysis.The tool enables the first element through a Web-based application that can be accessed by distributed teams to store and retrieve any type of digital investigation material in a secure environment. The second is accomplished by making the relationships between information explicit through the use of a semantic network-a structure that literally allows an investigator or team to "connect -the-dots." The third element, the significance of the correlated information, is established through causality and consistency tests using a number of different methods embedded within the tool, including fault trees, event sequences, and other accident models. And finally, the evidence gathered and structured within the tool can be directly, electronically archived to preserve the evidence and investigative reasoning.
NASA Astrophysics Data System (ADS)
Rachmat, Haris; Ibrahim, M. Rasidi; Hasan, Sulaiman bin
2017-04-01
On of high technology in machining is ultrasonic vibration assisted turning. The design of tool holder was a crucial step to make sure the tool holder is enough to handle all forces on turning process. Because of the direct experimental approach is expensive, the paper studied to predict feasibility of tool holder displacement and effective stress was used the computational in finite element simulation. SS201 and AISI 1045 materials were used with sharp and ramp corners flexure hinges on design. The result shows that AISI 1045 material and which has ramp corner flexure hinge was the best choice to be produced. The displacement is around 11.3 micron and effective stress is 1.71e+008 N/m2 and also the factor of safety is 3.10.
Aguiar, G F M; Batista, B L; Rodrigues, J L; Silva, L R S; Campiglia, A D; Barbosa, R M; Barbosa, F
2012-12-01
The reproductive performance of cattle may be influenced by several factors, but mineral imbalances are crucial in terms of direct effects on reproduction. Several studies have shown that elements such as calcium, copper, iron, magnesium, selenium, and zinc are essential for reproduction and can prevent oxidative stress. However, toxic elements such as lead, nickel, and arsenic can have adverse effects on reproduction. In this paper, we applied a simple and fast method of multi-element analysis to bovine semen samples from Zebu and European classes used in reproduction programs and artificial insemination. Samples were analyzed by inductively coupled plasma spectrometry (ICP-MS) using aqueous medium calibration and the samples were diluted in a proportion of 1:50 in a solution containing 0.01% (vol/vol) Triton X-100 and 0.5% (vol/vol) nitric acid. Rhodium, iridium, and yttrium were used as the internal standards for ICP-MS analysis. To develop a reliable method of tracing the class of bovine semen, we used data mining techniques that make it possible to classify unknown samples after checking the differentiation of known-class samples. Based on the determination of 15 elements in 41 samples of bovine semen, 3 machine-learning tools for classification were applied to determine cattle class. Our results demonstrate the potential of support vector machine (SVM), multilayer perceptron (MLP), and random forest (RF) chemometric tools to identify cattle class. Moreover, the selection tools made it possible to reduce the number of chemical elements needed from 15 to just 8. Copyright © 2012 American Dairy Science Association. Published by Elsevier Inc. All rights reserved.
Trades Between Opposition and Conjunction Class Trajectories for Early Human Missions to Mars
NASA Technical Reports Server (NTRS)
Mattfeld, Bryan; Stromgren, Chel; Shyface, Hilary; Komar, David R.; Cirillo, William; Goodliff, Kandyce
2014-01-01
Candidate human missions to Mars, including NASA's Design Reference Architecture 5.0, have focused on conjunction-class missions with long crewed durations and minimum energy trajectories to reduce total propellant requirements and total launch mass. However, in order to progressively reduce risk and gain experience in interplanetary mission operations, it may be desirable that initial human missions to Mars, whether to the surface or to Mars orbit, have shorter total crewed durations and minimal stay times at the destination. Opposition-class missions require larger total energy requirements relative to conjunction-class missions but offer the potential for much shorter mission durations, potentially reducing risk and overall systems performance requirements. This paper will present a detailed comparison of conjunction-class and opposition-class human missions to Mars vicinity with a focus on how such missions could be integrated into the initial phases of a Mars exploration campaign. The paper will present the results of a trade study that integrates trajectory/propellant analysis, element design, logistics and sparing analysis, and risk assessment to produce a comprehensive comparison of opposition and conjunction exploration mission constructs. Included in the trade study is an assessment of the risk to the crew and the trade offs between the mission duration and element, logistics, and spares mass. The analysis of the mission trade space was conducted using four simulation and analysis tools developed by NASA. Trajectory analyses for Mars destination missions were conducted using VISITOR (Versatile ImpulSive Interplanetary Trajectory OptimizeR), an in-house tool developed by NASA Langley Research Center. Architecture elements were evaluated using EXploration Architecture Model for IN-space and Earth-to-orbit (EXAMINE), a parametric modeling tool that generates exploration architectures through an integrated systems model. Logistics analysis was conducted using NASA's Human Exploration Logistics Model (HELM), and sparing allocation predictions were generated via the Exploration Maintainability Analysis Tool (EMAT), which is a probabilistic simulation engine that evaluates trades in spacecraft reliability and sparing requirements based on spacecraft system maintainability and reparability.
NASA Astrophysics Data System (ADS)
Sotokoba, Yasumasa; Okajima, Kenji; Iida, Toshiaki; Tanaka, Tadatsugu
We propose the trenchless box culvert construction method to construct box culverts in small covering soil layers while keeping roads or tracks open. When we use this construction method, it is necessary to clarify deformation and shear failure by excavation of grounds. In order to investigate the soil behavior, model experiments and elasto-plactic finite element analysis were performed. In the model experiments, it was shown that the shear failure was developed from the end of the roof to the toe of the boundary surface. In the finite element analysis, a shear band effect was introduced. Comparing the observed shear bands in model experiments with computed maximum shear strain contours, it was found that the observed direction of the shear band could be simulated reasonably by the finite element analysis. We may say that the finite element method used in this study is useful tool for this construction method.
Dynamic analysis and vibration testing of CFRP drive-line system used in heavy-duty machine tool
NASA Astrophysics Data System (ADS)
Yang, Mo; Gui, Lin; Hu, Yefa; Ding, Guoping; Song, Chunsheng
2018-03-01
Low critical rotary speed and large vibration in the metal drive-line system of heavy-duty machine tool affect the machining precision seriously. Replacing metal drive-line with the CFRP drive-line can effectively solve this problem. Based on the composite laminated theory and the transfer matrix method (TMM), this paper puts forward a modified TMM to analyze dynamic characteristics of CFRP drive-line system. With this modified TMM, the CFRP drive-line of a heavy vertical miller is analyzed. And the finite element modal analysis model of the shafting is established. The results of the modified TMM and finite element analysis (FEA) show that the modified TMM can effectively predict the critical rotary speed of CFRP drive-line. And the critical rotary speed of CFRP drive-line is 20% higher than that of the original metal drive-line. Then, the vibration of the CFRP and the metal drive-line were tested. The test results show that application of the CFRP drive shaft in the drive-line can effectively reduce the vibration of the heavy-duty machine tool.
Solernou, Albert
2018-01-01
Fluctuating Finite Element Analysis (FFEA) is a software package designed to perform continuum mechanics simulations of proteins and other globular macromolecules. It combines conventional finite element methods with stochastic thermal noise, and is appropriate for simulations of large proteins and protein complexes at the mesoscale (length-scales in the range of 5 nm to 1 μm), where there is currently a paucity of modelling tools. It requires 3D volumetric information as input, which can be low resolution structural information such as cryo-electron tomography (cryo-ET) maps or much higher resolution atomistic co-ordinates from which volumetric information can be extracted. In this article we introduce our open source software package for performing FFEA simulations which we have released under a GPLv3 license. The software package includes a C ++ implementation of FFEA, together with tools to assist the user to set up the system from Electron Microscopy Data Bank (EMDB) or Protein Data Bank (PDB) data files. We also provide a PyMOL plugin to perform basic visualisation and additional Python tools for the analysis of FFEA simulation trajectories. This manuscript provides a basic background to the FFEA method, describing the implementation of the core mechanical model and how intermolecular interactions and the solvent environment are included within this framework. We provide prospective FFEA users with a practical overview of how to set up an FFEA simulation with reference to our publicly available online tutorials and manuals that accompany this first release of the package. PMID:29570700
NASA Technical Reports Server (NTRS)
Abdul-Aziz, Ali; Baaklini, George Y.; Zagidulin, Dmitri; Rauser, Richard W.
2000-01-01
Capabilities and expertise related to the development of links between nondestructive evaluation (NDE) and finite element analysis (FEA) at Glenn Research Center (GRC) are demonstrated. Current tools to analyze data produced by computed tomography (CT) scans are exercised to help assess the damage state in high temperature structural composite materials. A utility translator was written to convert velocity (an image processing software) STL data file to a suitable CAD-FEA type file. Finite element analyses are carried out with MARC, a commercial nonlinear finite element code, and the analytical results are discussed. Modeling was established by building MSC/Patran (a pre and post processing finite element package) generated model and comparing it to a model generated by Velocity in conjunction with MSC/Patran Graphics. Modeling issues and results are discussed in this paper. The entire process that outlines the tie between the data extracted via NDE and the finite element modeling and analysis is fully described.
Method and apparatus for transmitting and receiving data to and from a downhole tool
Hall, David R.; Fox, Joe
2007-03-13
A transmission line network system for transmitting and/or receiving data from a downhole tool. The invention is achieved by providing one or more transceiving elements, preferably rings, at either end of a downhole tool. A conduit containing a coaxial cable capable of communicating an electrical signal is attached to the transceiving element and extends through a central bore of the downhole tool and through the central bore of any tool intermediate the first transceiving element and a second transceiving element. Upon receiving an electrical signal from the cable, the second transceiving element may convert such signal to a magnetic field. The magnetic field may be detected by a third transceiving element in close proximity to the second transceiving element. In this manner, many different tools may be included in a downhole transmission network without requiring substantial modification, if any, of any particular tool.
New Representation of Bearings in LS-DYNA
NASA Technical Reports Server (NTRS)
Carney, Kelly S.; Howard, Samuel A.; Miller, Brad A.; Benson, David J.
2014-01-01
Non-linear, dynamic, finite element analysis is used in various engineering disciplines to evaluate high-speed, dynamic impact and vibration events. Some of these applications require connecting rotating to stationary components. For example, bird impacts on rotating aircraft engine fan blades are a common analysis performed using this type of analysis tool. Traditionally, rotating machines utilize some type of bearing to allow rotation in one degree of freedom while offering constraints in the other degrees of freedom. Most times, bearings are modeled simply as linear springs with rotation. This is a simplification that is not necessarily accurate under the conditions of high-velocity, high-energy, dynamic events such as impact problems. For this reason, it is desirable to utilize a more realistic non-linear force-deflection characteristic of real bearings to model the interaction between rotating and non-rotating components during dynamic events. The present work describes a rolling element bearing model developed for use in non-linear, dynamic finite element analysis. This rolling element bearing model has been implemented in LS-DYNA as a new element, *ELEMENT_BEARING.
Towards an integral computer environment supporting system operations analysis and conceptual design
NASA Technical Reports Server (NTRS)
Barro, E.; Delbufalo, A.; Rossi, F.
1994-01-01
VITROCISET has in house developed a prototype tool named System Dynamic Analysis Environment (SDAE) to support system engineering activities in the initial definition phase of a complex space system. The SDAE goal is to provide powerful means for the definition, analysis, and trade-off of operations and design concepts for the space and ground elements involved in a mission. For this purpose SDAE implements a dedicated modeling methodology based on the integration of different modern (static and dynamic) analysis and simulation techniques. The resulting 'system model' is capable of representing all the operational, functional, and behavioral aspects of the system elements which are part of a mission. The execution of customized model simulations enables: the validation of selected concepts with respect to mission requirements; the in-depth investigation of mission specific operational and/or architectural aspects; and the early assessment of performances required by the system elements to cope with mission constraints and objectives. Due to its characteristics, SDAE is particularly tailored for nonconventional or highly complex systems, which require a great analysis effort in their early definition stages. SDAE runs under PC-Windows and is currently used by VITROCISET system engineering group. This paper describes the SDAE main features, showing some tool output examples.
NASA Astrophysics Data System (ADS)
Gritsan, Andrei V.; Röntsch, Raoul; Schulze, Markus; Xiao, Meng
2016-09-01
In this paper, we investigate anomalous interactions of the Higgs boson with heavy fermions, employing shapes of kinematic distributions. We study the processes p p →t t ¯+H , b b ¯+H , t q +H , and p p →H →τ+τ- and present applications of event generation, reweighting techniques for fast simulation of anomalous couplings, as well as matrix element techniques for optimal sensitivity. We extend the matrix element likelihood approach (MELA) technique, which proved to be a powerful matrix element tool for Higgs boson discovery and characterization during Run I of the LHC, and implement all analysis tools in the JHU generator framework. A next-to-leading-order QCD description of the p p →t t ¯+H process allows us to investigate the performance of the MELA in the presence of extra radiation. Finally, projections for LHC measurements through the end of Run III are presented.
Environmental influence on trace element levels in human hair
DOE Office of Scientific and Technical Information (OSTI.GOV)
Limic, N.; Valkovic, V.
1986-12-01
Trace element content of human hair depends on many factors. It has been shown by a large number of investigators that environmental factors play an important role. Elements from air particulates, water, shampoo or other media get incorporated into the hair structure. Here a model is proposed in which different contributions to trace element levels in human hair are factorized and the environmental contribution to the radial and longitudinal concentration profiles can be calculated. With the proper understanding of environmental contamination, hair analysis has better chances of being used as a diagnostic tool.
SINEBase: a database and tool for SINE analysis.
Vassetzky, Nikita S; Kramerov, Dmitri A
2013-01-01
SINEBase (http://sines.eimb.ru) integrates the revisited body of knowledge about short interspersed elements (SINEs). A set of formal definitions concerning SINEs was introduced. All available sequence data were screened through these definitions and the genetic elements misidentified as SINEs were discarded. As a result, 175 SINE families have been recognized in animals, flowering plants and green algae. These families were classified by the modular structure of their nucleotide sequences and the frequencies of different patterns were evaluated. These data formed the basis for the database of SINEs. The SINEBase website can be used in two ways: first, to explore the database of SINE families, and second, to analyse candidate SINE sequences using specifically developed tools. This article presents an overview of the database and the process of SINE identification and analysis.
SINEBase: a database and tool for SINE analysis
Vassetzky, Nikita S.; Kramerov, Dmitri A.
2013-01-01
SINEBase (http://sines.eimb.ru) integrates the revisited body of knowledge about short interspersed elements (SINEs). A set of formal definitions concerning SINEs was introduced. All available sequence data were screened through these definitions and the genetic elements misidentified as SINEs were discarded. As a result, 175 SINE families have been recognized in animals, flowering plants and green algae. These families were classified by the modular structure of their nucleotide sequences and the frequencies of different patterns were evaluated. These data formed the basis for the database of SINEs. The SINEBase website can be used in two ways: first, to explore the database of SINE families, and second, to analyse candidate SINE sequences using specifically developed tools. This article presents an overview of the database and the process of SINE identification and analysis. PMID:23203982
TACT: A Set of MSC/PATRAN- and MSC/NASTRAN- based Modal Correlation Tools
NASA Technical Reports Server (NTRS)
Marlowe, Jill M.; Dixon, Genevieve D.
1998-01-01
This paper describes the functionality and demonstrates the utility of the Test Analysis Correlation Tools (TACT), a suite of MSC/PATRAN Command Language (PCL) tools which automate the process of correlating finite element models to modal survey test data. The initial release of TACT provides a basic yet complete set of tools for performing correlation totally inside the PATRAN/NASTRAN environment. Features include a step-by-step menu structure, pre-test accelerometer set evaluation and selection, analysis and test result export/import in Universal File Format, calculation of frequency percent difference and cross-orthogonality correlation results using NASTRAN, creation and manipulation of mode pairs, and five different ways of viewing synchronized animations of analysis and test modal results. For the PATRAN-based analyst, TACT eliminates the repetitive, time-consuming and error-prone steps associated with transferring finite element data to a third-party modal correlation package, which allows the analyst to spend more time on the more challenging task of model updating. The usefulness of this software is presented using a case history, the correlation for a NASA Langley Research Center (LaRC) low aspect ratio research wind tunnel model. To demonstrate the improvements that TACT offers the MSC/PATRAN- and MSC/DIASTRAN- based structural analysis community, a comparison of the modal correlation process using TACT within PATRAN versus external third-party modal correlation packages is presented.
FIESTA ROC: A new finite element analysis program for solar cell simulation
NASA Technical Reports Server (NTRS)
Clark, Ralph O.
1991-01-01
The Finite Element Semiconductor Three-dimensional Analyzer by Ralph O. Clark (FIESTA ROC) is a computational tool for investigating in detail the performance of arbitrary solar cell structures. As its name indicates, it uses the finite element technique to solve the fundamental semiconductor equations in the cell. It may be used for predicting the performance (thereby dictating the design parameters) of a proposed cell or for investigating the limiting factors in an established design.
Water analysis via portable X-ray fluorescence spectrometry
NASA Astrophysics Data System (ADS)
Pearson, Delaina; Chakraborty, Somsubhra; Duda, Bogdan; Li, Bin; Weindorf, David C.; Deb, Shovik; Brevik, Eric; Ray, D. P.
2017-01-01
Rapid, in-situ elemental water analysis would be an invaluable tool in studying polluted and/or salt-impacted waters. Analysis of water salinity has commonly used electrical conductance (EC); however, the identity of the elements responsible for the salinity are not revealed using EC. Several studies have established the viability of using portable X-ray fluorescence (PXRF) spectrometry for elemental data analysis of soil, sediment, and other matrices. However, the accuracy of PXRF is known to be affected while scanning moisture-laden soil samples. This study used PXRF elemental data in water samples to predict water EC. A total of 256 water samples, from 10 different countries were collected and analyzed via PXRF, inductively coupled plasma atomic emission spectroscopy (ICP-AES), and a digital salinity bridge. The PXRF detected some elements more effectively than others, but overall results indicated that PXRF can successfully predict water EC via quantifying Cl in water samples (validation R2 and RMSE of 0.77 and 0.95 log μS cm-1, respectively). The findings of this study elucidated the potential of PXRF for future analysis of pollutant and/or metal contaminated waters.
Brand, Richard A; Stanford, Clark M; Swan, Colby C
2003-01-01
Joint implant design clearly affects long-term outcome. While many implant designs have been empirically-based, finite element analysis has the potential to identify beneficial and deleterious features prior to clinical trials. Finite element analysis is a powerful analytic tool allowing computation of the stress and strain distribution throughout an implant construct. Whether it is useful depends upon many assumptions and details of the model. Since ultimate failure is related to biological factors in addition to mechanical, and since the mechanical causes of failure are related to load history, rather than a few loading conditions, chief among them is whether the stresses or strains under limited loading conditions relate to outcome. Newer approaches can minimize this and the many other model limitations. If the surgeon is to critically and properly interpret the results in scientific articles and sales literature, he or she must have a fundamental understanding of finite element analysis. We outline here the major capabilities of finite element analysis, as well as the assumptions and limitations. PMID:14575244
Practical applications of surface analytic tools in tribology
NASA Technical Reports Server (NTRS)
Ferrante, J.
1980-01-01
Many of the currently, widely used tools available for surface analysis are described. Those which have the highest applicability for giving elemental and/or compound analysis for problems of interest in tribology and are truly surface sensitive (that is, less than 10 atomic layers) are presented. The latter group is evaluated in detail in terms of strengths and weaknesses. Emphasis is placed on post facto analysis of experiments performed under 'real' conditions (e.g., in air with lubricants). It is further indicated that such equipment could be used for screening and quality control.
Probabilistic Structural Analysis of the SRB Aft Skirt External Fitting Modification
NASA Technical Reports Server (NTRS)
Townsend, John S.; Peck, J.; Ayala, S.
1999-01-01
NASA has funded several major programs (the PSAM Project is an example) to develop Probabilistic Structural Analysis Methods and tools for engineers to apply in the design and assessment of aerospace hardware. A probabilistic finite element design tool, known as NESSUS, is used to determine the reliability of the Space Shuttle Solid Rocket Booster (SRB) aft skirt critical weld. An external bracket modification to the aft skirt provides a comparison basis for examining the details of the probabilistic analysis and its contributions to the design process.
Energy Dispersive X-ray (EDX) microanalysis: A powerful tool in biomedical research and diagnosis
Scimeca, Manuel; Bischetti, Simone; Lamsira, Harpreet Kaur; Bonfiglio, Rita; Bonanno, Elena
2018-01-01
The Energy Dispersive X-ray (EDX) microanalysis is a technique of elemental analysis associated to electron microscopy based on the generation of characteristic Xrays that reveals the presence of elements present in the specimens. The EDX microanalysis is used in different biomedical fields by many researchers and clinicians. Nevertheless, most of the scientific community is not fully aware of its possible applications. The spectrum of EDX microanalysis contains both semi-qualitative and semi-quantitative information. EDX technique is made useful in the study of drugs, such as in the study of drugs delivery in which the EDX is an important tool to detect nanoparticles (generally, used to improve the therapeutic performance of some chemotherapeutic agents). EDX is also used in the study of environmental pollution and in the characterization of mineral bioaccumulated in the tissues. In conclusion, the EDX can be considered as a useful tool in all works that require element determination, endogenous or exogenous, in the tissue, cell or any other sample. PMID:29569878
Terada, K; Sato, A; Ninomiya, K; Kawashima, Y; Shimomura, K; Yoshida, G; Kawai, Y; Osawa, T; Tachibana, S
2017-11-13
Electron- or X-ray-induced characteristic X-ray analysis has been widely used to determine chemical compositions of materials in vast research fields. In recent years, analysis of characteristic X-rays from muonic atoms, in which a muon is captured, has attracted attention because both a muon beam and a muon-induced characteristic X-ray have high transmission abilities. Here we report the first non-destructive elemental analysis of a carbonaceous chondrite using one of the world-leading intense direct current muon beam source (MuSIC; MUon Science Innovative Channel). We successfully detected characteristic muonic X-rays of Mg, Si, Fe, O, S and C from Jbilet Winselwan CM chondrite, of which carbon content is about 2 wt%, and the obtained elemental abundance pattern was consistent with that of CM chondrites. Because of its high sensitivity to carbon, non-destructive elemental analysis with a muon beam can be a novel powerful tool to characterize future retuned samples from carbonaceous asteroids.
Microanalysis of tool steel and glass with laser-induced breakdown spectroscopy
NASA Astrophysics Data System (ADS)
Loebe, Klaus; Uhl, Arnold; Lucht, Hartmut
2003-10-01
A laser microscope system for the microanalytical characterization of complex materials is described. The universal measuring principle of laser-induced breakdown spectroscopy (LIBS) in combination with echelle optics permits a fast simultaneous multielement analysis with a possible spatial resolution below 10 pm. The developed system features completely UV-transparent optics for the laser-microscope coupling and the emission beam path and enables parallel signal detection within the wavelength range of 200-800 nm with a spectral resolution of a few picometers. Investigations of glass defects and tool steels were performed. The characterization of a glass defect in a tumbler by a micro-LIBS line scan, with use of a 266-nm diode-pumped Nd:YAG laser for excitation, is possible by simple comparison of plasma spectra of the defect and the surrounding area. Variations in the main elemental composition as well as impurities by trace elements are detected at the same time. Through measurement of the calibration samples with the known concentration of the corresponding element, a correlation between the intensity of spectral lines and the element concentration was also achieved. The change of elemental composition at the transient stellite solder of tool steels has been determined by an area scan. The two-dimensional pictures show abrupt changes of the element distribution along the solder edge and allow fundamental researches of dynamic modifications (e.g., diffusion) in steel.
Use of LEED, Auger emission spectroscopy and field ion microscopy in microstructural studies
NASA Technical Reports Server (NTRS)
Ferrante, J.; Buckley, D. H.; Pepper, S. V.; Brainard, W. A.
1972-01-01
Surface research tools such as LEED, Auger emission spectroscopy analysis, and field ion microscopy are discussed. Examples of their use in studying adhesion, friction, wear, and lubrication presented. These tools have provided considerable insight into the basic nature of solid surface interactions. The transfer of metals from one surface to another at the atomic level has been observed and studied with each of these devices. The field ion microscope has been used to study polymer-metal interactions and Auger analysis to study the mechanism of polymer adhesion to metals. LEED and Auger analysis have identified surface segregation of alloying elements and indicated the influence of these elements in metallic adhesion. LEED and Auger analysis have assisted in adsorption studies in determining the structural arrangement and quantity of adsorbed species present in making an understanding of the influence of these species on adhesion possible. These devices are assisting in the furtherance of understanding of the fundamental mechanism involved in the adhesion, friction, wear, and lubrication processes.
Method of holding optical elements without deformation during their fabrication
Hed, P.P.
1997-04-29
An improved method for securing and removing an optical element to and from a blocking tool without causing deformation of the optical element is disclosed. A lens tissue is placed on the top surface of the blocking tool. Dots of UV cement are applied to the lens tissue without any of the dots contacting each other. An optical element is placed on top of the blocking tool with the lens tissue sandwiched therebetween. The UV cement is then cured. After subsequent fabrication steps, the bonded blocking tool, lens tissue, and optical element are placed in a debonding solution to soften the UV cement. The optical element is then removed from the blocking tool. 16 figs.
Method of holding optical elements without deformation during their fabrication
Hed, P. Paul
1997-01-01
An improved method for securing and removing an optical element to and from a blocking tool without causing deformation of the optical element. A lens tissue is placed on the top surface of the blocking tool. Dots of UV cement are applied to the lens tissue without any of the dots contacting each other. An optical element is placed on top of the blocking tool with the lens tissue sandwiched therebetween. The UV cement is then cured. After subsequent fabrication steps, the bonded blocking tool, lens tissue, and optical element are placed in a debonding solution to soften the UV cement. The optical element is then removed from the blocking tool.
Total reflection X-ray fluorescence as a tool for food screening
NASA Astrophysics Data System (ADS)
Borgese, Laura; Bilo, Fabjola; Dalipi, Rogerta; Bontempi, Elza; Depero, Laura E.
2015-11-01
This review provides a comprehensive overview of the applications of total reflection X-ray fluorescence (TXRF) in the field of food analysis. Elemental composition of food is of great importance, since food is the main source of essential, major and trace elements for animals and humans. Some potentially toxic elements, dangerous for human health may contaminate food, entering the food chain from the environment, processing, and storage. For this reason the elemental analysis of food is fundamental for safety assessment. Fast and sensitive analytical techniques, able to detect major and trace elements, are required as a result of the increasing demand on multi-elemental information and product screening. TXRF is suitable for elemental analysis of food, since it provides simultaneous multi-elemental identification in a wide dynamic range of concentrations. Several different matrices may be analyzed obtaining results with a good precision and accuracy. In this review, the most recent literature about the use of TXRF for the analysis of food is reported. The focus is placed on the applications within food quality monitoring of drinks, beverages, vegetables, fruits, cereals, animal derivatives and dietary supplements. Furthermore, this paper provides a critical outlook on the developments required to transfer these methods from research to the industrial and analytical laboratories contexts.
Contact Tools in Japanese Acupuncture: An Ethnography of Acupuncture Practitioners in Japan.
Chant, Benjamin Cw; Madison, Jeanne; Coop, Paul; Dieberg, Gudrun
2017-10-01
This study aimed to identify procedural elements of Japanese acupuncture, describe these elements in detail, and explain them in terms of the key thematic category of treatment principles. Between August 2012 and December 2016, ethnographic fieldwork was conducted in Japan. In total, 38 participants were recruited by chain referral and emergent sampling. Data was collected through participant observation, interviews, and by analyzing documents. A total of 22 participants agreed to clinical observation; 221 treatments were observed with 172 patients. Seventeen consented to formal interviews and 28 to informal interviews. Thematic analysis was used to critically evaluate data. One especially interesting theme was interpreted from the data: a variety of contact tools were applied in treatment and these were manipulated by adjusting elements of form, speed, repetition, and pressure. Tapping, holding, pressing/pushing, and stroking were the most important ways contact tools were used on patients. Contact tools are noninvasive, painless, can be applied in almost any environment, and may be easily accepted by patients worldwide. Contact tool theory and practice may be successfully integrated into acupuncture curricula outside of Japan, used to inform clinical trials, and contribute to an expanded repertoire of methods for practitioners to benefit individual patients in international contexts. Copyright © 2017. Published by Elsevier B.V.
Sinaci, A. Anil; Laleci Erturkmen, Gokce B.; Gonul, Suat; Yuksel, Mustafa; Invernizzi, Paolo; Thakrar, Bharat; Pacaci, Anil; Cinar, H. Alper; Cicekli, Nihan Kesim
2015-01-01
Postmarketing drug surveillance is a crucial aspect of the clinical research activities in pharmacovigilance and pharmacoepidemiology. Successful utilization of available Electronic Health Record (EHR) data can complement and strengthen postmarketing safety studies. In terms of the secondary use of EHRs, access and analysis of patient data across different domains are a critical factor; we address this data interoperability problem between EHR systems and clinical research systems in this paper. We demonstrate that this problem can be solved in an upper level with the use of common data elements in a standardized fashion so that clinical researchers can work with different EHR systems independently of the underlying information model. Postmarketing Safety Study Tool lets the clinical researchers extract data from different EHR systems by designing data collection set schemas through common data elements. The tool interacts with a semantic metadata registry through IHE data element exchange profile. Postmarketing Safety Study Tool and its supporting components have been implemented and deployed on the central data warehouse of the Lombardy region, Italy, which contains anonymized records of about 16 million patients with over 10-year longitudinal data on average. Clinical researchers in Roche validate the tool with real life use cases. PMID:26543873
FEAMAC/CARES Stochastic-Strength-Based Damage Simulation Tool for Ceramic Matrix Composites
NASA Technical Reports Server (NTRS)
Nemeth, Noel; Bednarcyk, Brett; Pineda, Evan; Arnold, Steven; Mital, Subodh; Murthy, Pappu; Bhatt, Ramakrishna
2016-01-01
Reported here is a coupling of two NASA developed codes: CARES (Ceramics Analysis and Reliability Evaluation of Structures) with the MAC/GMC (Micromechanics Analysis Code/ Generalized Method of Cells) composite material analysis code. The resulting code is called FEAMAC/CARES and is constructed as an Abaqus finite element analysis UMAT (user defined material). Here we describe the FEAMAC/CARES code and an example problem (taken from the open literature) of a laminated CMC in off-axis loading is shown. FEAMAC/CARES performs stochastic-strength-based damage simulation response of a CMC under multiaxial loading using elastic stiffness reduction of the failed elements.
Stochastic-Strength-Based Damage Simulation Tool for Ceramic Matrix Composite
NASA Technical Reports Server (NTRS)
Nemeth, Noel; Bednarcyk, Brett; Pineda, Evan; Arnold, Steven; Mital, Subodh; Murthy, Pappu
2015-01-01
Reported here is a coupling of two NASA developed codes: CARES (Ceramics Analysis and Reliability Evaluation of Structures) with the MAC/GMC (Micromechanics Analysis Code/ Generalized Method of Cells) composite material analysis code. The resulting code is called FEAMAC/CARES and is constructed as an Abaqus finite element analysis UMAT (user defined material). Here we describe the FEAMAC/CARES code and an example problem (taken from the open literature) of a laminated CMC in off-axis loading is shown. FEAMAC/CARES performs stochastic-strength-based damage simulation response of a CMC under multiaxial loading using elastic stiffness reduction of the failed elements.
Multi-element fingerprinting as a tool in origin authentication of four east China marine species.
Guo, Lipan; Gong, Like; Yu, Yanlei; Zhang, Hong
2013-12-01
The contents of 25 elements in 4 types of commercial marine species from the East China Sea were determined by inductively coupled plasma mass spectrometry and atomic absorption spectrometry. The elemental composition was used to differentiate marine species according to geographical origin by multivariate statistical analysis. The results showed that principal component analysis could distinguish samples from different areas and reveal the elements which played the most important role in origin diversity. The established models by partial least squares discriminant analysis (PLS-DA) and by probabilistic neural network (PNN) can both precisely predict the origin of the marine species. Further study indicated that PLS-DA and PNN were efficacious in regional discrimination. The models from these 2 statistical methods, with an accuracy of 97.92% and 100%, respectively, could both distinguish samples from different areas without the need for species differentiation. © 2013 Institute of Food Technologists®
Challenges Facing Design and Analysis Tools
NASA Technical Reports Server (NTRS)
Knight, Norman F., Jr.; Broduer, Steve (Technical Monitor)
2001-01-01
The design and analysis of future aerospace systems will strongly rely on advanced engineering analysis tools used in combination with risk mitigation procedures. The implications of such a trend place increased demands on these tools to assess off-nominal conditions, residual strength, damage propagation, and extreme loading conditions in order to understand and quantify these effects as they affect mission success. Advances in computer hardware such as CPU processing speed, memory, secondary storage, and visualization provide significant resources for the engineer to exploit in engineering design. The challenges facing design and analysis tools fall into three primary areas. The first area involves mechanics needs such as constitutive modeling, contact and penetration simulation, crack growth prediction, damage initiation and progression prediction, transient dynamics and deployment simulations, and solution algorithms. The second area involves computational needs such as fast, robust solvers, adaptivity for model and solution strategies, control processes for concurrent, distributed computing for uncertainty assessments, and immersive technology. Traditional finite element codes still require fast direct solvers which when coupled to current CPU power enables new insight as a result of high-fidelity modeling. The third area involves decision making by the analyst. This area involves the integration and interrogation of vast amounts of information - some global in character while local details are critical and often drive the design. The proposed presentation will describe and illustrate these areas using composite structures, energy-absorbing structures, and inflatable space structures. While certain engineering approximations within the finite element model may be adequate for global response prediction, they generally are inadequate in a design setting or when local response prediction is critical. Pitfalls to be avoided and trends for emerging analysis tools will be described.
Multi-color electron microscopy by element-guided identification of cells, organelles and molecules.
Scotuzzi, Marijke; Kuipers, Jeroen; Wensveen, Dasha I; de Boer, Pascal; Hagen, Kees C W; Hoogenboom, Jacob P; Giepmans, Ben N G
2017-04-07
Cellular complexity is unraveled at nanometer resolution using electron microscopy (EM), but interpretation of macromolecular functionality is hampered by the difficulty in interpreting grey-scale images and the unidentified molecular content. We perform large-scale EM on mammalian tissue complemented with energy-dispersive X-ray analysis (EDX) to allow EM-data analysis based on elemental composition. Endogenous elements, labels (gold and cadmium-based nanoparticles) as well as stains are analyzed at ultrastructural resolution. This provides a wide palette of colors to paint the traditional grey-scale EM images for composition-based interpretation. Our proof-of-principle application of EM-EDX reveals that endocrine and exocrine vesicles exist in single cells in Islets of Langerhans. This highlights how elemental mapping reveals unbiased biomedical relevant information. Broad application of EM-EDX will further allow experimental analysis on large-scale tissue using endogenous elements, multiple stains, and multiple markers and thus brings nanometer-scale 'color-EM' as a promising tool to unravel molecular (de)regulation in biomedicine.
Multi-color electron microscopy by element-guided identification of cells, organelles and molecules
Scotuzzi, Marijke; Kuipers, Jeroen; Wensveen, Dasha I.; de Boer, Pascal; Hagen, Kees (C.) W.; Hoogenboom, Jacob P.; Giepmans, Ben N. G.
2017-01-01
Cellular complexity is unraveled at nanometer resolution using electron microscopy (EM), but interpretation of macromolecular functionality is hampered by the difficulty in interpreting grey-scale images and the unidentified molecular content. We perform large-scale EM on mammalian tissue complemented with energy-dispersive X-ray analysis (EDX) to allow EM-data analysis based on elemental composition. Endogenous elements, labels (gold and cadmium-based nanoparticles) as well as stains are analyzed at ultrastructural resolution. This provides a wide palette of colors to paint the traditional grey-scale EM images for composition-based interpretation. Our proof-of-principle application of EM-EDX reveals that endocrine and exocrine vesicles exist in single cells in Islets of Langerhans. This highlights how elemental mapping reveals unbiased biomedical relevant information. Broad application of EM-EDX will further allow experimental analysis on large-scale tissue using endogenous elements, multiple stains, and multiple markers and thus brings nanometer-scale ‘color-EM’ as a promising tool to unravel molecular (de)regulation in biomedicine. PMID:28387351
Integrated Modeling Tools for Thermal Analysis and Applications
NASA Technical Reports Server (NTRS)
Milman, Mark H.; Needels, Laura; Papalexandris, Miltiadis
1999-01-01
Integrated modeling of spacecraft systems is a rapidly evolving area in which multidisciplinary models are developed to design and analyze spacecraft configurations. These models are especially important in the early design stages where rapid trades between subsystems can substantially impact design decisions. Integrated modeling is one of the cornerstones of two of NASA's planned missions in the Origins Program -- the Next Generation Space Telescope (NGST) and the Space Interferometry Mission (SIM). Common modeling tools for control design and opto-mechanical analysis have recently emerged and are becoming increasingly widely used. A discipline that has been somewhat less integrated, but is nevertheless of critical concern for high precision optical instruments, is thermal analysis and design. A major factor contributing to this mild estrangement is that the modeling philosophies and objectives for structural and thermal systems typically do not coincide. Consequently the tools that are used in these discplines suffer a degree of incompatibility, each having developed along their own evolutionary path. Although standard thermal tools have worked relatively well in the past. integration with other disciplines requires revisiting modeling assumptions and solution methods. Over the past several years we have been developing a MATLAB based integrated modeling tool called IMOS (Integrated Modeling of Optical Systems) which integrates many aspects of structural, optical, control and dynamical analysis disciplines. Recent efforts have included developing a thermal modeling and analysis capability, which is the subject of this article. Currently, the IMOS thermal suite contains steady state and transient heat equation solvers, and the ability to set up the linear conduction network from an IMOS finite element model. The IMOS code generates linear conduction elements associated with plates and beams/rods of the thermal network directly from the finite element structural model. Conductances for temperature varying materials are accommodated. This capability both streamlines the process of developing the thermal model from the finite element model, and also makes the structural and thermal models compatible in the sense that each structural node is associated with a thermal node. This is particularly useful when the purpose of the analysis is to predict structural deformations due to thermal loads. The steady state solver uses a restricted step size Newton method, and the transient solver is an adaptive step size implicit method applicable to general differential algebraic systems. Temperature dependent conductances and capacitances are accommodated by the solvers. In addition to discussing the modeling and solution methods. applications where the thermal modeling is "in the loop" with sensitivity analysis, optimization and optical performance drawn from our experiences with the Space Interferometry Mission (SIM), and the Next Generation Space Telescope (NGST) are presented.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Damiani, Rick
This manual summarizes the theory and preliminary verifications of the JacketSE module, which is an offshore jacket sizing tool that is part of the Wind-Plant Integrated System Design & Engineering Model toolbox. JacketSE is based on a finite-element formulation and on user-prescribed inputs and design standards' criteria (constraints). The physics are highly simplified, with a primary focus on satisfying ultimate limit states and modal performance requirements. Preliminary validation work included comparing industry data and verification against ANSYS, a commercial finite-element analysis package. The results are encouraging, and future improvements to the code are recommended in this manual.
Control/structure interaction conceptual design tool
NASA Technical Reports Server (NTRS)
Briggs, Hugh C.
1990-01-01
The JPL Control/Structure Interaction Program is developing new analytical methods for designing micro-precision spacecraft with controlled structures. One of these, the Conceptual Design Tool, will illustrate innovative new approaches to the integration of multi-disciplinary analysis and design methods. The tool will be used to demonstrate homogeneity of presentation, uniform data representation across analytical methods, and integrated systems modeling. The tool differs from current 'integrated systems' that support design teams most notably in its support for the new CSI multi-disciplinary engineer. The design tool will utilize a three dimensional solid model of the spacecraft under design as the central data organization metaphor. Various analytical methods, such as finite element structural analysis, control system analysis, and mechanical configuration layout, will store and retrieve data from a hierarchical, object oriented data structure that supports assemblies of components with associated data and algorithms. In addition to managing numerical model data, the tool will assist the designer in organizing, stating, and tracking system requirements.
Applying analysis tools in planning for operations
DOT National Transportation Integrated Search
2009-09-01
More and more, transportation system operators are seeing the benefits of strengthening links between planning and operations. A critical element in improving transportation decision-making and the effectiveness of transportation systems related to o...
NASA Astrophysics Data System (ADS)
Kiliclar, Yalin; Laurischkat, Roman; Vladimirov, Ivaylo N.; Reese, Stefanie
2011-08-01
The presented project deals with a robot based incremental sheet metal forming process, which is called roboforming and has been developed at the Chair of Production Systems. It is characterized by flexible shaping using a freely programmable path-synchronous movement of two industrial robots. The final shape is produced by the incremental infeed of the forming tool in depth direction and its movement along the part contour in lateral direction. However, the resulting geometries formed in roboforming deviate several millimeters from the reference geometry. This results from the compliance of the involved machine structures and the springback effects of the workpiece. The project aims to predict these deviations caused by resiliences and to carry out a compensative path planning based on this prediction. Therefore a planning tool is implemented which compensates the robots's compliance and the springback effects of the sheet metal. The forming process is simulated by means of a finite element analysis using a material model developed at the Institute of Applied Mechanics (IFAM). It is based on the multiplicative split of the deformation gradient in the context of hyperelasticity and combines nonlinear kinematic and isotropic hardening. Low-order finite elements used to simulate thin sheet structures, such as used for the experiments, have the major problem of locking, a nonphysical stiffening effect. For an efficient finite element analysis a special solid-shell finite element formulation based on reduced integration with hourglass stabilization has been developed. To circumvent different locking effects, the enhanced assumed strain (EAS) and the assumed natural strain (ANS) concepts are included in this formulation. Having such powerful tools available we obtain more accurate geometries.
Factors Influencing Progressive Failure Analysis Predictions for Laminated Composite Structure
NASA Technical Reports Server (NTRS)
Knight, Norman F., Jr.
2008-01-01
Progressive failure material modeling methods used for structural analysis including failure initiation and material degradation are presented. Different failure initiation criteria and material degradation models are described that define progressive failure formulations. These progressive failure formulations are implemented in a user-defined material model for use with a nonlinear finite element analysis tool. The failure initiation criteria include the maximum stress criteria, maximum strain criteria, the Tsai-Wu failure polynomial, and the Hashin criteria. The material degradation model is based on the ply-discounting approach where the local material constitutive coefficients are degraded. Applications and extensions of the progressive failure analysis material model address two-dimensional plate and shell finite elements and three-dimensional solid finite elements. Implementation details are described in the present paper. Parametric studies for laminated composite structures are discussed to illustrate the features of the progressive failure modeling methods that have been implemented and to demonstrate their influence on progressive failure analysis predictions.
Rojas, David; Grierson, Lawrence; Mylopoulos, Maria; Trbovich, Patricia; Bagli, Darius; Brydges, Ryan
2018-04-01
We evaluate programmes in health professions education (HPE) to determine their effectiveness and value. Programme evaluation has evolved from use of reductionist frameworks to those addressing the complex interactions between programme factors. Researchers in HPE have recently suggested a 'holistic programme evaluation' aiming to better describe and understand the implications of 'emergent processes and outcomes'. We propose a programme evaluation framework informed by principles and tools from systems engineering. Systems engineers conceptualise complexity and emergent elements in unique ways that may complement and extend contemporary programme evaluations in HPE. We demonstrate how the abstract decomposition space (ADS), an engineering knowledge elicitation tool, provides the foundation for a systems engineering informed programme evaluation designed to capture both planned and emergent programme elements. We translate the ADS tool to use education-oriented language, and describe how evaluators can use it to create a programme-specific ADS through iterative refinement. We provide a conceptualisation of emergent elements and an equation that evaluators can use to identify the emergent elements in their programme. Using our framework, evaluators can analyse programmes not as isolated units with planned processes and planned outcomes, but as unfolding, complex interactive systems that will exhibit emergent processes and emergent outcomes. Subsequent analysis of these emergent elements will inform the evaluator as they seek to optimise and improve the programme. Our proposed systems engineering informed programme evaluation framework provides principles and tools for analysing the implications of planned and emergent elements, as well as their potential interactions. We acknowledge that our framework is preliminary and will require application and constant refinement. We suggest that our framework will also advance our understanding of the construct of 'emergence' in HPE research. © 2017 John Wiley & Sons Ltd and The Association for the Study of Medical Education.
Begum, M. Sameena; Dinesh, M. R.; Tan, Kenneth F. H.; Jairaj, Vani; Md Khalid, K.; Singh, Varun Pratap
2015-01-01
The finite element method (FEM) is a powerful computational tool for solving stress-strain problems; its ability to handle material inhomogeneity and complex shapes makes the FEM, the most suitable method for the analysis of internal stress levels in the tooth, periodontium, and alveolar bone. This article intends to explain the steps involved in the generation of a three-dimensional finite element model of tooth, periodontal ligament (PDL) and alveolar bone, as the procedure of modeling is most important because the result is based on the nature of the modeling systems. Finite element analysis offers a means of determining strain-stress levels in the tooth, ligament, and bone structures for a broad range of orthodontic loading scenarios without producing tissue damage. PMID:26538895
Relating MBSE to Spacecraft Development: A NASA Pathfinder
NASA Technical Reports Server (NTRS)
Othon, Bill
2016-01-01
The NASA Engineering and Safety Center (NESC) has sponsored a Pathfinder Study to investigate how Model Based Systems Engineering (MBSE) and Model Based Engineering (MBE) techniques can be applied by NASA spacecraft development projects. The objectives of this Pathfinder Study included analyzing both the products of the modeling activity, as well as the process and tool chain through which the spacecraft design activities are executed. Several aspects of MBSE methodology and process were explored. Adoption and consistent use of the MBSE methodology within an existing development environment can be difficult. The Pathfinder Team evaluated the possibility that an "MBSE Template" could be developed as both a teaching tool as well as a baseline from which future NASA projects could leverage. Elements of this template include spacecraft system component libraries, data dictionaries and ontology specifications, as well as software services that do work on the models themselves. The Pathfinder Study also evaluated the tool chain aspects of development. Two chains were considered: 1. The Development tool chain, through which SysML model development was performed and controlled, and 2. The Analysis tool chain, through which both static and dynamic system analysis is performed. Of particular interest was the ability to exchange data between SysML and other engineering tools such as CAD and Dynamic Simulation tools. For this study, the team selected a Mars Lander vehicle as the element to be designed. The paper will discuss what system models were developed, how data was captured and exchanged, and what analyses were conducted.
NASA Technical Reports Server (NTRS)
Haber, Benjamin M.; Green, Joseph J.
2010-01-01
The GOATS Orbitology Component software was developed to specifically address the concerns presented by orbit analysis tools that are often written as stand-alone applications. These applications do not easily interface with standard JPL first-principles analysis tools, and have a steep learning curve due to their complicated nature. This toolset is written as a series of MATLAB functions, allowing seamless integration into existing JPL optical systems engineering modeling and analysis modules. The functions are completely open, and allow for advanced users to delve into and modify the underlying physics being modeled. Additionally, this software module fills an analysis gap, allowing for quick, high-level mission analysis trades without the need for detailed and complicated orbit analysis using commercial stand-alone tools. This software consists of a series of MATLAB functions to provide for geometric orbit-related analysis. This includes propagation of orbits to varying levels of generalization. In the simplest case, geosynchronous orbits can be modeled by specifying a subset of three orbit elements. The next case is a circular orbit, which can be specified by a subset of four orbit elements. The most general case is an arbitrary elliptical orbit specified by all six orbit elements. These orbits are all solved geometrically, under the basic problem of an object in circular (or elliptical) orbit around a rotating spheroid. The orbit functions output time series ground tracks, which serve as the basis for more detailed orbit analysis. This software module also includes functions to track the positions of the Sun, Moon, and arbitrary celestial bodies specified by right ascension and declination. Also included are functions to calculate line-of-sight geometries to ground-based targets, angular rotations and decompositions, and other line-of-site calculations. The toolset allows for the rapid execution of orbit trade studies at the level of detail required for the early stage of mission concept development.
The p-version of the finite element method in incremental elasto-plastic analysis
NASA Technical Reports Server (NTRS)
Holzer, Stefan M.; Yosibash, Zohar
1993-01-01
Whereas the higher-order versions of the finite elements method (the p- and hp-version) are fairly well established as highly efficient methods for monitoring and controlling the discretization error in linear problems, little has been done to exploit their benefits in elasto-plastic structural analysis. Aspects of incremental elasto-plastic finite element analysis which are particularly amenable to improvements by the p-version is discussed. These theoretical considerations are supported by several numerical experiments. First, an example for which an analytical solution is available is studied. It is demonstrated that the p-version performs very well even in cycles of elasto-plastic loading and unloading, not only as compared to the traditional h-version but also in respect to the exact solution. Finally, an example of considerable practical importance - the analysis of a cold-worked lug - is presented which demonstrates how the modeling tools offered by higher-order finite element techniques can contribute to an improved approximation of practical problems.
1988-11-01
264 ANALYSIS RESTART. ............. ..... ....... 269 1.0 TITLE CARD. .............. ............. 271 2.0 CONTROL CARDS...stress soil model will provide a tool for such analysis of waterfront structures. To understand the significance of liquefaction, it is important to note...Implementing this effective stress soil model into a finite element computer program would allow analysis of soil and structure together. TECHNICAL BACKGROUND
DOE Office of Scientific and Technical Information (OSTI.GOV)
Khaleel, Mohammad A.; Lin, Zijing; Singh, Prabhakar
2004-05-03
A 3D simulation tool for modeling solid oxide fuel cells is described. The tool combines the versatility and efficiency of a commercial finite element analysis code, MARC{reg_sign}, with an in-house developed robust and flexible electrochemical (EC) module. Based upon characteristic parameters obtained experimentally and assigned by the user, the EC module calculates the current density distribution, heat generation, and fuel and oxidant species concentration, taking the temperature profile provided by MARC{reg_sign} and operating conditions such as the fuel and oxidant flow rate and the total stack output voltage or current as the input. MARC{reg_sign} performs flow and thermal analyses basedmore » on the initial and boundary thermal and flow conditions and the heat generation calculated by the EC module. The main coupling between MARC{reg_sign} and EC is for MARC{reg_sign} to supply the temperature field to EC and for EC to give the heat generation profile to MARC{reg_sign}. The loosely coupled, iterative scheme is advantageous in terms of memory requirement, numerical stability and computational efficiency. The coupling is iterated to self-consistency for a steady-state solution. Sample results for steady states as well as the startup process for stacks with different flow designs are presented to illustrate the modeling capability and numerical performance characteristic of the simulation tool.« less
Wireless acceleration sensor of moving elements for condition monitoring of mechanisms
NASA Astrophysics Data System (ADS)
Sinitsin, Vladimir V.; Shestakov, Aleksandr L.
2017-09-01
Comprehensive analysis of the angular and linear accelerations of moving elements (shafts, gears) allows an increase in the quality of the condition monitoring of mechanisms. However, existing tools and methods measure either linear or angular acceleration with postprocessing. This paper suggests a new construction design of an angular acceleration sensor for moving elements. The sensor is mounted on a moving element and, among other things, the data transfer and electric power supply are carried out wirelessly. In addition, the authors introduce a method for processing the received information which makes it possible to divide the measured acceleration into the angular and linear components. The design has been validated by the results of laboratory tests of an experimental model of the sensor. The study has shown that this method provides a definite separation of the measured acceleration into linear and angular components, even in noise. This research contributes an advance in the range of methods and tools for condition monitoring of mechanisms.
NASA Technical Reports Server (NTRS)
Antonille, Scott R.; Miskey, Cherie L.; Ohl, Raymond G.; Rohrbach, Scott O.; Aronstein, David L.; Bartoszyk, Andrew E.; Bowers, Charles W.; Cofie, Emmanuel; Collins, Nicholas R.; Comber, Brian J.;
2016-01-01
NASA's James Webb Space Telescope (JWST) is a 6.6m diameter, segmented, deployable telescope for cryogenic IR space astronomy (40K). The JWST Observatory includes the Optical Telescope Element (OTE) and the Integrated Science Instrument Module (ISIM) that contains four science instruments (SI) and the fine guider. The SIs are mounted to a composite metering structure. The SI and guider units were integrated to the ISIM structure and optically tested at the NASA Goddard Space Flight Center as a suite using the Optical Telescope Element SIMulator (OSIM). OSIM is a full field, cryogenic JWST telescope simulator. SI performance, including alignment and wave front error, were evaluated using OSIM. We describe test and analysis methods for optical performance verification of the ISIM Element, with an emphasis on the processes used to plan and execute the test. The complexity of ISIM and OSIM drove us to develop a software tool for test planning that allows for configuration control of observations, associated scripts, and management of hardware and software limits and constraints, as well as tools for rapid data evaluation, and flexible re-planning in response to the unexpected. As examples of our test and analysis approach, we discuss how factors such as the ground test thermal environment are compensated in alignment. We describe how these innovative methods for test planning and execution and post-test analysis were instrumental in the verification program for the ISIM element, with enough information to allow the reader to consider these innovations and lessons learned in this successful effort in their future testing for other programs.
NASA Astrophysics Data System (ADS)
Antonille, Scott R.; Miskey, Cherie L.; Ohl, Raymond G.; Rohrbach, Scott O.; Aronstein, David L.; Bartoszyk, Andrew E.; Bowers, Charles W.; Cofie, Emmanuel; Collins, Nicholas R.; Comber, Brian J.; Eichhorn, William L.; Glasse, Alistair C.; Gracey, Renee; Hartig, George F.; Howard, Joseph M.; Kelly, Douglas M.; Kimble, Randy A.; Kirk, Jeffrey R.; Kubalak, David A.; Landsman, Wayne B.; Lindler, Don J.; Malumuth, Eliot M.; Maszkiewicz, Michael; Rieke, Marcia J.; Rowlands, Neil; Sabatke, Derek S.; Smith, Corbett T.; Smith, J. Scott; Sullivan, Joseph F.; Telfer, Randal C.; Te Plate, Maurice; Vila, M. Begoña.; Warner, Gerry D.; Wright, David; Wright, Raymond H.; Zhou, Julia; Zielinski, Thomas P.
2016-09-01
NASA's James Webb Space Telescope (JWST) is a 6.5m diameter, segmented, deployable telescope for cryogenic IR space astronomy. The JWST Observatory includes the Optical Telescope Element (OTE) and the Integrated Science Instrument Module (ISIM), that contains four science instruments (SI) and the Fine Guidance Sensor (FGS). The SIs are mounted to a composite metering structure. The SIs and FGS were integrated to the ISIM structure and optically tested at NASA's Goddard Space Flight Center using the Optical Telescope Element SIMulator (OSIM). OSIM is a full-field, cryogenic JWST telescope simulator. SI performance, including alignment and wavefront error, was evaluated using OSIM. We describe test and analysis methods for optical performance verification of the ISIM Element, with an emphasis on the processes used to plan and execute the test. The complexity of ISIM and OSIM drove us to develop a software tool for test planning that allows for configuration control of observations, implementation of associated scripts, and management of hardware and software limits and constraints, as well as tools for rapid data evaluation, and flexible re-planning in response to the unexpected. As examples of our test and analysis approach, we discuss how factors such as the ground test thermal environment are compensated in alignment. We describe how these innovative methods for test planning and execution and post-test analysis were instrumental in the verification program for the ISIM element, with enough information to allow the reader to consider these innovations and lessons learned in this successful effort in their future testing for other programs.
DCODE.ORG Anthology of Comparative Genomic Tools
DOE Office of Scientific and Technical Information (OSTI.GOV)
Loots, G G; Ovcharenko, I
2005-01-11
Comparative genomics provides the means to demarcate functional regions in anonymous DNA sequences. The successful application of this method to identifying novel genes is currently shifting to deciphering the noncoding encryption of gene regulation across genomes. To facilitate the use of comparative genomics to practical applications in genetics and genomics we have developed several analytical and visualization tools for the analysis of arbitrary sequences and whole genomes. These tools include two alignment tools: zPicture and Mulan; a phylogenetic shadowing tool: eShadow for identifying lineage- and species-specific functional elements; two evolutionary conserved transcription factor analysis tools: rVista and multiTF; a toolmore » for extracting cis-regulatory modules governing the expression of co-regulated genes, CREME; and a dynamic portal to multiple vertebrate and invertebrate genome alignments, the ECR Browser. Here we briefly describe each one of these tools and provide specific examples on their practical applications. All the tools are publicly available at the http://www.dcode.org/ web site.« less
ERIC Educational Resources Information Center
Castillo, Jose M.; March, Amanda L.; Stockslager, Kevin M.; Hines, Constance V.
2016-01-01
The "Perceptions of RtI Skills Survey" is a self-report measure that assesses educators' perceptions of their data-based problem-solving skills--a critical element of many Response-to-Intervention (RtI) models. Confirmatory factor analysis (CFA) was used to evaluate the underlying factor structure of this tool. Educators from 68 (n =…
In Search of a Time Efficient Approach to Crack and Delamination Growth Predictions in Composites
NASA Technical Reports Server (NTRS)
Krueger, Ronald; Carvalho, Nelson
2016-01-01
Analysis benchmarking was used to assess the accuracy and time efficiency of algorithms suitable for automated delamination growth analysis. First, the Floating Node Method (FNM) was introduced and its combination with a simple exponential growth law (Paris Law) and Virtual Crack Closure technique (VCCT) was discussed. Implementation of the method into a user element (UEL) in Abaqus/Standard(Registered TradeMark) was also presented. For the assessment of growth prediction capabilities, an existing benchmark case based on the Double Cantilever Beam (DCB) specimen was briefly summarized. Additionally, the development of new benchmark cases based on the Mixed-Mode Bending (MMB) specimen to assess the growth prediction capabilities under mixed-mode I/II conditions was discussed in detail. A comparison was presented, in which the benchmark cases were used to assess the existing low-cycle fatigue analysis tool in Abaqus/Standard(Registered TradeMark) in comparison to the FNM-VCCT fatigue growth analysis implementation. The low-cycle fatigue analysis tool in Abaqus/Standard(Registered TradeMark) was able to yield results that were in good agreement with the DCB benchmark example. Results for the MMB benchmark cases, however, only captured the trend correctly. The user element (FNM-VCCT) always yielded results that were in excellent agreement with all benchmark cases, at a fraction of the analysis time. The ability to assess the implementation of two methods in one finite element code illustrated the value of establishing benchmark solutions.
A drill-soil system modelization for future Mars exploration
NASA Astrophysics Data System (ADS)
Finzi, A. E.; Lavagna, M.; Rocchitelli, G.
2004-01-01
This paper presents a first approach to the problem of modeling a drilling process to be carried on in the space environment by a dedicated payload. Systems devoted to work in space present very strict requirements in many different fields such as thermal response, electric power demand, reliability and so on. Thus, models devoted to the operational behaviour simulation represent a fundamental help in the design phase and give a great improvement in the final product quality. As the required power is the crucial constraint within drilling devices, the tool-soil interaction modelization and simulation are finalized to the computation of the power demand as a function of both the drill and the soil parameters. An accurate study of the tool and the soil separately has been firstly carried on and, secondly their interaction has been analyzed. The Dee-Dri system, designed by Tecnospazio and to be part of the lander components in the NASA's Mars Sample Return Mission, has been taken as the tool reference. The Deep-Drill system is a complex rotary tool devoted to the soil perforation and sample collection; it has to operate in a Martian zone made of rocks similar to the terrestrial basalt, then the modelization is restricted to the interaction analysis between the tool and materials belonging to the rock set. The tool geometric modelization has been faced by a finite element approach with a Langrangian formulation: for the static analysis a refined model is assumed considering both the actual geometry of the head and the rod screws; a simplified model has been used to deal with the dynamic analysis. The soil representation is based on the Mohr-Coulomb crack criterion and an Eulerian approach has been selected to model it. However, software limitations in dealing with the tool-soil interface definition required assuming a Langrangian formulation for the soil too. The interaction between the soil and the tool has been modeled by extending the two-dimensional Nishimatsu's theory for rock cutting for rotating perforation tools. A fine analysis on f.e.m. element choice for each part of the tool is presented together with static analysis results. The dynamic analysis results are limited to the first impact phenomenon between the rock and the tool head. The validity of both the theoretical and numerical models is confirmed by the good agreement between simulation results and data coming from the experiments done within the Tecnospazio facilities.
Computational mechanics analysis tools for parallel-vector supercomputers
NASA Technical Reports Server (NTRS)
Storaasli, Olaf O.; Nguyen, Duc T.; Baddourah, Majdi; Qin, Jiangning
1993-01-01
Computational algorithms for structural analysis on parallel-vector supercomputers are reviewed. These parallel algorithms, developed by the authors, are for the assembly of structural equations, 'out-of-core' strategies for linear equation solution, massively distributed-memory equation solution, unsymmetric equation solution, general eigensolution, geometrically nonlinear finite element analysis, design sensitivity analysis for structural dynamics, optimization search analysis and domain decomposition. The source code for many of these algorithms is available.
NASTRAN analysis of Tokamak vacuum vessel using interactive graphics
NASA Technical Reports Server (NTRS)
Miller, A.; Badrian, M.
1978-01-01
Isoparametric quadrilateral and triangular elements were used to represent the vacuum vessel shell structure. For toroidally symmetric loadings, MPCs were employed across model boundaries and rigid format 24 was invoked. Nonsymmetric loadings required the use of the cyclic symmetry analysis available with rigid format 49. NASTRAN served as an important analysis tool in the Tokamak design effort by providing a reliable means for assessing structural integrity. Interactive graphics were employed in the finite element model generation and in the post-processing of results. It was felt that model generation and checkout with interactive graphics reduced the modelling effort and debugging man-hours significantly.
Finite Element Modeling Techniques for Analysis of VIIP
NASA Technical Reports Server (NTRS)
Feola, Andrew J.; Raykin, J.; Gleason, R.; Mulugeta, Lealem; Myers, Jerry G.; Nelson, Emily S.; Samuels, Brian C.; Ethier, C. Ross
2015-01-01
Visual Impairment and Intracranial Pressure (VIIP) syndrome is a major health concern for long-duration space missions. Currently, it is thought that a cephalad fluid shift in microgravity causes elevated intracranial pressure (ICP) that is transmitted along the optic nerve sheath (ONS). We hypothesize that this in turn leads to alteration and remodeling of connective tissue in the posterior eye which impacts vision. Finite element (FE) analysis is a powerful tool for examining the effects of mechanical loads in complex geometries. Our goal is to build a FE analysis framework to understand the response of the lamina cribrosa and optic nerve head to elevations in ICP in VIIP.
A new X-ray fluorescence spectroscopy for extraterrestrial materials using a muon beam
Terada, K.; Ninomiya, K.; Osawa, T.; Tachibana, S.; Miyake, Y.; Kubo, M. K.; Kawamura, N.; Higemoto, W.; Tsuchiyama, A.; Ebihara, M.; Uesugi, M.
2014-01-01
The recent development of the intense pulsed muon source at J-PARC MUSE, Japan Proton Accelerator Research Complex/MUon Science Establishment (106 s−1 for a momentum of 60 MeV/c), enabled us to pioneer a new frontier in analytical sciences. Here, we report a non-destructive elemental analysis using µ− capture. Controlling muon momentum from 32.5 to 57.5 MeV/c, we successfully demonstrate a depth-profile analysis of light elements (B, C, N, and O) from several mm-thick layered materials and non-destructive bulk analyses of meteorites containing organic materials. Muon beam analysis, enabling a bulk analysis of light to heavy elements without severe radioactivation, is a unique analytical method complementary to other non-destructive analyses. Furthermore, this technology can be used as a powerful tool to identify the content and distribution of organic components in future asteroidal return samples. PMID:24861282
NASA Technical Reports Server (NTRS)
Bogert, Philip B.; Satyanarayana, Arunkumar; Chunchu, Prasad B.
2006-01-01
Splitting, ultimate failure load and the damage path in center notched composite specimens subjected to in-plane tension loading are predicted using progressive failure analysis methodology. A 2-D Hashin-Rotem failure criterion is used in determining intra-laminar fiber and matrix failures. This progressive failure methodology has been implemented in the Abaqus/Explicit and Abaqus/Standard finite element codes through user written subroutines "VUMAT" and "USDFLD" respectively. A 2-D finite element model is used for predicting the intra-laminar damages. Analysis results obtained from the Abaqus/Explicit and Abaqus/Standard code show good agreement with experimental results. The importance of modeling delamination in progressive failure analysis methodology is recognized for future studies. The use of an explicit integration dynamics code for simple specimen geometry and static loading establishes a foundation for future analyses where complex loading and nonlinear dynamic interactions of damage and structure will necessitate it.
A new X-ray fluorescence spectroscopy for extraterrestrial materials using a muon beam.
Terada, K; Ninomiya, K; Osawa, T; Tachibana, S; Miyake, Y; Kubo, M K; Kawamura, N; Higemoto, W; Tsuchiyama, A; Ebihara, M; Uesugi, M
2014-05-27
The recent development of the intense pulsed muon source at J-PARC MUSE, Japan Proton Accelerator Research Complex/MUon Science Establishment (10(6) s(-1) for a momentum of 60 MeV/c), enabled us to pioneer a new frontier in analytical sciences. Here, we report a non-destructive elemental analysis using µ(-) capture. Controlling muon momentum from 32.5 to 57.5 MeV/c, we successfully demonstrate a depth-profile analysis of light elements (B, C, N, and O) from several mm-thick layered materials and non-destructive bulk analyses of meteorites containing organic materials. Muon beam analysis, enabling a bulk analysis of light to heavy elements without severe radioactivation, is a unique analytical method complementary to other non-destructive analyses. Furthermore, this technology can be used as a powerful tool to identify the content and distribution of organic components in future asteroidal return samples.
Participatory design of probability-based decision support tools for in-hospital nurses.
Jeffery, Alvin D; Novak, Laurie L; Kennedy, Betsy; Dietrich, Mary S; Mion, Lorraine C
2017-11-01
To describe nurses' preferences for the design of a probability-based clinical decision support (PB-CDS) tool for in-hospital clinical deterioration. A convenience sample of bedside nurses, charge nurses, and rapid response nurses (n = 20) from adult and pediatric hospitals completed participatory design sessions with researchers in a simulation laboratory to elicit preferred design considerations for a PB-CDS tool. Following theme-based content analysis, we shared findings with user interface designers and created a low-fidelity prototype. Three major themes and several considerations for design elements of a PB-CDS tool surfaced from end users. Themes focused on "painting a picture" of the patient condition over time, promoting empowerment, and aligning probability information with what a nurse already believes about the patient. The most notable design element consideration included visualizing a temporal trend of the predicted probability of the outcome along with user-selected overlapping depictions of vital signs, laboratory values, and outcome-related treatments and interventions. Participants expressed that the prototype adequately operationalized requests from the design sessions. Participatory design served as a valuable method in taking the first step toward developing PB-CDS tools for nurses. This information about preferred design elements of tools that support, rather than interrupt, nurses' cognitive workflows can benefit future studies in this field as well as nurses' practice. Published by Oxford University Press on behalf of the American Medical Informatics Association 2017. This work is written by US Government employees and is in the public domain in the United States.
Scalable Visual Analytics of Massive Textual Datasets
DOE Office of Scientific and Technical Information (OSTI.GOV)
Krishnan, Manoj Kumar; Bohn, Shawn J.; Cowley, Wendy E.
2007-04-01
This paper describes the first scalable implementation of text processing engine used in Visual Analytics tools. These tools aid information analysts in interacting with and understanding large textual information content through visual interfaces. By developing parallel implementation of the text processing engine, we enabled visual analytics tools to exploit cluster architectures and handle massive dataset. The paper describes key elements of our parallelization approach and demonstrates virtually linear scaling when processing multi-gigabyte data sets such as Pubmed. This approach enables interactive analysis of large datasets beyond capabilities of existing state-of-the art visual analytics tools.
When product designers use perceptually based color tools
NASA Astrophysics Data System (ADS)
Bender, Walter R.
1998-07-01
Palette synthesis and analysis tools have been built based upon a model of color experience. This model adjusts formal compositional elements such as hue, value, chroma, and their contrasts, as well as size and proportion. Clothing and household product designers were given these tools to give guidance to their selection of seasonal palettes for use in production of the private-label merchandise of a large retail chain. The designers chose base palettes. Accents to these palettes were generated with and without the aid of the color tools. These palettes are compared by using perceptual metrics and interviews. The results are presented.
When product designers use perceptually based color tools
NASA Astrophysics Data System (ADS)
Bender, Walter R.
2001-01-01
Palette synthesis and analysis tools have been built based upon a model of color experience. This model adjusts formal compositional elements such as hue, value, chroma, and their contrasts, as well as size and proportion. Clothing and household product designers were given these tools to guide their selection of seasonal palettes in the production of the private-label merchandise in a large retail chain. The designers chose base palettes. Accents to these palettes were generated with and without the aid of the color tools. These palettes are compared by using perceptual metrics and interviews. The results are presented.
On modelling three-dimensional piezoelectric smart structures with boundary spectral element method
NASA Astrophysics Data System (ADS)
Zou, Fangxin; Aliabadi, M. H.
2017-05-01
The computational efficiency of the boundary element method in elastodynamic analysis can be significantly improved by employing high-order spectral elements for boundary discretisation. In this work, for the first time, the so-called boundary spectral element method is utilised to formulate the piezoelectric smart structures that are widely used in structural health monitoring (SHM) applications. The resultant boundary spectral element formulation has been validated by the finite element method (FEM) and physical experiments. The new formulation has demonstrated a lower demand on computational resources and a higher numerical stability than commercial FEM packages. Comparing to the conventional boundary element formulation, a significant reduction in computational expenses has been achieved. In summary, the boundary spectral element formulation presented in this paper provides a highly efficient and stable mathematical tool for the development of SHM applications.
Status of the AIAA Modeling and Simulation Format Standard
NASA Technical Reports Server (NTRS)
Jackson, E. Bruce; Hildreth, Bruce L.
2008-01-01
The current draft AIAA Standard for flight simulation models represents an on-going effort to improve the productivity of practitioners of the art of digital flight simulation (one of the original digital computer applications). This initial release provides the capability for the efficient representation and exchange of an aerodynamic model in full fidelity; the DAVE-ML format can be easily imported (with development of site-specific import tools) in an unambiguous way with automatic verification. An attractive feature of the standard is the ability to coexist with existing legacy software or tools. The draft Standard is currently limited in scope to static elements of dynamic flight simulations; however, these static elements represent the bulk of typical flight simulation mathematical models. It is already seeing application within U.S. and Australian government agencies in an effort to improve productivity and reduce model rehosting overhead. An existing tool allows import of DAVE-ML models into a popular simulation modeling and analysis tool, and other community-contributed tools and libraries can simplify the use of DAVE-ML compliant models at compile- or run-time of high-fidelity flight simulation.
Structural Analysis Methods for Structural Health Management of Future Aerospace Vehicles
NASA Technical Reports Server (NTRS)
Tessler, Alexander
2007-01-01
Two finite element based computational methods, Smoothing Element Analysis (SEA) and the inverse Finite Element Method (iFEM), are reviewed, and examples of their use for structural health monitoring are discussed. Due to their versatility, robustness, and computational efficiency, the methods are well suited for real-time structural health monitoring of future space vehicles, large space structures, and habitats. The methods may be effectively employed to enable real-time processing of sensing information, specifically for identifying three-dimensional deformed structural shapes as well as the internal loads. In addition, they may be used in conjunction with evolutionary algorithms to design optimally distributed sensors. These computational tools have demonstrated substantial promise for utilization in future Structural Health Management (SHM) systems.
Determination of elements in hospital waste with neutron activation analysis method
NASA Astrophysics Data System (ADS)
Dwijananti, P.; Astuti, B.; Alwiyah; Fianti
2018-03-01
The producer of the biggest B3 waste is hospital. The waste is from medical and laboratory activities. The purpose of this study is to determine the elements contained in the liquid waste from hospital and calculate the levels of these elements. This research was done by analysis of the neutron activation conducted at BATAN Yogyakarta. The neutron activation analysis is divided into two stages: activation of the samples using neutron sources of reactor Kartini, then chopping by using a set of tools, gamma spectrometer with HPGe detector. Qualitative and quantitative analysis were done by matching the gamma spectrum peak to the Neutron Activation Table. The sample was taken from four points of the liquid waste treatment plant (WWTP) Bhakti Wira Tamtama Semarang hospital. The results showed that the samples containing elements of Cr, Zn, Fe, Co, and Na, with the levels of each element is Cr (0.033 - 0.075) mg/L, Zn (0.090 - 1.048) mg/L, Fe (2.937-37.743) mg/L, Co (0.005-0.023) mg/L, and Na (61.088-116.330) mg/L. Comparing to the standard value, the liquid is safe to the environment.
Dynamic analysis of flexible mechanical systems using LATDYN
NASA Technical Reports Server (NTRS)
Wu, Shih-Chin; Chang, Che-Wei; Housner, Jerrold M.
1989-01-01
A 3-D, finite element based simulation tool for flexible multibody systems is presented. Hinge degrees-of-freedom is built into equations of motion to reduce geometric constraints. The approach avoids the difficulty in selecting deformation modes for flexible components by using assumed mode method. The tool is applied to simulate a practical space structure deployment problem. Results of examples demonstrate the capability of the code and approach.
Analysis of past NBI ratings to determine future bridge preservation needs.
DOT National Transportation Integrated Search
2004-01-01
Bridge Management System (BMS) needs an analytical tool that can predict bridge element deterioration and answer questions related to bridge preservation. PONTIS, a comprehensive BMS software, was developed to serve this purpose. However, the intensi...
2010-02-27
investigated in more detail. The intermediate level of fidelity, though more expensive, is then used to refine the analysis , add geometric detail, and...design stage is used to further refine the analysis , narrowing the design to a handful of options. Figure 1. Integrated Hierarchical Framework. In...computational structural and computational fluid modeling. For the structural analysis tool we used McIntosh Structural Dynamics’ finite element code CNEVAL
The application of finite element analysis in the skull biomechanics and dentistry.
Prado, Felippe Bevilacqua; Rossi, Ana Cláudia; Freire, Alexandre Rodrigues; Ferreira Caria, Paulo Henrique
2014-01-01
Empirical concepts describe the direction of the masticatory stress dissipation in the skull. The scientific evidence of the trajectories and the magnitude of stress dissipation can help in the diagnosis of the masticatory alterations and the planning of oral rehabilitation in the different areas of Dentistry. The Finite Element Analysis (FEA) is a tool that may reproduce complex structures with irregular geometries of natural and artificial tissues of the human body because it uses mathematical functions that enable the understanding of the craniofacial biomechanics. The aim of this study was to review the literature on the advantages and limitations of FEA in the skull biomechanics and Dentistry study. The keywords of the selected original research articles were: Finite element analysis, biomechanics, skull, Dentistry, teeth, and implant. The literature review was performed in the databases, PUBMED, MEDLINE and SCOPUS. The selected books and articles were between the years 1928 and 2010. The FEA is an assessment tool whose application in different areas of the Dentistry has gradually increased over the past 10 years, but its application in the analysis of the skull biomechanics is scarce. The main advantages of the FEA are the realistic mode of approach and the possibility of results being based on analysis of only one model. On the other hand, the main limitation of the FEA studies is the lack of anatomical details in the modeling phase of the craniofacial structures and the lack of information about the material properties.
Evangelinos, Minoas; Anagnostopoulos, Gerasimos; Karvela-Kalogeraki, Iliana; Stathopoulou, Panagiota M; Scazzocchio, Claudio; Diallinas, George
2015-08-01
Transposons constitute powerful genetic tools for gene inactivation, exon or promoter trapping and genome analyses. The Minos element from Drosophila hydei, a Tc1/mariner-like transposon, has proved as a very efficient tool for heterologous transposition in several metazoa. In filamentous fungi, only a handful of fungal-specific transposable elements have been exploited as genetic tools, with the impala Tc1/mariner element from Fusarium oxysporum being the most successful. Here, we developed a two-component transposition system to manipulate Minos transposition in Aspergillus nidulans (AnMinos). Our system allows direct selection of transposition events based on re-activation of niaD, a gene necessary for growth on nitrate as a nitrogen source. On average, among 10(8) conidiospores, we obtain up to ∼0.8×10(2) transposition events leading to the expected revertant phenotype (niaD(+)), while ∼16% of excision events lead to AnMinos loss. Characterized excision footprints consisted of the four terminal bases of the transposon flanked by the TA target duplication and led to no major DNA rearrangements. AnMinos transposition depends on the presence of its homologous transposase. Its frequency was not significantly affected by temperature, UV irradiation or the transcription status of the original integration locus (niaD). Importantly, transposition is dependent on nkuA, encoding an enzyme essential for non-homologous end joining of DNA in double-strand break repair. AnMinos proved to be an efficient tool for functional analysis as it seems to transpose in different genomic loci positions in all chromosomes, including a high proportion of integration events within or close to genes. We have used Minos to obtain morphological and toxic analogue resistant mutants. Interestingly, among morphological mutants some seem to be due to Minos-elicited over-expression of specific genes, rather than gene inactivation. Copyright © 2015 Elsevier Inc. All rights reserved.
On Dynamics of Spinning Structures
NASA Technical Reports Server (NTRS)
Gupta, K. K.; Ibrahim, A.
2012-01-01
This paper provides details of developments pertaining to vibration analysis of gyroscopic systems, that involves a finite element structural discretization followed by the solution of the resulting matrix eigenvalue problem by a progressive, accelerated simultaneous iteration technique. Thus Coriolis, centrifugal and geometrical stiffness matrices are derived for shell and line elements, followed by the eigensolution details as well as solution of representative problems that demonstrates the efficacy of the currently developed numerical procedures and tools.
Lerat, Emmanuelle; Fablet, Marie; Modolo, Laurent; Lopez-Maestre, Hélène
2017-01-01
Abstract Over recent decades, substantial efforts have been made to understand the interactions between host genomes and transposable elements (TEs). The impact of TEs on the regulation of host genes is well known, with TEs acting as platforms of regulatory sequences. Nevertheless, due to their repetitive nature it is considerably hard to integrate TE analysis into genome-wide studies. Here, we developed a specific tool for the analysis of TE expression: TEtools. This tool takes into account the TE sequence diversity of the genome, it can be applied to unannotated or unassembled genomes and is freely available under the GPL3 (https://github.com/l-modolo/TEtools). TEtools performs the mapping of RNA-seq data obtained from classical mRNAs or small RNAs onto a list of TE sequences and performs differential expression analyses with statistical relevance. Using this tool, we analyzed TE expression from five Drosophila wild-type strains. Our data show for the first time that the activity of TEs is strictly linked to the activity of the genes implicated in the piwi-interacting RNA biogenesis and therefore fits an arms race scenario between TE sequences and host control genes. PMID:28204592
NASA Astrophysics Data System (ADS)
Godinho, R. M.; Cabrita, M. T.; Alves, L. C.; Pinheiro, T.
2015-04-01
Studies of the elemental composition of whole marine diatoms cells have high interest as they constitute a direct measurement of environmental changes, and allow anticipating consequences of anthropogenic alterations to organisms, ecosystems and global marine geochemical cycles. Nuclear microscopy is a powerful tool allowing direct measurement of whole cells giving qualitative imaging of distribution, and quantitative determination of intracellular concentration. Major obstacles to the analysis of marine microalgae are high medium salinity and the recurrent presence of extracellular exudates produced by algae to maintain colonies in natural media and in vitro. The objective of this paper was to optimize the methodology of sample preparation of marine unicellular algae for elemental analysis with nuclear microscopy, allowing further studies on cellular response to metals. Primary cultures of Coscinodiscus wailesii maintained in vitro were used to optimize protocols for elemental analysis with nuclear microscopy techniques. Adequate cell preparation procedures to isolate the cells from media components and exudates were established. The use of chemical agents proved to be inappropriate for elemental determination and for intracellular morphological analysis. The assessment of morphology and elemental partitioning in cell compartments obtained with nuclear microscopy techniques enabled to infer their function in natural environment and imbalances in exposure condition. Exposure to metal affected C. wailesii morphology and internal elemental distribution.
NASA Technical Reports Server (NTRS)
Lebiedzik, Catherine
1995-01-01
Development of design tools to furnish optimal acoustic environments for lightweight aircraft demands the ability to simulate the acoustic system on a workstation. In order to form an effective mathematical model of the phenomena at hand, we have begun by studying the propagation of acoustic waves inside closed spherical shells. Using a fully-coupled fluid-structure interaction model based upon variational principles, we have written a finite element analysis program and are in the process of examining several test cases. Future investigations are planned to increase model accuracy by incorporating non-linear and viscous effects.
RSAT 2018: regulatory sequence analysis tools 20th anniversary.
Nguyen, Nga Thi Thuy; Contreras-Moreira, Bruno; Castro-Mondragon, Jaime A; Santana-Garcia, Walter; Ossio, Raul; Robles-Espinoza, Carla Daniela; Bahin, Mathieu; Collombet, Samuel; Vincens, Pierre; Thieffry, Denis; van Helden, Jacques; Medina-Rivera, Alejandra; Thomas-Chollier, Morgane
2018-05-02
RSAT (Regulatory Sequence Analysis Tools) is a suite of modular tools for the detection and the analysis of cis-regulatory elements in genome sequences. Its main applications are (i) motif discovery, including from genome-wide datasets like ChIP-seq/ATAC-seq, (ii) motif scanning, (iii) motif analysis (quality assessment, comparisons and clustering), (iv) analysis of regulatory variations, (v) comparative genomics. Six public servers jointly support 10 000 genomes from all kingdoms. Six novel or refactored programs have been added since the 2015 NAR Web Software Issue, including updated programs to analyse regulatory variants (retrieve-variation-seq, variation-scan, convert-variations), along with tools to extract sequences from a list of coordinates (retrieve-seq-bed), to select motifs from motif collections (retrieve-matrix), and to extract orthologs based on Ensembl Compara (get-orthologs-compara). Three use cases illustrate the integration of new and refactored tools to the suite. This Anniversary update gives a 20-year perspective on the software suite. RSAT is well-documented and available through Web sites, SOAP/WSDL (Simple Object Access Protocol/Web Services Description Language) web services, virtual machines and stand-alone programs at http://www.rsat.eu/.
DEVELOPMENT OF DIAGNOSTIC ANALYTICAL AND MECHANICAL ABILITY TESTS THROUGH FACET DESIGN AND ANALYSIS.
ERIC Educational Resources Information Center
GUTTMAN, LOUIS,; SCHLESINGER, I.M.
METHODOLOGY BASED ON FACET THEORY (MODIFIED SET THEORY) WAS USED IN TEST CONSTRUCTION AND ANALYSIS TO PROVIDE AN EFFICIENT TOOL OF EVALUATION FOR VOCATIONAL GUIDANCE AND VOCATIONAL SCHOOL USE. THE TYPE OF TEST DEVELOPMENT UNDERTAKEN WAS LIMITED TO THE USE OF NONVERBAL PICTORIAL ITEMS. ITEMS FOR TESTING ABILITY TO IDENTIFY ELEMENTS BELONGING TO AN…
A Virtual World of Visualization
NASA Technical Reports Server (NTRS)
1998-01-01
In 1990, Sterling Software, Inc., developed the Flow Analysis Software Toolkit (FAST) for NASA Ames on contract. FAST is a workstation based modular analysis and visualization tool. It is used to visualize and animate grids and grid oriented data, typically generated by finite difference, finite element and other analytical methods. FAST is now available through COSMIC, NASA's software storehouse.
Solid rocket booster internal flow analysis by highly accurate adaptive computational methods
NASA Technical Reports Server (NTRS)
Huang, C. Y.; Tworzydlo, W.; Oden, J. T.; Bass, J. M.; Cullen, C.; Vadaketh, S.
1991-01-01
The primary objective of this project was to develop an adaptive finite element flow solver for simulating internal flows in the solid rocket booster. Described here is a unique flow simulator code for analyzing highly complex flow phenomena in the solid rocket booster. New methodologies and features incorporated into this analysis tool are described.
Cost-effective use of minicomputers to solve structural problems
NASA Technical Reports Server (NTRS)
Storaasli, O. O.; Foster, E. P.
1978-01-01
Minicomputers are receiving increased use throughout the aerospace industry. Until recently, their use focused primarily on process control and numerically controlled tooling applications, while their exposure to and the opportunity for structural calculations has been limited. With the increased availability of this computer hardware, the question arises as to the feasibility and practicality of carrying out comprehensive structural analysis on a minicomputer. This paper presents results on the potential for using minicomputers for structural analysis by (1) selecting a comprehensive, finite-element structural analysis system in use on large mainframe computers; (2) implementing the system on a minicomputer; and (3) comparing the performance of the minicomputers with that of a large mainframe computer for the solution to a wide range of finite element structural analysis problems.
NASA Technical Reports Server (NTRS)
Hou, Gene
1998-01-01
Sensitivity analysis is a technique for determining derivatives of system responses with respect to design parameters. Among many methods available for sensitivity analysis, automatic differentiation has been proven through many applications in fluid dynamics and structural mechanics to be an accurate and easy method for obtaining derivatives. Nevertheless, the method can be computational expensive and can require a high memory space. This project will apply an automatic differentiation tool, ADIFOR, to a p-version finite element code to obtain first- and second- order then-nal derivatives, respectively. The focus of the study is on the implementation process and the performance of the ADIFOR-enhanced codes for sensitivity analysis in terms of memory requirement, computational efficiency, and accuracy.
NASA Astrophysics Data System (ADS)
Nakashima, Hiroshi; Takatsu, Yuzuru
The goal of this study is to develop a practical and fast simulation tool for soil-tire interaction analysis, where finite element method (FEM) and discrete element method (DEM) are coupled together, and which can be realized on a desktop PC. We have extended our formerly proposed dynamic FE-DE method (FE-DEM) to include practical soil-tire system interaction, where not only the vertical sinkage of a tire, but also the travel of a driven tire was considered. Numerical simulation by FE-DEM is stable, and the relationships between variables, such as load-sinkage and sinkage-travel distance, and the gross tractive effort and running resistance characteristics, are obtained. Moreover, the simulation result is accurate enough to predict the maximum drawbar pull for a given tire, once the appropriate parameter values are provided. Therefore, the developed FE-DEM program can be applied with sufficient accuracy to interaction problems in soil-tire systems.
QuickMap: a public tool for large-scale gene therapy vector insertion site mapping and analysis.
Appelt, J-U; Giordano, F A; Ecker, M; Roeder, I; Grund, N; Hotz-Wagenblatt, A; Opelz, G; Zeller, W J; Allgayer, H; Fruehauf, S; Laufs, S
2009-07-01
Several events of insertional mutagenesis in pre-clinical and clinical gene therapy studies have created intense interest in assessing the genomic insertion profiles of gene therapy vectors. For the construction of such profiles, vector-flanking sequences detected by inverse PCR, linear amplification-mediated-PCR or ligation-mediated-PCR need to be mapped to the host cell's genome and compared to a reference set. Although remarkable progress has been achieved in mapping gene therapy vector insertion sites, public reference sets are lacking, as are the possibilities to quickly detect non-random patterns in experimental data. We developed a tool termed QuickMap, which uniformly maps and analyzes human and murine vector-flanking sequences within seconds (available at www.gtsg.org). Besides information about hits in chromosomes and fragile sites, QuickMap automatically determines insertion frequencies in +/- 250 kb adjacency to genes, cancer genes, pseudogenes, transcription factor and (post-transcriptional) miRNA binding sites, CpG islands and repetitive elements (short interspersed nuclear elements (SINE), long interspersed nuclear elements (LINE), Type II elements and LTR elements). Additionally, all experimental frequencies are compared with the data obtained from a reference set, containing 1 000 000 random integrations ('random set'). Thus, for the first time a tool allowing high-throughput profiling of gene therapy vector insertion sites is available. It provides a basis for large-scale insertion site analyses, which is now urgently needed to discover novel gene therapy vectors with 'safe' insertion profiles.
Improved accuracy for finite element structural analysis via an integrated force method
NASA Technical Reports Server (NTRS)
Patnaik, S. N.; Hopkins, D. A.; Aiello, R. A.; Berke, L.
1992-01-01
A comparative study was carried out to determine the accuracy of finite element analyses based on the stiffness method, a mixed method, and the new integrated force and dual integrated force methods. The numerical results were obtained with the following software: MSC/NASTRAN and ASKA for the stiffness method; an MHOST implementation method for the mixed method; and GIFT for the integrated force methods. The results indicate that on an overall basis, the stiffness and mixed methods present some limitations. The stiffness method generally requires a large number of elements in the model to achieve acceptable accuracy. The MHOST method tends to achieve a higher degree of accuracy for course models than does the stiffness method implemented by MSC/NASTRAN and ASKA. The two integrated force methods, which bestow simultaneous emphasis on stress equilibrium and strain compatibility, yield accurate solutions with fewer elements in a model. The full potential of these new integrated force methods remains largely unexploited, and they hold the promise of spawning new finite element structural analysis tools.
RSAT 2015: Regulatory Sequence Analysis Tools
Medina-Rivera, Alejandra; Defrance, Matthieu; Sand, Olivier; Herrmann, Carl; Castro-Mondragon, Jaime A.; Delerce, Jeremy; Jaeger, Sébastien; Blanchet, Christophe; Vincens, Pierre; Caron, Christophe; Staines, Daniel M.; Contreras-Moreira, Bruno; Artufel, Marie; Charbonnier-Khamvongsa, Lucie; Hernandez, Céline; Thieffry, Denis; Thomas-Chollier, Morgane; van Helden, Jacques
2015-01-01
RSAT (Regulatory Sequence Analysis Tools) is a modular software suite for the analysis of cis-regulatory elements in genome sequences. Its main applications are (i) motif discovery, appropriate to genome-wide data sets like ChIP-seq, (ii) transcription factor binding motif analysis (quality assessment, comparisons and clustering), (iii) comparative genomics and (iv) analysis of regulatory variations. Nine new programs have been added to the 43 described in the 2011 NAR Web Software Issue, including a tool to extract sequences from a list of coordinates (fetch-sequences from UCSC), novel programs dedicated to the analysis of regulatory variants from GWAS or population genomics (retrieve-variation-seq and variation-scan), a program to cluster motifs and visualize the similarities as trees (matrix-clustering). To deal with the drastic increase of sequenced genomes, RSAT public sites have been reorganized into taxon-specific servers. The suite is well-documented with tutorials and published protocols. The software suite is available through Web sites, SOAP/WSDL Web services, virtual machines and stand-alone programs at http://www.rsat.eu/. PMID:25904632
Finite Element Modelling and Analysis of Conventional Pultrusion Processes
NASA Astrophysics Data System (ADS)
Akishin, P.; Barkanov, E.; Bondarchuk, A.
2015-11-01
Pultrusion is one of many composite manufacturing techniques and one of the most efficient methods for producing fiber reinforced polymer composite parts with a constant cross-section. Numerical simulation is helpful for understanding the manufacturing process and developing scientific means for the pultrusion tooling design. Numerical technique based on the finite element method has been developed for the simulation of pultrusion processes. It uses the general purpose finite element software ANSYS Mechanical. It is shown that the developed technique predicts the temperature and cure profiles, which are in good agreement with those published in the open literature.
Rocket/launcher structural dynamics
NASA Technical Reports Server (NTRS)
Ferragut, N. J.
1976-01-01
The equations of motion describing the interactions between a rocket and a launcher were derived using Lagrange's Equation. A rocket launching was simulated. The motions of both the rocket and the launcher can be considered in detail. The model contains flexible elements and rigid elements. The rigid elements (masses) were judiciously utilized to simplify the derivation of the equations. The advantages of simultaneous shoe release were illustrated. Also, the loading history of the interstage structure of a boosted configuration was determined. The equations shown in this analysis could be used as a design tool during the modification of old launchers and the design of new launchers.
ERIC Educational Resources Information Center
de Hosson, Cecile; Kaminski, Wanda
2007-01-01
This paper describes the development, use, and analysis, of an educational tool inspired by the history of the optical mechanism of vision. We investigated 12-year-old students' reasoning about vision. Most of them explain it as the result of something coming either from the object or from the eye. Moreover, some of them think that light…
The Design and Analysis of Electrically Large Custom-Shaped Reflector Antennas
2013-06-01
GEO) satellite data are imported into STK and plotted to visualize the regions of the sky that the spherical reflector must have line of sight for...Magnetic Conductor PO Physical Optics STK Systems Tool Kit TE Transverse Electric xvii Acronym Definition TLE Two Line Element TM Transverse Magnetic...study for the spherical reflector, Systems Tool Kit ( STK ) software from Analytical Graphics Inc. (AGI) is used. In completing the cross-shaped
Finite element analysis on a medical implant.
Semenescu, Augustin; Radu-Ioniță, Florentina; Mateș, Ileana Mariana; Bădică, Petre; Batalu, Nicolae Dan; Negoita, Olivia Doina; Purcarea, Victor Lorin
2016-01-01
Several studies have shown a tight connection between several ocular pathologies and an increased risk of hip fractures due to falling, especially among elderly patients. The total replacement of the hip joint is a major surgical intervention that aims to restore the function of the affected hip by various factors, such as arthritis, injures, and others. A corkscrew-like femoral stem was designed in order to preserve the bone stock and to prevent the occurrence of iatrogenic fractures during the hammering of the implant. In this paper, the finite element analysis for the proposed design was applied, considering different loads and three types of materials. A finite element analysis is a powerful tool to simulate, optimize, design, and select suitable materials for new medical implants. The results showed that the best scenario was for Ti6Al4V alloy, although Ti and 316L stainless steel had a reasonable high safety factor.
Lashkari, A; Khalafi, H; Kazeminejad, H
2013-05-01
In this work, kinetic parameters of Tehran research reactor (TRR) mixed cores have been calculated. The mixed core configurations are made by replacement of the low enriched uranium control fuel elements with highly enriched uranium control fuel elements in the reference core. The MTR_PC package, a nuclear reactor analysis tool, is used to perform the analysis. Simulations were carried out to compute effective delayed neutron fraction and prompt neutron lifetime. Calculation of kinetic parameters is necessary for reactivity and power excursion transient analysis. The results of this research show that effective delayed neutron fraction decreases and prompt neutron lifetime increases with the fuels burn-up. Also, by increasing the number of highly enriched uranium control fuel elements in the reference core, the prompt neutron lifetime increases, but effective delayed neutron fraction does not show any considerable change.
Effective delayed neutron fraction and prompt neutron lifetime of Tehran research reactor mixed-core
Lashkari, A.; Khalafi, H.; Kazeminejad, H.
2013-01-01
In this work, kinetic parameters of Tehran research reactor (TRR) mixed cores have been calculated. The mixed core configurations are made by replacement of the low enriched uranium control fuel elements with highly enriched uranium control fuel elements in the reference core. The MTR_PC package, a nuclear reactor analysis tool, is used to perform the analysis. Simulations were carried out to compute effective delayed neutron fraction and prompt neutron lifetime. Calculation of kinetic parameters is necessary for reactivity and power excursion transient analysis. The results of this research show that effective delayed neutron fraction decreases and prompt neutron lifetime increases with the fuels burn-up. Also, by increasing the number of highly enriched uranium control fuel elements in the reference core, the prompt neutron lifetime increases, but effective delayed neutron fraction does not show any considerable change. PMID:24976672
Computational mechanics analysis tools for parallel-vector supercomputers
NASA Technical Reports Server (NTRS)
Storaasli, O. O.; Nguyen, D. T.; Baddourah, M. A.; Qin, J.
1993-01-01
Computational algorithms for structural analysis on parallel-vector supercomputers are reviewed. These parallel algorithms, developed by the authors, are for the assembly of structural equations, 'out-of-core' strategies for linear equation solution, massively distributed-memory equation solution, unsymmetric equation solution, general eigen-solution, geometrically nonlinear finite element analysis, design sensitivity analysis for structural dynamics, optimization algorithm and domain decomposition. The source code for many of these algorithms is available from NASA Langley.
NASA Technical Reports Server (NTRS)
Nguyen, Duc T.; Storaasli, Olaf O.; Qin, Jiangning; Qamar, Ramzi
1994-01-01
An automatic differentiation tool (ADIFOR) is incorporated into a finite element based structural analysis program for shape and non-shape design sensitivity analysis of structural systems. The entire analysis and sensitivity procedures are parallelized and vectorized for high performance computation. Small scale examples to verify the accuracy of the proposed program and a medium scale example to demonstrate the parallel vector performance on multiple CRAY C90 processors are included.
Practical applications of surface analytic tools in tribology
NASA Technical Reports Server (NTRS)
Ferrante, J.
1980-01-01
A brief description of many of the widely used tools is presented. Of this list, those which have the highest applicability for giving elemental and/or compound analysis for problems of interest in tribology along with being truly surface sensitive (that is less than 10 atomic layers) are presented. The latter group is critiqued in detail in terms of strengths and weaknesses. Emphasis is placed on post facto analysis of experiments performed under real conditions (e.g., in air with lubricants). It is further indicated that such equipment could be used for screening and quality control.
NASA Technical Reports Server (NTRS)
Xu, Xidong; Ulrey, Mike L.; Brown, John A.; Mast, James; Lapis, Mary B.
2013-01-01
NextGen is a complex socio-technical system and, in many ways, it is expected to be more complex than the current system. It is vital to assess the safety impact of the NextGen elements (technologies, systems, and procedures) in a rigorous and systematic way and to ensure that they do not compromise safety. In this study, the NextGen elements in the form of Operational Improvements (OIs), Enablers, Research Activities, Development Activities, and Policy Issues were identified. The overall hazard situation in NextGen was outlined; a high-level hazard analysis was conducted with respect to multiple elements in a representative NextGen OI known as OI-0349 (Automation Support for Separation Management); and the hazards resulting from the highly dynamic complexity involved in an OI-0349 scenario were illustrated. A selected but representative set of the existing safety methods, tools, processes, and regulations was then reviewed and analyzed regarding whether they are sufficient to assess safety in the elements of that OI and ensure that safety will not be compromised and whether they might incur intolerably high costs.
Ellingham, Sarah T D; Thompson, Tim J U; Islam, Meez
2018-03-01
This study investigates the use of Scanning electron microscopy-energy-dispersive X-ray (SEM-EDX) as a diagnostic tool for the determination of the osseous origin of samples subjected to different temperatures. Sheep (Ovis aries) ribs of two experimental groups (fleshed and defleshed) were burned at temperatures of between 100°C and 1100°C in 100°C increments and subsequently analyzed with the SEM-EDX to determine the atomic percentage of present elements. Three-factor ANOVA analysis showed that neither the exposure temperature, nor whether the burning occurred with or without soft tissue present had any significant influence on the bone's overall elemental makeup (p > 0.05). The Ca/P ratio remained in the osseous typical range of between 1.6 and 2.58 in all analyzed samples. This demonstrates that even faced with high temperatures, the overall gross elemental content and atomic percentage of elements in bone remain stable, creating a unique "fingerprint" for osseous material, even after exposure to extreme conditions. © 2017 American Academy of Forensic Sciences.
Efficient simulation of press hardening process through integrated structural and CFD analyses
DOE Office of Scientific and Technical Information (OSTI.GOV)
Palaniswamy, Hariharasudhan; Mondalek, Pamela; Wronski, Maciek
Press hardened steel parts are being increasingly used in automotive structures for their higher strength to meet safety standards while reducing vehicle weight to improve fuel consumption. However, manufacturing of sheet metal parts by press hardening process to achieve desired properties is extremely challenging as it involves complex interaction of plastic deformation, metallurgical change, thermal distribution, and fluid flow. Numerical simulation is critical for successful design of the process and to understand the interaction among the numerous process parameters to control the press hardening process in order to consistently achieve desired part properties. Until now there has been no integratedmore » commercial software solution that can efficiently model the complete process from forming of the blank, heat transfer between the blank and tool, microstructure evolution in the blank, heat loss from tool to the fluid that flows through water channels in the tools. In this study, a numerical solution based on Altair HyperWorks® product suite involving RADIOSS®, a non-linear finite element based structural analysis solver and AcuSolve®, an incompressible fluid flow solver based on Galerkin Least Square Finite Element Method have been utilized to develop an efficient solution for complete press hardening process design and analysis. RADIOSS is used to handle the plastic deformation, heat transfer between the blank and tool, and microstructure evolution in the blank during cooling. While AcuSolve is used to efficiently model heat loss from tool to the fluid that flows through water channels in the tools. The approach is demonstrated through some case studies.« less
NASA Astrophysics Data System (ADS)
Haddag, B.; Kagnaya, T.; Nouari, M.; Cutard, T.
2013-01-01
Modelling machining operations allows estimating cutting parameters which are difficult to obtain experimentally and in particular, include quantities characterizing the tool-workpiece interface. Temperature is one of these quantities which has an impact on the tool wear, thus its estimation is important. This study deals with a new modelling strategy, based on two steps of calculation, for analysis of the heat transfer into the cutting tool. Unlike the classical methods, considering only the cutting tool with application of an approximate heat flux at the cutting face, estimated from experimental data (e.g. measured cutting force, cutting power), the proposed approach consists of two successive 3D Finite Element calculations and fully independent on the experimental measurements; only the definition of the behaviour of the tool-workpiece couple is necessary. The first one is a 3D thermomechanical modelling of the chip formation process, which allows estimating cutting forces, chip morphology and its flow direction. The second calculation is a 3D thermal modelling of the heat diffusion into the cutting tool, by using an adequate thermal loading (applied uniform or non-uniform heat flux). This loading is estimated using some quantities obtained from the first step calculation, such as contact pressure, sliding velocity distributions and contact area. Comparisons in one hand between experimental data and the first calculation and at the other hand between measured temperatures with embedded thermocouples and the second calculation show a good agreement in terms of cutting forces, chip morphology and cutting temperature.
The novel approach to the biomonitor survey using one- and two-dimensional Kohonen networks.
Deljanin, Isidora; Antanasijević, Davor; Urošević, Mira Aničić; Tomašević, Milica; Perić-Grujić, Aleksandra; Ristić, Mirjana
2015-10-01
To compare the applicability of the leaves of horse chestnut (Aesculus hippocastanum) and linden (Tilia spp.) as biomonitors of trace element concentrations, a coupled approach of one- and two-dimensional Kohonen networks was applied for the first time. The self-organizing networks (SONs) and the self-organizing maps (SOMs) were applied on the database obtained for the element accumulation (Cr, Fe, Ni, Cu, Zn, Pb, V, As, Cd) and the SOM for the Pb isotopes in the leaves for a multiyear period (2002-2006). A. hippocastanum seems to be a more appropriate biomonitor since it showed more consistent results in the analysis of trace elements and Pb isotopes. The SOM proved to be a suitable and sensitive tool for assessing differences in trace element concentrations and for the Pb isotopic composition in leaves of different species. In addition, the SON provided more clear data on seasonal and temporal accumulation of trace elements in the leaves and could be recommended complementary to the SOM analysis of trace elements in biomonitoring studies.
DOT National Transportation Integrated Search
2009-09-01
More and more, transportation system operators are seeing the benefits of strengthening links between planning and operations. A critical element in improving transportation decision-making and the effectiveness of transportation systems related to o...
NASA Astrophysics Data System (ADS)
Byeon, J. H.; Ahmed, F.; Ko, T. J.; lee, D. K.; Kim, J. S.
2018-03-01
As the industry develops, miniaturization and refinement of products are important issues. Precise machining is required for cutting, which is a typical method of machining a product. The factor determining the workability of the cutting process is the material of the tool. Tool materials include carbon tool steel, alloy tool steel, high-speed steel, cemented carbide, and ceramics. In the case of a carbide material, the smaller the particle size, the better the mechanical properties with higher hardness, strength and toughness. The specific heat, density, and thermal diffusivity are also changed through finer particle size of the material. In this study, finite element analysis was performed to investigate the change of heat generation and cutting power depending on the physical properties (specific heat, density, thermal diffusivity) of tool material. The thermal conductivity coefficient was obtained by measuring the thermal diffusivity, specific heat, and density of the material (180 nm) in which the particle size was finer and the particle material (0.05 μm) in the conventional size. The coefficient of thermal conductivity was calculated as 61.33 for 180nm class material and 46.13 for 0.05μm class material. As a result of finite element analysis using this value, the average temperature of exothermic heat of micronized particle material (180nm) was 532.75 °C and the temperature of existing material (0.05μm) was 572.75 °C. Cutting power was also compared but not significant. Therefore, if the thermal conductivity is increased through particle refinement, the surface power can be improved and the tool life can be prolonged by lowering the temperature generated in the tool during machining without giving a great influence to the cutting power.
Tool Releases Optical Elements From Spring Brackets
NASA Technical Reports Server (NTRS)
Gum, J. S.
1984-01-01
Threaded hooks retract bracket arms holding element. Tool uses three hooks with threaded shanks mounted in ring-shaped holder to pull on tabs to release optical element. One person can easily insert or remove optical element (such as prism or lens) from spring holder or bracket with minimal risk of damage.
Elsaadany, Mostafa; Yan, Karen Chang; Yildirim-Ayan, Eda
2017-06-01
Successful tissue engineering and regenerative therapy necessitate having extensive knowledge about mechanical milieu in engineered tissues and the resident cells. In this study, we have merged two powerful analysis tools, namely finite element analysis and stochastic analysis, to understand the mechanical strain within the tissue scaffold and residing cells and to predict the cell viability upon applying mechanical strains. A continuum-based multi-length scale finite element model (FEM) was created to simulate the physiologically relevant equiaxial strain exposure on cell-embedded tissue scaffold and to calculate strain transferred to the tissue scaffold (macro-scale) and residing cells (micro-scale) upon various equiaxial strains. The data from FEM were used to predict cell viability under various equiaxial strain magnitudes using stochastic damage criterion analysis. The model validation was conducted through mechanically straining the cardiomyocyte-encapsulated collagen constructs using a custom-built mechanical loading platform (EQUicycler). FEM quantified the strain gradients over the radial and longitudinal direction of the scaffolds and the cells residing in different areas of interest. With the use of the experimental viability data, stochastic damage criterion, and the average cellular strains obtained from multi-length scale models, cellular viability was predicted and successfully validated. This methodology can provide a great tool to characterize the mechanical stimulation of bioreactors used in tissue engineering applications in providing quantification of mechanical strain and predicting cellular viability variations due to applied mechanical strain.
Sustainability of transport structures - some aspects of the nonlinear reliability assessment
NASA Astrophysics Data System (ADS)
Pukl, Radomír; Sajdlová, Tereza; Strauss, Alfred; Lehký, David; Novák, Drahomír
2017-09-01
Efficient techniques for both nonlinear numerical analysis of concrete structures and advanced stochastic simulation methods have been combined in order to offer an advanced tool for assessment of realistic behaviour, failure and safety assessment of transport structures. The utilized approach is based on randomization of the non-linear finite element analysis of the structural models. Degradation aspects such as carbonation of concrete can be accounted in order predict durability of the investigated structure and its sustainability. Results can serve as a rational basis for the performance and sustainability assessment based on advanced nonlinear computer analysis of the structures of transport infrastructure such as bridges or tunnels. In the stochastic simulation the input material parameters obtained from material tests including their randomness and uncertainty are represented as random variables or fields. Appropriate identification of material parameters is crucial for the virtual failure modelling of structures and structural elements. Inverse analysis using artificial neural networks and virtual stochastic simulations approach is applied to determine the fracture mechanical parameters of the structural material and its numerical model. Structural response, reliability and sustainability have been investigated on different types of transport structures made from various materials using the above mentioned methodology and tools.
User-Defined Material Model for Progressive Failure Analysis
NASA Technical Reports Server (NTRS)
Knight, Norman F. Jr.; Reeder, James R. (Technical Monitor)
2006-01-01
An overview of different types of composite material system architectures and a brief review of progressive failure material modeling methods used for structural analysis including failure initiation and material degradation are presented. Different failure initiation criteria and material degradation models are described that define progressive failure formulations. These progressive failure formulations are implemented in a user-defined material model (or UMAT) for use with the ABAQUS/Standard1 nonlinear finite element analysis tool. The failure initiation criteria include the maximum stress criteria, maximum strain criteria, the Tsai-Wu failure polynomial, and the Hashin criteria. The material degradation model is based on the ply-discounting approach where the local material constitutive coefficients are degraded. Applications and extensions of the progressive failure analysis material model address two-dimensional plate and shell finite elements and three-dimensional solid finite elements. Implementation details and use of the UMAT subroutine are described in the present paper. Parametric studies for composite structures are discussed to illustrate the features of the progressive failure modeling methods that have been implemented.
Improved Aerodynamic Analysis for Hybrid Wing Body Conceptual Design Optimization
NASA Technical Reports Server (NTRS)
Gern, Frank H.
2012-01-01
This paper provides an overview of ongoing efforts to develop, evaluate, and validate different tools for improved aerodynamic modeling and systems analysis of Hybrid Wing Body (HWB) aircraft configurations. Results are being presented for the evaluation of different aerodynamic tools including panel methods, enhanced panel methods with viscous drag prediction, and computational fluid dynamics. Emphasis is placed on proper prediction of aerodynamic loads for structural sizing as well as viscous drag prediction to develop drag polars for HWB conceptual design optimization. Data from transonic wind tunnel tests at the Arnold Engineering Development Center s 16-Foot Transonic Tunnel was used as a reference data set in order to evaluate the accuracy of the aerodynamic tools. Triangularized surface data and Vehicle Sketch Pad (VSP) models of an X-48B 2% scale wind tunnel model were used to generate input and model files for the different analysis tools. In support of ongoing HWB scaling studies within the NASA Environmentally Responsible Aviation (ERA) program, an improved finite element based structural analysis and weight estimation tool for HWB center bodies is currently under development. Aerodynamic results from these analyses are used to provide additional aerodynamic validation data.
Finite Element Simulation of Machining of Ti6Al4V Alloy
NASA Astrophysics Data System (ADS)
Rizzuti, S.; Umbrello, D.
2011-05-01
Titanium and its alloys are an important class of materials, especially for aerospace applications, due to their excellent combination of strength and fracture toughness as well as low density. However, these materials are generally regarded as difficult to machine because of their low thermal conductivity and high chemical reactivity with cutting tool materials. Moreover, the low thermal conductivity of Titanium inhibits dissipation of heat within the workpiece causing an higher temperature at the cutting edge and generating for higher cutting speed a rapid chipping at the cutting edge which leads to catastrophic failure. In addition, chip morphology significantly influences the thermo-mechanical behaviour at the workpiece/tool interface, which also affects the tool life. In this paper a finite element analysis of machining of TiAl6V4 is presented. In particular, cutting force, chip morphology and segmentation are taken into account due to their predominant roles to determine machinability and tool wear during the machining of these alloys. Results in terms of residual stresses are also presented. Moreover, the numerical results are compared with experimental ones.
ElemeNT: a computational tool for detecting core promoter elements.
Sloutskin, Anna; Danino, Yehuda M; Orenstein, Yaron; Zehavi, Yonathan; Doniger, Tirza; Shamir, Ron; Juven-Gershon, Tamar
2015-01-01
Core promoter elements play a pivotal role in the transcriptional output, yet they are often detected manually within sequences of interest. Here, we present 2 contributions to the detection and curation of core promoter elements within given sequences. First, the Elements Navigation Tool (ElemeNT) is a user-friendly web-based, interactive tool for prediction and display of putative core promoter elements and their biologically-relevant combinations. Second, the CORE database summarizes ElemeNT-predicted core promoter elements near CAGE and RNA-seq-defined Drosophila melanogaster transcription start sites (TSSs). ElemeNT's predictions are based on biologically-functional core promoter elements, and can be used to infer core promoter compositions. ElemeNT does not assume prior knowledge of the actual TSS position, and can therefore assist in annotation of any given sequence. These resources, freely accessible at http://lifefaculty.biu.ac.il/gershon-tamar/index.php/resources, facilitate the identification of core promoter elements as active contributors to gene expression.
NCC: A Multidisciplinary Design/Analysis Tool for Combustion Systems
NASA Technical Reports Server (NTRS)
Liu, Nan-Suey; Quealy, Angela
1999-01-01
A multi-disciplinary design/analysis tool for combustion systems is critical for optimizing the low-emission, high-performance combustor design process. Based on discussions between NASA Lewis Research Center and the jet engine companies, an industry-government team was formed in early 1995 to develop the National Combustion Code (NCC), which is an integrated system of computer codes for the design and analysis of combustion systems. NCC has advanced features that address the need to meet designer's requirements such as "assured accuracy", "fast turnaround", and "acceptable cost". The NCC development team is comprised of Allison Engine Company (Allison), CFD Research Corporation (CFDRC), GE Aircraft Engines (GEAE), NASA Lewis Research Center (LeRC), and Pratt & Whitney (P&W). This development team operates under the guidance of the NCC steering committee. The "unstructured mesh" capability and "parallel computing" are fundamental features of NCC from its inception. The NCC system is composed of a set of "elements" which includes grid generator, main flow solver, turbulence module, turbulence and chemistry interaction module, chemistry module, spray module, radiation heat transfer module, data visualization module, and a post-processor for evaluating engine performance parameters. Each element may have contributions from several team members. Such a multi-source multi-element system needs to be integrated in a way that facilitates inter-module data communication, flexibility in module selection, and ease of integration.
Nikkhoo, Mohammad; Hsu, Yu-Chun; Haghpanahi, Mohammad; Parnianpour, Mohamad; Wang, Jaw-Lin
2013-06-01
Finite element analysis is an effective tool to evaluate the material properties of living tissue. For an interactive optimization procedure, the finite element analysis usually needs many simulations to reach a reasonable solution. The meta-model analysis of finite element simulation can be used to reduce the computation of a structure with complex geometry or a material with composite constitutive equations. The intervertebral disc is a complex, heterogeneous, and hydrated porous structure. A poroelastic finite element model can be used to observe the fluid transferring, pressure deviation, and other properties within the disc. Defining reasonable poroelastic material properties of the anulus fibrosus and nucleus pulposus is critical for the quality of the simulation. We developed a material property updating protocol, which is basically a fitting algorithm consisted of finite element simulations and a quadratic response surface regression. This protocol was used to find the material properties, such as the hydraulic permeability, elastic modulus, and Poisson's ratio, of intact and degenerated porcine discs. The results showed that the in vitro disc experimental deformations were well fitted with limited finite element simulations and a quadratic response surface regression. The comparison of material properties of intact and degenerated discs showed that the hydraulic permeability significantly decreased but Poisson's ratio significantly increased for the degenerated discs. This study shows that the developed protocol is efficient and effective in defining material properties of a complex structure such as the intervertebral disc.
NASA Astrophysics Data System (ADS)
Xu, Guanshui
2000-12-01
A direct finite-element model is developed for the full-scale analysis of the electromechanical phenomena involved in surface acoustic wave (SAW) devices. The equations of wave propagation in piezoelectric materials are discretized using the Galerkin method, in which an implicit algorithm of the Newmark family with unconditional stability is implemented. The Rayleigh damping coefficients are included in the elements near the boundary to reduce the influence of the reflection of waves. The performance of the model is demonstrated by the analysis of the frequency response of a Y-Z lithium niobate filter with two uniform ports, with emphasis on the influence of the number of electrodes. The frequency response of the filter is obtained through the Fourier transform of the impulse response, which is solved directly from the finite-element simulation. It shows that the finite-element results are in good agreement with the characteristic frequency response of the filter predicted by the simple phase-matching argument. The ability of the method to evaluate the influence of the bulk waves at the high-frequency end of the filter passband and the influence of the number of electrodes on insertion loss is noteworthy. We conclude that the direct finite-element analysis of SAW devices can be used as an effective tool for the design of high-performance SAW devices. Some practical computational challenges of finite-element modeling of SAW devices are discussed.
DOT National Transportation Integrated Search
2009-09-01
More and more, transportation system operators are seeing the benefits of strengthening links between planning and operations. A critical element in improving transportation decision-making and the effectiveness of transportation systems related to o...
DOT National Transportation Integrated Search
2009-09-01
More and more, transportation system operators are seeing the benefi ts of strengthening links between : planning and operations. A critical element in improving transportation decision-making and the effectiveness : of transportation systems related...
NASA Astrophysics Data System (ADS)
Nilsson, E. J. C.; Pallon, J.; Przybylowicz, W. J.; Wang, Y. D.; Jönsson, K. I.
2014-08-01
Although heavy on labor and equipment, thus not often applied, cryoanalysis of frozen hydrated biological specimens can provide information that better reflects the living state of the organism, compared with analysis in the freeze-dried state. In this paper we report a study where the cryoanalysis facility with cryosectioning capabilities at Materials Research Department, iThemba LABS, South Africa was employed to evaluate the usefulness of combining three ion beam analytical methods (μPIXE, RBS and STIM) to analyze a biological target where a better elemental compositional description is needed - the tardigrade. Imaging as well as quantification results are of interest. In a previous study, the element composition and redistribution of elements in the desiccated and active states of two tardigrade species was investigated. This study included analysis of both whole and sectioned tardigrades, and the aim was to analyze each specimen twice; first frozen hydrated and later freeze-dried. The combination of the three analytical techniques proved useful: elements from C to Rb in the tardigrades could be determined and certain differences in distribution of elements between the frozen hydrated and the freeze-dried states were observed. RBS on frozen hydrated specimens provided knowledge of matrix elements.
Tracing Anthropogenic Pollution Through Dendrochemistry
NASA Astrophysics Data System (ADS)
Rocha, E.; Gunnarson, B. E.; Holzkaemper, S.
2017-12-01
The growing concern regarding pollution effects on the environment and human health demands new control strategies and monitoring tools. In this study we assess the potential of using dendrochemistry as a forensic tool to investigate chemical contamination patterns in the surroundings of a former glass factory in Southern Sweden. Tree-ring width chronologies were produced from exposed and non-exposed sites. Using energy disperse X-ray fluorescence (EDXRF) technique, tree cores of Scots Pine (Pinus sylvestris), Norway spruce (Picea Abies) and Populus tremula (European Aspen) were analysed for their elemental composition in accordance with previous soil analysis done in the area. Traces of barium and considerable alteration of the chlorine profiles were successfully detected confirming the potential of the method to record environmental releases. The dendrochemical analysis also highlighted the differences in the response of tree species to elements uptake (root sensitivity) and the importance of metals bioavailability. Finally, the adopted sampling strategy is of outmost importance to the success of the method.
A Multidisciplinary Tool for Systems Analysis of Planetary Entry, Descent, and Landing (SAPE)
NASA Technical Reports Server (NTRS)
Samareh, Jamshid A.
2009-01-01
SAPE is a Python-based multidisciplinary analysis tool for systems analysis of planetary entry, descent, and landing (EDL) for Venus, Earth, Mars, Jupiter, Saturn, Uranus, Neptune, and Titan. The purpose of SAPE is to provide a variable-fidelity capability for conceptual and preliminary analysis within the same framework. SAPE includes the following analysis modules: geometry, trajectory, aerodynamics, aerothermal, thermal protection system, and structural sizing. SAPE uses the Python language-a platform-independent open-source software for integration and for the user interface. The development has relied heavily on the object-oriented programming capabilities that are available in Python. Modules are provided to interface with commercial and government off-the-shelf software components (e.g., thermal protection systems and finite-element analysis). SAPE runs on Microsoft Windows and Apple Mac OS X and has been partially tested on Linux.
SAPNEW: Parallel finite element code for thin shell structures on the Alliant FX-80
NASA Astrophysics Data System (ADS)
Kamat, Manohar P.; Watson, Brian C.
1992-11-01
The finite element method has proven to be an invaluable tool for analysis and design of complex, high performance systems, such as bladed-disk assemblies in aircraft turbofan engines. However, as the problem size increase, the computation time required by conventional computers can be prohibitively high. Parallel processing computers provide the means to overcome these computation time limits. This report summarizes the results of a research activity aimed at providing a finite element capability for analyzing turbomachinery bladed-disk assemblies in a vector/parallel processing environment. A special purpose code, named with the acronym SAPNEW, has been developed to perform static and eigen analysis of multi-degree-of-freedom blade models built-up from flat thin shell elements. SAPNEW provides a stand alone capability for static and eigen analysis on the Alliant FX/80, a parallel processing computer. A preprocessor, named with the acronym NTOS, has been developed to accept NASTRAN input decks and convert them to the SAPNEW format to make SAPNEW more readily used by researchers at NASA Lewis Research Center.
DOE Office of Scientific and Technical Information (OSTI.GOV)
PANDOLFI, RONALD; KUMAR, DINESH; VENKATAKRISHNAN, SINGANALLUR
Xi-CAM aims to provide a community driven platform for multimodal analysis in synchrotron science. The platform core provides a robust plugin infrastructure for extensibility, allowing continuing development to simply add further functionality. Current modules include tools for characterization with (GI)SAXS, Tomography, and XAS. This will continue to serve as a development base as algorithms for multimodal analysis develop. Seamless remote data access, visualization and analysis are key elements of Xi-CAM, and will become critical to synchrotron data infrastructure as expectations for future data volume and acquisition rates rise with continuously increasing throughputs. The highly interactive design elements of Xi-cam willmore » similarly support a generation of users which depend on immediate data quality feedback during high-throughput or burst acquisition modes.« less
Automated array assembly task, phase 1
NASA Technical Reports Server (NTRS)
Carbajal, B. G.
1977-01-01
An assessment of state-of-the-art technologies that are applicable to silicon solar cell and solar cell module fabrication is provided. The assessment consists of a technical feasibility evaluation and a cost projection for high-volume production of silicon solar cell modules. The cost projection was approached from two directions; a design-to-cost analysis assigned cost goals to each major process element in the fabrication scheme, and a cost analysis built up projected costs for alternate technologies for each process element. A technical evaluation was used in combination with the cost analysis to identify a baseline low cost process. A novel approach to metal pattern design based on minimum power loss was developed. These design equations were used as a tool in the evaluation of metallization technologies.
The Mobile Element Locator Tool (MELT): population-scale mobile element discovery and biology
Gardner, Eugene J.; Lam, Vincent K.; Harris, Daniel N.; Chuang, Nelson T.; Scott, Emma C.; Pittard, W. Stephen; Mills, Ryan E.; Devine, Scott E.
2017-01-01
Mobile element insertions (MEIs) represent ∼25% of all structural variants in human genomes. Moreover, when they disrupt genes, MEIs can influence human traits and diseases. Therefore, MEIs should be fully discovered along with other forms of genetic variation in whole genome sequencing (WGS) projects involving population genetics, human diseases, and clinical genomics. Here, we describe the Mobile Element Locator Tool (MELT), which was developed as part of the 1000 Genomes Project to perform MEI discovery on a population scale. Using both Illumina WGS data and simulations, we demonstrate that MELT outperforms existing MEI discovery tools in terms of speed, scalability, specificity, and sensitivity, while also detecting a broader spectrum of MEI-associated features. Several run modes were developed to perform MEI discovery on local and cloud systems. In addition to using MELT to discover MEIs in modern humans as part of the 1000 Genomes Project, we also used it to discover MEIs in chimpanzees and ancient (Neanderthal and Denisovan) hominids. We detected diverse patterns of MEI stratification across these populations that likely were caused by (1) diverse rates of MEI production from source elements, (2) diverse patterns of MEI inheritance, and (3) the introgression of ancient MEIs into modern human genomes. Overall, our study provides the most comprehensive map of MEIs to date spanning chimpanzees, ancient hominids, and modern humans and reveals new aspects of MEI biology in these lineages. We also demonstrate that MELT is a robust platform for MEI discovery and analysis in a variety of experimental settings. PMID:28855259
Dental Videographic Analysis using Digital Age Media.
Agarwal, Anirudh; Seth, Karan; Parmar, Siddharaj; Jhawar, Rahul
2016-01-01
This study was to evaluate a new method of smile analysis using videographic and photographic softwares (as in this study Photoshop Elements X, Windows Movie Maker 2012) as primary assessment tools and to develop an index for malocclusion and treatment plan that could be used in assessing severity of maloc-clussion. Agarwal A, Seth K, Parmar S, Jhawar R. Dental Videographic Analysis using Digital Age Media. Int J Clin Pediatr Dent 2016;9(4):355-363.
Development of freight policy analysis tool for northeastern Illinois and the United States.
DOT National Transportation Integrated Search
2010-10-01
Freight transportation is a vital element in the economic prosperity of any country. According to the : nationwide commodity flow survey, over 12 billion tons of goods, valued at more than $11.6 trillion, were moved in : America in the year 2007(Bure...
The effects of dissecting tools on the trace element concentrations of fish and mussel tissues.
Heit, M; Klusek, C S
1982-06-01
A comparison of the effects of dissecting tools composed of various materials on the trace element content of the muscle of the marine bluefish, Pomatomus saltatrix, and the soft tissues of freshwater mussels, Eliptio complanatus and Lampsilus radiata, is presented. The fish were dissected with blades made of stainless steel, Lexan plastic, titanium, and Teflon-coated stainless steel. The mussels were dissected with stainless and Teflon tools only. Elements measured included As, Cd, Cr, Cu, Hg, Ni, Pb, Se, Sn, Te, V, and Zn. Significant concentration differences (P = 0.01) were not found for any element in fish or mussel samples dissected by the different tools.
Batson, Sarah; Score, Robert; Sutton, Alex J
2017-06-01
The aim of the study was to develop the three-dimensional (3D) evidence network plot system-a novel web-based interactive 3D tool to facilitate the visualization and exploration of covariate distributions and imbalances across evidence networks for network meta-analysis (NMA). We developed the 3D evidence network plot system within an AngularJS environment using a third party JavaScript library (Three.js) to create the 3D element of the application. Data used to enable the creation of the 3D element for a particular topic are inputted via a Microsoft Excel template spreadsheet that has been specifically formatted to hold these data. We display and discuss the findings of applying the tool to two NMA examples considering multiple covariates. These two examples have been previously identified as having potentially important covariate effects and allow us to document the various features of the tool while illustrating how it can be used. The 3D evidence network plot system provides an immediate, intuitive, and accessible way to assess the similarity and differences between the values of covariates for individual studies within and between each treatment contrast in an evidence network. In this way, differences between the studies, which may invalidate the usual assumptions of an NMA, can be identified for further scrutiny. Hence, the tool facilitates NMA feasibility/validity assessments and aids in the interpretation of NMA results. The 3D evidence network plot system is the first tool designed specifically to visualize covariate distributions and imbalances across evidence networks in 3D. This will be of primary interest to systematic review and meta-analysis researchers and, more generally, those assessing the validity and robustness of an NMA to inform reimbursement decisions. Copyright © 2017 Elsevier Inc. All rights reserved.
NASA Astrophysics Data System (ADS)
San-Blas, A. A.; Roca, J. M.; Cogollos, S.; Morro, J. V.; Boria, V. E.; Gimeno, B.
2016-06-01
In this work, a full-wave tool for the accurate analysis and design of compensated E-plane multiport junctions is proposed. The implemented tool is capable of evaluating the undesired effects related to the use of low-cost manufacturing techniques, which are mostly due to the introduction of rounded corners in the cross section of the rectangular waveguides of the device. The obtained results show that, although stringent mechanical effects are imposed, it is possible to compensate for the impact of the cited low-cost manufacturing techniques by redesigning the matching elements considered in the original device. Several new designs concerning a great variety of E-plane components (such as right-angled bends, T-junctions and magic-Ts) are presented, and useful design guidelines are provided. The implemented tool, which is mainly based on the boundary integral-resonant mode expansion technique, has been successfully validated by comparing the obtained results to simulated data provided by a commercial software based on the finite element method.
Design and development of progressive tool for manufacturing washer
NASA Astrophysics Data System (ADS)
Annigeri, Ulhas K.; Raghavendra Ravi Kiran, K.; Deepthi, Y. P.
2017-07-01
In a progressive tool the raw material is worked at different station to finally fabricate the component. A progressive tool is a lucrative tool for mass production of components. A lot of automobile and other transport industries develop progressive tool for the production of components. The design of tool involves lot of planning and the same amount of skill of process planning is required in the fabrication of the tool. The design also involves use of thumb rules and standard elements as per experience gained in practice. Manufacturing the press tool is a laborious task as special jigs and fixtures have to be designed for the purpose. Assembly of all the press tool elements is another task where use of accurate measuring instruments for alignment of various tool elements is important. In the present study, design and fabrication of progressive press tool for production of washer has been developed and the press tool has been tried out on a mechanical type of press. The components produced are to dimensions.
Improving finite element results in modeling heart valve mechanics.
Earl, Emily; Mohammadi, Hadi
2018-06-01
Finite element analysis is a well-established computational tool which can be used for the analysis of soft tissue mechanics. Due to the structural complexity of the leaflet tissue of the heart valve, the currently available finite element models do not adequately represent the leaflet tissue. A method of addressing this issue is to implement computationally expensive finite element models, characterized by precise constitutive models including high-order and high-density mesh techniques. In this study, we introduce a novel numerical technique that enhances the results obtained from coarse mesh finite element models to provide accuracy comparable to that of fine mesh finite element models while maintaining a relatively low computational cost. Introduced in this study is a method by which the computational expense required to solve linear and nonlinear constitutive models, commonly used in heart valve mechanics simulations, is reduced while continuing to account for large and infinitesimal deformations. This continuum model is developed based on the least square algorithm procedure coupled with the finite difference method adhering to the assumption that the components of the strain tensor are available at all nodes of the finite element mesh model. The suggested numerical technique is easy to implement, practically efficient, and requires less computational time compared to currently available commercial finite element packages such as ANSYS and/or ABAQUS.
Dynamic Analysis of Darrieus Vertical Axis Wind Turbine Rotors
NASA Technical Reports Server (NTRS)
Lobitz, D. W.
1981-01-01
The dynamic response characteristics of the vertical axis wind turbine (VAWT) rotor are important factors governing the safety and fatigue life of VAWT systems. The principal problems are the determination of critical rotor speeds (resonances) and the assessment of forced vibration response amplitudes. The solution to these problems is complicated by centrifugal and Coriolis effects which can have substantial influence on rotor resonant frequencies and mode shapes. The primary tools now in use for rotor analysis are described and discussed. These tools include a lumped spring mass model (VAWTDYN) and also finite-element based approaches. The accuracy and completeness of current capabilities are also discussed.
DeAngelo, Jacob; Shervais, John W.; Glen, Jonathan; Nielson, Dennis L.; Garg, Sabodh; Dobson, Patrick; Gasperikova, Erika; Sonnenthal, Eric; Visser, Charles; Liberty, Lee M.; Siler, Drew; Evans, James P.; Santellanes, Sean
2016-01-01
Play fairway analysis in geothermal exploration derives from a systematic methodology originally developed within the petroleum industry and is based on a geologic and hydrologic framework of identified geothermal systems. We are tailoring this methodology to study the geothermal resource potential of the Snake River Plain and surrounding region. This project has contributed to the success of this approach by cataloging the critical elements controlling exploitable hydrothermal systems, establishing risk matrices that evaluate these elements in terms of both probability of success and level of knowledge, and building automated tools to process results. ArcGIS was used to compile a range of different data types, which we refer to as ‘elements’ (e.g., faults, vents, heatflow…), with distinct characteristics and confidence values. Raw data for each element were transformed into data layers with a common format. Because different data types have different uncertainties, each evidence layer had an accompanying confidence layer, which reflects spatial variations in these uncertainties. Risk maps represent the product of evidence and confidence layers, and are the basic building blocks used to construct Common Risk Segment (CRS) maps for heat, permeability, and seal. CRS maps quantify the variable risk associated with each of these critical components. In a final step, the three CRS maps were combined into a Composite Common Risk Segment (CCRS) map for analysis that reveals favorable areas for geothermal exploration. Python scripts were developed to automate data processing and to enhance the flexibility of the data analysis. Python scripting provided the structure that makes a custom workflow possible. Nearly every tool available in the ArcGIS ArcToolbox can be executed using commands in the Python programming language. This enabled the construction of a group of tools that could automate most of the processing for the project. Currently, our tools are repeatable, scalable, modifiable, and transferrable, allowing us to automate the task of data analysis and the production of CRS and CCRS maps. Our ultimate goal is to produce a toolkit that can be imported into ArcGIS and applied to any geothermal play type, with fully tunable parameters that will allow for the production of multiple versions of the CRS and CCRS maps in order to better test for sensitivity and to validate results.
A computer program for cyclic plasticity and structural fatigue analysis
NASA Technical Reports Server (NTRS)
Kalev, I.
1980-01-01
A computerized tool for the analysis of time independent cyclic plasticity structural response, life to crack initiation prediction, and crack growth rate prediction for metallic materials is described. Three analytical items are combined: the finite element method with its associated numerical techniques for idealization of the structural component, cyclic plasticity models for idealization of the material behavior, and damage accumulation criteria for the fatigue failure.
LAMPAT and LAMPATNL User’s Manual
2012-09-01
nonlinearity. These tools are implemented as subroutines in the finite element software ABAQUS . This user’s manual provides information on the proper...model either through the General tab of the Edit Job dialog box in Abaqus /CAE or the command line with user=( subroutine filename). Table 1...Selection of software product and subroutine . Static Analysis With Abaqus /Standard Dynamic Analysis With Abaqus /Explicit Linear, uncoupled
Lee, Jaehoon; Hulse, Nathan C; Wood, Grant M; Oniki, Thomas A; Huff, Stanley M
2016-01-01
In this study we developed a Fast Healthcare Interoperability Resources (FHIR) profile to support exchanging a full pedigree based family health history (FHH) information across multiple systems and applications used by clinicians, patients, and researchers. We used previously developed clinical element models (CEMs) that are capable of representing the FHH information, and derived essential data elements including attributes, constraints, and value sets. We analyzed gaps between the FHH CEM elements and existing FHIR resources. Based on the analysis, we developed a profile that consists of 1) FHIR resources for essential FHH data elements, 2) extensions for additional elements that were not covered by the resources, and 3) a structured definition to integrate patient and family member information in a FHIR message. We implemented the profile using an open-source based FHIR framework and validated it using patient-entered FHH data that was captured through a locally developed FHH tool.
1983-09-01
Reyer’s research 12 . I -’. .. - n -l l--" (11:1), of "scrounging" from another worker’s tool box. As a consequence, control and accounting procedures...PPOGPAM $ELEMENT. PROJECT, TASK School of Systems and Logistics APEA A WORK UNIT NUMBERS Air Force Institute of Technology, WPAFBO1 II. CONTROLLING ...thesis analyzes those attitudes by evaluating the collected data from the AFLMC questionnaire and this research team’s telephone interviews. The
Improved accuracy for finite element structural analysis via a new integrated force method
NASA Technical Reports Server (NTRS)
Patnaik, Surya N.; Hopkins, Dale A.; Aiello, Robert A.; Berke, Laszlo
1992-01-01
A comparative study was carried out to determine the accuracy of finite element analyses based on the stiffness method, a mixed method, and the new integrated force and dual integrated force methods. The numerical results were obtained with the following software: MSC/NASTRAN and ASKA for the stiffness method; an MHOST implementation method for the mixed method; and GIFT for the integrated force methods. The results indicate that on an overall basis, the stiffness and mixed methods present some limitations. The stiffness method generally requires a large number of elements in the model to achieve acceptable accuracy. The MHOST method tends to achieve a higher degree of accuracy for course models than does the stiffness method implemented by MSC/NASTRAN and ASKA. The two integrated force methods, which bestow simultaneous emphasis on stress equilibrium and strain compatibility, yield accurate solutions with fewer elements in a model. The full potential of these new integrated force methods remains largely unexploited, and they hold the promise of spawning new finite element structural analysis tools.
Tracking Radionuclide Fractionation in the First Atomic Explosion Using Stable Elements
Bonamici, Chloë E.; Hervig, Richard L.; Kinman, William S.
2017-08-25
Compositional analysis of postdetonation fallout is a tool for forensic identification of nuclear devices. However, the relationship between device composition and fallout composition is difficult to interpret because of the complex combination of physical mixing, nuclear reactions, and chemical fractionations that occur in the chaotic nuclear fireball. By using a combination of in situ microanalytical techniques (electron microprobe analysis and secondary ion mass spectrometry), we show that some heavy stable elements (Rb, Sr, Zr, Ba, Cs, Ba, La, Ce, Nd, Sm, Dy, Lu, U, Th) in glassy fallout from the first nuclear test, Trinity, are reliable chemical proxies for radionuclidesmore » generated during the explosion. Stable-element proxies show that radionuclides from the Trinity device were chemically, but not isotopically, fractionated by condensation. Moreover, stable-element proxies delineate chemical fractionation trends that can be used to connect present-day fallout composition to past fireball composition. Stable-element proxies therefore offer a novel approach for elucidating the phenomenology of the nuclear fireball as it relates to the formation of debris and the fixation of device materials within debris.« less
Evaluation of Macro- and Microelement Levels in Black Tea in View of Its Geographical Origin.
Brzezicha-Cirocka, Justyna; Grembecka, Małgorzata; Ciesielski, Tomasz; Flaten, Trond Peder; Szefer, Piotr
2017-04-01
The aim of this study was to evaluate the elemental composition of black tea samples and their infusions in view of their geographical origin. In total, 14 elements were analyzed, 13 (Ca, K, Mg, Na, Mn, Fe, Zn, Cu, Cr, Ni, Co, Cd, and Pb) by flame atomic absorption spectrometry, and P by UV-Vis spectrometry, after mineralization of samples. It was found that K was the most abundant macroelement in the analyzed samples, whereas among microelements, the highest concentration was found for Mn. Based on the obtained data, the percentage of elements leached into the infusions as well as the daily elemental intake from tea were calculated. The daily intake from tea was compared to the recommended daily allowances (RDAs), and the highest percentages of the RDAs were found for Mn (15 %) and Co (10 %). To study the relations between elemental composition and country of origin of samples, factor analysis and cluster analysis were applied. These multivariate techniques proved to be efficient tools able to differentiate samples according to their provenance as well as plantation within the common regions.
Tracking Radionuclide Fractionation in the First Atomic Explosion Using Stable Elements
DOE Office of Scientific and Technical Information (OSTI.GOV)
Bonamici, Chloë E.; Hervig, Richard L.; Kinman, William S.
Compositional analysis of postdetonation fallout is a tool for forensic identification of nuclear devices. However, the relationship between device composition and fallout composition is difficult to interpret because of the complex combination of physical mixing, nuclear reactions, and chemical fractionations that occur in the chaotic nuclear fireball. By using a combination of in situ microanalytical techniques (electron microprobe analysis and secondary ion mass spectrometry), we show that some heavy stable elements (Rb, Sr, Zr, Ba, Cs, Ba, La, Ce, Nd, Sm, Dy, Lu, U, Th) in glassy fallout from the first nuclear test, Trinity, are reliable chemical proxies for radionuclidesmore » generated during the explosion. Stable-element proxies show that radionuclides from the Trinity device were chemically, but not isotopically, fractionated by condensation. Moreover, stable-element proxies delineate chemical fractionation trends that can be used to connect present-day fallout composition to past fireball composition. Stable-element proxies therefore offer a novel approach for elucidating the phenomenology of the nuclear fireball as it relates to the formation of debris and the fixation of device materials within debris.« less
Tracking Radionuclide Fractionation in the First Atomic Explosion Using Stable Elements.
Bonamici, Chloë E; Hervig, Richard L; Kinman, William S
2017-09-19
Compositional analysis of postdetonation fallout is a tool for forensic identification of nuclear devices. However, the relationship between device composition and fallout composition is difficult to interpret because of the complex combination of physical mixing, nuclear reactions, and chemical fractionations that occur in the chaotic nuclear fireball. Using a combination of in situ microanalytical techniques (electron microprobe analysis and secondary ion mass spectrometry), we show that some heavy stable elements (Rb, Sr, Zr, Ba, Cs, Ba, La, Ce, Nd, Sm, Dy, Lu, U, Th) in glassy fallout from the first nuclear test, Trinity, are reliable chemical proxies for radionuclides generated during the explosion. Stable-element proxies show that radionuclides from the Trinity device were chemically, but not isotopically, fractionated by condensation. Furthermore, stable-element proxies delineate chemical fractionation trends that can be used to connect present-day fallout composition to past fireball composition. Stable-element proxies therefore offer a novel approach for elucidating the phenomenology of the nuclear fireball as it relates to the formation of debris and the fixation of device materials within debris.
RSAT 2015: Regulatory Sequence Analysis Tools.
Medina-Rivera, Alejandra; Defrance, Matthieu; Sand, Olivier; Herrmann, Carl; Castro-Mondragon, Jaime A; Delerce, Jeremy; Jaeger, Sébastien; Blanchet, Christophe; Vincens, Pierre; Caron, Christophe; Staines, Daniel M; Contreras-Moreira, Bruno; Artufel, Marie; Charbonnier-Khamvongsa, Lucie; Hernandez, Céline; Thieffry, Denis; Thomas-Chollier, Morgane; van Helden, Jacques
2015-07-01
RSAT (Regulatory Sequence Analysis Tools) is a modular software suite for the analysis of cis-regulatory elements in genome sequences. Its main applications are (i) motif discovery, appropriate to genome-wide data sets like ChIP-seq, (ii) transcription factor binding motif analysis (quality assessment, comparisons and clustering), (iii) comparative genomics and (iv) analysis of regulatory variations. Nine new programs have been added to the 43 described in the 2011 NAR Web Software Issue, including a tool to extract sequences from a list of coordinates (fetch-sequences from UCSC), novel programs dedicated to the analysis of regulatory variants from GWAS or population genomics (retrieve-variation-seq and variation-scan), a program to cluster motifs and visualize the similarities as trees (matrix-clustering). To deal with the drastic increase of sequenced genomes, RSAT public sites have been reorganized into taxon-specific servers. The suite is well-documented with tutorials and published protocols. The software suite is available through Web sites, SOAP/WSDL Web services, virtual machines and stand-alone programs at http://www.rsat.eu/. © The Author(s) 2015. Published by Oxford University Press on behalf of Nucleic Acids Research.
NASA Technical Reports Server (NTRS)
Lung, Shun-fat; Pak, Chan-gi
2008-01-01
Updating the finite element model using measured data is a challenging problem in the area of structural dynamics. The model updating process requires not only satisfactory correlations between analytical and experimental results, but also the retention of dynamic properties of structures. Accurate rigid body dynamics are important for flight control system design and aeroelastic trim analysis. Minimizing the difference between analytical and experimental results is a type of optimization problem. In this research, a multidisciplinary design, analysis, and optimization (MDAO) tool is introduced to optimize the objective function and constraints such that the mass properties, the natural frequencies, and the mode shapes are matched to the target data as well as the mass matrix being orthogonalized.
NASA Technical Reports Server (NTRS)
Lung, Shun-fat; Pak, Chan-gi
2008-01-01
Updating the finite element model using measured data is a challenging problem in the area of structural dynamics. The model updating process requires not only satisfactory correlations between analytical and experimental results, but also the retention of dynamic properties of structures. Accurate rigid body dynamics are important for flight control system design and aeroelastic trim analysis. Minimizing the difference between analytical and experimental results is a type of optimization problem. In this research, a multidisciplinary design, analysis, and optimization [MDAO] tool is introduced to optimize the objective function and constraints such that the mass properties, the natural frequencies, and the mode shapes are matched to the target data as well as the mass matrix being orthogonalized.
Update on HCDstruct - A Tool for Hybrid Wing Body Conceptual Design and Structural Optimization
NASA Technical Reports Server (NTRS)
Gern, Frank H.
2015-01-01
HCDstruct is a Matlab® based software tool to rapidly build a finite element model for structural optimization of hybrid wing body (HWB) aircraft at the conceptual design level. The tool uses outputs from a Flight Optimization System (FLOPS) performance analysis together with a conceptual outer mold line of the vehicle, e.g. created by Vehicle Sketch Pad (VSP), to generate a set of MSC Nastran® bulk data files. These files can readily be used to perform a structural optimization and weight estimation using Nastran’s® Solution 200 multidisciplinary optimization solver. Initially developed at NASA Langley Research Center to perform increased fidelity conceptual level HWB centerbody structural analyses, HCDstruct has grown into a complete HWB structural sizing and weight estimation tool, including a fully flexible aeroelastic loads analysis. Recent upgrades to the tool include the expansion to a full wing tip-to-wing tip model for asymmetric analyses like engine out conditions and dynamic overswings, as well as a fully actuated trailing edge, featuring up to 15 independently actuated control surfaces and twin tails. Several example applications of the HCDstruct tool are presented.
USING STABLE ISOTOPES FOR FISH DIETARY ANALYSES: COPING WITH TOO MANY SOURCES
Stable isotope analysis can provide a useful tool for determining time-integrated measures of proportional food source contributions to fish diets. Ratios of stable (non-radioactive) isotopes of common elements (e.g., C,N,S) vary among food sources, and tissues of consumers (e.g...
Risk Management for Weapon Systems Acquisition: A Decision Support System
1985-02-28
includes the program evaluation and review technique (PERT) for network analysis, the PMRM for quantifying risk , an optimization package for generating...Despite the inclusion of uncertainty in time, PERT can at best be considered as a tool for quantifying risk with regard to the time element only. Moreover
USDA-ARS?s Scientific Manuscript database
Although slowly abandoned in developed countries, furrow irrigation systems continue to be a dominant irrigation method in developing countries. Numerical models represent powerful tools to assess irrigation and fertigation efficiency. While several models have been proposed in the past, the develop...
Anti-flammable properties of capable phosphorus-nitrogen containing triazine derivatives on cotton
USDA-ARS?s Scientific Manuscript database
Most of new phosphorus-nitrogen containing compounds were prepared by organic reactions of cyanuric chloride and phosphonates. They were characterized by analytical tools such as proton (1H), carbon (13C), and phosphorus (31P) nuclear magnetic resonance (NMR) spectroscopy and elemental analysis (EA)...
Application of the finite element method in orthopedic implant design.
Saha, Subrata; Roychowdhury, Amit
2009-01-01
The finite element method (FEM) was first introduced to the field of orthopedic biomechanics in the early 1970s to evaluate stresses in human bones. By the early 1980s, the method had become well established as a tool for basic research and design analysis. Since the late 1980s and early 1990s, FEM has also been used to study bone remodeling. Today, it is one of the most reliable simulation tools for evaluating wear, fatigue, crack propagation, and so forth, and is used in many types of preoperative testing. Since the introduction of FEM to orthopedic biomechanics, there have been rapid advances in computer processing speeds, the finite element and other numerical methods, understanding of mechanical properties of soft and hard tissues and their modeling, and image-processing techniques. In light of these advances, it is accepted today that FEM will continue to contribute significantly to further progress in the design and development of orthopedic implants, as well as in the understanding of other complex systems of the human body. In the following article, different main application areas of finite element simulation will be reviewed including total hip joint arthroplasty, followed by the knee, spine, shoulder, and elbow, respectively.
Spatial and temporal regulation of the forisome gene for1 in the phloem during plant development.
Noll, Gundula A; Fontanellaz, Maria E; Rüping, Boris; Ashoub, Ahmed; van Bel, Aart J E; Fischer, Rainer; Knoblauch, Michael; Prüfer, Dirk
2007-10-01
Forisomes are protein aggregates found uniquely in the sieve elements of Fabaceaen plants. Upon wounding they undergo a reversible, calcium-dependent conformational switch which enables them to act as cellular stopcocks. Forisomes begin to form in young sieve elements at an early stage of metaphloem differentiation. Genes encoding forisome components could therefore be useful as markers of early sieve element development. Here we present a comprehensive analysis of the developmental expression profile of for1, which encodes such a forisome component. The for1 gene is highly conserved among Fabaceaen species and appears to be unique to this phylogenetic lineage since no orthologous genes have been found in other plants, including Arabidopsis and rice. Even so, transgenic tobacco plants expressing reporter genes under the control of the for1 promoter display reporter activity exclusively in immature sieve elements. This suggests that the regulation of sieve element development is highly conserved even in plants where mature forisomes have not been detected. The promoter system could therefore provide a powerful tool for the detailed analysis of differentiation in metaphloem sieve elements in an unexpectedly broad range of plant species.
Gamma Spectroscopy by Artificial Neural Network Coupled with MCNP
NASA Astrophysics Data System (ADS)
Sahiner, Huseyin
While neutron activation analysis is widely used in many areas, sensitivity of the analysis depends on how the analysis is conducted. Even though the sensitivity of the techniques carries error, compared to chemical analysis, its range is in parts per million or sometimes billion. Due to this sensitivity, the use of neutron activation analysis becomes important when analyzing bio-samples. Artificial neural network is an attractive technique for complex systems. Although there are neural network applications on spectral analysis, training by simulated data to analyze experimental data has not been made. This study offers an improvement on spectral analysis and optimization on neural network for the purpose. The work considers five elements that are considered as trace elements for bio-samples. However, the system is not limited to five elements. The only limitation of the study comes from data library availability on MCNP. A perceptron network was employed to identify five elements from gamma spectra. In quantitative analysis, better results were obtained when the neural fitting tool in MATLAB was used. As a training function, Levenberg-Marquardt algorithm was used with 23 neurons in the hidden layer with 259 gamma spectra in the input. Because the interest of the study deals with five elements, five neurons representing peak counts of five isotopes in the input layer were used. Five output neurons revealed mass information of these elements from irradiated kidney stones. Results showing max error of 17.9% in APA, 24.9% in UA, 28.2% in COM, 27.9% in STRU type showed the success of neural network approach in analyzing gamma spectra. This high error was attributed to Zn that has a very long decay half-life compared to the other elements. The simulation and experiments were made under certain experimental setup (3 hours irradiation, 96 hours decay time, 8 hours counting time). Nevertheless, the approach is subject to be generalized for different setups.
Tools for Communication: Novel infrastructure to address patient-perceived gaps in oncology care .
McMullen, Suzanne; Szabo, Shelagh; Halbert, Ronald J; Lai, Catherine; Parikh, Aparna; Bunce, Mikele; Khoury, Raya; Small, Art; Masaquel, Anthony
2017-04-01
Healthcare providers (HCPs) and patient communication are integral to high-quality oncology care. The patient and HCP perspectives are needed to identify gaps in care and develop communication tools. . This study aimed to understand patient- and HCP-perceived elements of and gaps in high-quality care to develop novel communication tools to improve care. . Qualitative interviews were conducted among 16 patients with cancer and 10 HCPs in the United States. Trained interviewers elicited patients' and HCPs' concerns, views, and perceived needs for communication tools. A thematic analysis was used to identify four quality of care domains, depicted in a conceptual model, and two draft communication tools were developed to address identified gaps. . No patients reported previously using a communication tool, and gaps in communication regarding treatment aims and education were evident. Two tools were developed to assess patients' life and treatment goals and the importance of ongoing education.
Design and analysis of tactile optical sensor for endovascular surgery
NASA Astrophysics Data System (ADS)
Qasaimeh, M. A.; Dargahi, J.; Kahrizi, M.; Packirisamy, M.
2007-06-01
In this paper, design and Finite Element analysis of a new tactile optical sensor for the measurement of contact-pressure and tissue compliance in endovascular surgeries are presented. Using Micro-Electro-Mechanical-Systems (MEMS) technology, this sensor can be fabricated and integrated with the medical tools for endovascular surgeries such as Catheter tool. The designed sensor is capable of detecting the magnitude of the applied forces, the pressure distribution on contact objects, and also estimating the compliance of the contact tissue. The designed sensor is made of three layers, the upper layer is fabricated from monocrystalline silicon to form silicon membranes, the middle layer which is the supporting element is fabricated from both silicon and silicone rubber as a soft material and the lower layer is a supporting Plexiglas substrate to connect the designed sensor to the optical fibers. Simulation results show that for the given contact forces, the magnitude and the distribution of contacting tissues pressure along with tissue compliance can be determined. This sensor as proposed is a good candidate for batch micromachining, which is yet another commercial advantage for this design. Because of its less expensive cost, the surgeon can use it as a disposal part of the endovascular tools, requiring no re-sterilization and reducing the cost of surgery.
Design, Optimization and Evaluation of Integrally Stiffened Al 7050 Panel with Curved Stiffeners
NASA Technical Reports Server (NTRS)
Slemp, Wesley C. H.; Bird, R. Keith; Kapania, Rakesh K.; Havens, David; Norris, Ashley; Olliffe, Robert
2011-01-01
A curvilinear stiffened panel was designed, manufactured, and tested in the Combined Load Test Fixture at NASA Langley Research Center. The panel was optimized for minimum mass subjected to constraints on buckling load, yielding, and crippling or local stiffener failure using a new analysis tool named EBF3PanelOpt. The panel was designed for a combined compression-shear loading configuration that is a realistic load case for a typical aircraft wing panel. The panel was loaded beyond buckling and strains and out-of-plane displacements were measured. The experimental data were compared with the strains and out-of-plane deflections from a high fidelity nonlinear finite element analysis and linear elastic finite element analysis of the panel/test-fixture assembly. The numerical results indicated that the panel buckled at the linearly elastic buckling eigenvalue predicted for the panel/test-fixture assembly. The experimental strains prior to buckling compared well with both the linear and nonlinear finite element model.
Composite structural materials
NASA Technical Reports Server (NTRS)
Ansell, G. S.; Loewy, R. G.; Wiberley, S. E.
1981-01-01
The composite aircraft program component (CAPCOMP) is a graduate level project conducted in parallel with a composite structures program. The composite aircraft program glider (CAPGLIDE) is an undergraduate demonstration project which has as its objectives the design, fabrication, and testing of a foot launched ultralight glider using composite structures. The objective of the computer aided design (COMPAD) portion of the composites project is to provide computer tools for the analysis and design of composite structures. The major thrust of COMPAD is in the finite element area with effort directed at implementing finite element analysis capabilities and developing interactive graphics preprocessing and postprocessing capabilities. The criteria for selecting research projects to be conducted under the innovative and supporting research (INSURE) program are described.
NASA Technical Reports Server (NTRS)
Harris, Charles E.; Starnes, James H., Jr.; Newman, James C., Jr.
1995-01-01
NASA is developing a 'tool box' that includes a number of advanced structural analysis computer codes which, taken together, represent the comprehensive fracture mechanics capability required to predict the onset of widespread fatigue damage. These structural analysis tools have complementary and specialized capabilities ranging from a finite-element-based stress-analysis code for two- and three-dimensional built-up structures with cracks to a fatigue and fracture analysis code that uses stress-intensity factors and material-property data found in 'look-up' tables or from equations. NASA is conducting critical experiments necessary to verify the predictive capabilities of the codes, and these tests represent a first step in the technology-validation and industry-acceptance processes. NASA has established cooperative programs with aircraft manufacturers to facilitate the comprehensive transfer of this technology by making these advanced structural analysis codes available to industry.
NASA Technical Reports Server (NTRS)
Hairr, John W.; Huang, Jui-Ten; Ingram, J. Edward; Shah, Bharat M.
1992-01-01
The ISPAN Program (Interactive Stiffened Panel Analysis) is an interactive design tool that is intended to provide a means of performing simple and self contained preliminary analysis of aircraft primary structures made of composite materials. The program combines a series of modules with the finite element code DIAL as its backbone. Four ISPAN Modules were developed and are documented. These include: (1) flat stiffened panel; (2) curved stiffened panel; (3) flat tubular panel; and (4) curved geodesic panel. Users are instructed to input geometric and material properties, load information and types of analysis (linear, bifurcation buckling, or post-buckling) interactively. The program utilizing this information will generate finite element mesh and perform analysis. The output in the form of summary tables of stress or margins of safety, contour plots of loads or stress, and deflected shape plots may be generalized and used to evaluate specific design.
Introduction to the computational structural mechanics testbed
NASA Technical Reports Server (NTRS)
Lotts, C. G.; Greene, W. H.; Mccleary, S. L.; Knight, N. F., Jr.; Paulson, S. S.; Gillian, R. E.
1987-01-01
The Computational Structural Mechanics (CSM) testbed software system based on the SPAR finite element code and the NICE system is described. This software is denoted NICE/SPAR. NICE was developed at Lockheed Palo Alto Research Laboratory and contains data management utilities, a command language interpreter, and a command language definition for integrating engineering computational modules. SPAR is a system of programs used for finite element structural analysis developed for NASA by Lockheed and Engineering Information Systems, Inc. It includes many complementary structural analysis, thermal analysis, utility functions which communicate through a common database. The work on NICE/SPAR was motivated by requirements for a highly modular and flexible structural analysis system to use as a tool in carrying out research in computational methods and exploring computer hardware. Analysis examples are presented which demonstrate the benefits gained from a combination of the NICE command language with a SPAR computational modules.
Tools for Understanding Identity
DOE Office of Scientific and Technical Information (OSTI.GOV)
Creese, Sadie; Gibson-Robinson, Thomas; Goldsmith, Michael
Identity attribution and enrichment is critical to many aspects of law-enforcement and intelligence gathering; this identity typically spans a number of domains in the natural-world such as biographic information (factual information – e.g. names, addresses), biometric information (e.g. fingerprints) and psychological information. In addition to these natural-world projections of identity, identity elements are projected in the cyber-world. Conversely, undesirable elements may use similar techniques to target individuals for spear-phishing attacks (or worse), and potential targets or their organizations may want to determine how to minimize the attack surface exposed. Our research has been exploring the construction of a mathematical modelmore » for identity that supports such holistic identities. The model captures the ways in which an identity is constructed through a combination of data elements (e.g. a username on a forum, an address, a telephone number). Some of these elements may allow new characteristics to be inferred, hence enriching the holistic view of the identity. An example use-case would be the inference of real names from usernames, the ‘path’ created by inferring new elements of identity is highlighted in the ‘critical information’ panel. Individual attribution exercises can be understood as paths through a number of elements. Intuitively the entire realizable ‘capability’ can be modeled as a directed graph, where the elements are nodes and the inferences are represented by links connecting one or more antecedents with a conclusion. The model can be operationalized with two levels of tool support described in this paper, the first is a working prototype, the second is expected to reach prototype by July 2013: Understanding the Model The tool allows a user to easily determine, given a particular set of inferences and attributes, which elements or inferences are of most value to an investigator (or an attacker). The tool is also able to take into account the difficulty of the inferences, allowing the user to consider different scenarios depending on the perceived resources of the attacker, or to prioritize lines of investigation. It also has a number of interesting visualizations that are designed to aid the user in understanding the model. The tool works by considering the inferences as a graph and runs various graph-theoretic algorithms, with some novel adaptations, in order to deduce various properties. Using the Model To help investigators exploit the model to perform identity attribution, we have developed the Identity Map visualization. For a user-provided set of known starting elements and a set of desired target elements for a given identity, the Identity Map generates investigative workflows as paths through the model. Each path consists of a series of elements and inferences between them that connect the input and output elements. Each path also has an associated confidence level that estimates the reliability of the resulting attribution. Identity Map can help investigators understand the possible ways to make an identification decision and guide them toward the data-collection or analysis steps required to reach that decision.« less
Zhang, H-H; Shen, Y-H; Xu, H-E; Liang, H-Y; Han, M-J; Zhang, Z
2013-10-01
Comparative analysis of transposable elements (TEs) from different species can make it possible to reconstruct their history over evolutionary time. In this study, we identified a novel hAT element in Bombyx mori and Rhodnius prolixus with characteristic GGGCGGCA repeats in its subterminal region. Meanwhile, phylogenetic analysis demonstrated that the elements in these two species might represent a separate cluster of the hAT superfamily. Strikingly, a previously identified miniature inverted repeat transposable element (MITE) shared high identity with this autonomous element across the entire length, supporting the hypothesis that MITEs are derived from the internal deletion of DNA transposons. Interestingly, identity of the consensus sequences of this novel hAT element between B. mori and R. prolixus, which diverged about 370 million years ago, was as high as 96.5% over their full length (about 3.6 kb) at the nucleotide level. The patchy distribution amongst species, coupled with overall lack of intense purifying selection acting on this element, suggest that this novel hAT element might have experienced horizontal transfer between the ancestors of B. mori and R. prolixus. Our results highlight that this novel hAT element could be used as a potential tool for germline transformation of R. prolixus to control the transmission of Trypanosoma cruzi, which causes Chagas disease. © 2013 Royal Entomological Society.
Approaches to Fungal Genome Annotation
Haas, Brian J.; Zeng, Qiandong; Pearson, Matthew D.; Cuomo, Christina A.; Wortman, Jennifer R.
2011-01-01
Fungal genome annotation is the starting point for analysis of genome content. This generally involves the application of diverse methods to identify features on a genome assembly such as protein-coding and non-coding genes, repeats and transposable elements, and pseudogenes. Here we describe tools and methods leveraged for eukaryotic genome annotation with a focus on the annotation of fungal nuclear and mitochondrial genomes. We highlight the application of the latest technologies and tools to improve the quality of predicted gene sets. The Broad Institute eukaryotic genome annotation pipeline is described as one example of how such methods and tools are integrated into a sequencing center’s production genome annotation environment. PMID:22059117
Planar Inlet Design and Analysis Process (PINDAP)
NASA Technical Reports Server (NTRS)
Slater, John W.; Gruber, Christopher R.
2005-01-01
The Planar Inlet Design and Analysis Process (PINDAP) is a collection of software tools that allow the efficient aerodynamic design and analysis of planar (two-dimensional and axisymmetric) inlets. The aerodynamic analysis is performed using the Wind-US computational fluid dynamics (CFD) program. A major element in PINDAP is a Fortran 90 code named PINDAP that can establish the parametric design of the inlet and efficiently model the geometry and generate the grid for CFD analysis with design changes to those parameters. The use of PINDAP is demonstrated for subsonic, supersonic, and hypersonic inlets.
NASA Astrophysics Data System (ADS)
Naga Raju, G. J.; Sarita, P.; Murthy, K. S. R.
2017-08-01
Particle Induced X-ray Emission (PIXE), an accelerator based analytical technique has been employed in this work for the analysis of trace elements in the cancerous and non-cancerous tissues of rectal cancer patients. A beam of 3 MeV protons generated from 3 MV Pelletron accelerator at the Ion Beam Laboratory of Institute of Physics, Bhubaneswar, India was used as projectile to excite the atoms present in the tissues samples. PIXE technique, with its capability to detect simultaneously several elements present at very low concentrations, offers an excellent tool for trace element analysis. The characteristic X-rays emitted by the samples were recorded by a high resolution Si (Li) detector. On the basis of the PIXE spectrum obtained for each sample, the elements Cl, K, Ca, Ti, V, Cr, Mn, Fe, Co, Ni, Cu, Zn, As, and Br were identified and their relative concentrations were estimated in the cancerous and non-cancerous tissues of rectum. The levels of Mn, Fe, Co, Cu, Zn, and As were higher (p < 0.005) while the levels of Ca, Cr and Ni were lower (p < 0.005) in the cancer tissues relative to the normal tissues. The alterations in the levels of the trace elements observed in the present work are discussed in this paper with respect to their potential role in the initiation, promotion and inhibition of cancer of the rectum.
Vibration characteristics of teak wood filled steel tubes
NASA Astrophysics Data System (ADS)
Danawade, Bharatesh Adappa; Malagi, Ravindra Rachappa
2018-05-01
The objective of this paper is to determine fundamental frequency and damping ratio of teak wood filled steel tubes. Mechanically bonded teak wood filled steel tubes have been evaluated by experimental impact hammer test using modal analysis. The results of impact hammer test were verified and validated by finite element tool ANSYS using harmonic analysis. The error between the two methods was observed to be within acceptable limit.
NASA Astrophysics Data System (ADS)
Liu, Bilin; Chen, Xinjun; Fang, Zhou; Hu, Song; Song, Qian
2015-12-01
We applied solution-based ICP-MS method to quantify the trace-elemental signatures in statoliths of jumbo flying squid, Dosidius gigas, which were collected from the waters off northern and central Chile during the scientific surveys carried out by Chinese squid jigging vessels in 2007 and 2008. The age and spawning date of the squid were back-calculated based on daily increments in statoliths. Eight elemental ratios (Sr/Ca, Ba/Ca, Mg/Ca, Mn/Ca, Na/Ca, Fe/Ca, Cu/Ca and Zn/Ca) were analyzed. It was found that Sr is the second most abundant element next to Ca, followed by Na, Fe, Mg, Zn, Cu, Ba and Mn. There was no significant relationship between element/Ca and sea surface temperature (SST) and sea surface salinity (SSS), although weak negative or positive tendency was found. MANOVA analysis showed that multivariate elemental signatures did not differ among the cohorts spawned in spring, autumn and winter, and no significant difference was found between the northern and central sampling locations. Classification results showed that all individuals of each spawned cohorts were correctly classified. This study demonstrates that the elemental signatures in D. gigas statoliths are potentially a useful tool to improve our understanding of its population structure and habitat environment.
IGA: A Simplified Introduction and Implementation Details for Finite Element Users
NASA Astrophysics Data System (ADS)
Agrawal, Vishal; Gautam, Sachin S.
2018-05-01
Isogeometric analysis (IGA) is a recently introduced technique that employs the Computer Aided Design (CAD) concept of Non-uniform Rational B-splines (NURBS) tool to bridge the substantial bottleneck between the CAD and finite element analysis (FEA) fields. The simplified transition of exact CAD models into the analysis alleviates the issues originating from geometrical discontinuities and thus, significantly reduces the design-to-analysis time in comparison to traditional FEA technique. Since its origination, the research in the field of IGA is accelerating and has been applied to various problems. However, the employment of CAD tools in the area of FEA invokes the need of adapting the existing implementation procedure for the framework of IGA. Also, the usage of IGA requires the in-depth knowledge of both the CAD and FEA fields. This can be overwhelming for a beginner in IGA. Hence, in this paper, a simplified introduction and implementation details for the incorporation of NURBS based IGA technique within the existing FEA code is presented. It is shown that with little modifications, the available standard code structure of FEA can be adapted for IGA. For the clear and concise explanation of these modifications, step-by-step implementation of a benchmark plate with a circular hole under the action of in-plane tension is included.
NASA Technical Reports Server (NTRS)
Panda, Binayak
2009-01-01
Modern analytical tools can yield invaluable results during materials characterization and failure analysis. Scanning electron microscopes (SEMs) provide significant analytical capabilities, including angstrom-level resolution. These systems can be equipped with a silicon drift detector (SDD) for very fast yet precise analytical mapping of phases, as well as electron back-scattered diffraction (EBSD) units to map grain orientations, chambers that admit large samples, variable pressure for wet samples, and quantitative analysis software to examine phases. Advanced solid-state electronics have also improved surface and bulk analysis instruments: Secondary ion mass spectroscopy (SIMS) can quantitatively determine and map light elements such as hydrogen, lithium, and boron - with their isotopes. Its high sensitivity detects impurities at parts per billion (ppb) levels. X-ray photo-electron spectroscopy (XPS) can determine oxidation states of elements, as well as identifying polymers and measuring film thicknesses on coated composites. This technique is also known as electron spectroscopy for chemical analysis (ESCA). Scanning Auger electron spectroscopy (SAM) combines surface sensitivity, spatial lateral resolution (10 nm), and depth profiling capabilities to describe elemental compositions of near and below surface regions down to the chemical state of an atom.
The aggregated unfitted finite element method for elliptic problems
NASA Astrophysics Data System (ADS)
Badia, Santiago; Verdugo, Francesc; Martín, Alberto F.
2018-07-01
Unfitted finite element techniques are valuable tools in different applications where the generation of body-fitted meshes is difficult. However, these techniques are prone to severe ill conditioning problems that obstruct the efficient use of iterative Krylov methods and, in consequence, hinders the practical usage of unfitted methods for realistic large scale applications. In this work, we present a technique that addresses such conditioning problems by constructing enhanced finite element spaces based on a cell aggregation technique. The presented method, called aggregated unfitted finite element method, is easy to implement, and can be used, in contrast to previous works, in Galerkin approximations of coercive problems with conforming Lagrangian finite element spaces. The mathematical analysis of the new method states that the condition number of the resulting linear system matrix scales as in standard finite elements for body-fitted meshes, without being affected by small cut cells, and that the method leads to the optimal finite element convergence order. These theoretical results are confirmed with 2D and 3D numerical experiments.
HOLST, Alexandra Ioana; HOLST, Stefan; HIRSCHFELDER, Ursula; von SECKENDORFF, Volker
2012-01-01
Objective The objective of this study was to investigate the applicability of micro-analytical methods with high spatial resolution to the characterization of the composition and corrosion behavior of two bracket systems. Material and methods The surfaces of six nickel-free brackets and six nickel-containing brackets were examined for signs of corrosion and qualitative surface analysis using an electron probe microanalyzer (EPMA), prior to bonding to patient's tooth surfaces and four months after clinical use. The surfaces were characterized qualitatively by secondary electron (SE) images and back scattered electron (BSE) images in both compositional and topographical mode. Qualitative and quantitative wavelength-dispersive analyses were performed for different elements, and by utilizing qualitative analysis the relative concentration of selected elements was mapped two-dimensionally. The absolute concentration of the elements was determined in specially prepared brackets by quantitative analysis using pure element standards for calibration and calculating correction-factors (ZAF). Results Clear differences were observed between the different bracket types. The nickel-containing stainless steel brackets consist of two separate pieces joined by a brazing alloy. Compositional analysis revealed two different alloy compositions, and reaction zones on both sides of the brazing alloy. The nickel-free bracket was a single piece with only slight variation in element concentration, but had a significantly rougher surface. After clinical use, no corrosive phenomena were detectable with the methods applied. Traces of intraoral wear at the contact areas between the bracket slot and the arch wire were verified. Conclusion Electron probe microanalysis is a valuable tool for the characterization of element distribution and quantitative analysis for corrosion studies. PMID:23032212
Quantifying Volcanic Emissions of Trace Elements to the Atmosphere: Ideas Based on Past Studies
NASA Astrophysics Data System (ADS)
Rose, W. I.
2003-12-01
Extensive data exist from volcanological and geochemical studies about exotic elemental enrichments in volcanic emissions to the atmosphere but quantitative data are quite rare. Advanced, highly sensitive techniques of analysis are needed to detect low concentrations of some minor elements, especially during major eruptions. I will present data from studies done during low levels of activity (incrustations and silica tube sublimates at high temperature fumaroles, from SEM studies of particle samples collected in volcanic plumes and volcanic clouds, from geochemical analysis of volcanic gas condensates, from analysis of treated particle and gas filter packs) and a much smaller number that could reflect explosive activity (from fresh ashfall leachate geochemistry, and from thermodynamic codes modeling volatile emissions from magma). This data describes a highly variable pattern of elemental enrichments which are difficult to quantify, generalize and understand. Sampling in a routine way is difficult, and work in active craters has heightened our awareness of danger, which appropriately inhibits some sampling. There are numerous localized enrichments of minor elements that can be documented and others can be expected or inferred. There is a lack of systematic tools to measure minor element abundances in volcanic emissions. The careful combination of several methodologies listed above for the same volcanic vents can provide redundant data on multiple elements which could lead to overall quantification of minor element fluxes but there are challenging issues about detection. For quiescent plumes we can design combinations of measurements to quantify minor element emission rates. Doing a comparable methodology to succeed in measuring minor element fluxes for significant eruptions will require new strategies and/or ideas.
Tools for Designing and Analyzing Structures
NASA Technical Reports Server (NTRS)
Luz, Paul L.
2005-01-01
Structural Design and Analysis Toolset is a collection of approximately 26 Microsoft Excel spreadsheet programs, each of which performs calculations within a different subdiscipline of structural design and analysis. These programs present input and output data in user-friendly, menu-driven formats. Although these programs cannot solve complex cases like those treated by larger finite element codes, these programs do yield quick solutions to numerous common problems more rapidly than the finite element codes, thereby making it possible to quickly perform multiple preliminary analyses - e.g., to establish approximate limits prior to detailed analyses by the larger finite element codes. These programs perform different types of calculations, as follows: 1. determination of geometric properties for a variety of standard structural components; 2. analysis of static, vibrational, and thermal- gradient loads and deflections in certain structures (mostly beams and, in the case of thermal-gradients, mirrors); 3. kinetic energies of fans; 4. detailed analysis of stress and buckling in beams, plates, columns, and a variety of shell structures; and 5. temperature dependent properties of materials, including figures of merit that characterize strength, stiffness, and deformation response to thermal gradients
NASA Technical Reports Server (NTRS)
Norga, Koenraad K.; Gurganus, Marjorie C.; Dilda, Christy L.; Yamamoto, Akihiko; Lyman, Richard F.; Patel, Prajal H.; Rubin, Gerald M.; Hoskins, Roger A.; Mackay, Trudy F.; Bellen, Hugo J.
2003-01-01
BACKGROUND: The identification of the function of all genes that contribute to specific biological processes and complex traits is one of the major challenges in the postgenomic era. One approach is to employ forward genetic screens in genetically tractable model organisms. In Drosophila melanogaster, P element-mediated insertional mutagenesis is a versatile tool for the dissection of molecular pathways, and there is an ongoing effort to tag every gene with a P element insertion. However, the vast majority of P element insertion lines are viable and fertile as homozygotes and do not exhibit obvious phenotypic defects, perhaps because of the tendency for P elements to insert 5' of transcription units. Quantitative genetic analysis of subtle effects of P element mutations that have been induced in an isogenic background may be a highly efficient method for functional genome annotation. RESULTS: Here, we have tested the efficacy of this strategy by assessing the extent to which screening for quantitative effects of P elements on sensory bristle number can identify genes affecting neural development. We find that such quantitative screens uncover an unusually large number of genes that are known to function in neural development, as well as genes with yet uncharacterized effects on neural development, and novel loci. CONCLUSIONS: Our findings establish the use of quantitative trait analysis for functional genome annotation through forward genetics. Similar analyses of quantitative effects of P element insertions will facilitate our understanding of the genes affecting many other complex traits in Drosophila.
Augmenting Conceptual Design Trajectory Tradespace Exploration with Graph Theory
NASA Technical Reports Server (NTRS)
Dees, Patrick D.; Zwack, Mathew R.; Steffens, Michael; Edwards, Stephen
2016-01-01
Within conceptual design changes occur rapidly due to a combination of uncertainty and shifting requirements. To stay relevant in this fluid time, trade studies must also be performed rapidly. In order to drive down analysis time while improving the information gained by these studies, surrogate models can be created to represent the complex output of a tool or tools within a specified tradespace. In order to create this model however, a large amount of data must be collected in a short amount of time. By this method, the historical approach of relying on subject matter experts to generate the data required is schedule infeasible. However, by implementing automation and distributed analysis the required data can be generated in a fraction of the time. Previous work focused on setting up a tool called multiPOST capable of orchestrating many simultaneous runs of an analysis tool assessing these automated analyses utilizing heuristics gleaned from the best practices of current subject matter experts. In this update to the previous work, elements of graph theory are included to further drive down analysis time by leveraging data previously gathered. It is shown to outperform the previous method in both time required, and the quantity and quality of data produced.
NASA Technical Reports Server (NTRS)
Knight, Norman F., Jr.; Nemeth, Michael P.; Hilburger, Mark W.
2004-01-01
A technology review and assessment of modeling and analysis efforts underway in support of a safe return to flight of the thermal protection system (TPS) for the Space Shuttle external tank (ET) are summarized. This review and assessment effort focuses on the structural modeling and analysis practices employed for ET TPS foam design and analysis and on identifying analysis capabilities needed in the short-term and long-term. The current understanding of the relationship between complex flight environments and ET TPS foam failure modes are reviewed as they relate to modeling and analysis. A literature review on modeling and analysis of TPS foam material systems is also presented. Finally, a review of modeling and analysis tools employed in the Space Shuttle Program is presented for the ET TPS acreage and close-out foam regions. This review includes existing simplified engineering analysis tools are well as finite element analysis procedures.
Tools for Data Analysis in the Middle School Classroom: A Teacher Professional Development Program
NASA Astrophysics Data System (ADS)
Ledley, T. S.; Haddad, N.; McAuliffe, C.; Dahlman, L.
2006-12-01
In order for students to learn how to engage with scientific data to answer questions about the real world, it is imperative that their teachers are 1) comfortable with the data and the tools used to analyze it, and 2) feel prepared to support their students in this complex endeavor. TERC's Tools for Data Analysis in the Middle School Classroom (DataTools) professional development program, funded by NSF's ITEST program, prepares middle school teachers to integrate Web-based scientific data and analysis tools into their existing curricula. This 13-month program supports teachers in using a set of freely or commonly available tools with a wide range of data. It also gives them an opportunity to practice teaching these skills to students before teaching in their own classrooms. The ultimate goal of the program is to increase the number of middle school students who work directly with scientific data, who use the tools of technology to import, manipulate, visualize and analyze the data, who come to understand the power of data-based arguments, and who will consider pursuing a career in technical and scientific fields. In this session, we will describe the elements of the DataTools program and the Earth Exploration Toolbook (EET, http://serc.carleton.edu/eet), a Web-based resource that supports Earth system education for teachers and students in grades 6 through 16. The EET provides essential support to DataTools teachers as they use it to learn to locate and download Web-based data and use data analysis tools. We will also share what we have learned during the first year of this three-year program.
Axisymmetric analysis of a tube-type acoustic levitator by a finite element method.
Hatano, H
1994-01-01
A finite element approach was taken for the study of the sound field and positioning force in a tube-type acoustic levitator. An axisymmetric model, where a rigid sphere is suspended on the tube axis, was introduced to model a cylindrical chamber of a levitation tube furnace. Distributions of velocity potential, magnitudes of positioning force, and resonance frequency shifts of the chamber due to the presence of the sphere were numerically estimated in relation to the sphere's position and diameter. Experiments were additionally made to compare with the simulation. The finite element method proved to be a useful tool for analyzing and designing the tube-type levitator.
An introduction to Space Weather Integrated Modeling
NASA Astrophysics Data System (ADS)
Zhong, D.; Feng, X.
2012-12-01
The need for a software toolkit that integrates space weather models and data is one of many challenges we are facing with when applying the models to space weather forecasting. To meet this challenge, we have developed Space Weather Integrated Modeling (SWIM) that is capable of analysis and visualizations of the results from a diverse set of space weather models. SWIM has a modular design and is written in Python, by using NumPy, matplotlib, and the Visualization ToolKit (VTK). SWIM provides data management module to read a variety of spacecraft data products and a specific data format of Solar-Interplanetary Conservation Element/Solution Element MHD model (SIP-CESE MHD model) for the study of solar-terrestrial phenomena. Data analysis, visualization and graphic user interface modules are also presented in a user-friendly way to run the integrated models and visualize the 2-D and 3-D data sets interactively. With these tools we can locally or remotely analysis the model result rapidly, such as extraction of data on specific location in time-sequence data sets, plotting interplanetary magnetic field lines, multi-slicing of solar wind speed, volume rendering of solar wind density, animation of time-sequence data sets, comparing between model result and observational data. To speed-up the analysis, an in-situ visualization interface is used to support visualizing the data 'on-the-fly'. We also modified some critical time-consuming analysis and visualization methods with the aid of GPU and multi-core CPU. We have used this tool to visualize the data of SIP-CESE MHD model in real time, and integrated the Database Model of shock arrival, Shock Propagation Model, Dst forecasting model and SIP-CESE MHD model developed by SIGMA Weather Group at State Key Laboratory of Space Weather/CAS.
Rapid State Space Modeling Tool for Rectangular Wing Aeroservoelastic Studies
NASA Technical Reports Server (NTRS)
Suh, Peter M.; Conyers, Howard J.; Mavris, Dimitri N.
2014-01-01
This paper introduces a modeling and simulation tool for aeroservoelastic analysis of rectangular wings with trailing edge control surfaces. The inputs to the code are planform design parameters such as wing span, aspect ratio and number of control surfaces. A doublet lattice approach is taken to compute generalized forces. A rational function approximation is computed. The output, computed in a few seconds, is a state space aeroservoelastic model which can be used for analysis and control design. The tool is fully parameterized with default information so there is little required interaction with the model developer. Although, all parameters can be easily modified if desired.The focus of this paper is on tool presentation, verification and validation. This process is carried out in stages throughout the paper. The rational function approximation is verified against computed generalized forces for a plate model. A model composed of finite element plates is compared to a modal analysis from commercial software and an independently conducted experimental ground vibration test analysis. Aeroservoelastic analysis is the ultimate goal of this tool. Therefore the flutter speed and frequency for a clamped plate are computed using V-g and V-f analysis. The computational results are compared to a previously published computational analysis and wind tunnel results for the same structure. Finally a case study of a generic wing model with a single control surface is presented. Verification of the state space model is presented in comparison to V-g and V-f analysis. This also includes the analysis of the model in response to a 1-cos gust.
Rapid State Space Modeling Tool for Rectangular Wing Aeroservoelastic Studies
NASA Technical Reports Server (NTRS)
Suh, Peter M.; Conyers, Howard J.; Mavris, Dimitri N.
2015-01-01
This paper introduces a modeling and simulation tool for aeroservoelastic analysis of rectangular wings with trailing-edge control surfaces. The inputs to the code are planform design parameters such as wing span, aspect ratio, and number of control surfaces. Using this information, the generalized forces are computed using the doublet-lattice method. Using Roger's approximation, a rational function approximation is computed. The output, computed in a few seconds, is a state space aeroservoelastic model which can be used for analysis and control design. The tool is fully parameterized with default information so there is little required interaction with the model developer. All parameters can be easily modified if desired. The focus of this paper is on tool presentation, verification, and validation. These processes are carried out in stages throughout the paper. The rational function approximation is verified against computed generalized forces for a plate model. A model composed of finite element plates is compared to a modal analysis from commercial software and an independently conducted experimental ground vibration test analysis. Aeroservoelastic analysis is the ultimate goal of this tool, therefore, the flutter speed and frequency for a clamped plate are computed using damping-versus-velocity and frequency-versus-velocity analysis. The computational results are compared to a previously published computational analysis and wind-tunnel results for the same structure. A case study of a generic wing model with a single control surface is presented. Verification of the state space model is presented in comparison to damping-versus-velocity and frequency-versus-velocity analysis, including the analysis of the model in response to a 1-cos gust.
Rapid State Space Modeling Tool for Rectangular Wing Aeroservoelastic Studies
NASA Technical Reports Server (NTRS)
Suh, Peter M.; Conyers, Howard Jason; Mavris, Dimitri N.
2015-01-01
This report introduces a modeling and simulation tool for aeroservoelastic analysis of rectangular wings with trailing-edge control surfaces. The inputs to the code are planform design parameters such as wing span, aspect ratio, and number of control surfaces. Using this information, the generalized forces are computed using the doublet-lattice method. Using Roger's approximation, a rational function approximation is computed. The output, computed in a few seconds, is a state space aeroservoelastic model which can be used for analysis and control design. The tool is fully parameterized with default information so there is little required interaction with the model developer. All parameters can be easily modified if desired. The focus of this report is on tool presentation, verification, and validation. These processes are carried out in stages throughout the report. The rational function approximation is verified against computed generalized forces for a plate model. A model composed of finite element plates is compared to a modal analysis from commercial software and an independently conducted experimental ground vibration test analysis. Aeroservoelastic analysis is the ultimate goal of this tool, therefore, the flutter speed and frequency for a clamped plate are computed using damping-versus-velocity and frequency-versus-velocity analysis. The computational results are compared to a previously published computational analysis and wind-tunnel results for the same structure. A case study of a generic wing model with a single control surface is presented. Verification of the state space model is presented in comparison to damping-versus-velocity and frequency-versus-velocity analysis, including the analysis of the model in response to a 1-cos gust.
Dental Videographic Analysis using Digital Age Media
Seth, Karan; Parmar, Siddharaj; Jhawar, Rahul
2016-01-01
Aims and objectives This study was to evaluate a new method of smile analysis using videographic and photographic softwares (as in this study Photoshop Elements X, Windows Movie Maker 2012) as primary assessment tools and to develop an index for malocclusion and treatment plan that could be used in assessing severity of maloc-clussion. How to cite this article Agarwal A, Seth K, Parmar S, Jhawar R. Dental Videographic Analysis using Digital Age Media. Int J Clin Pediatr Dent 2016;9(4):355-363. PMID:28127169
1988-10-01
Structured Analysis involves building a logical (non-physical) model of a system, using graphic techniques which enable users, analysts, and designers to... Design uses tools, especially graphic ones, to render systems readily understandable. 8 Ř. Structured Design offers a set of strategies for...in the overall systems design process, and an overview of the assessment procedures, as well as a guide to the overall assessment. 20. DISTRIBUTION
Advanced Usage of Vehicle Sketch Pad for CFD-Based Conceptual Design
NASA Technical Reports Server (NTRS)
Ordaz, Irian; Li, Wu
2013-01-01
Conceptual design is the most fluid phase of aircraft design. It is important to be able to perform large scale design space exploration of candidate concepts that can achieve the design intent to avoid more costly configuration changes in later stages of design. This also means that conceptual design is highly dependent on the disciplinary analysis tools to capture the underlying physics accurately. The required level of analysis fidelity can vary greatly depending on the application. Vehicle Sketch Pad (VSP) allows the designer to easily construct aircraft concepts and make changes as the design matures. More recent development efforts have enabled VSP to bridge the gap to high-fidelity analysis disciplines such as computational fluid dynamics and structural modeling for finite element analysis. This paper focuses on the current state-of-the-art geometry modeling for the automated process of analysis and design of low-boom supersonic concepts using VSP and several capability-enhancing design tools.
Introducing Graduate Students to High-Resolution Mass Spectrometry (HRMS) Using a Hands-On Approach
ERIC Educational Resources Information Center
Stock, Naomi L.
2017-01-01
High-resolution mass spectrometry (HRMS) features both high resolution and high mass accuracy and is a powerful tool for the analysis and quantitation of compounds, determination of elemental compositions, and identification of unknowns. A hands-on laboratory experiment for upper-level undergraduate and graduate students to investigate HRMS is…
Got Tools? The Blended Learning Analysis and Design Expediter
ERIC Educational Resources Information Center
Elsenheimer, Jim
2006-01-01
Blended learning is an approach to instructional design that seeks to maximize learning potential by applying the most effective form of instruction for a given program element. The term "blended learning" should not refer to just the mixing of training delivery methods (as it is often defined) but to the orchestrated application and integration…
Overview Of Recent Enhancements To The Bumper-II Meteoroid and Orbital Debris Risk Assessment Tool
NASA Technical Reports Server (NTRS)
Hyde, James L.; Christiansen, Eric L.; Lear, Dana M.; Prior, Thomas G.
2006-01-01
Discussion includes recent enhancements to the BUMPER-II program and input files in support of Shuttle Return to Flight. Improvements to the mesh definitions of the finite element input model will be presented. A BUMPER-II analysis process that was used to estimate statistical uncertainty is introduced.
Beam position monitor engineering
NASA Astrophysics Data System (ADS)
Smith, Stephen R.
1997-01-01
The design of beam position monitors often involves challenging system design choices. Position transducers must be robust, accurate, and generate adequate position signal without unduly disturbing the beam. Electronics must be reliable and affordable, usually while meeting tough requirements on precision, accuracy, and dynamic range. These requirements may be difficult to achieve simultaneously, leading the designer into interesting opportunities for optimization or compromise. Some useful techniques and tools are shown. Both finite element analysis and analytic techniques will be used to investigate quasi-static aspects of electromagnetic fields such as the impedance of and the coupling of beam to striplines or buttons. Finite-element tools will be used to understand dynamic aspects of the electromagnetic fields of beams, such as wake fields and transmission-line and cavity effects in vacuum-to-air feedthroughs. Mathematical modeling of electrical signals through a processing chain will be demonstrated, in particular to illuminate areas where neither a pure time-domain nor a pure frequency-domain analysis is obviously advantageous. Emphasis will be on calculational techniques, in particular on using both time domain and frequency domain approaches to the applicable parts of interesting problems.
Finite element analysis as a design tool for thermoplastic vulcanizate glazing seals
DOE Office of Scientific and Technical Information (OSTI.GOV)
Gase, K.M.; Hudacek, L.L.; Pesevski, G.T.
1998-12-31
There are three materials that are commonly used in commercial glazing seals: EPDM, silicone and thermoplastic vulcanizates (TPVs). TPVs are a high performance class of thermoplastic elastomers (TPEs), where TPEs have elastomeric properties with thermoplastic processability. TPVs have emerged as materials well suited for use in glazing seals due to ease of processing, economics and part design flexibility. The part design and development process is critical to ensure that the chosen TPV provides economics, quality and function in demanding environments. In the design and development process, there is great value in utilizing dual durometer systems to capitalize on the benefitsmore » of soft and rigid materials. Computer-aided design tools, such as Finite Element Analysis (FEA), are effective in minimizing development time and predicting system performance. Examples of TPV glazing seals will illustrate the benefits of utilizing FEA to take full advantage of the material characteristics, which results in functional performance and quality while reducing development iterations. FEA will be performed on two glazing seal profiles to confirm optimum geometry.« less
Structural Embeddings: Mechanization with Method
NASA Technical Reports Server (NTRS)
Munoz, Cesar; Rushby, John
1999-01-01
The most powerful tools for analysis of formal specifications are general-purpose theorem provers and model checkers, but these tools provide scant methodological support. Conversely, those approaches that do provide a well-developed method generally have less powerful automation. It is natural, therefore, to try to combine the better-developed methods with the more powerful general-purpose tools. An obstacle is that the methods and the tools often employ very different logics. We argue that methods are separable from their logics and are largely concerned with the structure and organization of specifications. We, propose a technique called structural embedding that allows the structural elements of a method to be supported by a general-purpose tool, while substituting the logic of the tool for that of the method. We have found this technique quite effective and we provide some examples of its application. We also suggest how general-purpose systems could be restructured to support this activity better.
Pressure distribution under flexible polishing tools. I - Conventional aspheric optics
NASA Astrophysics Data System (ADS)
Mehta, Pravin K.; Hufnagel, Robert E.
1990-10-01
The paper presents a mathematical model, based on Kirchoff's thin flat plate theory, developed to determine polishing pressure distribution for a flexible polishing tool. A two-layered tool in which bending and compressive stiffnesses are equal is developed, which is formulated as a plate on a linearly elastic foundation. An equivalent eigenvalue problem and solution for a free-free plate are created from the plate formulation. For aspheric, anamorphic optical surfaces, the tool misfit is derived; it is defined as the result of movement from the initial perfect fit on the optic to any other position. The Polisher Design (POD) software for circular tools on aspheric optics is introduced. NASTRAN-based finite element analysis results are compared with the POD software, showing high correlation. By employing existing free-free eigenvalues and eigenfunctions, the work may be extended to rectangular polishing tools as well.
Toward designing for trust in database automation
DOE Office of Scientific and Technical Information (OSTI.GOV)
Duez, P. P.; Jamieson, G. A.
Appropriate reliance on system automation is imperative for safe and productive work, especially in safety-critical systems. It is unsafe to rely on automation beyond its designed use; conversely, it can be both unproductive and unsafe to manually perform tasks that are better relegated to automated tools. Operator trust in automated tools mediates reliance, and trust appears to affect how operators use technology. As automated agents become more complex, the question of trust in automation is increasingly important. In order to achieve proper use of automation, we must engender an appropriate degree of trust that is sensitive to changes in operatingmore » functions and context. In this paper, we present research concerning trust in automation in the domain of automated tools for relational databases. Lee and See have provided models of trust in automation. One model developed by Lee and See identifies three key categories of information about the automation that lie along a continuum of attributional abstraction. Purpose-, process-and performance-related information serve, both individually and through inferences between them, to describe automation in such a way as to engender r properly-calibrated trust. Thus, one can look at information from different levels of attributional abstraction as a general requirements analysis for information key to appropriate trust in automation. The model of information necessary to engender appropriate trust in automation [1] is a general one. Although it describes categories of information, it does not provide insight on how to determine the specific information elements required for a given automated tool. We have applied the Abstraction Hierarchy (AH) to this problem in the domain of relational databases. The AH serves as a formal description of the automation at several levels of abstraction, ranging from a very abstract purpose-oriented description to a more concrete description of the resources involved in the automated process. The connection between an AH for an automated tool and a list of information elements at the three levels of attributional abstraction is then direct, providing a method for satisfying information requirements for appropriate trust in automation. In this paper, we will present our method for developing specific information requirements for an automated tool, based on a formal analysis of that tool and the models presented by Lee and See. We will show an example of the application of the AH to automation, in the domain of relational database automation, and the resulting set of specific information elements for appropriate trust in the automated tool. Finally, we will comment on the applicability of this approach to the domain of nuclear plant instrumentation. (authors)« less
NASA Technical Reports Server (NTRS)
Litvin, Faydor L.; Nava, Alessandro; Fan, Qi; Fuentes, Alfonso
2002-01-01
New geometry of face worm gear drives with conical and cylindrical worms is proposed. The generation of the face worm-gear is based on application of a tilted head-cutter (grinding tool) instead of application of a hob applied at present. The generation of a conjugated worm is based on application of a tilted head-cutter (grinding tool) as well. The bearing contact of the gear drive is localized and is oriented longitudinally. A predesigned parabolic function of transmission errors for reduction of noise and vibration is provided. The stress analysis of the gear drive is performed using a three-dimensional finite element analysis. The contacting model is automatically generated. The developed theory is illustrated with numerical examples.
NASA Technical Reports Server (NTRS)
Beck, Louisa R.; Rodriquez, Mario H.; Dister, Sheri W.; Rodriquez, Americo D.; Rejmankova, Eliska; Ulloa, Armando; Meza, Rosa A.; Roberts, Donald R.; Paris, Jack F.; Spanner, Michael A.;
1994-01-01
A landscape approach using remote sensing and Geographic Information System (GIS) technologies was developed to discriminate between villages at high and low risk for malaria transmission, as defined by adult Anopheles albimanus abundance. Satellite data for an area in southern Chiapas, Mexico were digitally processed to generate a map of landscape elements. The GIS processes were used to determine the proportion of mapped landscape elements surrounding 40 villages where An. albimanus data had been collected. The relationships between vector abundance and landscape element proportions were investigated using stepwise discriminant analysis and stepwise linear regression. Both analyses indicated that the most important landscape elements in terms of explaining vector abundance were transitional swamp and unmanaged pasture. Discriminant functions generated for these two elements were able to correctly distinguish between villages with high ind low vector abundance, with an overall accuracy of 90%. Regression results found both transitional swamp and unmanaged pasture proportions to be predictive of vector abundance during the mid-to-late wet season. This approach, which integrates remotely sensed data and GIS capabilities to identify villages with high vector-human contact risk, provides a promising tool for malaria surveillance programs that depend on labor-intensive field techniques. This is particularly relevant in areas where the lack of accurate surveillance capabilities may result in no malaria control action when, in fact, directed action is necessary. In general, this landscape approach could be applied to other vector-borne diseases in areas where: 1. the landscape elements critical to vector survival are known and 2. these elements can be detected at remote sensing scales.
The Application of Ultrasonic Inspection to Crimped Electrical Connections
NASA Technical Reports Server (NTRS)
Cramer, K. Elliott; Perey, Daniel F.; Yost, William T.
2010-01-01
The development of a new ultrasonic measurement technique to quantitatively assess wire crimp terminations is discussed. The development of a prototype instrument, based on a modified, commercially available, crimp tool, is demonstrated for applying this technique when wire crimps are installed. The crimp tool has three separate crimping locations that accommodate the three different ferrule diameters. The crimp tool in this study is capable of crimping wire diameters ranging from 12 to 26 American Wire Gauge (AWG). A transducer design is presented that allows for interrogation of each of the three crimp locations on the crimp tool without reconfiguring the device. An analysis methodology, based on transmitted ultrasonic energy and timing of the first received pulse is shown to correlate to both crimp location in the tool and the AWG of the crimp/ferrule combination. The detectability of a number of the crimp failure pathologies, such as missing strands, partially inserted wires and incomplete crimp compression, is discussed. A wave propagation model, solved by finite element analysis, describes the compressional ultrasonic wave propagation through the junction during the crimping process.
Jonnalagadda, Siddhartha; Gonzalez, Graciela
2010-11-13
BioSimplify is an open source tool written in Java that introduces and facilitates the use of a novel model for sentence simplification tuned for automatic discourse analysis and information extraction (as opposed to sentence simplification for improving human readability). The model is based on a "shot-gun" approach that produces many different (simpler) versions of the original sentence by combining variants of its constituent elements. This tool is optimized for processing biomedical scientific literature such as the abstracts indexed in PubMed. We tested our tool on its impact to the task of PPI extraction and it improved the f-score of the PPI tool by around 7%, with an improvement in recall of around 20%. The BioSimplify tool and test corpus can be downloaded from https://biosimplify.sourceforge.net.
Development of methodology for horizontal axis wind turbine dynamic analysis
NASA Technical Reports Server (NTRS)
Dugundji, J.
1982-01-01
Horizontal axis wind turbine dynamics were studied. The following findings are summarized: (1) review of the MOSTAS computer programs for dynamic analysis of horizontal axis wind turbines; (2) review of various analysis methods for rotating systems with periodic coefficients; (3) review of structural dynamics analysis tools for large wind turbine; (4) experiments for yaw characteristics of a rotating rotor; (5) development of a finite element model for rotors; (6) development of simple models for aeroelastics; and (7) development of simple models for stability and response of wind turbines on flexible towers.
Design, analysis and testing of a new piezoelectric tool actuator for elliptical vibration turning
NASA Astrophysics Data System (ADS)
Lin, Jieqiong; Han, Jinguo; Lu, Mingming; Yu, Baojun; Gu, Yan
2017-08-01
A new piezoelectric tool actuator (PETA) for elliptical vibration turning has been developed based on a hybrid flexure hinge connection. Two double parallel four-bar linkage mechanisms and two right circular flexure hinges were chosen to guide the motion. The two input displacement directional stiffness were modeled according to the principle of virtual work modeling method and the kinematic analysis was conducted theoretically. Finite element analysis was used to carry out static and dynamic analyses. To evaluate the performance of the developed PETA, off-line experimental tests were carried out to investigate the step responses, motion strokes, resolutions, parasitic motions, and natural frequencies of the PETA along the two input directions. The relationship between input displacement and output displacement, as well as the tool tip’s elliptical trajectory in different phase shifts was analyzed. By using the developed PETA mechanism, micro-dimple patterns were generated as the preliminary application to demonstrate the feasibility and efficiency of PETA for elliptical vibration turning.
Neutron Activation Analysis of the Rare Earth Elements (REE) - With Emphasis on Geological Materials
NASA Astrophysics Data System (ADS)
Stosch, Heinz-Günter
2016-08-01
Neutron activation analysis (NAA) has been the analytical method of choice for rare earth element (REE) analysis from the early 1960s through the 1980s. At that time, irradiation facilitieswere widely available and fairly easily accessible. The development of high-resolution gamma-ray detectors in the mid-1960s eliminated, formany applications, the need for chemical separation of the REE from the matrix material, making NAA a reliable and effective analytical tool. While not as precise as isotopedilution mass spectrometry, NAA was competitive by being sensitive for the analysis of about half of the rare earths (La, Ce, Nd, Sm, Eu, Tb, Yb, Lu). The development of inductively coupled plasma mass spectrometry since the 1980s, together with decommissioning of research reactors and the lack of installation of new ones in Europe and North America has led to the rapid decline of NAA.
A comparison of experimental and calculated thin-shell leading-edge buckling due to thermal stresses
NASA Technical Reports Server (NTRS)
Jenkins, Jerald M.
1988-01-01
High-temperature thin-shell leading-edge buckling test data are analyzed using NASA structural analysis (NASTRAN) as a finite element tool for predicting thermal buckling characteristics. Buckling points are predicted for several combinations of edge boundary conditions. The problem of relating the appropriate plate area to the edge stress distribution and the stress gradient is addressed in terms of analysis assumptions. Local plasticity was found to occur on the specimen analyzed, and this tended to simplify the basic problem since it effectively equalized the stress gradient from loaded edge to loaded edge. The initial loading was found to be difficult to select for the buckling analysis because of the transient nature of thermal stress. Multiple initial model loadings are likely required for complicated thermal stress time histories before a pertinent finite element buckling analysis can be achieved. The basic mode shapes determined from experimentation were correctly identified from computation.
Analysis of design tool attributes with regards to sustainability benefits
NASA Astrophysics Data System (ADS)
Zain, S.; Ismail, A. F.; Ahmad, Z.; Adesta, E. Y. T.
2018-01-01
The trend of global manufacturing competitiveness has shown a significant shift from profit and customer driven business to a more harmonious sustainability paradigm. This new direction, which emphasises the interests of three pillars of sustainability, i.e., social, economic and environment dimensions, has changed the ways products are designed. As a result, the roles of design tools in the product development stage of manufacturing in adapting to the new strategy are vital and increasingly challenging. The aim of this paper is to review the literature on the attributes of design tools with regards to the sustainability perspective. Four well-established design tools are selected, namely Quality Function Deployment (QFD), Failure Mode and Element Analysis (FMEA), Design for Six Sigma (DFSS) and Design for Environment (DfE). By analysing previous studies, the main attributes of each design tool and its benefits with respect to each sustainability dimension throughout four stages of product lifecycle are discussed. From this study, it is learnt that each of the design tools contributes to the three pillars of sustainability either directly or indirectly, but they are unbalanced and not holistic. Therefore, the prospective of improving and optimising the design tools is projected, and the possibility of collaboration between the different tools is discussed.
Engine Structural Analysis Software
NASA Technical Reports Server (NTRS)
McKnight, R. L.; Maffeo, R. J.; Schrantz, S.; Hartle, M. S.; Bechtel, G. S.; Lewis, K.; Ridgway, M.; Chamis, Christos C. (Technical Monitor)
2001-01-01
The report describes the technical effort to develop: (1) geometry recipes for nozzles, inlets, disks, frames, shafts, and ducts in finite element form, (2) component design tools for nozzles, inlets, disks, frames, shafts, and ducts which utilize the recipes and (3) an integrated design tool which combines the simulations of the nozzles, inlets, disks, frames, shafts, and ducts with the previously developed combustor, turbine blade, and turbine vane models for a total engine representation. These developments will be accomplished in cooperation and in conjunction with comparable efforts of NASA Glenn Research Center.
3D Feature Extraction for Unstructured Grids
NASA Technical Reports Server (NTRS)
Silver, Deborah
1996-01-01
Visualization techniques provide tools that help scientists identify observed phenomena in scientific simulation. To be useful, these tools must allow the user to extract regions, classify and visualize them, abstract them for simplified representations, and track their evolution. Object Segmentation provides a technique to extract and quantify regions of interest within these massive datasets. This article explores basic algorithms to extract coherent amorphous regions from two-dimensional and three-dimensional scalar unstructured grids. The techniques are applied to datasets from Computational Fluid Dynamics and those from Finite Element Analysis.
Quantification of indium in steel using PIXE
NASA Astrophysics Data System (ADS)
Oliver, A.; Miranda, J.; Rickards, J.; Cheang, J. C.
1989-04-01
The quantitative analysis of steel for endodontics tools was carried out using low-energy protons (≤ 700 keV). A computer program for a thick-target analysis which includes enhancement due to secondary fluorescence was used. In this experiment the L-lines of indium are enhanced due to the proximity of other elements' K-lines to the indium absorption edge. The results show that the ionization cross section expression employed to evaluate this magnitude is important.
NESSUS/EXPERT - An expert system for probabilistic structural analysis methods
NASA Technical Reports Server (NTRS)
Millwater, H.; Palmer, K.; Fink, P.
1988-01-01
An expert system (NESSUS/EXPERT) is presented which provides assistance in using probabilistic structural analysis methods. NESSUS/EXPERT is an interactive menu-driven expert system that provides information to assist in the use of the probabilistic finite element code NESSUS/FEM and the fast probability integrator. NESSUS/EXPERT was developed with a combination of FORTRAN and CLIPS, a C language expert system tool, to exploit the strengths of each language.
Block, Annette; Debode, Frédéric; Grohmann, Lutz; Hulin, Julie; Taverniers, Isabel; Kluga, Linda; Barbau-Piednoir, Elodie; Broeders, Sylvia; Huber, Ingrid; Van den Bulcke, Marc; Heinze, Petra; Berben, Gilbert; Busch, Ulrich; Roosens, Nancy; Janssen, Eric; Žel, Jana; Gruden, Kristina; Morisset, Dany
2013-08-22
Since their first commercialization, the diversity of taxa and the genetic composition of transgene sequences in genetically modified plants (GMOs) are constantly increasing. To date, the detection of GMOs and derived products is commonly performed by PCR-based methods targeting specific DNA sequences introduced into the host genome. Information available regarding the GMOs' molecular characterization is dispersed and not appropriately organized. For this reason, GMO testing is very challenging and requires more complex screening strategies and decision making schemes, demanding in return the use of efficient bioinformatics tools relying on reliable information. The GMOseek matrix was built as a comprehensive, online open-access tabulated database which provides a reliable, comprehensive and user-friendly overview of 328 GMO events and 247 different genetic elements (status: 18/07/2013). The GMOseek matrix is aiming to facilitate GMO detection from plant origin at different phases of the analysis. It assists in selecting the targets for a screening analysis, interpreting the screening results, checking the occurrence of a screening element in a group of selected GMOs, identifying gaps in the available pool of GMO detection methods, and designing a decision tree. The GMOseek matrix is an independent database with effective functionalities in a format facilitating transferability to other platforms. Data were collected from all available sources and experimentally tested where detection methods and certified reference materials (CRMs) were available. The GMOseek matrix is currently a unique and very valuable tool with reliable information on GMOs from plant origin and their present genetic elements that enables further development of appropriate strategies for GMO detection. It is flexible enough to be further updated with new information and integrated in different applications and platforms.
2013-01-01
Background Since their first commercialization, the diversity of taxa and the genetic composition of transgene sequences in genetically modified plants (GMOs) are constantly increasing. To date, the detection of GMOs and derived products is commonly performed by PCR-based methods targeting specific DNA sequences introduced into the host genome. Information available regarding the GMOs’ molecular characterization is dispersed and not appropriately organized. For this reason, GMO testing is very challenging and requires more complex screening strategies and decision making schemes, demanding in return the use of efficient bioinformatics tools relying on reliable information. Description The GMOseek matrix was built as a comprehensive, online open-access tabulated database which provides a reliable, comprehensive and user-friendly overview of 328 GMO events and 247 different genetic elements (status: 18/07/2013). The GMOseek matrix is aiming to facilitate GMO detection from plant origin at different phases of the analysis. It assists in selecting the targets for a screening analysis, interpreting the screening results, checking the occurrence of a screening element in a group of selected GMOs, identifying gaps in the available pool of GMO detection methods, and designing a decision tree. The GMOseek matrix is an independent database with effective functionalities in a format facilitating transferability to other platforms. Data were collected from all available sources and experimentally tested where detection methods and certified reference materials (CRMs) were available. Conclusions The GMOseek matrix is currently a unique and very valuable tool with reliable information on GMOs from plant origin and their present genetic elements that enables further development of appropriate strategies for GMO detection. It is flexible enough to be further updated with new information and integrated in different applications and platforms. PMID:23965170
A spectroscopic tool for identifying sources of origin for materials of military interest
NASA Astrophysics Data System (ADS)
Miziolek, Andrzej W.; De Lucia, Frank C.
2014-05-01
There is a need to identify the source of origin for many items of military interest, including ammunition and weapons that may be circulated and traded in illicit markets. Both fieldable systems (man-portable or handheld) as well as benchtop systems in field and home base laboratories are desired for screening and attribution purposes. Laser Induced Breakdown Spectroscopy (LIBS) continues to show significant capability as a promising new tool for materials identification, matching, and provenance. With the use of the broadband, high resolution spectrometer systems, the LIBS devices can not only determine the elemental inventory of the sample, but they are also capable of elemental fingerprinting to signify sources of origin of various materials. We present the results of an initial study to differentiate and match spent cartridges from different manufacturers and countries. We have found that using Partial Least Squares Discriminant Analysis (PLS-DA) we are able to achieve on average 93.3% True Positives and 5.3% False Positives. These results add to the large body of publications that have demonstrated that LIBS is a particularly suitable tool for source of origin determinations.
NASA Technical Reports Server (NTRS)
Towner, Robert L.; Band, Jonathan L.
2012-01-01
An analysis technique was developed to compare and track mode shapes for different Finite Element Models. The technique may be applied to a variety of structural dynamics analyses, including model reduction validation (comparing unreduced and reduced models), mode tracking for various parametric analyses (e.g., launch vehicle model dispersion analysis to identify sensitivities to modal gain for Guidance, Navigation, and Control), comparing models of different mesh fidelity (e.g., a coarse model for a preliminary analysis compared to a higher-fidelity model for a detailed analysis) and mode tracking for a structure with properties that change over time (e.g., a launch vehicle from liftoff through end-of-burn, with propellant being expended during the flight). Mode shapes for different models are compared and tracked using several numerical indicators, including traditional Cross-Orthogonality and Modal Assurance Criteria approaches, as well as numerical indicators obtained by comparing modal strain energy and kinetic energy distributions. This analysis technique has been used to reliably identify correlated mode shapes for complex Finite Element Models that would otherwise be difficult to compare using traditional techniques. This improved approach also utilizes an adaptive mode tracking algorithm that allows for automated tracking when working with complex models and/or comparing a large group of models.
Adaptation of a Control Center Development Environment for Industrial Process Control
NASA Technical Reports Server (NTRS)
Killough, Ronnie L.; Malik, James M.
1994-01-01
In the control center, raw telemetry data is received for storage, display, and analysis. This raw data must be combined and manipulated in various ways by mathematical computations to facilitate analysis, provide diversified fault detection mechanisms, and enhance display readability. A development tool called the Graphical Computation Builder (GCB) has been implemented which provides flight controllers with the capability to implement computations for use in the control center. The GCB provides a language that contains both general programming constructs and language elements specifically tailored for the control center environment. The GCB concept allows staff who are not skilled in computer programming to author and maintain computer programs. The GCB user is isolated from the details of external subsystem interfaces and has access to high-level functions such as matrix operators, trigonometric functions, and unit conversion macros. The GCB provides a high level of feedback during computation development that improves upon the often cryptic errors produced by computer language compilers. An equivalent need can be identified in the industrial data acquisition and process control domain: that of an integrated graphical development tool tailored to the application to hide the operating system, computer language, and data acquisition interface details. The GCB features a modular design which makes it suitable for technology transfer without significant rework. Control center-specific language elements can be replaced by elements specific to industrial process control.
Design of High Field Solenoids made of High Temperature Superconductors
DOE Office of Scientific and Technical Information (OSTI.GOV)
Bartalesi, Antonio; /Pisa U.
2010-12-01
This thesis starts from the analytical mechanical analysis of a superconducting solenoid, loaded by self generated Lorentz forces. Also, a finite element model is proposed and verified with the analytical results. To study the anisotropic behavior of a coil made by layers of superconductor and insulation, a finite element meso-mechanic model is proposed and designed. The resulting material properties are then used in the main solenoid analysis. In parallel, design work is performed as well: an existing Insert Test Facility (ITF) is adapted and structurally verified to support a coil made of YBa{sub 2}Cu{sub 3}O{sub 7}, a High Temperature Superconductormore » (HTS). Finally, a technological winding process was proposed and the required tooling is designed.« less
An Advanced One-Dimensional Finite Element Model for Incompressible Thermally Expandable Flow
Hu, Rui
2017-03-27
Here, this paper provides an overview of a new one-dimensional finite element flow model for incompressible but thermally expandable flow. The flow model was developed for use in system analysis tools for whole-plant safety analysis of sodium fast reactors. Although the pressure-based formulation was implemented, the use of integral equations in the conservative form ensured the conservation laws of the fluid. A stabilization scheme based on streamline-upwind/Petrov-Galerkin and pressure-stabilizing/Petrov-Galerkin formulations is also introduced. The flow model and its implementation have been verified by many test problems, including density wave propagation, steep gradient problems, discharging between tanks, and the conjugate heatmore » transfer in a heat exchanger.« less
An Advanced One-Dimensional Finite Element Model for Incompressible Thermally Expandable Flow
DOE Office of Scientific and Technical Information (OSTI.GOV)
Hu, Rui
Here, this paper provides an overview of a new one-dimensional finite element flow model for incompressible but thermally expandable flow. The flow model was developed for use in system analysis tools for whole-plant safety analysis of sodium fast reactors. Although the pressure-based formulation was implemented, the use of integral equations in the conservative form ensured the conservation laws of the fluid. A stabilization scheme based on streamline-upwind/Petrov-Galerkin and pressure-stabilizing/Petrov-Galerkin formulations is also introduced. The flow model and its implementation have been verified by many test problems, including density wave propagation, steep gradient problems, discharging between tanks, and the conjugate heatmore » transfer in a heat exchanger.« less
Drivelos, Spiros A; Higgins, Kevin; Kalivas, John H; Haroutounian, Serkos A; Georgiou, Constantinos A
2014-12-15
"Fava Santorinis", is a protected designation of origin (PDO) yellow split pea species growing only in the island of Santorini in Greece. Due to its nutritional quality and taste, it has gained a high monetary value. Thus, it is prone to adulteration with other yellow split peas. In order to discriminate "Fava Santorinis" from other yellow split peas, four classification methods utilising rare earth elements (REEs) measured through inductively coupled plasma-mass spectrometry (ICP-MS) are studied. The four classification processes are orthogonal projection analysis (OPA), Mahalanobis distance (MD), partial least squares discriminant analysis (PLS-DA) and k nearest neighbours (KNN). Since it is known that trace elements are often useful to determine geographical origin of food products, we further quantitated for trace elements using ICP-MS. Presented in this paper are results using the four classification processes based on the fusion of the REEs data with the trace element data. Overall, the OPA method was found to perform best with up to 100% accuracy using the fused data. Copyright © 2014 Elsevier Ltd. All rights reserved.
NASA Technical Reports Server (NTRS)
Schweikhard, W. G.; Singnoi, W. N.
1985-01-01
A two axis thrust measuring system was analyzed by using a finite a element computer program to determine the sensitivities of the thrust vectoring nozzle system to misalignment of the load cells and applied loads, and the stiffness of the structural members. Three models were evaluated: (1) the basic measuring element and its internal calibration load cells; (2) the basic measuring element and its external load calibration equipment; and (3) the basic measuring element, external calibration load frame and the altitude facility support structure. Alignment of calibration loads was the greatest source of error for multiaxis thrust measuring systems. Uniform increases or decreases in stiffness of the members, which might be caused by the selection of the materials, have little effect on the accuracy of the measurements. It is found that the POLO-FINITE program is a viable tool for designing and analyzing multiaxis thrust measurement systems. The response of the test stand to step inputs that might be encountered with thrust vectoring tests was determined. The dynamic analysis show a potential problem for measuring the dynamic response characteristics of thrust vectoring systems because of the inherently light damping of the test stand.
Transterm: a database to aid the analysis of regulatory sequences in mRNAs
Jacobs, Grant H.; Chen, Augustine; Stevens, Stewart G.; Stockwell, Peter A.; Black, Michael A.; Tate, Warren P.; Brown, Chris M.
2009-01-01
Messenger RNAs, in addition to coding for proteins, may contain regulatory elements that affect how the protein is translated. These include protein and microRNA-binding sites. Transterm (http://mRNA.otago.ac.nz/Transterm.html) is a database of regions and elements that affect translation with two major unique components. The first is integrated results of analysis of general features that affect translation (initiation, elongation, termination) for species or strains in Genbank, processed through a standard pipeline. The second is curated descriptions of experimentally determined regulatory elements that function as translational control elements in mRNAs. Transterm focuses on protein binding sites, particularly those in 3′-untranslated regions (3′-UTR). For this release the interface has been extensively updated based on user feedback. The data is now accessible by strain rather than species, for example there are 10 Escherichia coli strains (genomes) analysed separately. In addition to providing a repository of data, the database also provides tools for users to query their own mRNA sequences. Users can search sequences for Transterm or user defined regulatory elements, including protein or miRNA targets. Transterm also provides a central core of links to related resources for complementary analyses. PMID:18984623
Kappen, Peter; Tröger, Larc; Materlik, Gerhard; Reckleben, Christian; Hansen, Karsten; Grunwaldt, Jan-Dierk; Clausen, Bjerne S
2002-07-01
A silicon drift detector (SDD) was used for ex situ and time-resolved in situ fluorescence X-ray absorption fine structure (XAFS) on low-concentrated catalyst samples. For a single-element and a seven-element SDD the energy resolution and the peak-to-background ratio were verified at high count rates, sufficient for fluorescence XAFS. An experimental set-up including the seven-element SDD without any cooling and an in situ cell with gas supply and on-line gas analysis was developed. With this set-up the reduction and oxidation of a zeolite supported catalyst containing 0.3 wt% platinum was followed by fluorescence near-edge scans with a time resolution of 10 min each. From ex situ experiments on low-concentrated platinum- and gold-based catalysts fluorescence XAFS scans could be obtained with sufficient statistical quality for a quantitative analysis. Structural information on the gold and platinum particles could be extracted by both the Fourier transforms and the near-edge region of the XAFS spectra. Moreover, it was found that with the seven-element SDD concentrations of the element of interest as low as 100 ppm can be examined by fluorescence XAFS.
Guo, Liyuan; Wang, Jing
2018-01-04
Here, we present the updated rSNPBase 3.0 database (http://rsnp3.psych.ac.cn), which provides human SNP-related regulatory elements, element-gene pairs and SNP-based regulatory networks. This database is the updated version of the SNP regulatory annotation database rSNPBase and rVarBase. In comparison to the last two versions, there are both structural and data adjustments in rSNPBase 3.0: (i) The most significant new feature is the expansion of analysis scope from SNP-related regulatory elements to include regulatory element-target gene pairs (E-G pairs), therefore it can provide SNP-based gene regulatory networks. (ii) Web function was modified according to data content and a new network search module is provided in the rSNPBase 3.0 in addition to the previous regulatory SNP (rSNP) search module. The two search modules support data query for detailed information (related-elements, element-gene pairs, and other extended annotations) on specific SNPs and SNP-related graphic networks constructed by interacting transcription factors (TFs), miRNAs and genes. (3) The type of regulatory elements was modified and enriched. To our best knowledge, the updated rSNPBase 3.0 is the first data tool supports SNP functional analysis from a regulatory network prospective, it will provide both a comprehensive understanding and concrete guidance for SNP-related regulatory studies. © The Author(s) 2017. Published by Oxford University Press on behalf of Nucleic Acids Research.
EnGeoMAP - geological applications within the EnMAP hyperspectral satellite science program
NASA Astrophysics Data System (ADS)
Boesche, N. K.; Mielke, C.; Rogass, C.; Guanter, L.
2016-12-01
Hyperspectral investigations from near field to space substantially contribute to geological exploration and mining monitoring of raw material and mineral deposits. Due to their spectral characteristics, large mineral occurrences and minefields can be identified from space and the spatial distribution of distinct proxy minerals be mapped. In the frame of the EnMAP hyperspectral satellite science program a mineral and elemental mapping tool was developed - the EnGeoMAP. It contains a basic mineral mapping and a rare earth element mapping approach. This study shows the performance of EnGeoMAP based on simulated EnMAP data of the rare earth element bearing Mountain Pass Carbonatite Complex, USA, and the Rodalquilar and Lomilla Calderas, Spain, which host the economically relevant gold-silver, lead-zinc-silver-gold and alunite deposits. The mountain pass image data was simulated on the basis of AVIRIS Next Generation images, while the Rodalquilar data is based on HyMap images. The EnGeoMAP - Base approach was applied to both images, while the mountain pass image data were additionally analysed using the EnGeoMAP - REE software tool. The results are mineral and elemental maps that serve as proxies for the regional lithology and deposit types. The validation of the maps is based on chemical analyses of field samples. Current airborne sensors meet the spatial and spectral requirements for detailed mineral mapping and future hyperspectral space borne missions will additionally provide a large coverage. For those hyperspectral missions, EnGeoMAP is a rapid data analysis tool that is provided to spectral geologists working in mineral exploration.
Applications of NASTRAN to nuclear problems
NASA Technical Reports Server (NTRS)
Spreeuw, E.
1972-01-01
The extent to which suitable solutions may be obtained for one physics problem and two engineering type problems is traced. NASTRAN appears to be a practical tool to solve one-group steady-state neutron diffusion equations. Transient diffusion analysis may be performed after new levels that allow time-dependent temperature calculations are developed. NASTRAN piecewise linear anlaysis may be applied to solve those plasticity problems for which a smooth stress-strain curve can be used to describe the nonlinear material behavior. The accuracy decreases when sharp transitions in the stress-strain relations are involved. Improved NASTRAN usefulness will be obtained when nonlinear material capabilities are extended to axisymmetric elements and to include provisions for time-dependent material properties and creep analysis. Rigid formats 3 and 5 proved to be very convenient for the buckling and normal-mode analysis of a nuclear fuel element.
Torres-Dowdall, J.; Farmer, A.H.; Abril, M.; Bucher, E.H.; Ridley, I.
2010-01-01
Trace-element analysis has been suggested as a tool for the study of migratory connectivity because (1) trace-element abundance varies spatially in the environment, (2) trace elements are assimilated into animals' tissues through the diet, and (3) current technology permits the analysis of multiple trace elements in a small tissue sample, allowing the simultaneous exploration of several elements. We explored the potential of trace elements (B, Na, Mg, Al, Si, P, S, K, Ca, Ti, Cr, Mn, Ni, Cu, Zn, As, Sr, Cs, Hg, Tl, Pb, Bi, Th, and U) to clarify the migratory connectivity of shorebirds that breed in North America and winter in southern South America. We collected 66 recently replaced secondary feathers from Red Knots (Calidris canutus) at three sites in Patagonia and 76 from White-rumped Sandpipers (C. fuscicollis) at nine sites across Argentina. There were significant differences in trace-element abundance in shorebird feathers grown at different nonbreeding sites, and annual variability within a site was small compared to variability among sites. Across Argentina, there was no large-scale gradient in trace elements. The lack of such a gradient restricts the application of this technique to questions concerning the origin of shorebirds to a small number of discrete sites. Furthermore, our results including three additional species, the Pectoral Sandpiper (C. melanotos), Wilson's Phalarope (Phalaropus tricolor), and Collared Plover (Charadrius collaris), suggest that trace-element profiles change as feathers age. Temporal instability of trace-element values could undermine their application to the study of migratory connectivity in shorebirds. ?? The Cooper Ornithological Society 2010.
INTELLIGENT COMPUTING SYSTEM FOR RESERVOIR ANALYSIS AND RISK ASSESSMENT OF THE RED RIVER FORMATION
DOE Office of Scientific and Technical Information (OSTI.GOV)
Kenneth D. Luff
2002-06-30
Integrated software has been written that comprises the tool kit for the Intelligent Computing System (ICS). Luff Exploration Company is applying these tools for analysis of carbonate reservoirs in the southern Williston Basin. The integrated software programs are designed to be used by small team consisting of an engineer, geologist and geophysicist. The software tools are flexible and robust, allowing application in many environments for hydrocarbon reservoirs. Keystone elements of the software tools include clustering and neural-network techniques. The tools are used to transform seismic attribute data to reservoir characteristics such as storage (phi-h), probable oil-water contacts, structural depths andmore » structural growth history. When these reservoir characteristics are combined with neural network or fuzzy logic solvers, they can provide a more complete description of the reservoir. This leads to better estimates of hydrocarbons in place, areal limits and potential for infill or step-out drilling. These tools were developed and tested using seismic, geologic and well data from the Red River Play in Bowman County, North Dakota and Harding County, South Dakota. The geologic setting for the Red River Formation is shallow-shelf carbonate at a depth from 8000 to 10,000 ft.« less
INTELLIGENT COMPUTING SYSTEM FOR RESERVOIR ANALYSIS AND RISK ASSESSMENT OF THE RED RIVER FORMATION
DOE Office of Scientific and Technical Information (OSTI.GOV)
Kenneth D. Luff
2002-09-30
Integrated software has been written that comprises the tool kit for the Intelligent Computing System (ICS). Luff Exploration Company is applying these tools for analysis of carbonate reservoirs in the southern Williston Basin. The integrated software programs are designed to be used by small team consisting of an engineer, geologist and geophysicist. The software tools are flexible and robust, allowing application in many environments for hydrocarbon reservoirs. Keystone elements of the software tools include clustering and neural-network techniques. The tools are used to transform seismic attribute data to reservoir characteristics such as storage (phi-h), probable oil-water contacts, structural depths andmore » structural growth history. When these reservoir characteristics are combined with neural network or fuzzy logic solvers, they can provide a more complete description of the reservoir. This leads to better estimates of hydrocarbons in place, areal limits and potential for infill or step-out drilling. These tools were developed and tested using seismic, geologic and well data from the Red River Play in Bowman County, North Dakota and Harding County, South Dakota. The geologic setting for the Red River Formation is shallow-shelf carbonate at a depth from 8000 to 10,000 ft.« less
Total Diet Studies as a Tool for Ensuring Food Safety
Lee, Joon-Goo; Kim, Sheen-Hee; Kim, Hae-Jung
2015-01-01
With the diversification and internationalization of the food industry and the increased focus on health from a majority of consumers, food safety policies are being implemented based on scientific evidence. Risk analysis represents the most useful scientific approach for making food safety decisions. Total diet study (TDS) is often used as a risk assessment tool to evaluate exposure to hazardous elements. Many countries perform TDSs to screen for chemicals in foods and analyze exposure trends to hazardous elements. TDSs differ from traditional food monitoring in two major aspects: chemicals are analyzed in food in the form in which it will be consumed and it is cost-effective in analyzing composite samples after processing multiple ingredients together. In Korea, TDSs have been conducted to estimate dietary intakes of heavy metals, pesticides, mycotoxins, persistent organic pollutants, and processing contaminants. TDSs need to be carried out periodically to ensure food safety. PMID:26483881
DOE Office of Scientific and Technical Information (OSTI.GOV)
Hasan, Iftekhar; Husain, Tausif; Sozer, Yilmaz
This paper proposes an analytical machine design tool using magnetic equivalent circuit (MEC)-based particle swarm optimization (PSO) for a double-sided, flux-concentrating transverse flux machine (TFM). The magnetic equivalent circuit method is applied to analytically establish the relationship between the design objective and the input variables of prospective TFM designs. This is computationally less intensive and more time efficient than finite element solvers. A PSO algorithm is then used to design a machine with the highest torque density within the specified power range along with some geometric design constraints. The stator pole length, magnet length, and rotor thickness are the variablesmore » that define the optimization search space. Finite element analysis (FEA) was carried out to verify the performance of the MEC-PSO optimized machine. The proposed analytical design tool helps save computation time by at least 50% when compared to commercial FEA-based optimization programs, with results found to be in agreement with less than 5% error.« less
NASA Technical Reports Server (NTRS)
Lyle, Karen H.
2008-01-01
The Space Shuttle Columbia Accident Investigation Board recommended that NASA develop, validate, and maintain a modeling tool capable of predicting the damage threshold for debris impacts on the Space Shuttle Reinforced Carbon-Carbon (RCC) wing leading edge and nosecap assembly. The results presented in this paper are one part of a multi-level approach that supported the development of the predictive tool used to recertify the shuttle for flight following the Columbia Accident. The assessment of predictive capability was largely based on test analysis comparisons for simpler component structures. This paper provides comparisons of finite element simulations with test data for external tank foam debris impacts onto 6-in. square RCC flat panels. Both quantitative displacement and qualitative damage assessment correlations are provided. The comparisons show good agreement and provided the Space Shuttle Program with confidence in the predictive tool.
Framework for Multidisciplinary Analysis, Design, and Optimization with High-Fidelity Analysis Tools
NASA Technical Reports Server (NTRS)
Orr, Stanley A.; Narducci, Robert P.
2009-01-01
A plan is presented for the development of a high fidelity multidisciplinary optimization process for rotorcraft. The plan formulates individual disciplinary design problems, identifies practical high-fidelity tools and processes that can be incorporated in an automated optimization environment, and establishes statements of the multidisciplinary design problem including objectives, constraints, design variables, and cross-disciplinary dependencies. Five key disciplinary areas are selected in the development plan. These are rotor aerodynamics, rotor structures and dynamics, fuselage aerodynamics, fuselage structures, and propulsion / drive system. Flying qualities and noise are included as ancillary areas. Consistency across engineering disciplines is maintained with a central geometry engine that supports all multidisciplinary analysis. The multidisciplinary optimization process targets the preliminary design cycle where gross elements of the helicopter have been defined. These might include number of rotors and rotor configuration (tandem, coaxial, etc.). It is at this stage that sufficient configuration information is defined to perform high-fidelity analysis. At the same time there is enough design freedom to influence a design. The rotorcraft multidisciplinary optimization tool is built and substantiated throughout its development cycle in a staged approach by incorporating disciplines sequentially.
Finite element modelling for mode-I fracture behaviour of CFRP
NASA Astrophysics Data System (ADS)
Chetan, H. C.; Kattimani, Subhaschandra; Murigendrappa, S. M.
2018-04-01
Debonding is a major failure mechanism in Carbon Fiber Reinforced Polymer (CFRP) due to presence of many adhesion joins, in between many layers. In the current study a finite element simulation is carried out using Virtual Crack Closure Technique (VCCT) and Cohesive Zone Modelling (CZM) using Abaqus as analysis tool. A comparative study is performed in to order analyze convergence of results from CZM and VCCT. It was noted that CZM results matched well with published literature. The results from VCCT were also in good comparison with experimental data of published literature, but were seen to be overestimated. Parametric study is performed to evaluate the variation of input parameters like initial stiffness, element size, peak stress and energy release rate `G'. From the numerical evaluation, it was noted that CZM simulation relies largely on element size and peak stress.
Orbit Design Based on the Global Maps of Telecom Metrics
NASA Technical Reports Server (NTRS)
Lee, Charles H.; Cheung, Kar-Ming; Edwards, Chad; Noreen, Gary K.; Vaisnys, Arvydas
2004-01-01
In this paper we describe an orbit design aide tool, called Telecom Orbit Analysis and Simulation Tool(TOAST). Although it can be used for studying and selecting orbits for any planet, we solely concentrate on its use for Mars. By specifying the six orbital elements for an orbit, a time frame of interest, a horizon mask angle, and some telecom parameters such as the transmitting power, frequency, antenna gains, antenna losses, link margin, received threshold powers for the rates, etc. this tool enables the user to view the animation of the orbit in two and three-dimensional different telecom metrics at any point on the Mars, namely the global planetary map.
Gil, F; Hernández, A F
2015-06-01
Human biomonitoring has become an important tool for the assessment of internal doses of metallic and metalloid elements. These elements are of great significance because of their toxic properties and wide distribution in environmental compartments. Although blood and urine are the most used and accepted matrices for human biomonitoring, other non-conventional samples (saliva, placenta, meconium, hair, nails, teeth, breast milk) may have practical advantages and would provide additional information on health risk. Nevertheless, the analysis of these compounds in biological matrices other than blood and urine has not yet been accepted as a useful tool for biomonitoring. The validation of analytical procedures is absolutely necessary for a proper implementation of non-conventional samples in biomonitoring programs. However, the lack of reliable and useful analytical methodologies to assess exposure to metallic elements, and the potential interference of external contamination and variation in biological features of non-conventional samples are important limitations for setting health-based reference values. The influence of potential confounding factors on metallic concentration should always be considered. More research is needed to ascertain whether or not non-conventional matrices offer definitive advantages over the traditional samples and to broaden the available database for establishing worldwide accepted reference values in non-exposed populations. Copyright © 2015 Elsevier Ltd. All rights reserved.
The intervals method: a new approach to analyse finite element outputs using multivariate statistics
De Esteban-Trivigno, Soledad; Püschel, Thomas A.; Fortuny, Josep
2017-01-01
Background In this paper, we propose a new method, named the intervals’ method, to analyse data from finite element models in a comparative multivariate framework. As a case study, several armadillo mandibles are analysed, showing that the proposed method is useful to distinguish and characterise biomechanical differences related to diet/ecomorphology. Methods The intervals’ method consists of generating a set of variables, each one defined by an interval of stress values. Each variable is expressed as a percentage of the area of the mandible occupied by those stress values. Afterwards these newly generated variables can be analysed using multivariate methods. Results Applying this novel method to the biological case study of whether armadillo mandibles differ according to dietary groups, we show that the intervals’ method is a powerful tool to characterize biomechanical performance and how this relates to different diets. This allows us to positively discriminate between specialist and generalist species. Discussion We show that the proposed approach is a useful methodology not affected by the characteristics of the finite element mesh. Additionally, the positive discriminating results obtained when analysing a difficult case study suggest that the proposed method could be a very useful tool for comparative studies in finite element analysis using multivariate statistical approaches. PMID:29043107
NASA Technical Reports Server (NTRS)
Genuardi, Michael T.
1993-01-01
One strategy for machine-aided indexing (MAI) is to provide a concept-level analysis of the textual elements of documents or document abstracts. In such systems, natural-language phrases are analyzed in order to identify and classify concepts related to a particular subject domain. The overall performance of these MAI systems is largely dependent on the quality and comprehensiveness of their knowledge bases. These knowledge bases function to (1) define the relations between a controlled indexing vocabulary and natural language expressions; (2) provide a simple mechanism for disambiguation and the determination of relevancy; and (3) allow the extension of concept-hierarchical structure to all elements of the knowledge file. After a brief description of the NASA Machine-Aided Indexing system, concerns related to the development and maintenance of MAI knowledge bases are discussed. Particular emphasis is given to statistically-based text analysis tools designed to aid the knowledge base developer. One such tool, the Knowledge Base Building (KBB) program, presents the domain expert with a well-filtered list of synonyms and conceptually-related phrases for each thesaurus concept. Another tool, the Knowledge Base Maintenance (KBM) program, functions to identify areas of the knowledge base affected by changes in the conceptual domain (for example, the addition of a new thesaurus term). An alternate use of the KBM as an aid in thesaurus construction is also discussed.
Bearing tester data compilation, analysis, and reporting and bearing math modeling
NASA Technical Reports Server (NTRS)
1983-01-01
The Shaberth bearing analysis computer program was developed for the analysis of jet engine shaft/bearing systems operating above room temperature with normal hydrocarbon lubricants. It is also possible to use this tool to evaluate the shaft bearing systems operating in cryogenics. Effects such as fluid drag, radial temperature gradients, outer race misalignments and clearance changes were simulated and evaluated. In addition, the speed and preload effects on bearing radial stiffness was evaluated. The Shaberth program was also used to provide contact stresses from which contact geometry was calculated to support other analyses such as the determination of cryogenic fluid film thickness in the contacts and evaluation of surface and subsurface stresses necessary for bearing failure evaluation. This program was a vital tool for the thermal analysis of the bearing in that it provides the heat generation rates at the rolling element/race contacts for input into a thermal model of the bearing/shaft assembly.
A system view and analysis of essential hypertension.
Botzer, Alon; Grossman, Ehud; Moult, John; Unger, Ron
2018-05-01
The goal of this study was to investigate genes associated with essential hypertension from a system perspective, making use of bioinformatic tools to gain insights that are not evident when focusing at a detail-based resolution. Using various databases (pathways, Genome Wide Association Studies, knockouts etc.), we compiled a set of about 200 genes that play a major role in hypertension and identified the interactions between them. This enabled us to create a protein-protein interaction network graph, from which we identified key elements, based on graph centrality analysis. Enriched gene regulatory elements (transcription factors and microRNAs) were extracted by motif finding techniques and knowledge-based tools. We found that the network is composed of modules associated with functions such as water retention, endothelial vasoconstriction, sympathetic activity and others. We identified the transcription factor SP1 and the two microRNAs miR27 (a and b) and miR548c-3p that seem to play a major role in regulating the network as they exert their control over several modules and are not restricted to specific functions. We also noticed that genes involved in metabolic diseases (e.g. insulin) are central to the network. We view the blood-pressure regulation mechanism as a system-of-systems, composed of several contributing subsystems and pathways rather than a single module. The system is regulated by distributed elements. Understanding this mode of action can lead to a more precise treatment and drug target discovery. Our analysis suggests that insulin plays a primary role in hypertension, highlighting the tight link between essential hypertension and diseases associated with the metabolic syndrome.
Queering Student Perspectives: Gender, Sexuality and Activism in School
ERIC Educational Resources Information Center
Elliott, Kathleen O.
2016-01-01
This paper examines the work of a gay-straight alliance and the ways in which members use elements of queer theory to understand their own and others' identities, both to make sense of their experiences and to support their activist efforts. The analysis identifies queer perspectives on gender and sexual identity as useful tools for supporting…
eCAF: A New Tool for the Conversational Analysis of Electronic Communication
ERIC Educational Resources Information Center
Duncan-Howell, Jennifer
2009-01-01
Electronic communication is characteristically concerned with "the message" (eM), those who send them (S), and those who receive and read them (R). This relationship could be simplified into the equation eM = S + R. When this simple equation is applied to electronic communication, several elements are added that make this straightforward act of…
2005-06-01
test, the entire turbulence model was changed from standard k- epsilon to Spalart- Allmaras. Using these different tools of turbulence models, a few...this research, leaving only pre-existing finite element models to be used. At some point a NASTRAN model was developed for vibrations analysis but
Analysing the Theme of Pollution in Portuguese Geography and Biology Textbooks
ERIC Educational Resources Information Center
Tracana, Rosa Branca; Ferreira, Claudia; Ferreira, Maria Eduarda; Carvalho, Graca S.
2008-01-01
Environmental education has been seen as a basic tool to contribute to the change of conceptions, values and attitudes. Textbook analysis is a major element in the evaluation of how the educational goals (at the legislative level of national programmes) are implemented at the school level. The aim of the present study was to analyse the…
"Bridging" Engineering & Art: An Outreach Approach for Middle and High School Students
ERIC Educational Resources Information Center
Asiabanpour, Bahram; DesChamps-Benke, Nicole; Wilson, Thomas; Loerwald, Matthew; Gourgey, Hannah
2010-01-01
This paper describes a novel outreach approach to high school and middle school students to familiarize them with engineering functions and methods. In this approach students participated in a seven-day summer research camp and learned many engineering skills and tools such as CAD solid modeling, finite element analysis, rapid prototyping,…
Alzheimer's Disease under Scrutiny: Short Newspaper Articles as a Case Study Tool.
ERIC Educational Resources Information Center
Hudecki, Michael S.
2001-01-01
After reading a newspaper article on Alzheimer disease, an incurable medical problem involving gradual and debilitating loss of memory, students examine the key elements of the scientific method as conveyed in the story. During their analysis students explore the workings of the nervous system and consider the role of animal model systems in…
The Impact of a Semiotic Analysis Theory-Based Writing Activity on Students' Writing Skills
ERIC Educational Resources Information Center
Sarar Kuzu, Tulay
2016-01-01
Problem Statement: In entering the world intellectually and affectively equipped, humans develop in a systemic way that encompasses both thought and art education and in which written, oral, and visual texts are important tools. In particular, visual literacy, which refers to the interpretation of elements other than written text, including…
The finite element method in low speed aerodynamics
NASA Technical Reports Server (NTRS)
Baker, A. J.; Manhardt, P. D.
1975-01-01
The finite element procedure is shown to be of significant impact in design of the 'computational wind tunnel' for low speed aerodynamics. The uniformity of the mathematical differential equation description, for viscous and/or inviscid, multi-dimensional subsonic flows about practical aerodynamic system configurations, is utilized to establish the general form of the finite element algorithm. Numerical results for inviscid flow analysis, as well as viscous boundary layer, parabolic, and full Navier Stokes flow descriptions verify the capabilities and overall versatility of the fundamental algorithm for aerodynamics. The proven mathematical basis, coupled with the distinct user-orientation features of the computer program embodiment, indicate near-term evolution of a highly useful analytical design tool to support computational configuration studies in low speed aerodynamics.
Human alteration of the rural landscape: Variations in visual perception
DOE Office of Scientific and Technical Information (OSTI.GOV)
Cloquell-Ballester, Vicente-Agustin, E-mail: cloquell@dpi.upv.es; Carmen Torres-Sibille, Ana del; Cloquell-Ballester, Victor-Andres
2012-01-15
The objective of this investigation is to evaluate how visual perception varies as the rural landscape is altered by human interventions of varying character. An experiment is carried out using Semantic Differential Analysis to analyse the effect of the character and the type of the intervention on perception. Interventions are divided into elements of 'permanent industrial character', 'elements of permanent rural character' and 'elements of temporary character', and these categories are sub-divided into smaller groups according to the type of development. To increase the reliability of the results, the Intraclass Correlation Coefficient tool, is applied to validate the semantic spacemore » of the perceptual responses and to determine the number of subjects required for a reliable evaluation of the scenes.« less
Development of an integrated BEM approach for hot fluid structure interaction
NASA Technical Reports Server (NTRS)
Dargush, Gary F.; Banerjee, Prasanta K.; Honkala, Keith A.
1988-01-01
In the present work, the boundary element method (BEM) is chosen as the basic analysis tool, principally because the definition of temperature, flux, displacement and traction are very precise on a boundary-based discretization scheme. One fundamental difficulty is, of course, that a BEM formulation requires a considerable amount of analytical work, which is not needed in the other numerical methods. Progress made toward the development of a boundary element formulation for the study of hot fluid-structure interaction in Earth-to-Orbit engine hot section components is reported. The primary thrust of the program to date has been directed quite naturally toward the examination of fluid flow, since boundary element methods for fluids are at a much less developed state.
The Use Of Computational Human Performance Modeling As Task Analysis Tool
DOE Office of Scientific and Technical Information (OSTI.GOV)
Jacuqes Hugo; David Gertman
2012-07-01
During a review of the Advanced Test Reactor safety basis at the Idaho National Laboratory, human factors engineers identified ergonomic and human reliability risks involving the inadvertent exposure of a fuel element to the air during manual fuel movement and inspection in the canal. There were clear indications that these risks increased the probability of human error and possible severe physical outcomes to the operator. In response to this concern, a detailed study was conducted to determine the probability of the inadvertent exposure of a fuel element. Due to practical and safety constraints, the task network analysis technique was employedmore » to study the work procedures at the canal. Discrete-event simulation software was used to model the entire procedure as well as the salient physical attributes of the task environment, such as distances walked, the effect of dropped tools, the effect of hazardous body postures, and physical exertion due to strenuous tool handling. The model also allowed analysis of the effect of cognitive processes such as visual perception demands, auditory information and verbal communication. The model made it possible to obtain reliable predictions of operator performance and workload estimates. It was also found that operator workload as well as the probability of human error in the fuel inspection and transfer task were influenced by the concurrent nature of certain phases of the task and the associated demand on cognitive and physical resources. More importantly, it was possible to determine with reasonable accuracy the stages as well as physical locations in the fuel handling task where operators would be most at risk of losing their balance and falling into the canal. The model also provided sufficient information for a human reliability analysis that indicated that the postulated fuel exposure accident was less than credible.« less
Frequency Response Studies using Receptance Coupling Approach in High Speed Spindles
NASA Astrophysics Data System (ADS)
Shaik, Jakeer Hussain; Ramakotaiah, K.; Srinivas, J.
2018-01-01
In order to assess the stability of high speed machining, estimate the frequency response at the end of tool tip is of great importance. Evaluating dynamic response of several combinations of integrated spindle-tool holder-tool will consume a lot of time. This paper presents coupled field dynamic response at tool tip for the entire integrated spindle tool unit. The spindle unit is assumed to be relying over the front and rear bearings and investigated using the Timoshenko beam theory to arrive the receptances at different locations of the spindle-tool unit. The responses are further validated with conventional finite element model as well as with the experiments. This approach permits quick outputs without losing accuracy of solution and further these methods are utilized to analyze the various design variables on system dynamics. The results obtained through this analysis are needed to design the better spindle unit in an attempt to reduce the frequency amplitudes at the tool tip to improvise the milling stability during cutting process.
The analysis of bottom forming process for hybrid heating device
NASA Astrophysics Data System (ADS)
Bałon, Paweł; Świątoniowski, Andrzej; Kiełbasa, Bartłomiej
2017-10-01
In this paper the authors present an unusual method for bottom forming applicable for various industrial purposes including the manufacture of water heaters or pressure equipment. This method allows for the fabrication of the bottom of a given piece of stainless steel into a pre-determined shape conforming to the DIN standard which determines the most advantageous dimensions for the bottom cross section in terms of working pressure loading. The authors checked the validity of the method in a numerical and experimental way generating a tool designed to produce bottoms of specified geometry. Many problems are encountered during the design and production of parts, especially excessive sheet wrinkling over a large area of the part. The experiment showed that a lack of experience and numerical analysis in the design of such elements would result in the production of highly wrinkled parts. This defect would render the parts impossible to assemble with the cylindrical part. Many tool shops employ a method for drawing elements with a spherical surface which involves additional spinning, stamping, and grading operations, which greatly increases the cost of parts production. The authors present and compare two forming methods for spherical and parabolic objects, and experimentally confirm the validity of the sheet reversing method with adequate pressure force. The applied method produces parts in one drawing operation and in a following operation that is based on laser or water cutting to obtain a round blank. This reduces the costs of tooling manufacturing by requiring just one tool which can be placed on any hydraulic press with a minimum force of 2 000 kN.
Signal connection for a downhole tool string
Hall, David R.; Hall, Jr., H. Tracy; Pixton, David S.; Bradford, Kline; Fox, Joe; Briscoe, Michael
2006-08-29
A signal transmission connection for a tool string used in exploration and production of natural resources, namely: oil, gas, and geothermal energy resources. The connection comprises first and second annular elements deployed in cooperative association with each other. The respective elements comprise inductive transducers that are capable of two-way signal transmission between each other, with downhole components of the tool string, and with ground-level equipment. The respective inductive transducers comprise one or more conductive loops housed within ferrite troughs, or within ferrite trough segments. When energized, the conductive loops produce a magnetic field suitable for transmitting the signal. The second element may be rotational in drilling applications. The respective elements may be fitted with electronic equipment to aid and manipulate the transmission of the signal. The first element may also be in communication with the World Wide Web.
Statistical analysis and interpolation of compositional data in materials science.
Pesenson, Misha Z; Suram, Santosh K; Gregoire, John M
2015-02-09
Compositional data are ubiquitous in chemistry and materials science: analysis of elements in multicomponent systems, combinatorial problems, etc., lead to data that are non-negative and sum to a constant (for example, atomic concentrations). The constant sum constraint restricts the sampling space to a simplex instead of the usual Euclidean space. Since statistical measures such as mean and standard deviation are defined for the Euclidean space, traditional correlation studies, multivariate analysis, and hypothesis testing may lead to erroneous dependencies and incorrect inferences when applied to compositional data. Furthermore, composition measurements that are used for data analytics may not include all of the elements contained in the material; that is, the measurements may be subcompositions of a higher-dimensional parent composition. Physically meaningful statistical analysis must yield results that are invariant under the number of composition elements, requiring the application of specialized statistical tools. We present specifics and subtleties of compositional data processing through discussion of illustrative examples. We introduce basic concepts, terminology, and methods required for the analysis of compositional data and utilize them for the spatial interpolation of composition in a sputtered thin film. The results demonstrate the importance of this mathematical framework for compositional data analysis (CDA) in the fields of materials science and chemistry.
NASA Astrophysics Data System (ADS)
Potyrailo, Radislav A.; Chisholm, Bret J.; Olson, Daniel R.; Brennan, Michael J.; Molaison, Chris A.
2002-02-01
Design, validation, and implementation of an optical spectroscopic system for high-throughput analysis of combinatorially developed protective organic coatings are reported. Our approach replaces labor-intensive coating evaluation steps with an automated system that rapidly analyzes 8x6 arrays of coating elements that are deposited on a plastic substrate. Each coating element of the library is 10 mm in diameter and 2 to 5 micrometers thick. Performance of coatings is evaluated with respect to their resistance to wear abrasion because this parameter is one of the primary considerations in end-use applications. Upon testing, the organic coatings undergo changes that are impossible to quantitatively predict using existing knowledge. Coatings are abraded using industry-accepted abrasion test methods at single-or multiple-abrasion conditions, followed by high- throughput analysis of abrasion-induced light scatter. The developed automated system is optimized for the analysis of diffusively scattered light that corresponds to 0 to 30% haze. System precision of 0.1 to 2.5% relative standard deviation provides capability for the reliable ranking of coatings performance. While the system was implemented for high-throughput screening of combinatorially developed organic protective coatings for automotive applications, it can be applied to a variety of other applications where materials ranking can be achieved using optical spectroscopic tools.
NASA Astrophysics Data System (ADS)
Piga, Giampaolo; Brunetti, Antonio; Lasio, Barbara; Enzo, Stefano; Malgosa, Assumpció
2014-03-01
We conducted an X-Ray Fluorescence investigation on bone fragments belonging to King Peter III of Aragon and Queen Blanche of Anjou. The spectroscopic analysis was carried out in selected points of the bone fragments. Several transitional elements normally unexpected in the bone composition have been found at varying level of concentration. The presence of these elements was interpreted in relation to chemical treatments for mummification of bodies as well as to dietary habits, including tools used for cooking and for the consumption of food.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Rest, J.; Hofman, G.L.
1997-12-01
The Dispersion Analysis Research Tool (DART) contains models for fission-gas-induced fuel swelling, interaction of fuel with the matrix aluminum, for the resultant reaction-product swelling, and for the calculation of the stress gradient within the fuel particle. The effects of an aluminide shell on fuel particle swelling are evaluated. Validation of the model is demonstrated by a comparison of DART calculations of fuel swelling of U{sub 3}SiAl-Al and U{sub 3}Si{sub 2}-Al for various dispersion fuel element designs with the data.
Analysis of the Requirements Generation Process for the Logistics Analysis and Wargame Support Tool
2017-06-01
For instance, the requirements for a pen seem straight forward; however, they may vary depending on the context in which the pen will be used...the interactions between the operational elements, specify which tasks are dependent on others and the order of executing task, and estimate how...configuration file to call that spreadsheet. This requirement can be met depending on the situation. If the nodes and arcs are pre-defined and readily
NASA Astrophysics Data System (ADS)
Pedersen, N. L.
2015-06-01
The strength of a gear is typically defined relative to durability (pitting) and load capacity (tooth-breakage). Tooth-breakage is controlled by the root shape and this gear part can be designed because there is no contact between gear pairs here. The shape of gears is generally defined by different standards, with the ISO standard probably being the most common one. Gears are manufactured using two principally different tools: rack tools and gear tools. In this work, the bending stress of involute teeth is minimized by shape optimization made directly on the final gear. This optimized shape is then used to find the cutting tool (the gear envelope) that can create this optimized gear shape. A simple but sufficiently flexible root parameterization is applied and emphasis is put on the importance of separating the shape parameterization from the finite element analysis of stresses. Large improvements in the stress level are found.
Transportation Infrastructure Design and Construction \\0x16 Virtual Training Tools
DOT National Transportation Integrated Search
2003-09-01
This project will develop 3D interactive computer-training environments for a major element of transportation infrastructure : hot mix asphalt paving. These tools will include elements of hot mix design (including laboratory equipment) and constructi...
A System for Fault Management and Fault Consequences Analysis for NASA's Deep Space Habitat
NASA Technical Reports Server (NTRS)
Colombano, Silvano; Spirkovska, Liljana; Baskaran, Vijaykumar; Aaseng, Gordon; McCann, Robert S.; Ossenfort, John; Smith, Irene; Iverson, David L.; Schwabacher, Mark
2013-01-01
NASA's exploration program envisions the utilization of a Deep Space Habitat (DSH) for human exploration of the space environment in the vicinity of Mars and/or asteroids. Communication latencies with ground control of as long as 20+ minutes make it imperative that DSH operations be highly autonomous, as any telemetry-based detection of a systems problem on Earth could well occur too late to assist the crew with the problem. A DSH-based development program has been initiated to develop and test the automation technologies necessary to support highly autonomous DSH operations. One such technology is a fault management tool to support performance monitoring of vehicle systems operations and to assist with real-time decision making in connection with operational anomalies and failures. Toward that end, we are developing Advanced Caution and Warning System (ACAWS), a tool that combines dynamic and interactive graphical representations of spacecraft systems, systems modeling, automated diagnostic analysis and root cause identification, system and mission impact assessment, and mitigation procedure identification to help spacecraft operators (both flight controllers and crew) understand and respond to anomalies more effectively. In this paper, we describe four major architecture elements of ACAWS: Anomaly Detection, Fault Isolation, System Effects Analysis, and Graphic User Interface (GUI), and how these elements work in concert with each other and with other tools to provide fault management support to both the controllers and crew. We then describe recent evaluations and tests of ACAWS on the DSH testbed. The results of these tests support the feasibility and strength of our approach to failure management automation and enhanced operational autonomy
2018-01-01
Abstract Here, we present the updated rSNPBase 3.0 database (http://rsnp3.psych.ac.cn), which provides human SNP-related regulatory elements, element-gene pairs and SNP-based regulatory networks. This database is the updated version of the SNP regulatory annotation database rSNPBase and rVarBase. In comparison to the last two versions, there are both structural and data adjustments in rSNPBase 3.0: (i) The most significant new feature is the expansion of analysis scope from SNP-related regulatory elements to include regulatory element–target gene pairs (E–G pairs), therefore it can provide SNP-based gene regulatory networks. (ii) Web function was modified according to data content and a new network search module is provided in the rSNPBase 3.0 in addition to the previous regulatory SNP (rSNP) search module. The two search modules support data query for detailed information (related-elements, element-gene pairs, and other extended annotations) on specific SNPs and SNP-related graphic networks constructed by interacting transcription factors (TFs), miRNAs and genes. (3) The type of regulatory elements was modified and enriched. To our best knowledge, the updated rSNPBase 3.0 is the first data tool supports SNP functional analysis from a regulatory network prospective, it will provide both a comprehensive understanding and concrete guidance for SNP-related regulatory studies. PMID:29140525
Landslide risk models for decision making.
Bonachea, Jaime; Remondo, Juan; de Terán, José Ramón Díaz; González-Díez, Alberto; Cendrero, Antonio
2009-11-01
This contribution presents a quantitative procedure for landslide risk analysis and zoning considering hazard, exposure (or value of elements at risk), and vulnerability. The method provides the means to obtain landslide risk models (expressing expected damage due to landslides on material elements and economic activities in monetary terms, according to different scenarios and periods) useful to identify areas where mitigation efforts will be most cost effective. It allows identifying priority areas for the implementation of actions to reduce vulnerability (elements) or hazard (processes). The procedure proposed can also be used as a preventive tool, through its application to strategic environmental impact analysis (SEIA) of land-use plans. The underlying hypothesis is that reliable predictions about hazard and risk can be made using models based on a detailed analysis of past landslide occurrences in connection with conditioning factors and data on past damage. The results show that the approach proposed and the hypothesis formulated are essentially correct, providing estimates of the order of magnitude of expected losses for a given time period. Uncertainties, strengths, and shortcomings of the procedure and results obtained are discussed and potential lines of research to improve the models are indicated. Finally, comments and suggestions are provided to generalize this type of analysis.
Ensembles of NLP Tools for Data Element Extraction from Clinical Notes
Kuo, Tsung-Ting; Rao, Pallavi; Maehara, Cleo; Doan, Son; Chaparro, Juan D.; Day, Michele E.; Farcas, Claudiu; Ohno-Machado, Lucila; Hsu, Chun-Nan
2016-01-01
Natural Language Processing (NLP) is essential for concept extraction from narrative text in electronic health records (EHR). To extract numerous and diverse concepts, such as data elements (i.e., important concepts related to a certain medical condition), a plausible solution is to combine various NLP tools into an ensemble to improve extraction performance. However, it is unclear to what extent ensembles of popular NLP tools improve the extraction of numerous and diverse concepts. Therefore, we built an NLP ensemble pipeline to synergize the strength of popular NLP tools using seven ensemble methods, and to quantify the improvement in performance achieved by ensembles in the extraction of data elements for three very different cohorts. Evaluation results show that the pipeline can improve the performance of NLP tools, but there is high variability depending on the cohort. PMID:28269947
Ensembles of NLP Tools for Data Element Extraction from Clinical Notes.
Kuo, Tsung-Ting; Rao, Pallavi; Maehara, Cleo; Doan, Son; Chaparro, Juan D; Day, Michele E; Farcas, Claudiu; Ohno-Machado, Lucila; Hsu, Chun-Nan
2016-01-01
Natural Language Processing (NLP) is essential for concept extraction from narrative text in electronic health records (EHR). To extract numerous and diverse concepts, such as data elements (i.e., important concepts related to a certain medical condition), a plausible solution is to combine various NLP tools into an ensemble to improve extraction performance. However, it is unclear to what extent ensembles of popular NLP tools improve the extraction of numerous and diverse concepts. Therefore, we built an NLP ensemble pipeline to synergize the strength of popular NLP tools using seven ensemble methods, and to quantify the improvement in performance achieved by ensembles in the extraction of data elements for three very different cohorts. Evaluation results show that the pipeline can improve the performance of NLP tools, but there is high variability depending on the cohort.
Review of quality assessment tools for the evaluation of pharmacoepidemiological safety studies
Neyarapally, George A; Hammad, Tarek A; Pinheiro, Simone P; Iyasu, Solomon
2012-01-01
Objectives Pharmacoepidemiological studies are an important hypothesis-testing tool in the evaluation of postmarketing drug safety. Despite the potential to produce robust value-added data, interpretation of findings can be hindered due to well-recognised methodological limitations of these studies. Therefore, assessment of their quality is essential to evaluating their credibility. The objective of this review was to evaluate the suitability and relevance of available tools for the assessment of pharmacoepidemiological safety studies. Design We created an a priori assessment framework consisting of reporting elements (REs) and quality assessment attributes (QAAs). A comprehensive literature search identified distinct assessment tools and the prespecified elements and attributes were evaluated. Primary and secondary outcome measures The primary outcome measure was the percentage representation of each domain, RE and QAA for the quality assessment tools. Results A total of 61 tools were reviewed. Most tools were not designed to evaluate pharmacoepidemiological safety studies. More than 50% of the reviewed tools considered REs under the research aims, analytical approach, outcome definition and ascertainment, study population and exposure definition and ascertainment domains. REs under the discussion and interpretation, results and study team domains were considered in less than 40% of the tools. Except for the data source domain, quality attributes were considered in less than 50% of the tools. Conclusions Many tools failed to include critical assessment elements relevant to observational pharmacoepidemiological safety studies and did not distinguish between REs and QAAs. Further, there is a lack of considerations on the relative weights of different domains and elements. The development of a quality assessment tool would facilitate consistent, objective and evidence-based assessments of pharmacoepidemiological safety studies. PMID:23015600
Biased insert for installing data transmission components in downhole drilling pipe
Hall, David R [Provo, UT; Briscoe, Michael A [Lehi, UT; Garner, Kory K [Payson, UT; Wilde, Tyson J [Spanish Fork, UT
2007-04-10
An apparatus for installing data transmission hardware in downhole tools includes an insert insertable into the box end or pin end of drill tool, such as a section of drill pipe. The insert typically includes a mount portion and a slide portion. A data transmission element is mounted in the slide portion of the insert. A biasing element is installed between the mount portion and the slide portion and is configured to create a bias between the slide portion and the mount portion. This biasing element is configured to compensate for varying tolerances encountered in different types of downhole tools. In selected embodiments, the biasing element is an elastomeric material, a spring, compressed gas, or a combination thereof.
Kellogg, Harvey J.; Holm, Robert O.
1983-01-01
A groove refinishing tool which utilizes a finishing wheel which is controlled by an air grinder motor. The air grinder motor is mounted on a main body section which is pivotally attached to a shoe element. The shoe element contains guide pins which guide the shoe element on the groove to be refinished. Application of pressure on the main body element compresses a weight counterbalance spring to extend the finishing wheel through the shoe element to refinish the groove surface. A window is provided for viewing the refinishing operation. Milling operations can also be performed by replacing the finishing wheel with a milling wheel.
Static aeroelastic analysis and tailoring of a single-element racing car wing
NASA Astrophysics Data System (ADS)
Sadd, Christopher James
This thesis presents the research from an Engineering Doctorate research programme in collaboration with Reynard Motorsport Ltd, a manufacturer of racing cars. Racing car wing design has traditionally considered structures to be rigid. However, structures are never perfectly rigid and the interaction between aerodynamic loading and structural flexibility has a direct impact on aerodynamic performance. This interaction is often referred to as static aeroelasticity and the focus of this research has been the development of a computational static aeroelastic analysis method to improve the design of a single-element racing car wing. A static aeroelastic analysis method has been developed by coupling a Reynolds-Averaged Navier-Stokes CFD analysis method with a Finite Element structural analysis method using an iterative scheme. Development of this method has included assessment of CFD and Finite Element analysis methods and development of data transfer and mesh deflection methods. Experimental testing was also completed to further assess the computational analyses. The computational and experimental results show a good correlation and these studies have also shown that a Navier-Stokes static aeroelastic analysis of an isolated wing can be performed at an acceptable computational cost. The static aeroelastic analysis tool was used to assess methods of tailoring the structural flexibility of the wing to increase its aerodynamic performance. These tailoring methods were then used to produce two final wing designs to increase downforce and reduce drag respectively. At the average operating dynamic pressure of the racing car, the computational analysis predicts that the downforce-increasing wing has a downforce of C[1]=-1.377 in comparison to C[1]=-1.265 for the original wing. The computational analysis predicts that the drag-reducing wing has a drag of C[d]=0.115 in comparison to C[d]=0.143 for the original wing.
A new method for measuring the rotational accuracy of rolling element bearings
NASA Astrophysics Data System (ADS)
Chen, Ye; Zhao, Xiangsong; Gao, Weiguo; Hu, Gaofeng; Zhang, Shizhen; Zhang, Dawei
2016-12-01
The rotational accuracy of a machine tool spindle has critical influence upon the geometric shape and surface roughness of finished workpiece. The rotational performance of the rolling element bearings is a main factor which affects the spindle accuracy, especially in the ultra-precision machining. In this paper, a new method is developed to measure the rotational accuracy of rolling element bearings of machine tool spindles. Variable and measurable axial preload is applied to seat the rolling elements in the bearing races, which is used to simulate the operating conditions. A high-precision (radial error is less than 300 nm) and high-stiffness (radial stiffness is 600 N/μm) hydrostatic reference spindle is adopted to rotate the inner race of the test bearing. To prevent the outer race from rotating, a 2-degrees of freedom flexure hinge mechanism (2-DOF FHM) is designed. Correction factors by using stiffness analysis are adopted to eliminate the influences of 2-DOF FHM in the radial direction. Two capacitive displacement sensors with nano-resolution (the highest resolution is 9 nm) are used to measure the radial error motion of the rolling element bearing, without separating the profile error as the traditional rotational accuracy metrology of the spindle. Finally, experimental measurements are performed at different spindle speeds (100-4000 rpm) and axial preloads (75-780 N). Synchronous and asynchronous error motion values are evaluated to demonstrate the feasibility and repeatability of the developed method and instrument.
Comparisons of Kinematics and Dynamics Simulation Software Tools
NASA Technical Reports Server (NTRS)
Shiue, Yeu-Sheng Paul
2002-01-01
Kinematic and dynamic analyses for moving bodies are essential to system engineers and designers in the process of design and validations. 3D visualization and motion simulation plus finite element analysis (FEA) give engineers a better way to present ideas and results. Marshall Space Flight Center (MSFC) system engineering researchers are currently using IGRIP from DELMIA Inc. as a kinematic simulation tool for discrete bodies motion simulations. Although IGRIP is an excellent tool for kinematic simulation with some dynamic analysis capabilities in robotic control, explorations of other alternatives with more powerful dynamic analysis and FEA capabilities are necessary. Kinematics analysis will only examine the displacement, velocity, and acceleration of the mechanism without considering effects from masses of components. With dynamic analysis and FEA, effects such as the forces or torques at the joint due to mass and inertia of components can be identified. With keen market competition, ALGOR Mechanical Event Simulation (MES), MSC visualNastran 4D, Unigraphics Motion+, and Pro/MECHANICA were chosen for explorations. In this study, comparisons between software tools were presented in terms of following categories: graphical user interface (GUI), import capability, tutorial availability, ease of use, kinematic simulation capability, dynamic simulation capability, FEA capability, graphical output, technical support, and cost. Propulsion Test Article (PTA) with Fastrac engine model exported from IGRIP and an office chair mechanism were used as examples for simulations.
Geometric modeling for computer aided design
NASA Technical Reports Server (NTRS)
Schwing, James L.; Olariu, Stephen
1995-01-01
The primary goal of this grant has been the design and implementation of software to be used in the conceptual design of aerospace vehicles particularly focused on the elements of geometric design, graphical user interfaces, and the interaction of the multitude of software typically used in this engineering environment. This has resulted in the development of several analysis packages and design studies. These include two major software systems currently used in the conceptual level design of aerospace vehicles. These tools are SMART, the Solid Modeling Aerospace Research Tool, and EASIE, the Environment for Software Integration and Execution. Additional software tools were designed and implemented to address the needs of the engineer working in the conceptual design environment. SMART provides conceptual designers with a rapid prototyping capability and several engineering analysis capabilities. In addition, SMART has a carefully engineered user interface that makes it easy to learn and use. Finally, a number of specialty characteristics have been built into SMART which allow it to be used efficiently as a front end geometry processor for other analysis packages. EASIE provides a set of interactive utilities that simplify the task of building and executing computer aided design systems consisting of diverse, stand-alone, analysis codes. Resulting in a streamlining of the exchange of data between programs reducing errors and improving the efficiency. EASIE provides both a methodology and a collection of software tools to ease the task of coordinating engineering design and analysis codes.
Development of thermoplastic composite aircraft structures
NASA Technical Reports Server (NTRS)
Renieri, Michael P.; Burpo, Steven J.; Roundy, Lance M.; Todd, Stephanie A.; Kim, H. J.
1992-01-01
Efforts focused on the use of thermoplastic composite materials in the development of structural details associated with an advanced fighter fuselage section with applicability to transport design. In support of these designs, mechanics developments were conducted in two areas. First, a dissipative strain energy approach to material characterization and failure prediction, developed at the Naval Research Laboratory, was evaluated as a design/analysis tool. Second, a finite element formulation for thick composites was developed and incorporated into a lug analysis method which incorporates pin bending effects. Manufacturing concepts were developed for an upper fuel cell cover. A detailed trade study produced two promising concepts: fiber placement and single-step diaphragm forming. Based on the innovative design/manufacturing concepts for the fuselage section primary structure, elements were designed, fabricated, and structurally tested. These elements focused on key issues such as thick composite lugs and low cost forming of fastenerless, stiffener/moldine concepts. Manufacturing techniques included autoclave consolidation, single diaphragm consolidation (SDCC) and roll-forming.
Unification of color postprocessing techniques for 3-dimensional computational mechanics
NASA Technical Reports Server (NTRS)
Bailey, Bruce Charles
1985-01-01
To facilitate the understanding of complex three-dimensional numerical models, advanced interactive color postprocessing techniques are introduced. These techniques are sufficiently flexible so that postprocessing difficulties arising from model size, geometric complexity, response variation, and analysis type can be adequately overcome. Finite element, finite difference, and boundary element models may be evaluated with the prototype postprocessor. Elements may be removed from parent models to be studied as independent subobjects. Discontinuous responses may be contoured including responses which become singular, and nonlinear color scales may be input by the user for the enhancement of the contouring operation. Hit testing can be performed to extract precise geometric, response, mesh, or material information from the database. In addition, stress intensity factors may be contoured along the crack front of a fracture model. Stepwise analyses can be studied, and the user can recontour responses repeatedly, as if he were paging through the response sets. As a system, these tools allow effective interpretation of complex analysis results.
Li, Haiyun; Wang, Zheng
2006-01-01
In this paper, a 3D geometric model of the intervertebral and lumbar disks has been presented, which integrated the spine CT and MRI data-based anatomical structure. Based on the geometric model, a 3D finite element model of an L1-L2 segment was created. Loads, which simulate the pressure from above were applied to the FEM, while a boundary condition describing the relative L1-L2 displacement is imposed on the FEM to account for 3D physiological states. The simulation calculation illustrates the stress and strain distribution and deformation of the spine. The method has two characteristics compared to previous studies: first, the finite element model of the lumbar are based on the data directly derived from medical images such as CTs and MRIs. Second, the result of analysis will be more accurate than using the data of geometric parameters. The FEM provides a promising tool in clinical diagnosis and for optimizing individual therapy in the intervertebral disc herniation.
Finite Element Analysis of Flexural Vibrations in Hard Disk Drive Spindle Systems
NASA Astrophysics Data System (ADS)
LIM, SEUNGCHUL
2000-06-01
This paper is concerned with the flexural vibration analysis of the hard disk drive (HDD) spindle system by means of the finite element method. In contrast to previous research, every system component is here analytically modelled taking into account its structural flexibility and also the centrifugal effect particularly on the disk. To prove the effectiveness and accuracy of the formulated models, commercial HDD systems with two and three identical disks are selected as examples. Then their major natural modes are computed with only a small number of element meshes as the shaft rotational speed is varied, and subsequently compared with the existing numerical results obtained using other methods and newly acquired experimental ones. Based on such a series of studies, the proposed method can be concluded as a very promising tool for the design of HDDs and various other high-performance computer disk drives such as floppy disk drives, CD ROM drives, and their variations having spindle mechanisms similar to those of HDDs.
FAST Mast Structural Response to Axial Loading: Modeling and Verification
NASA Technical Reports Server (NTRS)
Knight, Norman F., Jr.; Elliott, Kenny B.; Templeton, Justin D.; Song, Kyongchan; Rayburn, Jeffery T.
2012-01-01
The International Space Station s solar array wing mast shadowing problem is the focus of this paper. A building-block approach to modeling and analysis is pursued for the primary structural components of the solar array wing mast structure. Starting with an ANSYS (Registered Trademark) finite element model, a verified MSC.Nastran (Trademark) model is established for a single longeron. This finite element model translation requires the conversion of several modeling and analysis features for the two structural analysis tools to produce comparable results for the single-longeron configuration. The model is then reconciled using test data. The resulting MSC.Nastran (Trademark) model is then extended to a single-bay configuration and verified using single-bay test data. Conversion of the MSC. Nastran (Trademark) single-bay model to Abaqus (Trademark) is also performed to simulate the elastic-plastic longeron buckling response of the single bay prior to folding.
Structural Analysis of Composite Flywheels: an Integrated NDE and FEM Approach
NASA Technical Reports Server (NTRS)
Abdul-Aziz, Ali; Baaklini, George; Trudell, Jeffrey
2001-01-01
A structural assessment by integrating finite-element methods (FEM) and a nondestructive evaluation (NDE) of two flywheel rotor assemblies is presented. Composite rotor A is pancake-like with a solid hub design, and composite rotor B is cylindrical with a hollow hub design. Detailed analyses under combined centrifugal and interference-fit loading are performed. Two- and three-dimensional stress analyses and two-dimensional fracture mechanics analyses are conducted. A comparison of the structural analysis results obtained with those extracted via NDE findings is reported. Contact effects due to press-fit conditions are evaluated. Stress results generated from the finite-element analyses were corroborated with the analytical solution. Cracks due to rotational loading up to 48,000 rpm for rotor A and 34,000 rpm for rotor B were successfully imaged with NDE and predicted with FEM and fracture mechanics analyses. A procedure that extends current structural analysis to a life prediction tool is also defined.
NASA Technical Reports Server (NTRS)
Pak, Chan-Gi; Truong, Samson S.
2014-01-01
Small modeling errors in the finite element model will eventually induce errors in the structural flexibility and mass, thus propagating into unpredictable errors in the unsteady aerodynamics and the control law design. One of the primary objectives of Multi Utility Technology Test Bed, X-56A, aircraft is the flight demonstration of active flutter suppression, and therefore in this study, the identification of the primary and secondary modes for the structural model tuning based on the flutter analysis of X-56A. The ground vibration test validated structural dynamic finite element model of the X-56A is created in this study. The structural dynamic finite element model of the X-56A is improved using a model tuning tool. In this study, two different weight configurations of the X-56A have been improved in a single optimization run.
Toward high-speed 3D nonlinear soft tissue deformation simulations using Abaqus software.
Idkaidek, Ashraf; Jasiuk, Iwona
2015-12-01
We aim to achieve a fast and accurate three-dimensional (3D) simulation of a porcine liver deformation under a surgical tool pressure using the commercial finite element software Abaqus. The liver geometry is obtained using magnetic resonance imaging, and a nonlinear constitutive law is employed to capture large deformations of the tissue. Effects of implicit versus explicit analysis schemes, element type, and mesh density on computation time are studied. We find that Abaqus explicit and implicit solvers are capable of simulating nonlinear soft tissue deformations accurately using first-order tetrahedral elements in a relatively short time by optimizing the element size. This study provides new insights and guidance on accurate and relatively fast nonlinear soft tissue simulations. Such simulations can provide force feedback during robotic surgery and allow visualization of tissue deformations for surgery planning and training of surgical residents.
Jantzi, Sarah C; Almirall, José R
2014-01-01
Elemental analysis of soil is a useful application of both laser ablation inductively coupled plasma mass spectrometry (LA-ICP-MS) and laser-induced breakdown spectroscopy (LIBS) in geological, agricultural, environmental, archeological, planetary, and forensic sciences. In forensic science, the question to be answered is often whether soil specimens found on objects (e.g., shoes, tires, or tools) originated from the crime scene or other location of interest. Elemental analysis of the soil from the object and the locations of interest results in a characteristic elemental profile of each specimen, consisting of the amount of each element present. Because multiple elements are measured, multivariate statistics can be used to compare the elemental profiles in order to determine whether the specimen from the object is similar to one of the locations of interest. Previous work involved milling and pressing 0.5 g of soil into pellets before analysis using LA-ICP-MS and LIBS. However, forensic examiners prefer techniques that require smaller samples, are less time consuming, and are less destructive, allowing for future analysis by other techniques. An alternative sample introduction method was developed to meet these needs while still providing quantitative results suitable for multivariate comparisons. The tape-mounting method involved deposition of a thin layer of soil onto double-sided adhesive tape. A comparison of tape-mounting and pellet method performance is reported for both LA-ICP-MS and LIBS. Calibration standards and reference materials, prepared using the tape method, were analyzed by LA-ICP-MS and LIBS. As with the pellet method, linear calibration curves were achieved with the tape method, as well as good precision and low bias. Soil specimens from Miami-Dade County were prepared by both the pellet and tape methods and analyzed by LA-ICP-MS and LIBS. Principal components analysis and linear discriminant analysis were applied to the multivariate data. Results from both the tape method and the pellet method were nearly identical, with clear groupings and correct classification rates of >94%.
Views of general practitioners on the use of STOPP&START in primary care: a qualitative study.
Dalleur, O; Feron, J-M; Spinewine, A
2014-08-01
STOPP (Screening Tool of Older Person's Prescriptions) and START (Screening Tool to Alert Doctors to Right Treatment) criteria aim at detecting potentially inappropriate prescribing in older people. The objective was to explore general practitioners' (GPs) perceptions regarding the use of the STOPP&START tool in their practice. We conducted three focus groups which were conveniently sampled. Vignettes with clinical cases were provided for discussion as well as a full version of the STOPP&START tool. Knowledge, strengths and weaknesses of the tool and its implementation were discussed. Two researchers independently performed content analysis, classifying quotes and creating new categories for emerging themes. Discussions highlighted incentives (e.g. systematic procedure for medication review) and barriers (e.g. time-consuming application) influencing the use of STOPP&START in primary care. Usefulness, comprehensiveness, and relevance of the tool were also questioned. Another important category emerging from the content analysis was the projected use of the tool. The GPs imagined key elements for the implementation in daily practice: computerized clinical decision support system, education, and multidisciplinary collaborations, especially at care transitions and in nursing homes. Despite variables views on the usefulness, comprehensiveness, and relevance of STOPP&START, GPs suggest the implementation of this tool in primary care within computerized clinical decision support systems, through education, and used as part of multidisciplinary collaborations.
Using experimental modal analysis to assess the behaviour of timber elements
NASA Astrophysics Data System (ADS)
Kouroussis, Georges; Fekih, Lassaad Ben; Descamps, Thierry
2018-03-01
Timber frameworks are one of the most important and widespread types of structures. Their configurations and joints are usually complex and require a high level of craftsmanship to assemble. In the field of restoration, a good understanding of the structural behaviour is necessary and is often based on assessment techniques dedicated to wood characterisation. This paper presents the use of experimental modal analysis for finite element updating. To do this, several timber beams in a free supported condition were analysed in order to extract their bending natural characteristics (frequency, damping and mode shapes). Corresponding ABAQUS finite element models were derived which included the effects of local defects (holes, cracks and wood nodes), moisture and structural decay. To achieve the modal updating, additional simulations were performed in order to study the sensitivity of the mechanical parameters. With the intent to estimate their mechanical properties, a procedure of modal updating was carried out in MatLab with a Python script. This was created to extract the modal information from the ABAQUS modal analysis results to be compared with the experimental results. The updating was based on a minimum of unconstrained multivariable function using a derivative-free method. The objective function was selected from the conventional comparison tools (absolute or relative frequency difference, and/or modal assurance criterion). This testing technique was used to determine the dynamic mechanical properties of timber beams, such as the anisotropic Young's Moduli and damping ratio. To verify the modulus, a series of static 4-point bending tests and STS04 classifications were conducted. The results also revealed that local defects have a negligible influence on natural frequencies. The results demonstrate that this assessment tool offers an effective method to obtain the mechanical properties of timber elements, especially when on-site and non-destructive techniques are needed, for example when retrofitting an existing structure.
An expert system for spectroscopic analysis of rocket engine plumes
NASA Technical Reports Server (NTRS)
Reese, Greg; Valenti, Elizabeth; Alphonso, Keith; Holladay, Wendy
1991-01-01
The expert system described in this paper analyzes spectral emissions of rocket engine exhaust plumes and shows major promise for use in engine health diagnostics. Plume emission spectroscopy is an important tool for diagnosing engine anomalies, but it is time-consuming and requires highly skilled personnel. The expert system was created to alleviate such problems. The system accepts a spectral plot in the form of wavelength vs intensity pairs and finds the emission peaks in the spectrum, lists the elemental emitters present in the data and deduces the emitter that produced each peak. The system consists of a conventional language component and a commercially available inference engine that runs on an Apple Macintosh computer. The expert system has undergone limited preliminary testing. It detects elements well and significantly decreases analysis time.
Measuring Surface Bulk Elemental Composition on Venus
NASA Technical Reports Server (NTRS)
Schweitzer, Jeffrey S.; Parsons, Ann M.; Grau, Jim; Lawrence, David J.; McCclanahan, Timothy P.; Miles, Jeffrey; Peplowski, Patrick; Perkins, Luke; Starr, Richard
2017-01-01
The extreme surface environment (462 C, 93 bars pressure) of Venus makes subsurface measurements of its bulk elemental composition extremely challenging. Instruments landed on the surface of Venus must be enclosed in a pressure vessel. The high surface temperatures also require a thermal control system to keep the instrumentation temperatures within their operational range for as long as possible. Since Venus surface probes can currently operate for only a few hours, it is crucial that the lander instrumentation be able to make statistically significant measurements in a short time. An instrument is described that can achieve such a measurement over a volume of thousands of cubic centimeters of material by using high energy penetrating neutron and gamma radiation. The instrument consists of a Pulsed Neutron Generator (PNG) and a Gamma-Ray Spectrometer (GRS). The PNG emits isotropic pulses of 14.1 MeV neutrons that penetrate the pressure vessel walls, the dense atmosphere and the surface rock. The neutrons induce nuclear reactions in the rock to produce gamma rays with energies specific to the element and nuclear process involved. Thus the energies of the detected gamma rays identify the elements present and their intensities provide the abundance of each element. The GRS spectra are analyzed to determine the Venus elemental composition from the spectral signature of individual major, minor, and trace radioactive elements. As a test of such an instrument, a Schlumberger Litho Scanner oil well logging tool was used in a series of experiments at NASA's Goddard Space Flight Center. The Litho Scanner tool was mounted above large (1.8 m x 1.8 m x.9 m) granite and basalt monuments and made a series of one-hour elemental composition measurements in a planar geometry more similar to a planetary lander measurement. Initial analysis of the results shows good agreement with target elemental assays
Measuring Surface Bulk Elemental Composition on Venus
NASA Astrophysics Data System (ADS)
Schweitzer, Jeffrey S.; Parsons, Ann M.; Grau, Jim; Lawrence, David J.; McClanahan, Timothy P.; Miles, Jeffrey; Peplowski, Patrick; Perkins, Luke; Starr, Richard
Bulk elemental composition measurements of the subsurface of Venus are challenging because of the extreme surface environment (462 ˚C, 93 bars pressure). Instruments provided by landed probes on the surface of Venus must therefore be enclosed in a pressure vessel. The high surface temperatures require a thermal control system that keeps the instrumentation and electronics within their operating temperature range for as long as possible. Currently, Venus surface probes can operate for only a few hours. It is therefore crucial that the lander instrumentation be able to make statistically significant measurements in a short time. An instrument is described that can achieve such a measurement over a volume of thousands of cubic centimeters of material by using high energy penetrating neutron and gamma radiation. The instrument consists of a Pulsed Neutron Generator (PNG) and a Gamma-Ray Spectrometer (GRS). The PNG emits isotropic pulses of 14.1 MeV neutrons that penetrate the pressure vessel walls, the dense atmosphere and the surface rock. The neutrons induce nuclear reactions in the rock to produce gamma rays with energies specific to the element and nuclear process involved. Thus the energies of the detected gamma rays identify the elements present and their intensities provide the abundance of each element. The GRS spectra are analyzed to determine the Venus elemental composition from the spectral signature of individual major, minor, and trace radioactive elements. As a test of such an instrument, a Schlumberger Litho Scanner1 oil well logging tool was used in a series of experiments at NASA's Goddard Space Flight Center. The Litho Scanner tool was mounted above large (1.8 m x 1.8 m x .9 m) granite and basalt monuments and made a series of one-hour elemental composition measurements in a planar geometry more similar to a planetary lander measurement. Initial analysis of the results shows good agreement with target elemental assays.
Mashburn, Shana L.; Smith, S. Jerrod
2007-01-01
The U.S. Geological Survey, in cooperation with the Absentee Shawnee Tribe of Oklahoma, began a reconnaissance study of a site in Pottawatomie County, Oklahoma, in 2005 by testing soil, shallow ground water, and plant material for the presence of trace elements and semivolatile organic compounds. Chemical analysis of plant material at the site was investigated as a preliminary tool to determine the extent of contamination at the site. Thirty soil samples were collected from 15 soil cores during October 2005 and analyzed for trace elements and semivolatile organic compounds. Five small-diameter, polyvinyl-chloride-cased wells were installed and ground-water samples were collected during December 2005 and May 2006 and analyzed for trace elements and semivolatile organic compounds. Thirty Johnsongrass samples and 16 Coralberry samples were collected during September 2005 and analyzed for 53 constituents, including trace elements. Results of the soil, ground-water, and plant data indicate that the areas of trace element and semivolatile organic compound contamination are located in the shallow (A-horizon) soils near the threading barn. Most of the trace-element concentrations in the soils on the study site were either similar to or less than trace-element concentrations in background soils. Several trace elements and semivolatile organic compounds exceeded the U.S. Environmental Protection Agency, Region 6, Human Health Medium-Specific Screening Levels 2007 for Tap Water, Residential Soils, Industrial Indoor Soils, and Industrial Outdoor Soils. There was little or no correlation between the plant and soil sample concentrations and the plant and ground-water concentrations based on the current sample size and study design. The lack of correlation between trace-element concentrations in plants and soils, and plants and ground water indicate that plant sampling was not useful as a preliminary tool to assess contamination at the study site.
Tnt1 Retrotransposon Mutagenesis: A Tool for Soybean Functional Genomics1[W][OA
Cui, Yaya; Barampuram, Shyam; Stacey, Minviluz G.; Hancock, C. Nathan; Findley, Seth; Mathieu, Melanie; Zhang, Zhanyuan; Parrott, Wayne A.; Stacey, Gary
2013-01-01
Insertional mutagenesis is a powerful tool for determining gene function in both model and crop plant species. Tnt1, the transposable element of tobacco (Nicotiana tabacum) cell type 1, is a retrotransposon that replicates via an RNA copy that is reverse transcribed and integrated elsewhere in the plant genome. Based on studies in a variety of plants, Tnt1 appears to be inactive in normal plant tissue but can be reactivated by tissue culture. Our goal was to evaluate the utility of the Tnt1 retrotransposon as a mutagenesis strategy in soybean (Glycine max). Experiments showed that the Tnt1 element was stably transformed into soybean plants by Agrobacterium tumefaciens-mediated transformation. Twenty-seven independent transgenic lines carrying Tnt1 insertions were generated. Southern-blot analysis revealed that the copy number of transposed Tnt1 elements ranged from four to 19 insertions, with an average of approximately eight copies per line. These insertions showed Mendelian segregation and did not transpose under normal growth conditions. Analysis of 99 Tnt1 flanking sequences revealed insertions into 62 (62%) annotated genes, indicating that the element preferentially inserts into protein-coding regions. Tnt1 insertions were found in all 20 soybean chromosomes, indicating that Tnt1 transposed throughout the soybean genome. Furthermore, fluorescence in situ hybridization experiments validated that Tnt1 inserted into multiple chromosomes. Passage of transgenic lines through two different tissue culture treatments resulted in Tnt1 transposition, significantly increasing the number of insertions per line. Thus, our data demonstrate the Tnt1 retrotransposon to be a powerful system that can be used for effective large-scale insertional mutagenesis in soybean. PMID:23124322
Thyssen, G M; Holtkamp, M; Kaulfürst-Soboll, H; Wehe, C A; Sperling, M; von Schaewen, A; Karst, U
2017-06-21
Laser ablation-inductively coupled plasma-optical emission spectroscopy (LA-ICP-OES) is presented as a valuable tool for elemental bioimaging of alkali and earth alkali elements in plants. Whereas LA-ICP-OES is commonly used for micro analysis of solid samples, laser ablation-inductively coupled plasma-mass spectrometry (LA-ICP-MS) has advanced to the gold standard for bioimaging. However, especially for easily excitable and ubiquitous elements such as alkali and earth alkali elements, LA-ICP-OES holds some advantages regarding simultaneous detection, costs, contamination, and user-friendliness. This is demonstrated by determining the calcium, sodium and potassium distribution in tobacco plant stem and leaf petiole tissues. A quantification of the calcium contents in a concentration range up to 1000 μg g -1 using matrix-matched standards is presented as well. The method is directly compared to a LA-ICP-MS approach by analyzing parallel slices of the same samples.
2016-12-01
based on life expectancy and the TSP account selected. The TSP Growth and Annuity Element also estimates how taxes will increase at the time the service...the BRS. 14. SUBJECT TERMS military retirement, blended retirement, HIGH-36, thrift savings plan, investment risk, retirement taxes , net present...20 2. Tax Impacts
Box compression analysis of world-wide data spanning 46 years
Thomas J. Urbanik; Benjamin Frank
2006-01-01
The state of the art among most industry citations of box compression estimation is the equation by McKee developed in 1963. Because of limitations in computing tools at the time the McKee equation was developed, the equation is a simplification, with many constraints, of a more general relationship. By applying the results of sophisticated finite element modeling, in...
Integrating Space Systems Operations at the Marine Expeditionary Force Level
2015-06-01
Electromagnetic Interference ENVI Environment for Visualizing Images EW Electronic Warfare xvi FA40 Space Operations Officer FEC Fires and Effects...Information Facility SFE Space Force Enhancement SIGINT Signals Intelligence SSA Space Situational Awareness SSE Space Support Element STK Systems...April 23, 2015. 65 • GPS Interference and Navigation Tool (GIANT) for providing GPS accuracy prediction reports • Systems Toolkit ( STK ) Analysis
ERIC Educational Resources Information Center
Asiabanpour, Bahram
2010-01-01
In this paper a novel outreach approach to high school students to familiarize them with engineering functions and methods is explained. In this approach students participated in a seven days research camp and learned many engineering skills and tools such as CAD solid modeling, finite element analysis, rapid prototyping, mechanical tests, team…
Design and Analysis of Bionic Cutting Blades Using Finite Element Method.
Li, Mo; Yang, Yuwang; Guo, Li; Chen, Donghui; Sun, Hongliang; Tong, Jin
2015-01-01
Praying mantis is one of the most efficient predators in insect world, which has a pair of powerful tools, two sharp and strong forelegs. Its femur and tibia are both armed with a double row of strong spines along their posterior edges which can firmly grasp the prey, when the femur and tibia fold on each other in capturing. These spines are so sharp that they can easily and quickly cut into the prey. The geometrical characteristic of the praying mantis's foreleg, especially its tibia, has important reference value for the design of agricultural soil-cutting tools. Learning from the profile and arrangement of these spines, cutting blades with tooth profile were designed in this work. Two different sizes of tooth structure and arrangement were utilized in the design on the cutting edge. A conventional smooth-edge blade was used to compare with the bionic serrate-edge blades. To compare the working efficiency of conventional blade and bionic blades, 3D finite element simulation analysis and experimental measurement were operated in present work. Both the simulation and experimental results indicated that the bionic serrate-edge blades showed better performance in cutting efficiency.
Design and Analysis of Bionic Cutting Blades Using Finite Element Method
Li, Mo; Yang, Yuwang; Guo, Li; Chen, Donghui; Sun, Hongliang; Tong, Jin
2015-01-01
Praying mantis is one of the most efficient predators in insect world, which has a pair of powerful tools, two sharp and strong forelegs. Its femur and tibia are both armed with a double row of strong spines along their posterior edges which can firmly grasp the prey, when the femur and tibia fold on each other in capturing. These spines are so sharp that they can easily and quickly cut into the prey. The geometrical characteristic of the praying mantis's foreleg, especially its tibia, has important reference value for the design of agricultural soil-cutting tools. Learning from the profile and arrangement of these spines, cutting blades with tooth profile were designed in this work. Two different sizes of tooth structure and arrangement were utilized in the design on the cutting edge. A conventional smooth-edge blade was used to compare with the bionic serrate-edge blades. To compare the working efficiency of conventional blade and bionic blades, 3D finite element simulation analysis and experimental measurement were operated in present work. Both the simulation and experimental results indicated that the bionic serrate-edge blades showed better performance in cutting efficiency. PMID:27019583
Expressions of cultural safety in public health nursing practice.
Richardson, Anna; Yarwood, Judy; Richardson, Sandra
2017-01-01
Cultural safety is an essential concept within New Zealand nursing that is formally linked to registration and competency-based practice certification. Despite its centrality to New Zealand nursing philosophies and the stated expectation of cultural safety as a practice element, there is limited evidence of its application in the literature. This research presents insight into public health nurse's (PHN) experiences, demonstrating the integration of cultural safety principles into practice. These findings emerged following secondary analysis of data from a collaborative, educative research project where PHNs explored the use of family assessment tools. In particular, the 15-minute interview tool was introduced and used by the PHNs when working with families. Critical analysis of transcribed data from PHN interviews, utilising a cultural safety lens, illuminated practical ways in which cultural safety concepts infused PHN practice with families. The themes that emerged reflected the interweaving of the principles of cultural safety with the application of the five components of the 15-minute interview. This highlights elements of PHN work with individuals and families not previously acknowledged. Examples of culturally safe nursing practice resonated throughout the PHN conversations as they grappled with the increasing complexity of working with a diverse range of families. © 2016 John Wiley & Sons Ltd.
NASA Astrophysics Data System (ADS)
Thibault, Fradet; Grégory, Quenet; Kevin, Manchuel
2014-05-01
Historical studies, including historical seismicity analysis, deal with historical documents. Numerous factors, such as culture, social condition, demography, political situations and opinions or religious ones influence the way the events are transcribed in the archives. As a consequence, it is crucial to contextualize and compare the historical documents reporting on a given event in order to reduce the uncertainties affecting their analysis and interpretation. When studying historical seismic events it is often tricky to have a global view of all the information provided by the historical documents. It is also difficult to extract cross-correlated information from the documents and draw a precise historical context. Use of cartographic and geographic tools in GIS software is the best tool for the synthesis, interpretation and contextualization of the historical material. The main goal is to produce the most complete dataset of available information, in order to take into account all the components of the historical context and consequently improve the macroseismic analysis. The Entre-Deux-Mers earthquake (1759, Iepc= VII-VIII) [SISFRANCE 2013 - EDF-IRSN-BRGM] is well documented but has never benefited from a cross-analysis of historical documents and historical context elements. The map of available intensity data from SISFRANCE highlights a gap in macroseismic information within the estimated epicentral area. The aim of this study is to understand the origin of this gap by making a cartographic compilation of both, archive information and historical context elements. The results support the hypothesis that the lack of documents and macroseismic data in the epicentral area is related to a low human activity rather than low seismic effects in this zone. Topographic features, geographical position, flood hazard, roads and pathways locations, vineyards distribution and the forester coverage, mentioned in the archives and reported on the Cassini's map confirm this hypothesis. The location of the recently explored documentary sources evidence that there was no notarial activity in this particular area at that time. The importance of the economic and political dominance of Bordeaux during the XVIIth-XVIIIth centuries has to be taken into account in order to apprehend the way the earthquake was reported by the population at the regional scale. Elements related to chimneys forms or construction techniques could in turn help in identifying regional peculiarities allowing better quantifying the vulnerability aspects of the region.
Kennedy Space Center Orion Processing Team Planning for Ground Operations
NASA Technical Reports Server (NTRS)
Letchworth, Gary; Schlierf, Roland
2011-01-01
Topics in this presentation are: Constellation Ares I/Orion/Ground Ops Elements Orion Ground Operations Flow Orion Operations Planning Process and Toolset Overview, including: 1 Orion Concept of Operations by Phase 2 Ops Analysis Capabilities Overview 3 Operations Planning Evolution 4 Functional Flow Block Diagrams 5 Operations Timeline Development 6 Discrete Event Simulation (DES) Modeling 7 Ground Operations Planning Document Database (GOPDb) Using Operations Planning Tools for Operability Improvements includes: 1 Kaizen/Lean Events 2 Mockups 3 Human Factors Analysis
Hu, Zhi-yu; Zhang, Lei; Ma, Wei-guang; Yan, Xiao-juan; Li, Zhi-xin; Zhang, Yong-zhi; Wang, Le; Dong, Lei; Yin, Wang-bao; Jia, Suo-tang
2012-03-01
Self-designed identifying software for LIBS spectral line was introduced. Being integrated with LabVIEW, the soft ware can smooth spectral lines and pick peaks. The second difference and threshold methods were employed. Characteristic spectrum of several elements matches the NIST database, and realizes automatic spectral line identification and qualitative analysis of the basic composition of sample. This software can analyze spectrum handily and rapidly. It will be a useful tool for LIBS.
Engine dynamic analysis with general nonlinear finite element codes
NASA Technical Reports Server (NTRS)
Adams, M. L.; Padovan, J.; Fertis, D. G.
1991-01-01
A general engine dynamic analysis as a standard design study computational tool is described for the prediction and understanding of complex engine dynamic behavior. Improved definition of engine dynamic response provides valuable information and insights leading to reduced maintenance and overhaul costs on existing engine configurations. Application of advanced engine dynamic simulation methods provides a considerable cost reduction in the development of new engine designs by eliminating some of the trial and error process done with engine hardware development.
1988-11-01
system, using graphic techniques which enable users, analysts, and designers to get a clear and common picture of the system and how its parts fit...boxes into hierarchies suitable for computer implementation. ŗ. Structured Design uses tools, especially graphic ones, to render systems readily...LSA, PROCESSES, DATA FLOWS, DATA STORES, EX"RNAL ENTITIES, OVERALL SYSTEMS DESIGN PROCESS, over 19, ABSTRACT (Continue on reverse if necessary and
1981-05-01
represented as a Winkler foundation. The program can treat any number of slabs connected by steel bars or other load trans- fer devices at the joints...dimensional finite element method. The inherent flexibility of such an approach permits the analysis of a rigid pavement with steel bars and stabilized...layers and provides an efficient tool for analyzing stress conditions at the joint. Unfor- tunately, such a procedure would require a tremendously
ImageParser: a tool for finite element generation from three-dimensional medical images
Yin, HM; Sun, LZ; Wang, G; Yamada, T; Wang, J; Vannier, MW
2004-01-01
Background The finite element method (FEM) is a powerful mathematical tool to simulate and visualize the mechanical deformation of tissues and organs during medical examinations or interventions. It is yet a challenge to build up an FEM mesh directly from a volumetric image partially because the regions (or structures) of interest (ROIs) may be irregular and fuzzy. Methods A software package, ImageParser, is developed to generate an FEM mesh from 3-D tomographic medical images. This software uses a semi-automatic method to detect ROIs from the context of image including neighboring tissues and organs, completes segmentation of different tissues, and meshes the organ into elements. Results The ImageParser is shown to build up an FEM model for simulating the mechanical responses of the breast based on 3-D CT images. The breast is compressed by two plate paddles under an overall displacement as large as 20% of the initial distance between the paddles. The strain and tangential Young's modulus distributions are specified for the biomechanical analysis of breast tissues. Conclusion The ImageParser can successfully exact the geometry of ROIs from a complex medical image and generate the FEM mesh with customer-defined segmentation information. PMID:15461787
The ENCODE Project at UC Santa Cruz.
Thomas, Daryl J; Rosenbloom, Kate R; Clawson, Hiram; Hinrichs, Angie S; Trumbower, Heather; Raney, Brian J; Karolchik, Donna; Barber, Galt P; Harte, Rachel A; Hillman-Jackson, Jennifer; Kuhn, Robert M; Rhead, Brooke L; Smith, Kayla E; Thakkapallayil, Archana; Zweig, Ann S; Haussler, David; Kent, W James
2007-01-01
The goal of the Encyclopedia Of DNA Elements (ENCODE) Project is to identify all functional elements in the human genome. The pilot phase is for comparison of existing methods and for the development of new methods to rigorously analyze a defined 1% of the human genome sequence. Experimental datasets are focused on the origin of replication, DNase I hypersensitivity, chromatin immunoprecipitation, promoter function, gene structure, pseudogenes, non-protein-coding RNAs, transcribed RNAs, multiple sequence alignment and evolutionarily constrained elements. The ENCODE project at UCSC website (http://genome.ucsc.edu/ENCODE) is the primary portal for the sequence-based data produced as part of the ENCODE project. In the pilot phase of the project, over 30 labs provided experimental results for a total of 56 browser tracks supported by 385 database tables. The site provides researchers with a number of tools that allow them to visualize and analyze the data as well as download data for local analyses. This paper describes the portal to the data, highlights the data that has been made available, and presents the tools that have been developed within the ENCODE project. Access to the data and types of interactive analysis that are possible are illustrated through supplemental examples.
Using task analysis to understand the Data System Operations Team
NASA Technical Reports Server (NTRS)
Holder, Barbara E.
1994-01-01
The Data Systems Operations Team (DSOT) currently monitors the Multimission Ground Data System (MGDS) at JPL. The MGDS currently supports five spacecraft and within the next five years, it will support ten spacecraft simultaneously. The ground processing element of the MGDS consists of a distributed UNIX-based system of over 40 nodes and 100 processes. The MGDS system provides operators with little or no information about the system's end-to-end processing status or end-to-end configuration. The lack of system visibility has become a critical issue in the daily operation of the MGDS. A task analysis was conducted to determine what kinds of tools were needed to provide DSOT with useful status information and to prioritize the tool development. The analysis provided the formality and structure needed to get the right information exchange between development and operations. How even a small task analysis can improve developer-operator communications is described, and the challenges associated with conducting a task analysis in a real-time mission operations environment are examined.
Power flow as a complement to statistical energy analysis and finite element analysis
NASA Technical Reports Server (NTRS)
Cuschieri, J. M.
1987-01-01
Present methods of analysis of the structural response and the structure-borne transmission of vibrational energy use either finite element (FE) techniques or statistical energy analysis (SEA) methods. The FE methods are a very useful tool at low frequencies where the number of resonances involved in the analysis is rather small. On the other hand SEA methods can predict with acceptable accuracy the response and energy transmission between coupled structures at relatively high frequencies where the structural modal density is high and a statistical approach is the appropriate solution. In the mid-frequency range, a relatively large number of resonances exist which make finite element method too costly. On the other hand SEA methods can only predict an average level form. In this mid-frequency range a possible alternative is to use power flow techniques, where the input and flow of vibrational energy to excited and coupled structural components can be expressed in terms of input and transfer mobilities. This power flow technique can be extended from low to high frequencies and this can be integrated with established FE models at low frequencies and SEA models at high frequencies to form a verification of the method. This method of structural analysis using power flo and mobility methods, and its integration with SEA and FE analysis is applied to the case of two thin beams joined together at right angles.
Mammographic x-ray unit kilovoltage test tool based on k-edge absorption effect.
Napolitano, Mary E; Trueblood, Jon H; Hertel, Nolan E; David, George
2002-09-01
A simple tool to determine the peak kilovoltage (kVp) of a mammographic x-ray unit has been designed. Tool design is based on comparing the effect of k-edge discontinuity of the attenuation coefficient for a series of element filters. Compatibility with the mammography accreditation phantom (MAP) to obtain a single quality control film is a second design objective. When the attenuation of a series of sequential elements is studied simultaneously, differences in the absorption characteristics due to the k-edge discontinuities are more evident. Specifically, when the incident photon energy is higher than the k-edge energy of a number of the elements and lower than the remainder, an inflection may be seen in the resulting attenuation data. The maximum energy of the incident photon spectra may be determined based on this inflection point for a series of element filters. Monte Carlo photon transport analysis was used to estimate the photon transmission probabilities for each of the sequential k-edge filter elements. The photon transmission corresponds directly to optical density recorded on mammographic x-ray film. To observe the inflection, the element filters chosen must have k-edge energies that span a range greater than the expected range of the end point energies to be determined. For the design, incident x-ray spectra ranging from 25 to 40 kVp were assumed to be from a molybdenum target. Over this range, the k-edge energy changes by approximately 1.5 keV between sequential elements. For this design 21 elements spanning an energy range from 20 to 50 keV were chosen. Optimum filter element thicknesses were calculated to maximize attenuation differences at the k-edge while maintaining optical densities between 0.10 and 3.00. Calculated relative transmission data show that the kVp could be determined to within +/-1 kV. To obtain experimental data, a phantom was constructed containing 21 different elements placed in an acrylic holder. MAP images were used to determine appropriate exposure techniques for a series of end point energies from 25 to 35 kVp. The average difference between the kVp determination and the calibrated dial setting was 0.8 and 1.0 kV for a Senographe 600 T and a Senographe DMR, respectively. Since the k-edge absorption energies of the filter materials are well known, independent calibration or a series of calibration curves is not required.
Split-tapered joint clamping device
Olsen, Max J.; Schwartz, Jr., John F.
1988-01-01
This invention relates to a clamping device for removably attaching a tool element to a bracket element wherein a bracket element is disposed in a groove in the tool and a clamping member is disposed in said groove and in engagement with a clamping face of the bracket and a wall of the groove and with the clamping member having pivot means engaging the bracket and about which the clamping member rotates.
Electromagnetic variable degrees of freedom actuator systems and methods
Montesanti, Richard C [Pleasanton, CA; Trumper, David L [Plaistow, NH; Kirtley, Jr., James L.
2009-02-17
The present invention provides a variable reluctance actuator system and method that can be adapted for simultaneous rotation and translation of a moving element by applying a normal-direction magnetic flux on the moving element. In a beneficial example arrangement, the moving element includes a swing arm that carries a cutting tool at a set radius from an axis of rotation so as to produce a rotary fast tool servo that provides a tool motion in a direction substantially parallel to the surface-normal of a workpiece at the point of contact between the cutting tool and workpiece. An actuator rotates a swing arm such that a cutting tool moves toward and away from a mounted rotating workpiece in a controlled manner in order to machine the workpiece. Position sensors provide rotation and displacement information for a swing arm to a control system. A control system commands and coordinates motion of the fast tool servo with the motion of a spindle, rotating table, cross-feed slide, and in feed slide of a precision lathe.
Pressure distribution under flexible polishing tools. II - Cylindrical (conical) optics
NASA Astrophysics Data System (ADS)
Mehta, Pravin K.
1990-10-01
A previously developed eigenvalue model is extended to determine polishing pressure distribution by rectangular tools with unequal stiffness in two directions on cylindrical optics. Tool misfit is divided into two simplified one-dimensional problems and one simplified two-dimensional problem. Tools with nonuniform cross-sections are treated with a new one-dimensional eigenvalue algorithm, permitting evaluation of tool designs where the edge is more flexible than the interior. This maintains edge pressure variations within acceptable parameters. Finite element modeling is employed to resolve upper bounds, which handle pressure changes in the two-dimensional misfit element. Paraboloids and hyperboloids from the NASA AXAF system are treated with the AXAFPOD software for this method, and are verified with NASTRAN finite element analyses. The maximum deviation from the one-dimensional azimuthal pressure variation is predicted to be 10 percent and 20 percent for paraboloids and hyperboloids, respectively.
Prototype Development of a Tradespace Analysis Tool for Spaceflight Medical Resources.
Antonsen, Erik L; Mulcahy, Robert A; Rubin, David; Blue, Rebecca S; Canga, Michael A; Shah, Ronak
2018-02-01
The provision of medical care in exploration-class spaceflight is limited by mass, volume, and power constraints, as well as limitations of available skillsets of crewmembers. A quantitative means of exploring the risks and benefits of inclusion or exclusion of onboard medical capabilities may help to inform the development of an appropriate medical system. A pilot project was designed to demonstrate the utility of an early tradespace analysis tool for identifying high-priority resources geared toward properly equipping an exploration mission medical system. Physician subject matter experts identified resources, tools, and skillsets required, as well as associated criticality scores of the same, to meet terrestrial, U.S.-specific ideal medical solutions for conditions concerning for exploration-class spaceflight. A database of diagnostic and treatment actions and resources was created based on this input and weighed against the probabilities of mission-specific medical events to help identify common and critical elements needed in a future exploration medical capability. Analysis of repository data demonstrates the utility of a quantitative method of comparing various medical resources and skillsets for future missions. Directed database queries can provide detailed comparative estimates concerning likelihood of resource utilization within a given mission and the weighted utility of tangible and intangible resources. This prototype tool demonstrates one quantitative approach to the complex needs and limitations of an exploration medical system. While this early version identified areas for refinement in future version development, more robust analysis tools may help to inform the development of a comprehensive medical system for future exploration missions.Antonsen EL, Mulcahy RA, Rubin D, Blue RS, Canga MA, Shah R. Prototype development of a tradespace analysis tool for spaceflight medical resources. Aerosp Med Hum Perform. 2018; 89(2):108-114.
NASA Technical Reports Server (NTRS)
Quinlan, Jesse R.; Gern, Frank H.
2016-01-01
Simultaneously achieving the fuel consumption and noise reduction goals set forth by NASA's Environmentally Responsible Aviation (ERA) project requires innovative and unconventional aircraft concepts. In response, advanced hybrid wing body (HWB) aircraft concepts have been proposed and analyzed as a means of meeting these objectives. For the current study, several HWB concepts were analyzed using the Hybrid wing body Conceptual Design and structural optimization (HCDstruct) analysis code. HCDstruct is a medium-fidelity finite element based conceptual design and structural optimization tool developed to fill the critical analysis gap existing between lower order structural sizing approaches and detailed, often finite element based sizing methods for HWB aircraft concepts. Whereas prior versions of the tool used a half-model approach in building the representative finite element model, a full wing-tip-to-wing-tip modeling capability was recently added to HCDstruct, which alleviated the symmetry constraints at the model centerline in place of a free-flying model and allowed for more realistic center body, aft body, and wing loading and trim response. The latest version of HCDstruct was applied to two ERA reference cases, including the Boeing Open Rotor Engine Integration On an HWB (OREIO) concept and the Boeing ERA-0009H1 concept, and results agreed favorably with detailed Boeing design data and related Flight Optimization System (FLOPS) analyses. Following these benchmark cases, HCDstruct was used to size NASA's ERA HWB concepts and to perform a related scaling study.
Casco-Robles, Martin Miguel; Miura, Tomoya; Chiba, Chikafumi
2015-06-01
The adult newt has the ability to regenerate the neural retina following injury, a process achieved primarily by the retinal pigment epithelium (RPE). To deliver exogenous genes to the RPE for genetic manipulation of regenerative events, we isolated the newt RPE65 promoter region by genome walking. First, we cloned the 2.8 kb RPE65 promoter from the newt, Cynops pyrrhogaster. Sequence analysis revealed several conserved regulatory elements described previously in mouse and human RPE65 promoters. Second, having previously established an I-SceI-mediated transgenic protocol for the newt, we used it here to examine the -657 bp proximal promoter of RPE65. The promoter assay used with F0 transgenic newts confirmed transgene expression of mCherry fluorescent protein in the RPE. Using bioinformatic tools and the TRANSFAC database, we identified a 340 bp CpG island located between -635 and -296 bp in the promoter; this region contains response elements for the microphthalmia-associated transcription factor known as MITF (CACGTG, CATGTG), and E-boxes (CANNTG). Sex-determining region box 9 (or SOX9) response element previously reported in the regulation of RPE genes (including RPE65) was also identified in the newt RPE65 promoter. Third, we identified DNA motif boxes in the newt RPE65 promoter that are conserved among other vertebrates. The newt RPE65 promoter is an invaluable tool for site-specific delivery of exogenous genes or genetic manipulation systems for the study of retinal regeneration in this animal.
NASA Astrophysics Data System (ADS)
Griesel, S.; Mundry, R.; Kakuschke, A.; Fonfara, S.; Siebert, U.; Prange, A.
2006-11-01
Mineral and essential trace elements are involved in numerous physiological processes in mammals. Often, diseases are associated with an imbalance of the electrolyte homeostasis. In this study, the concentrations of mineral elements (P, S, K, Ca) and essential trace elements (Fe, Cu, Zn, Se, Rb, Sr) in whole blood of harbor seals ( Phoca vitulina) were determined using total-reflection X-ray fluorescence spectrometry (TXRF). Samples from 81 free-ranging harbor seals from the North Sea and two captive seals were collected during 2003-2005. Reference ranges and element correlations for health status determination were derived for P, S, K, Ca, Fe, Cu, and Zn level in whole blood. Grouping the seals by age, gender and sample location the concentration levels of the elements were compared. The blood from two captive seals with signs of diseases and four free-ranging seals showed reduced element levels of P, S, and Ca and differences in element correlation of electrolytes were ascertained. Thus, simultaneous measurements of several elements in only 500 μL volumes of whole blood provide the possibility to obtain information on both, the electrolyte balance and the hydration status of the seals. The method could therefore serve as an additional biomonitoring tool for the health assessment.
Zhang, Chaosheng
2006-08-01
Galway is a small but rapidly growing tourism city in western Ireland. To evaluate its environmental quality, a total of 166 surface soil samples (0-10 cm depth) were collected from parks and grasslands at the density of 1 sample per 0.25 km2 at the end of 2004. All samples were analysed using ICP-AES for the near-total concentrations of 26 chemical elements. Multivariate statistics and GIS techniques were applied to classify the elements and to identify elements influenced by human activities. Cluster analysis (CA) and principal component analysis (PCA) classified the elements into two groups: the first group predominantly derived from natural sources, the second being influenced by human activities. GIS mapping is a powerful tool in identifying the possible sources of pollutants. Relatively high concentrations of Cu, Pb and Zn were found in the city centre, old residential areas, and along major traffic routes, showing significant effects of traffic pollution. The element As is enriched in soils of the old built-up areas, which can be attributed to coal and peat combustion for home heating. Such significant spatial patterns of pollutants displayed by urban soils may imply potential health threat to residents of the contaminated areas of the city.
Konaté, Ahmed Amara; Ma, Huolin; Pan, Heping; Qin, Zhen; Ahmed, Hafizullah Abba; Dembele, N'dji Dit Jacques
2017-10-01
The availability of a deep well that penetrates deep into the Ultra High Pressure (UHP) metamorphic rocks is unusual and consequently offers a unique chance to study the metamorphic rocks. One such borehole is located in the southern part of Donghai County in the Sulu UHP metamorphic belt of Eastern China, from the Chinese Continental Scientific Drilling Main hole. This study reports the results obtained from the analysis of oxide log data. A geochemical logging tool provides in situ, gamma ray spectroscopy measurements of major and trace elements in the borehole. Dry weight percent oxide concentration logs obtained for this study were SiO 2 , K 2 O, TiO 2 , H 2 O, CO 2 , Na 2 O, Fe 2 O 3 , FeO, CaO, MnO, MgO, P 2 O 5 and Al 2 O 3 . Cross plot and Principal Component Analysis methods were applied for lithology characterization and mineralogy description respectively. Cross plot analysis allows lithological variations to be characterized. Principal Component Analysis shows that the oxide logs can be summarized by two components related to the feldspar and hydrous minerals. This study has shown that geochemical logging tool data is accurate and adequate to be tremendously useful in UHP metamorphic rocks analysis. Copyright © 2017 Elsevier Ltd. All rights reserved.
Lee, Haofu; Nguyen, Alan; Hong, Christine; Hoang, Paul; Pham, John; Ting, Kang
2017-01-01
Introduction The aims of this study were to evaluate the effects of rapid palatal expansion on the craniofacial skeleton of a patient with unilateral cleft lip and palate (UCLP) and to predict the points of force application for optimal expansion using a 3-dimensional finite element model. Methods A 3-dimensional finite element model of the craniofacial complex with UCLP was generated from spiral computed tomographic scans with imaging software (Mimics, version 13.1; Materialise, Leuven, Belgium). This model was imported into the finite element solver (version 12.0; ANSYS, Canonsburg, Pa) to evaluate transverse expansion forces from rapid palatal expansion. Finite element analysis was performed with transverse expansion to achieve 5 mm of anterolateral expansion of the collapsed minor segment to simulate correction of the anterior crossbite in a patient with UCLP. Results High-stress concentrations were observed at the body of the sphenoid, medial to the orbit, and at the inferior area of the zygomatic process of the maxilla. The craniofacial stress distribution was asymmetric, with higher stress levels on the cleft side. When forces were applied more anteriorly on the collapsed minor segment and more posteriorly on the major segment, there was greater expansion of the anterior region of the minor segment with minimal expansion of the major segment. Conclusions The transverse expansion forces from rapid palatal expansion are distributed to the 3 maxillary buttresses. Finite element analysis is an appropriate tool to study and predict the points of force application for better controlled expansion in patients with UCLP. PMID:27476365
Lee, Haofu; Nguyen, Alan; Hong, Christine; Hoang, Paul; Pham, John; Ting, Kang
2016-08-01
The aims of this study were to evaluate the effects of rapid palatal expansion on the craniofacial skeleton of a patient with unilateral cleft lip and palate (UCLP) and to predict the points of force application for optimal expansion using a 3-dimensional finite element model. A 3-dimensional finite element model of the craniofacial complex with UCLP was generated from spiral computed tomographic scans with imaging software (Mimics, version 13.1; Materialise, Leuven, Belgium). This model was imported into the finite element solver (version 12.0; ANSYS, Canonsburg, Pa) to evaluate transverse expansion forces from rapid palatal expansion. Finite element analysis was performed with transverse expansion to achieve 5 mm of anterolateral expansion of the collapsed minor segment to simulate correction of the anterior crossbite in a patient with UCLP. High-stress concentrations were observed at the body of the sphenoid, medial to the orbit, and at the inferior area of the zygomatic process of the maxilla. The craniofacial stress distribution was asymmetric, with higher stress levels on the cleft side. When forces were applied more anteriorly on the collapsed minor segment and more posteriorly on the major segment, there was greater expansion of the anterior region of the minor segment with minimal expansion of the major segment. The transverse expansion forces from rapid palatal expansion are distributed to the 3 maxillary buttresses. Finite element analysis is an appropriate tool to study and predict the points of force application for better controlled expansion in patients with UCLP. Copyright © 2016 American Association of Orthodontists. Published by Elsevier Inc. All rights reserved.
Wisenthal, Adeena; Krupa, Terry
2014-12-12
Mental health related work disability leaves are increasing at alarming rates with depression emerging as the most common mental disorder in the workforce. Treatments are available to alleviate depressive symptoms and associated functional impacts; however, they are not specifically aimed at preparing people to return to work. Cognitive work hardening (CWH) is a novel intervention that addresses this gap in the health care system. This paper presents a theoretical analysis of the components and underlying mechanisms of CWH using Intervention Mapping (IM) as a tool to deconstruct its elements. The cognitive sequelae of depression and their relevance to return-to-work (RTW) are examined together with interpersonal skills and other work-related competencies that affect work ability. IM, a tool typically used to create programs, is used to deconstruct an existing program, namely CWH, into its component parts and link them to theories and models in the literature. CWH has been deconstructed into intervention elements which are linked to program performance objectives through underlying theoretical models. In this way, linkages are made between tools and materials of the intervention and the overall program objective of 'successful RTW for people with depression'. An empirical study of the efficacy of CWH is currently underway which should provide added insight and understanding into this intervention. The application of IM to CWH illustrates the theoretical underpinnings of the treatment intervention and assists with better understanding the linkage between intervention elements and intervention objective. Applying IM to deconstruct an existing program (rather than create a program) presents an alternate application of the IM tool which can have implications for other programs in terms of enhancing understanding, grounding in theoretical foundations, communicating program design, and establishing a basis for program evaluation and improvement.
A software platform for continuum modeling of ion channels based on unstructured mesh
NASA Astrophysics Data System (ADS)
Tu, B.; Bai, S. Y.; Chen, M. X.; Xie, Y.; Zhang, L. B.; Lu, B. Z.
2014-01-01
Most traditional continuum molecular modeling adopted finite difference or finite volume methods which were based on a structured mesh (grid). Unstructured meshes were only occasionally used, but an increased number of applications emerge in molecular simulations. To facilitate the continuum modeling of biomolecular systems based on unstructured meshes, we are developing a software platform with tools which are particularly beneficial to those approaches. This work describes the software system specifically for the simulation of a typical, complex molecular procedure: ion transport through a three-dimensional channel system that consists of a protein and a membrane. The platform contains three parts: a meshing tool chain for ion channel systems, a parallel finite element solver for the Poisson-Nernst-Planck equations describing the electrodiffusion process of ion transport, and a visualization program for continuum molecular modeling. The meshing tool chain in the platform, which consists of a set of mesh generation tools, is able to generate high-quality surface and volume meshes for ion channel systems. The parallel finite element solver in our platform is based on the parallel adaptive finite element package PHG which wass developed by one of the authors [1]. As a featured component of the platform, a new visualization program, VCMM, has specifically been developed for continuum molecular modeling with an emphasis on providing useful facilities for unstructured mesh-based methods and for their output analysis and visualization. VCMM provides a graphic user interface and consists of three modules: a molecular module, a meshing module and a numerical module. A demonstration of the platform is provided with a study of two real proteins, the connexin 26 and hemolysin ion channels.
NASA Astrophysics Data System (ADS)
Zhang, Airui; Jin, Axiang; Wang, Hai; Wang, Xiaokang; Zha, Pengfei; Wang, Meiling; Song, Xiaoping; Gao, Sitian
2018-03-01
Quantitative determination of trace elements like S, Fe, Cu, Mn and Pb in gasoline and S in diesel is of great importance due to the growing concerns over air pollution, human health and engine failure caused by utilization of gasoline and diesel with these harmful elements. A method of total reflection X-ray fluorescence (TXRF) was developed to measure these harmful trace elements in gasoline and diesel. A variety of factors to affect measurement results, including TXRF parameters, microwave-assisted digestion conditions and internal standard element and its addition, were examined to optimize these experimental procedures. The hydrophobic treatment of the surface of quartz reflectors to support the analyte with neutral silicone solutions could prepare thin films of gasoline and diesel digestion solutions for subsequent TXRF analysis. The proposed method shows good potential and reliability to determine the content of harmful trace elements in gasoline and diesel with high sensitivity and accuracy without drawing different standard calibration curves, and can be easily employed to screen gasoline and diesel in routine quality control and assurance.
Experimental analysis of synchronization and dynamics in an automobile as a complex system
NASA Astrophysics Data System (ADS)
González-Cruz, C. A.; Jáuregui-Correa, J. C.; López-Cajún, C.; Sen, M.; Domínguez-González, A.
2015-08-01
A complex system is composed of many interacting elements, and its behavior, as a whole, can be quite different from that of the individual elements. An automobile is an example of a common mechanical system composed of a large number of individual elements. These elements are connected through soft and hard linkages that transmit motion to each other. This paper proposes a variety of analytical tools to study experimental data from complex systems using two elements of an automobile as an example. Accelerometer measurements were taken from two elements within an automobile: the door and the dashboard. Two types of data were collected: response to impact loading, and response to road excitation of the automobile driven at different speeds. The signals were processed via Fourier and wavelet transforms, cross-correlation coefficients, Hilbert transform, and Kuramoto order parameters. A new parameter, called the order-deficit parameter, is introduced. Considerable, but not complete, synchronization can be found between the accelerations measured at these two locations in the automobile, and the degree of synchronization is quantified using the order-deficit parameter.
Trace element study in scallop shells by laser ablation ICP-MS: the example of Ba/Ca ratios
NASA Astrophysics Data System (ADS)
Lorrain, A.; Pécheyran, C.; Paulet, Y.-M.; Chauvaud, L.; Amouroux, D.; Krupp, E.; Donard, O.
2003-04-01
As scallop shells grow incrementally at a rate of one line per day, environmental changes could then be evidenced on a daily basis. As an example for trace element incorporation studies, barium is a geochemical tracer that can be directly related to oceanic primary productivity. Hence, monitoring Ba/Ca variations in a scallop shell should give information about phytoplanktonic events encountered day by day during its life. The very high spatial resolution (typically 40 - 200 µm) and the high elemental sensitivity required can only be achieved by the combination of laser ablation coupled to inductively coupled plasma mass spectrometry. This study demonstrates that Laser ablation coupled to ICP-MS determination is a relevant tool for high resolution distribution measurement of trace elements in calcite matrix. The ablation strategy related to single line rastering and calcium normalisation were found to be the best analytical conditions in terms of reproducibility and sensitivity. The knowledge of P. maximus growth rings periodicity (daily), combined with LA-ICP-MS micro analysis allows the acquisition of time dated profiles with high spatial and thus temporal resolution. This resolution makes P. maximus a potential tool for environmental reconstruction and especially for accurate calibration of proxies. However, the relations among Ba/Ca peaks and phytoplanktonic events differed according to the animals and some inter-annual discrepancies complexify the interpretation.
NASA Astrophysics Data System (ADS)
Carrera, E.; Miglioretti, F.; Petrolo, M.
2011-11-01
This paper compares and evaluates various plate finite elements to analyse the static response of thick and thin plates subjected to different loading and boundary conditions. Plate elements are based on different assumptions for the displacement distribution along the thickness direction. Classical (Kirchhoff and Reissner-Mindlin), refined (Reddy and Kant), and other higher-order displacement fields are implemented up to fourth-order expansion. The Unified Formulation UF by the first author is used to derive finite element matrices in terms of fundamental nuclei which consist of 3×3 arrays. The MITC4 shear-locking free type formulation is used for the FE approximation. Accuracy of a given plate element is established in terms of the error vs. thickness-to-length parameter. A significant number of finite elements for plates are implemented and compared using displacement and stress variables for various plate problems. Reduced models that are able to detect the 3D solution are built and a Best Plate Diagram (BPD) is introduced to give guidelines for the construction of plate theories based on a given accuracy and number of terms. It is concluded that the UF is a valuable tool to establish, for a given plate problem, the most accurate FE able to furnish results within a certain accuracy range. This allows us to obtain guidelines and recommendations in building refined elements in the bending analysis of plates for various geometries, loadings, and boundary conditions.
'Big Data' Collaboration: Exploring, Recording and Sharing Enterprise Knowledge
DOE Office of Scientific and Technical Information (OSTI.GOV)
Sukumar, Sreenivas R; Ferrell, Regina Kay
2013-01-01
As data sources and data size proliferate, knowledge discovery from "Big Data" is starting to pose several challenges. In this paper, we address a specific challenge in the practice of enterprise knowledge management while extracting actionable nuggets from diverse data sources of seemingly-related information. In particular, we address the challenge of archiving knowledge gained through collaboration, dissemination and visualization as part of the data analysis, inference and decision-making lifecycle. We motivate the implementation of an enterprise data-discovery and knowledge recorder tool, called SEEKER based on real world case-study. We demonstrate SEEKER capturing schema and data-element relationships, tracking the data elementsmore » of value based on the queries and the analytical artifacts that are being created by analysts as they use the data. We show how the tool serves as digital record of institutional domain knowledge and a documentation for the evolution of data elements, queries and schemas over time. As a knowledge management service, a tool like SEEKER saves enterprise resources and time by avoiding analytic silos, expediting the process of multi-source data integration and intelligently documenting discoveries from fellow analysts.« less
Two dimensional finite element thermal model of laser surface glazing for H13 tool steel
NASA Astrophysics Data System (ADS)
Kabir, I. R.; Yin, D.; Naher, S.
2016-10-01
A two dimensional (2D) transient thermal model with line-heat-source was developed by Finite Element Method (FEM) for laser surface glazing of H13 tool steel using commercial software-ANSYS 15. The geometry of the model was taken as a transverse circular cross-section of cylindrical specimen. Two different power levels (300W, 200W) were used with 0.2mm width of laser beam and 0.15ms exposure time. Temperature distribution, heating and cooling rates, and the dimensions of modified surface were analysed. The maximum temperatures achieved were 2532K (2259°C) and 1592K (1319°C) for laser power 300W and 200W respectively. The maximum cooling rates were 4.2×107 K/s for 300W and 2×107 K/s for 200W. Depths of modified zone increased with increasing laser power. From this analysis, it can be predicted that for 0.2mm beam width and 0.15ms time exposer melting temperature of H13 tool steel is achieved within 200-300W power range of laser beam in laser surface glazing.
Mass spectrometry in Earth sciences: the precise and accurate measurement of time.
Schaltegger, Urs; Wotzlaw, Jörn-Frederik; Ovtcharova, Maria; Chiaradia, Massimo; Spikings, Richard
2014-01-01
Precise determinations of the isotopic compositions of a variety of elements is a widely applied tool in Earth sciences. Isotope ratios are used to quantify rates of geological processes that occurred during the previous 4.5 billion years, and also at the present time. An outstanding application is geochronology, which utilizes the production of radiogenic daughter isotopes by the radioactive decay of parent isotopes. Geochronological tools, involving isotopic analysis of selected elements from smallest volumes of minerals by thermal ionization mass spectrometry, provide precise and accurate measurements of time throughout the geological history of our planet over nine orders of magnitude, from the accretion of the proto-planetary disk, to the timing of the last glaciation. This article summarizes the recent efforts of the Isotope Geochemistry, Geochronology and Thermochronology research group at the University of Geneva to advance the U-Pb geochronological tool to achieve unprecedented precision and accuracy, and presents two examples of its application to two significant open questions in Earth sciences: what are the triggers and timescales of volcanic supereruptions, and what were the causes of mass extinctions in the geological past, driven by global climatic and environmental deterioration?
MAPA: Implementation of the Standard Interchange Format and use for analyzing lattices
NASA Astrophysics Data System (ADS)
Shasharina, Svetlana G.; Cary, John R.
1997-05-01
MAPA (Modular Accelerator Physics Analysis) is an object oriented application for accelerator design and analysis with a Motif based graphical user interface. MAPA has been ported to AIX, Linux, HPUX, Solaris, and IRIX. MAPA provides an intuitive environment for accelerator study and design. The user can bring up windows for fully nonlinear analysis of accelerator lattices in any number of dimensions. The current graphical analysis methods of Lifetime plots and Surfaces of Section have been used to analyze the improved lattice designs of Wan, Cary, and Shasharina (this conference). MAPA can now read and write Standard Interchange Format (MAD) accelerator description files and it has a general graphical user interface for adding, changing, and deleting elements. MAPA's consistency checks prevent deletion of used elements and prevent creation of recursive beam lines. Plans include development of a richer set of modeling tools and the ability to invoke existing modeling codes through the MAPA interface. MAPA will be demonstrated on a Pentium 150 laptop running Linux.
Life Predicted in a Probabilistic Design Space for Brittle Materials With Transient Loads
NASA Technical Reports Server (NTRS)
Nemeth, Noel N.; Palfi, Tamas; Reh, Stefan
2005-01-01
Analytical techniques have progressively become more sophisticated, and now we can consider the probabilistic nature of the entire space of random input variables on the lifetime reliability of brittle structures. This was demonstrated with NASA s CARES/Life (Ceramic Analysis and Reliability Evaluation of Structures/Life) code combined with the commercially available ANSYS/Probabilistic Design System (ANSYS/PDS), a probabilistic analysis tool that is an integral part of the ANSYS finite-element analysis program. ANSYS/PDS allows probabilistic loads, component geometry, and material properties to be considered in the finite-element analysis. CARES/Life predicts the time dependent probability of failure of brittle material structures under generalized thermomechanical loading--such as that found in a turbine engine hot-section. Glenn researchers coupled ANSYS/PDS with CARES/Life to assess the effects of the stochastic variables of component geometry, loading, and material properties on the predicted life of the component for fully transient thermomechanical loading and cyclic loading.
Multielemental speciation analysis by advanced hyphenated technique - HPLC/ICP-MS: A review.
Marcinkowska, Monika; Barałkiewicz, Danuta
2016-12-01
Speciation analysis has become an invaluable tool in human health risk assessment, environmental monitoring or food quality control. Another step is to develop reliable multielemental speciation methodologies, to reduce costs, waste and time needed for the analysis. Separation and detection of species of several elements in a single analytical run can be accomplished by high performance liquid chromatography hyphenated to inductively coupled plasma mass spectrometry (HPLC/ICP-MS). Our review assembles articles concerning multielemental speciation determination of: As, Se, Cr, Sb, I, Br, Pb, Hg, V, Mo, Te, Tl, Cd and W in environmental, biological, food and clinical samples analyzed with HPLC/ICP-MS. It addresses the procedures in terms of following issues: sample collection and pretreatment, selection of optimal conditions for elements species separation by HPLC and determination using ICP-MS as well as metrological approach. The presented work is the first review article concerning multielemental speciation analysis by advanced hyphenated technique HPLC/ICP-MS. Copyright © 2016 Elsevier B.V. All rights reserved.
NASA Astrophysics Data System (ADS)
Ortolano, Gaetano; Visalli, Roberto; Godard, Gaston; Cirrincione, Rosolino
2018-06-01
We present a new ArcGIS®-based tool developed in the Python programming language for calibrating EDS/WDS X-ray element maps, with the aim of acquiring quantitative information of petrological interest. The calibration procedure is based on a multiple linear regression technique that takes into account interdependence among elements and is constrained by the stoichiometry of minerals. The procedure requires an appropriate number of spot analyses for use as internal standards and provides several test indexes for a rapid check of calibration accuracy. The code is based on an earlier image-processing tool designed primarily for classifying minerals in X-ray element maps; the original Python code has now been enhanced to yield calibrated maps of mineral end-members or the chemical parameters of each classified mineral. The semi-automated procedure can be used to extract a dataset that is automatically stored within queryable tables. As a case study, the software was applied to an amphibolite-facies garnet-bearing micaschist. The calibrated images obtained for both anhydrous (i.e., garnet and plagioclase) and hydrous (i.e., biotite) phases show a good fit with corresponding electron microprobe analyses. This new GIS-based tool package can thus find useful application in petrology and materials science research. Moreover, the huge quantity of data extracted opens new opportunities for the development of a thin-section microchemical database that, using a GIS platform, can be linked with other major global geoscience databases.
Optimization of an Advanced Hybrid Wing Body Concept Using HCDstruct Version 1.2
NASA Technical Reports Server (NTRS)
Quinlan, Jesse R.; Gern, Frank H.
2016-01-01
Hybrid Wing Body (HWB) aircraft concepts continue to be promising candidates for achieving the simultaneous fuel consumption and noise reduction goals set forth by NASA's Environmentally Responsible Aviation (ERA) project. In order to evaluate the projected benefits, improvements in structural analysis at the conceptual design level were necessary; thus, NASA researchers developed the Hybrid wing body Conceptual Design and structural optimization (HCDstruct) tool to perform aeroservoelastic structural optimizations of advanced HWB concepts. In this paper, the authors present substantial updates to the HCDstruct tool and related analysis, including: the addition of four inboard and eight outboard control surfaces and two all-movable tail/rudder assemblies, providing a full aeroservoelastic analysis capability; the implementation of asymmetric load cases for structural sizing applications; and a methodology for minimizing control surface actuation power using NASTRAN SOL 200 and HCDstruct's aeroservoelastic finite-element model (FEM).
Discrete event simulation tool for analysis of qualitative models of continuous processing systems
NASA Technical Reports Server (NTRS)
Malin, Jane T. (Inventor); Basham, Bryan D. (Inventor); Harris, Richard A. (Inventor)
1990-01-01
An artificial intelligence design and qualitative modeling tool is disclosed for creating computer models and simulating continuous activities, functions, and/or behavior using developed discrete event techniques. Conveniently, the tool is organized in four modules: library design module, model construction module, simulation module, and experimentation and analysis. The library design module supports the building of library knowledge including component classes and elements pertinent to a particular domain of continuous activities, functions, and behavior being modeled. The continuous behavior is defined discretely with respect to invocation statements, effect statements, and time delays. The functionality of the components is defined in terms of variable cluster instances, independent processes, and modes, further defined in terms of mode transition processes and mode dependent processes. Model construction utilizes the hierarchy of libraries and connects them with appropriate relations. The simulation executes a specialized initialization routine and executes events in a manner that includes selective inherency of characteristics through a time and event schema until the event queue in the simulator is emptied. The experimentation and analysis module supports analysis through the generation of appropriate log files and graphics developments and includes the ability of log file comparisons.
ExGUtils: A Python Package for Statistical Analysis With the ex-Gaussian Probability Density.
Moret-Tatay, Carmen; Gamermann, Daniel; Navarro-Pardo, Esperanza; Fernández de Córdoba Castellá, Pedro
2018-01-01
The study of reaction times and their underlying cognitive processes is an important field in Psychology. Reaction times are often modeled through the ex-Gaussian distribution, because it provides a good fit to multiple empirical data. The complexity of this distribution makes the use of computational tools an essential element. Therefore, there is a strong need for efficient and versatile computational tools for the research in this area. In this manuscript we discuss some mathematical details of the ex-Gaussian distribution and apply the ExGUtils package, a set of functions and numerical tools, programmed for python, developed for numerical analysis of data involving the ex-Gaussian probability density. In order to validate the package, we present an extensive analysis of fits obtained with it, discuss advantages and differences between the least squares and maximum likelihood methods and quantitatively evaluate the goodness of the obtained fits (which is usually an overlooked point in most literature in the area). The analysis done allows one to identify outliers in the empirical datasets and criteriously determine if there is a need for data trimming and at which points it should be done.
ExGUtils: A Python Package for Statistical Analysis With the ex-Gaussian Probability Density
Moret-Tatay, Carmen; Gamermann, Daniel; Navarro-Pardo, Esperanza; Fernández de Córdoba Castellá, Pedro
2018-01-01
The study of reaction times and their underlying cognitive processes is an important field in Psychology. Reaction times are often modeled through the ex-Gaussian distribution, because it provides a good fit to multiple empirical data. The complexity of this distribution makes the use of computational tools an essential element. Therefore, there is a strong need for efficient and versatile computational tools for the research in this area. In this manuscript we discuss some mathematical details of the ex-Gaussian distribution and apply the ExGUtils package, a set of functions and numerical tools, programmed for python, developed for numerical analysis of data involving the ex-Gaussian probability density. In order to validate the package, we present an extensive analysis of fits obtained with it, discuss advantages and differences between the least squares and maximum likelihood methods and quantitatively evaluate the goodness of the obtained fits (which is usually an overlooked point in most literature in the area). The analysis done allows one to identify outliers in the empirical datasets and criteriously determine if there is a need for data trimming and at which points it should be done. PMID:29765345
NASA Astrophysics Data System (ADS)
Gardner, Robin P.; Xu, Libai
2009-10-01
The Center for Engineering Applications of Radioisotopes (CEAR) has been working for over a decade on the Monte Carlo library least-squares (MCLLS) approach for treating non-linear radiation analyzer problems including: (1) prompt gamma-ray neutron activation analysis (PGNAA) for bulk analysis, (2) energy-dispersive X-ray fluorescence (EDXRF) analyzers, and (3) carbon/oxygen tool analysis in oil well logging. This approach essentially consists of using Monte Carlo simulation to generate the libraries of all the elements to be analyzed plus any other required background libraries. These libraries are then used in the linear library least-squares (LLS) approach with unknown sample spectra to analyze for all elements in the sample. Iterations of this are used until the LLS values agree with the composition used to generate the libraries. The current status of the methods (and topics) necessary to implement the MCLLS approach is reported. This includes: (1) the Monte Carlo codes such as CEARXRF, CEARCPG, and CEARCO for forward generation of the necessary elemental library spectra for the LLS calculation for X-ray fluorescence, neutron capture prompt gamma-ray analyzers, and carbon/oxygen tools; (2) the correction of spectral pulse pile-up (PPU) distortion by Monte Carlo simulation with the code CEARIPPU; (3) generation of detector response functions (DRF) for detectors with linear and non-linear responses for Monte Carlo simulation of pulse-height spectra; and (4) the use of the differential operator (DO) technique to make the necessary iterations for non-linear responses practical. In addition to commonly analyzed single spectra, coincidence spectra or even two-dimensional (2-D) coincidence spectra can also be used in the MCLLS approach and may provide more accurate results.
Single-cell analysis by ICP-MS/MS as a fast tool for cellular bioavailability studies of arsenite.
Meyer, S; López-Serrano, A; Mitze, H; Jakubowski, N; Schwerdtle, T
2018-01-24
Single-cell inductively coupled plasma mass spectrometry (SC-ICP-MS) has become a powerful and fast tool to evaluate the elemental composition at a single-cell level. In this study, the cellular bioavailability of arsenite (incubation of 25 and 50 μM for 0-48 h) has been successfully assessed by SC-ICP-MS/MS for the first time directly after re-suspending the cells in water. This procedure avoids the normally arising cell membrane permeabilization caused by cell fixation methods (e.g. methanol fixation). The reliability and feasibility of this SC-ICP-MS/MS approach with a limit of detection of 0.35 fg per cell was validated by conventional bulk ICP-MS/MS analysis after cell digestion and parallel measurement of sulfur and phosphorus.
Lawryk, Nicholas J; Feng, H Amy; Chen, Bean T
2009-07-01
Recent advances in field-portable X-ray fluorescence (FP XRF) spectrometer technology have made it a potentially valuable screening tool for the industrial hygienist to estimate worker exposures to airborne metals. Although recent studies have shown that FP XRF technology may be better suited for qualitative or semiquantitative analysis of airborne lead in the workplace, these studies have not extensively addressed its ability to measure other elements. This study involved a laboratory-based evaluation of a representative model FP XRF spectrometer to measure elements commonly encountered in workplace settings that may be collected on air sample filter media, including chromium, copper, iron, manganese, nickel, lead, and zinc. The evaluation included assessments of (1) response intensity with respect to location on the probe window, (2) limits of detection for five different filter media, (3) limits of detection as a function of analysis time, and (4) bias, precision, and accuracy estimates. Teflon, polyvinyl chloride, polypropylene, and mixed cellulose ester filter media all had similarly low limits of detection for the set of elements examined. Limits of detection, bias, and precision generally improved with increasing analysis time. Bias, precision, and accuracy estimates generally improved with increasing element concentration. Accuracy estimates met the National Institute for Occupational Safety and Health criterion for nearly all the element and concentration combinations. Based on these results, FP XRF spectrometry shows potential to be useful in the assessment of worker inhalation exposures to other metals in addition to lead.
RESPIZZI, STEFANO; COVELLI, ELISABETTA
2015-01-01
The emotional coaching model uses quantitative and qualitative elements to demonstrate some assumptions relevant to new methods of treatment in physical rehabilitation, considering emotional, cognitive and behavioral aspects in patients, whether or not they are sportsmen. Through quantitative tools (Tampa Kinesiophobia Scale, Emotional Interview Test, Previous Re-Injury Test, and reports on test scores) and qualitative tools (training contracts and relationships of emotional alliance or “contagion”), we investigate initial assumptions regarding: the presence of a cognitive and emotional mental state of impasse in patients at the beginning of the rehabilitation pathway; the curative value of the emotional alliance or “emotional contagion” relationship between healthcare provider and patient; the link between the patient’s pathology and type of contact with his own body and emotions; analysis of the psychosocial variables for the prediction of possible cases of re-injury for patients who have undergone or are afraid to undergo reconstruction of the anterior cruciate ligament (ACL). Although this approach is still in the experimental stage, the scores of the administered tests show the possibility of integrating quantitative and qualitative tools to investigate and develop a patient’s physical, mental and emotional resources during the course of his rehabilitation. Furthermore, it seems possible to identify many elements characterizing patients likely to undergo episodes of re-injury or to withdraw totally from sporting activity. In particular, such patients are competitive athletes, who fear or have previously undergone ACL reconstruction. The theories referred to (the transactional analysis theory, self-determination theory) and the tools used demonstrate the usefulness of continuing this research in order to build a shared coaching model treatment aimed at all patients, sportspeople or otherwise, which is not only physical but also emotional, cognitive and behavioral. PMID:26904525
NASA Technical Reports Server (NTRS)
Cruse, T. A.
1987-01-01
The objective is the development of several modular structural analysis packages capable of predicting the probabilistic response distribution for key structural variables such as maximum stress, natural frequencies, transient response, etc. The structural analysis packages are to include stochastic modeling of loads, material properties, geometry (tolerances), and boundary conditions. The solution is to be in terms of the cumulative probability of exceedance distribution (CDF) and confidence bounds. Two methods of probability modeling are to be included as well as three types of structural models - probabilistic finite-element method (PFEM); probabilistic approximate analysis methods (PAAM); and probabilistic boundary element methods (PBEM). The purpose in doing probabilistic structural analysis is to provide the designer with a more realistic ability to assess the importance of uncertainty in the response of a high performance structure. Probabilistic Structural Analysis Method (PSAM) tools will estimate structural safety and reliability, while providing the engineer with information on the confidence that should be given to the predicted behavior. Perhaps most critically, the PSAM results will directly provide information on the sensitivity of the design response to those variables which are seen to be uncertain.
NASA Technical Reports Server (NTRS)
Cruse, T. A.; Burnside, O. H.; Wu, Y.-T.; Polch, E. Z.; Dias, J. B.
1988-01-01
The objective is the development of several modular structural analysis packages capable of predicting the probabilistic response distribution for key structural variables such as maximum stress, natural frequencies, transient response, etc. The structural analysis packages are to include stochastic modeling of loads, material properties, geometry (tolerances), and boundary conditions. The solution is to be in terms of the cumulative probability of exceedance distribution (CDF) and confidence bounds. Two methods of probability modeling are to be included as well as three types of structural models - probabilistic finite-element method (PFEM); probabilistic approximate analysis methods (PAAM); and probabilistic boundary element methods (PBEM). The purpose in doing probabilistic structural analysis is to provide the designer with a more realistic ability to assess the importance of uncertainty in the response of a high performance structure. Probabilistic Structural Analysis Method (PSAM) tools will estimate structural safety and reliability, while providing the engineer with information on the confidence that should be given to the predicted behavior. Perhaps most critically, the PSAM results will directly provide information on the sensitivity of the design response to those variables which are seen to be uncertain.
A validated finite element model of a soft artificial muscle motor
NASA Astrophysics Data System (ADS)
Tse, Tony Chun H.; O'Brien, Benjamin; McKay, Thomas; Anderson, Iain A.
2011-04-01
The Biomimetics Laboratory has developed a soft artificial muscle motor based on Dielectric Elastomers. The motor, 'Flexidrive', is light-weight and has low system complexity. It works by gripping and turning a shaft with a soft gear, like we would with our fingers. The motor's performance depends on many factors, such as actuation waveform, electrode patterning, geometries and contact tribology between the shaft and gear. We have developed a finite element model (FEM) of the motor as a study and design tool. Contact interaction was integrated with previous material and electromechanical coupling models in ABAQUS. The model was experimentally validated through a shape and blocked force analysis.
Low-speed single-element airfoil synthesis
NASA Technical Reports Server (NTRS)
Mcmasters, J. H.; Henderson, M. L.
1979-01-01
The use of recently developed airfoil analysis/design computational tools to clarify, enrich and extend the existing experimental data base on low-speed, single element airfoils is demonstrated. A discussion of the problem of tailoring an airfoil for a specific application at its appropriate Reynolds number is presented. This problem is approached by use of inverse (or synthesis) techniques, wherein a desirable set of boundary layer characteristics, performance objectives, and constraints are specified, which then leads to derivation of a corresponding viscous flow pressure distribution. Examples are presented which demonstrate the synthesis approach, following presentation of some historical information and background data which motivate the basic synthesis process.
Development and analysis of educational technologies for a blended organic chemistry course
NASA Astrophysics Data System (ADS)
Evans, Michael James
Blended courses incorporate elements of both face-to-face and online instruction. The extent to which blended courses are conducted online, and the proper role of the online components of blended courses, have been debated and may vary. What can be said in general, however, is that online tools for blended courses are typically culled together from a variety of sources, are often very large scale, and may present distractions for students that decrease their utility as teaching tools. Furthermore, large-scale educational technologies may not be amenable to rigorous, detailed study, limiting evaluation of their effectiveness. Small-scale educational technologies run from the instructor's own server have the potential to mitigate many of these issues. Such tools give the instructor or researcher direct access to all available data, facilitating detailed analysis of student use. Code modification is simple and rapid if errors arise, since code is stored where the instructor can easily access it. Finally, the design of a small-scale tool can target a very specific application. With these ideas in mind, this work describes several projects aimed at exploring the use of small-scale, web-based software in a blended organic chemistry course. A number of activities were developed and evaluated using the Student Assessment of Learning Gains survey, and data from the activities were analyzed using quantitative methods of statistics and social network analysis methods. Findings from this work suggest that small-scale educational technologies provide significant learning benefits for students of organic chemistry---with the important caveat that instructors must offer appropriate levels of technical and pedagogical support for students. Most notably, students reported significant learning gains from activities that included collaborative learning supported by novel online tools. For the particular context of organic chemistry, which has a unique semantic language (Lewis structures), the incorporation of shared video was a novel but important element of these activities. In fields for which mere text would not provide enough information in communications between students, video offers an appealing medium for student-student interaction.
NASA Astrophysics Data System (ADS)
Madajewski, Marek; Nowakowski, Zbigniew
2017-01-01
This paper presents analysis of flank wear influence on forces in orthogonal turning of 42CrMo4 steel and evaluates capacity of finite element model to provide such force values. Data about magnitude of feed and cutting force were obtained from measurements with force tensiometer in experimental test as well as from finite element analysis of chip formation process in ABAQUS/Explicit software. For studies an insert with complex rake face was selected and flank wear was simulated by grinding operation on its flank face. The aim of grinding inset surface was to obtain even flat wear along cutting edge, which after the measurement could be modeled with CAD program and applied in FE analysis for selected range of wear width. By comparing both sets of force values as function of flank wear in given cutting conditions FEA model was validated and it was established that it can be applied to analyze other physical aspects of machining. Force analysis found that progression of wear causes increase in cutting force magnitude and steep boost to feed force magnitude. Analysis of Fc/Ff force ratio revealed that flank wear has significant impact on resultant force in orthogonal cutting and magnitude of this force components in cutting and feed direction. Surge in force values can result in transfer of substantial loads to machine-tool interface.
Space Launch System Upper Stage Technology Assessment
NASA Technical Reports Server (NTRS)
Holladay, Jon; Hampton, Bryan; Monk, Timothy
2014-01-01
The Space Launch System (SLS) is envisioned as a heavy-lift vehicle that will provide the foundation for future beyond low-Earth orbit (LEO) exploration missions. Previous studies have been performed to determine the optimal configuration for the SLS and the applicability of commercial off-the-shelf in-space stages for Earth departure. Currently NASA is analyzing the concept of a Dual Use Upper Stage (DUUS) that will provide LEO insertion and Earth departure burns. This paper will explore candidate in-space stages based on the DUUS design for a wide range of beyond LEO missions. Mission payloads will range from small robotic systems up to human systems with deep space habitats and landers. Mission destinations will include cislunar space, Mars, Jupiter, and Saturn. Given these wide-ranging mission objectives, a vehicle-sizing tool has been developed to determine the size of an Earth departure stage based on the mission objectives. The tool calculates masses for all the major subsystems of the vehicle including propellant loads, avionics, power, engines, main propulsion system components, tanks, pressurization system and gases, primary structural elements, and secondary structural elements. The tool uses an iterative sizing algorithm to determine the resulting mass of the stage. Any input into one of the subsystem sizing routines or the mission parameters can be treated as a parametric sweep or as a distribution for use in Monte Carlo analysis. Taking these factors together allows for multi-variable, coupled analysis runs. To increase confidence in the tool, the results have been verified against two point-of-departure designs of the DUUS. The tool has also been verified against Apollo moon mission elements and other manned space systems. This paper will focus on trading key propulsion technologies including chemical, Nuclear Thermal Propulsion (NTP), and Solar Electric Propulsion (SEP). All of the key performance inputs and relationships will be presented and discussed in light of the various missions. For each mission there are several trajectory options and each will be discussed in terms of delta-v required and transit duration. Each propulsion system will be modeled, sized, and judged based on their applicability to the whole range of beyond LEO missions. Criteria for scoring will include the resulting dry mass of the stage, resulting propellant required, time to destination, and an assessment of key enabling technologies. In addition to the larger metrics, this paper will present the results of several coupled sensitivity studies. The ultimate goals of these tools and studies are to provide NASA with the most mass-, technology-, and cost-effective in-space stage for its future exploration missions.
A finite element head and neck model as a supportive tool for deformable image registration.
Kim, Jihun; Saitou, Kazuhiro; Matuszak, Martha M; Balter, James M
2016-07-01
A finite element (FE) head and neck model was developed as a tool to aid investigations and development of deformable image registration and patient modeling in radiation oncology. Useful aspects of a FE model for these purposes include ability to produce realistic deformations (similar to those seen in patients over the course of treatment) and a rational means of generating new configurations, e.g., via the application of force and/or displacement boundary conditions. The model was constructed based on a cone-beam computed tomography image of a head and neck cancer patient. The three-node triangular surface meshes created for the bony elements (skull, mandible, and cervical spine) and joint elements were integrated into a skeletal system and combined with the exterior surface. Nodes were additionally created inside the surface structures which were composed of the three-node triangular surface meshes, so that four-node tetrahedral FE elements were created over the whole region of the model. The bony elements were modeled as a homogeneous linear elastic material connected by intervertebral disks. The surrounding tissues were modeled as a homogeneous linear elastic material. Under force or displacement boundary conditions, FE analysis on the model calculates approximate solutions of the displacement vector field. A FE head and neck model was constructed that skull, mandible, and cervical vertebrae were mechanically connected by disks. The developed FE model is capable of generating realistic deformations that are strain-free for the bony elements and of creating new configurations of the skeletal system with the surrounding tissues reasonably deformed. The FE model can generate realistic deformations for skeletal elements. In addition, the model provides a way of evaluating the accuracy of image alignment methods by producing a ground truth deformation and correspondingly simulated images. The ability to combine force and displacement conditions provides flexibility for simulating realistic anatomic configurations.
NASA Technical Reports Server (NTRS)
Statler, Irving C. (Editor)
2007-01-01
The Aviation System Monitoring and Modeling (ASMM) Project was one of the projects within NASA s Aviation Safety Program from 1999 through 2005. The objective of the ASMM Project was to develop the technologies to enable the aviation industry to undertake a proactive approach to the management of its system-wide safety risks. The ASMM Project entailed four interdependent elements: (1) Data Analysis Tools Development - develop tools to convert numerical and textual data into information; (2) Intramural Monitoring - test and evaluate the data analysis tools in operational environments; (3) Extramural Monitoring - gain insight into the aviation system performance by surveying its front-line operators; and (4) Modeling and Simulations - provide reliable predictions of the system-wide hazards, their causal factors, and their operational risks that may result from the introduction of new technologies, new procedures, or new operational concepts. This report is a documentation of the history of this highly successful project and of its many accomplishments and contributions to improved safety of the aviation system.
Design/Analysis of Metal/Composite Bonded Joints for Survivability at Cryogenic Temperatures
NASA Technical Reports Server (NTRS)
Bartoszyk, Andrew E.
2004-01-01
A major design and analysis challenge for the JWST ISM structure is the metal/composite bonded joints that will be required to survive down to an operational ultra-low temperature of 30K (-405 F). The initial and current baseline design for the plug-type joint consists of a titanium thin walled fitting (1-3mm thick) bonded to the interior surface of an M555/954-6 composite truss square tube with an axially stiff biased lay-up. Metallic fittings are required at various nodes of the truss structure to accommodate instrument and lift-point bolted interfaces. Analytical experience and design work done on metal/composite bonded joints at temperatures below liquid nitrogen are limited and important analysis tools, material properties, and failure criteria for composites at cryogenic temperatures are virtually nonexistent. Increasing the challenge is the difficulty in testing for these required tools and parameters at 30K. A preliminary finite element analysis shows that failure due to CTE mismatch between the biased composite and titanium or aluminum is likely. Failure is less likely with Invar, however an initial mass estimate of Invar fittings demonstrates that Invar is not an automatic alternative. In order to gain confidence in analyzing and designing the ISM joints, a comprehensive joint development testing program has been planned and is currently running. The test program is designed for the correlation of the analysis methodology, including tuning finite element model parameters, and developing a composite failure criterion for the effect of multi-axial composite stresses on the strength of a bonded joint at 30K. The testing program will also consider stress mitigation using compliant composite layers and potential strength degradation due to multiple thermal cycles. Not only will the finite element analysis be correlated to the test data, but the FEA will be used to guide the design of the test. The first phase of the test program has been completed and the preliminary analysis has been revisited based on the test data In this work, we present an overview of the test plan, results today, and resulting design improvements.
An advanced environment for hybrid modeling of biological systems based on modelica.
Pross, Sabrina; Bachmann, Bernhard
2011-01-20
Biological systems are often very complex so that an appropriate formalism is needed for modeling their behavior. Hybrid Petri Nets, consisting of time-discrete Petri Net elements as well as continuous ones, have proven to be ideal for this task. Therefore, a new Petri Net library was implemented based on the object-oriented modeling language Modelica which allows the modeling of discrete, stochastic and continuous Petri Net elements by differential, algebraic and discrete equations. An appropriate Modelica-tool performs the hybrid simulation with discrete events and the solution of continuous differential equations. A special sub-library contains so-called wrappers for specific reactions to simplify the modeling process. The Modelica-models can be connected to Simulink-models for parameter optimization, sensitivity analysis and stochastic simulation in Matlab. The present paper illustrates the implementation of the Petri Net component models, their usage within the modeling process and the coupling between the Modelica-tool Dymola and Matlab/Simulink. The application is demonstrated by modeling the metabolism of Chinese Hamster Ovary Cells.
Analytical Model-Based Design Optimization of a Transverse Flux Machine
DOE Office of Scientific and Technical Information (OSTI.GOV)
Hasan, Iftekhar; Husain, Tausif; Sozer, Yilmaz
This paper proposes an analytical machine design tool using magnetic equivalent circuit (MEC)-based particle swarm optimization (PSO) for a double-sided, flux-concentrating transverse flux machine (TFM). The magnetic equivalent circuit method is applied to analytically establish the relationship between the design objective and the input variables of prospective TFM designs. This is computationally less intensive and more time efficient than finite element solvers. A PSO algorithm is then used to design a machine with the highest torque density within the specified power range along with some geometric design constraints. The stator pole length, magnet length, and rotor thickness are the variablesmore » that define the optimization search space. Finite element analysis (FEA) was carried out to verify the performance of the MEC-PSO optimized machine. The proposed analytical design tool helps save computation time by at least 50% when compared to commercial FEA-based optimization programs, with results found to be in agreement with less than 5% error.« less
Can we use high precision metal isotope analysis to improve our understanding of cancer?
Larner, Fiona
2016-01-01
High precision natural isotope analyses are widely used in geosciences to trace elemental transport pathways. The use of this analytical tool is increasing in nutritional and disease-related research. In recent months, a number of groups have shown the potential this technique has in providing new observations for various cancers when applied to trace metal metabolism. The deconvolution of isotopic signatures, however, relies on mathematical models and geochemical data, which are not representative of the system under investigation. In addition to relevant biochemical studies of protein-metal isotopic interactions, technological development both in terms of sample throughput and detection sensitivity of these elements is now needed to translate this novel approach into a mainstream analytical tool. Following this, essential background healthy population studies must be performed, alongside observational, cross-sectional disease-based studies. Only then can the sensitivity and specificity of isotopic analyses be tested alongside currently employed methods, and important questions such as the influence of cancer heterogeneity and disease stage on isotopic signatures be addressed.
Eminaga, Okyaz; Hinkelammert, Reemt; Semjonow, Axel; Neumann, Joerg; Abbas, Mahmoud; Koepke, Thomas; Bettendorf, Olaf; Eltze, Elke; Dugas, Martin
2010-11-15
The pathology report of radical prostatectomy specimens plays an important role in clinical decisions and the prognostic evaluation in Prostate Cancer (PCa). The anatomical schema is a helpful tool to document PCa extension for clinical and research purposes. To achieve electronic documentation and analysis, an appropriate documentation model for anatomical schemas is needed. For this purpose we developed cMDX. The document architecture of cMDX was designed according to Open Packaging Conventions by separating the whole data into template data and patient data. Analogue custom XML elements were considered to harmonize the graphical representation (e.g. tumour extension) with the textual data (e.g. histological patterns). The graphical documentation was based on the four-layer visualization model that forms the interaction between different custom XML elements. Sensible personal data were encrypted with a 256-bit cryptographic algorithm to avoid misuse. In order to assess the clinical value, we retrospectively analysed the tumour extension in 255 patients after radical prostatectomy. The pathology report with cMDX can represent pathological findings of the prostate in schematic styles. Such reports can be integrated into the hospital information system. "cMDX" documents can be converted into different data formats like text, graphics and PDF. Supplementary tools like cMDX Editor and an analyser tool were implemented. The graphical analysis of 255 prostatectomy specimens showed that PCa were mostly localized in the peripheral zone (Mean: 73% ± 25). 54% of PCa showed a multifocal growth pattern. cMDX can be used for routine histopathological reporting of radical prostatectomy specimens and provide data for scientific analysis.
2010-01-01
Background The pathology report of radical prostatectomy specimens plays an important role in clinical decisions and the prognostic evaluation in Prostate Cancer (PCa). The anatomical schema is a helpful tool to document PCa extension for clinical and research purposes. To achieve electronic documentation and analysis, an appropriate documentation model for anatomical schemas is needed. For this purpose we developed cMDX. Methods The document architecture of cMDX was designed according to Open Packaging Conventions by separating the whole data into template data and patient data. Analogue custom XML elements were considered to harmonize the graphical representation (e.g. tumour extension) with the textual data (e.g. histological patterns). The graphical documentation was based on the four-layer visualization model that forms the interaction between different custom XML elements. Sensible personal data were encrypted with a 256-bit cryptographic algorithm to avoid misuse. In order to assess the clinical value, we retrospectively analysed the tumour extension in 255 patients after radical prostatectomy. Results The pathology report with cMDX can represent pathological findings of the prostate in schematic styles. Such reports can be integrated into the hospital information system. "cMDX" documents can be converted into different data formats like text, graphics and PDF. Supplementary tools like cMDX Editor and an analyser tool were implemented. The graphical analysis of 255 prostatectomy specimens showed that PCa were mostly localized in the peripheral zone (Mean: 73% ± 25). 54% of PCa showed a multifocal growth pattern. Conclusions cMDX can be used for routine histopathological reporting of radical prostatectomy specimens and provide data for scientific analysis. PMID:21078179
Mathematical model of ambulance resources in Saint-Petersburg
NASA Astrophysics Data System (ADS)
Shavidze, G. G.; Balykina, Y. E.; Lejnina, E. A.; Svirkin, M. V.
2016-06-01
Emergency medical system is one of the main elements in city infrastructure. The article contains analysis of existing system of ambulance resource distribution. Paper considers the idea of using multiperiodicity as a tool to increase the efficiency of the Emergency Medical Services. The program developed in programming environment Matlab helps to evaluate the changes in the functioning of the system of emergency medical service.
Incorporating psychological influences in probabilistic cost analysis
DOE Office of Scientific and Technical Information (OSTI.GOV)
Kujawski, Edouard; Alvaro, Mariana; Edwards, William
2004-01-08
Today's typical probabilistic cost analysis assumes an ''ideal'' project that is devoid of the human and organizational considerations that heavily influence the success and cost of real-world projects. In the real world ''Money Allocated Is Money Spent'' (MAIMS principle); cost underruns are rarely available to protect against cost overruns while task overruns are passed on to the total project cost. Realistic cost estimates therefore require a modified probabilistic cost analysis that simultaneously models the cost management strategy including budget allocation. Psychological influences such as overconfidence in assessing uncertainties and dependencies among cost elements and risks are other important considerations thatmore » are generally not addressed. It should then be no surprise that actual project costs often exceed the initial estimates and are delivered late and/or with a reduced scope. This paper presents a practical probabilistic cost analysis model that incorporates recent findings in human behavior and judgment under uncertainty, dependencies among cost elements, the MAIMS principle, and project management practices. Uncertain cost elements are elicited from experts using the direct fractile assessment method and fitted with three-parameter Weibull distributions. The full correlation matrix is specified in terms of two parameters that characterize correlations among cost elements in the same and in different subsystems. The analysis is readily implemented using standard Monte Carlo simulation tools such as {at}Risk and Crystal Ball{reg_sign}. The analysis of a representative design and engineering project substantiates that today's typical probabilistic cost analysis is likely to severely underestimate project cost for probability of success values of importance to contractors and procuring activities. The proposed approach provides a framework for developing a viable cost management strategy for allocating baseline budgets and contingencies. Given the scope and magnitude of the cost-overrun problem, the benefits are likely to be significant.« less
Sidhu, Navdeep S; Edwards, Morgan
2018-04-27
We conducted a scoping review of tools designed to add structure to clinical teaching, with a thematic analysis to establish definitional clarity. Six thousand and forty nine citations were screened, 434 reviewed for eligibility, and 230 identified as meeting study inclusion criteria. Eighty-nine names and 51 definitions were identified. Based on a post facto thematic analysis, we propose that these tools be named "deliberate teaching tools" (DTTs) and defined as "frameworks that enable clinicians to have a purposeful and considered approach to teaching encounters by incorporating elements identified with good teaching practice." We identified 46 DTTs in the literature, with 38 (82.6%) originally described for the medical setting. Forty justification articles consisted of 16 feedback surveys, 13 controlled trials, seven pre-post intervention studies with no control group, and four observation studies. Current evidence of efficacy is not entirely conclusive, and many studies contain methodology flaws. Forty-nine clarification articles comprised 12 systematic reviews and 37 narrative reviews. The most number of DTTs described by any review was four. A common design theme was identified in approximately three-quarters of DTTs. Applicability of DTTs to specific alternate settings should be considered in context, and appropriately designed justification studies are warranted to demonstrate efficacy.
Staccini, Pascal M.; Joubert, Michel; Quaranta, Jean-Francois; Fieschi, Marius
2002-01-01
Growing attention is being given to the use of process modeling methodology for user requirements elicitation. In the analysis phase of hospital information systems, the usefulness of care-process models has been investigated to evaluate the conceptual applicability and practical understandability by clinical staff and members of users teams. Nevertheless, there still remains a gap between users and analysts in their mutual ability to share conceptual views and vocabulary, keeping the meaning of clinical context while providing elements for analysis. One of the solutions for filling this gap is to consider the process model itself in the role of a hub as a centralized means of facilitating communication between team members. Starting with a robust and descriptive technique for process modeling called IDEF0/SADT, we refined the basic data model by extracting concepts from ISO 9000 process analysis and from enterprise ontology. We defined a web-based architecture to serve as a collaborative tool and implemented it using an object-oriented database. The prospects of such a tool are discussed notably regarding to its ability to generate data dictionaries and to be used as a navigation tool through the medium of hospital-wide documentation. PMID:12463921
Staccini, Pascal M; Joubert, Michel; Quaranta, Jean-Francois; Fieschi, Marius
2002-01-01
Growing attention is being given to the use of process modeling methodology for user requirements elicitation. In the analysis phase of hospital information systems, the usefulness of care-process models has been investigated to evaluate the conceptual applicability and practical understandability by clinical staff and members of users teams. Nevertheless, there still remains a gap between users and analysts in their mutual ability to share conceptual views and vocabulary, keeping the meaning of clinical context while providing elements for analysis. One of the solutions for filling this gap is to consider the process model itself in the role of a hub as a centralized means of facilitating communication between team members. Starting with a robust and descriptive technique for process modeling called IDEF0/SADT, we refined the basic data model by extracting concepts from ISO 9000 process analysis and from enterprise ontology. We defined a web-based architecture to serve as a collaborative tool and implemented it using an object-oriented database. The prospects of such a tool are discussed notably regarding to its ability to generate data dictionaries and to be used as a navigation tool through the medium of hospital-wide documentation.
Dfam: a database of repetitive DNA based on profile hidden Markov models.
Wheeler, Travis J; Clements, Jody; Eddy, Sean R; Hubley, Robert; Jones, Thomas A; Jurka, Jerzy; Smit, Arian F A; Finn, Robert D
2013-01-01
We present a database of repetitive DNA elements, called Dfam (http://dfam.janelia.org). Many genomes contain a large fraction of repetitive DNA, much of which is made up of remnants of transposable elements (TEs). Accurate annotation of TEs enables research into their biology and can shed light on the evolutionary processes that shape genomes. Identification and masking of TEs can also greatly simplify many downstream genome annotation and sequence analysis tasks. The commonly used TE annotation tools RepeatMasker and Censor depend on sequence homology search tools such as cross_match and BLAST variants, as well as Repbase, a collection of known TE families each represented by a single consensus sequence. Dfam contains entries corresponding to all Repbase TE entries for which instances have been found in the human genome. Each Dfam entry is represented by a profile hidden Markov model, built from alignments generated using RepeatMasker and Repbase. When used in conjunction with the hidden Markov model search tool nhmmer, Dfam produces a 2.9% increase in coverage over consensus sequence search methods on a large human benchmark, while maintaining low false discovery rates, and coverage of the full human genome is 54.5%. The website provides a collection of tools and data views to support improved TE curation and annotation efforts. Dfam is also available for download in flat file format or in the form of MySQL table dumps.
General Mission Analysis Tool (GMAT) Architectural Specification. Draft
NASA Technical Reports Server (NTRS)
Hughes, Steven P.; Conway, Darrel, J.
2007-01-01
Early in 2002, Goddard Space Flight Center (GSFC) began to identify requirements for the flight dynamics software needed to fly upcoming missions that use formations of spacecraft to collect data. These requirements ranged from low level modeling features to large scale interoperability requirements. In 2003 we began work on a system designed to meet these requirement; this system is GMAT. The General Mission Analysis Tool (GMAT) is a general purpose flight dynamics modeling tool built on open source principles. The GMAT code is written in C++, and uses modern C++ constructs extensively. GMAT can be run through either a fully functional Graphical User Interface (GUI) or as a command line program with minimal user feedback. The system is built and runs on Microsoft Windows, Linux, and Macintosh OS X platforms. The GMAT GUI is written using wxWidgets, a cross platform library of components that streamlines the development and extension of the user interface Flight dynamics modeling is performed in GMAT by building components that represent the players in the analysis problem that is being modeled. These components interact through the sequential execution of instructions, embodied in the GMAT Mission Sequence. A typical Mission Sequence will model the trajectories of a set of spacecraft evolving over time, calculating relevant parameters during this propagation, and maneuvering individual spacecraft to maintain a set of mission constraints as established by the mission analyst. All of the elements used in GMAT for mission analysis can be viewed in the GMAT GUI or through a custom scripting language. Analysis problems modeled in GMAT are saved as script files, and these files can be read into GMAT. When a script is read into the GMAT GUI, the corresponding user interface elements are constructed in the GMAT GUI. The GMAT system was developed from the ground up to run in a platform agnostic environment. The source code compiles on numerous different platforms, and is regularly exercised running on Windows, Linux and Macintosh computers by the development and analysis teams working on the project. The system can be run using either a graphical user interface, written using the open source wxWidgets framework, or from a text console. The GMAT source code was written using open source tools. GSFC has released the code using the NASA open source license.
Prioritizing Health: A Systematic Approach to Scoping Determinants in Health Impact Assessment.
McCallum, Lindsay C; Ollson, Christopher A; Stefanovic, Ingrid L
2016-01-01
The determinants of health are those factors that have the potential to affect health, either positively or negatively, and include a range of personal, social, economic, and environmental factors. In the practice of health impact assessment (HIA), the stage at which the determinants of health are considered for inclusion is during the scoping step. The scoping step is intended to identify how the HIA will be carried out and to set the boundaries (e.g., temporal and geographical) for the assessment. There are several factors that can help to inform the scoping process, many of which are considered in existing HIA tools and guidance; however, a systematic method of prioritizing determinants was found to be lacking. In order to analyze existing HIA scoping tools that are available, a systematic literature review was conducted, including both primary and gray literature. A total of 10 HIA scoping tools met the inclusion/exclusion criteria and were carried forward for comparative analysis. The analysis focused on minimum elements and practice standards of HIA scoping that have been established in the field. The analysis determined that existing approaches lack a clear, systematic method of prioritization of health determinants for inclusion in HIA. This finding led to the development of a Systematic HIA Scoping tool that addressed this gap. The decision matrix tool uses factors, such as impact, public concern, and data availability, to prioritize health determinants. Additionally, the tool allows for identification of data gaps and provides a transparent method for budget allocation and assessment planning. In order to increase efficiency and improve utility, the tool was programed into Microsoft Excel. Future work in the area of HIA methodology development is vital to the ongoing success of the practice and utilization of HIA as a reliable decision-making tool.
Analyzing Human-Landscape Interactions: Tools That Integrate
NASA Astrophysics Data System (ADS)
Zvoleff, Alex; An, Li
2014-01-01
Humans have transformed much of Earth's land surface, giving rise to loss of biodiversity, climate change, and a host of other environmental issues that are affecting human and biophysical systems in unexpected ways. To confront these problems, environmental managers must consider human and landscape systems in integrated ways. This means making use of data obtained from a broad range of methods (e.g., sensors, surveys), while taking into account new findings from the social and biophysical science literatures. New integrative methods (including data fusion, simulation modeling, and participatory approaches) have emerged in recent years to address these challenges, and to allow analysts to provide information that links qualitative and quantitative elements for policymakers. This paper brings attention to these emergent tools while providing an overview of the tools currently in use for analysis of human-landscape interactions. Analysts are now faced with a staggering array of approaches in the human-landscape literature—in an attempt to bring increased clarity to the field, we identify the relative strengths of each tool, and provide guidance to analysts on the areas to which each tool is best applied. We discuss four broad categories of tools: statistical methods (including survival analysis, multi-level modeling, and Bayesian approaches), GIS and spatial analysis methods, simulation approaches (including cellular automata, agent-based modeling, and participatory modeling), and mixed-method techniques (such as alternative futures modeling and integrated assessment). For each tool, we offer an example from the literature of its application in human-landscape research. Among these tools, participatory approaches are gaining prominence for analysts to make the broadest possible array of information available to researchers, environmental managers, and policymakers. Further development of new approaches of data fusion and integration across sites or disciplines pose an important challenge for future work in integrating human and landscape components.
Helping decision makers frame, analyze, and implement decisions
Runge, Michael C.; McDonald-Madden, Eve
2018-01-01
All decisions have the same recognizable elements. Context, objectives, alternatives, consequences, and deliberation. Decision makers and analysts familiar with these elements can quickly see the underlying structure of a decision.There are only a small number of classes of decisions. These classes differ in the cognitive and scientific challenge they present to the decision maker; the ability to recognize the class of decision leads a decision maker to tools to aid in the analysis.Sometimes we need more information, sometimes we don’t. The role of science in a decision-making process is to provide the predictions that link the alternative actions to the desired outcomes. Investing in more science is only valuable if it helps to choose a better action.Implementation. The successful integration of decision analysis into environmental decisions requires careful attention to the decision, the people, and the institutions involved.
Integrated Aero-Propulsion CFD Methodology for the Hyper-X Flight Experiment
NASA Technical Reports Server (NTRS)
Cockrell, Charles E., Jr.; Engelund, Walter C.; Bittner, Robert D.; Dilley, Arthur D.; Jentink, Tom N.; Frendi, Abdelkader
2000-01-01
Computational fluid dynamics (CFD) tools have been used extensively in the analysis and development of the X-43A Hyper-X Research Vehicle (HXRV). A significant element of this analysis is the prediction of integrated vehicle aero-propulsive performance, which includes an integration of aerodynamic and propulsion flow fields. This paper describes analysis tools used and the methodology for obtaining pre-flight predictions of longitudinal performance increments. The use of higher-fidelity methods to examine flow-field characteristics and scramjet flowpath component performance is also discussed. Limited comparisons with available ground test data are shown to illustrate the approach used to calibrate methods and assess solution accuracy. Inviscid calculations to evaluate lateral-directional stability characteristics are discussed. The methodology behind 3D tip-to-tail calculations is described and the impact of 3D exhaust plume expansion in the afterbody region is illustrated. Finally, future technology development needs in the area of hypersonic propulsion-airframe integration analysis are discussed.
Ubukata, Masaaki; Jobst, Karl J; Reiner, Eric J; Reichenbach, Stephen E; Tao, Qingping; Hang, Jiliang; Wu, Zhanpin; Dane, A John; Cody, Robert B
2015-05-22
Comprehensive two-dimensional gas chromatography (GC×GC) and high-resolution mass spectrometry (HRMS) offer the best possible separation of their respective techniques. Recent commercialization of combined GC×GC-HRMS systems offers new possibilities for the analysis of complex mixtures. However, such experiments yield enormous data sets that require new informatics tools to facilitate the interpretation of the rich information content. This study reports on the analysis of dust obtained from an electronics recycling facility by using GC×GC in combination with a new high-resolution time-of-flight (TOF) mass spectrometer. New software tools for (non-traditional) Kendrick mass defect analysis were developed in this research and greatly aided in the identification of compounds containing chlorine and bromine, elements that feature in most persistent organic pollutants (POPs). In essence, the mass defect plot serves as a visual aid from which halogenated compounds are recognizable on the basis of their mass defect and isotope patterns. Mass chromatograms were generated based on specific ions identified in the plots as well as region of the plot predominantly occupied by halogenated contaminants. Tentative identification was aided by database searches, complementary electron-capture negative ionization experiments and elemental composition determinations from the exact mass data. These included known and emerging flame retardants, such as polybrominated diphenyl ethers (PBDEs), hexabromobenzene, tetrabromo bisphenol A and tris (1-chloro-2-propyl) phosphate (TCPP), as well as other legacy contaminants such as polychlorinated biphenyls (PCBs) and polychlorinated terphenyls (PCTs). Copyright © 2015 Elsevier B.V. All rights reserved.
Grissa, Ibtissem; Vergnaud, Gilles; Pourcel, Christine
2007-01-01
Background In Archeae and Bacteria, the repeated elements called CRISPRs for "clustered regularly interspaced short palindromic repeats" are believed to participate in the defence against viruses. Short sequences called spacers are stored in-between repeated elements. In the current model, motifs comprising spacers and repeats may target an invading DNA and lead to its degradation through a proposed mechanism similar to RNA interference. Analysis of intra-species polymorphism shows that new motifs (one spacer and one repeated element) are added in a polarised fashion. Although their principal characteristics have been described, a lot remains to be discovered on the way CRISPRs are created and evolve. As new genome sequences become available it appears necessary to develop automated scanning tools to make available CRISPRs related information and to facilitate additional investigations. Description We have produced a program, CRISPRFinder, which identifies CRISPRs and extracts the repeated and unique sequences. Using this software, a database is constructed which is automatically updated monthly from newly released genome sequences. Additional tools were created to allow the alignment of flanking sequences in search for similarities between different loci and to build dictionaries of unique sequences. To date, almost six hundred CRISPRs have been identified in 475 published genomes. Two Archeae out of thirty-seven and about half of Bacteria do not possess a CRISPR. Fine analysis of repeated sequences strongly supports the current view that new motifs are added at one end of the CRISPR adjacent to the putative promoter. Conclusion It is hoped that availability of a public database, regularly updated and which can be queried on the web will help in further dissecting and understanding CRISPR structure and flanking sequences evolution. Subsequent analyses of the intra-species CRISPR polymorphism will be facilitated by CRISPRFinder and the dictionary creator. CRISPRdb is accessible at PMID:17521438
ELM - A SIMPLE TOOL FOR THERMAL-HYDRAULIC ANALYSIS OF SOLID-CORE NUCLEAR ROCKET FUEL ELEMENTS
NASA Technical Reports Server (NTRS)
Walton, J. T.
1994-01-01
ELM is a simple computational tool for modeling the steady-state thermal-hydraulics of propellant flow through fuel element coolant channels in nuclear thermal rockets. Written for the nuclear propulsion project of the Space Exploration Initiative, ELM evaluates the various heat transfer coefficient and friction factor correlations available for turbulent pipe flow with heat addition. In the past, these correlations were found in different reactor analysis codes, but now comparisons are possible within one program. The logic of ELM is based on the one-dimensional conservation of energy in combination with Newton's Law of Cooling to determine the bulk flow temperature and the wall temperature across a control volume. Since the control volume is an incremental length of tube, the corresponding pressure drop is determined by application of the Law of Conservation of Momentum. The size, speed, and accuracy of ELM make it a simple tool for use in fuel element parametric studies. ELM is a machine independent program written in FORTRAN 77. It has been successfully compiled on an IBM PC compatible running MS-DOS using Lahey FORTRAN 77, a DEC VAX series computer running VMS, and a Sun4 series computer running SunOS UNIX. ELM requires 565K of RAM under SunOS 4.1, 360K of RAM under VMS 5.4, and 406K of RAM under MS-DOS. Because this program is machine independent, no executable is provided on the distribution media. The standard distribution medium for ELM is one 5.25 inch 360K MS-DOS format diskette. ELM was developed in 1991. DEC, VAX, and VMS are trademarks of Digital Equipment Corporation. Sun4 and SunOS are trademarks of Sun Microsystems, Inc. IBM PC is a registered trademark of International Business Machines. MS-DOS is a registered trademark of Microsoft Corporation.
Mulcahy, Nicholas J; Call, Josep; Dunbar, Robin I M
2005-02-01
Two important elements in problem solving are the abilities to encode relevant task features and to combine multiple actions to achieve the goal. The authors investigated these 2 elements in a task in which gorillas (Gorilla gorilla) and orangutans (Pongo pygmaeus) had to use a tool to retrieve an out-of-reach reward. Subjects were able to select tools of an appropriate length to reach the reward even when the position of the reward and tools were not simultaneously visible. When presented with tools that were too short to retrieve the reward, subjects were more likely to refuse to use them than when tools were the appropriate length. Subjects were proficient at using tools in sequence to retrieve the reward.
Finite Element Simulations of Micro Turning of Ti-6Al-4V using PCD and Coated Carbide tools
NASA Astrophysics Data System (ADS)
Jagadesh, Thangavel; Samuel, G. L.
2017-02-01
The demand for manufacturing axi-symmetric Ti-6Al-4V implants is increasing in biomedical applications and it involves micro turning process. To understand the micro turning process, in this work, a 3D finite element model has been developed for predicting the tool chip interface temperature, cutting, thrust and axial forces. Strain gradient effect has been included in the Johnson-Cook material model to represent the flow stress of the work material. To verify the simulation results, experiments have been conducted at four different feed rates and at three different cutting speeds. Since titanium alloy has low Young's modulus, spring back effect is predominant for higher edge radius coated carbide tool which leads to the increase in the forces. Whereas, polycrystalline diamond (PCD) tool has smaller edge radius that leads to lesser forces and decrease in tool chip interface temperature due to high thermal conductivity. Tool chip interface temperature increases by increasing the cutting speed, however the increase is less for PCD tool as compared to the coated carbide tool. When uncut chip thickness decreases, there is an increase in specific cutting energy due to material strengthening effects. Surface roughness is higher for coated carbide tool due to ploughing effect when compared with PCD tool. The average prediction error of finite element model for cutting and thrust forces are 11.45 and 14.87 % respectively.
Progress on the Multiphysics Capabilities of the Parallel Electromagnetic ACE3P Simulation Suite
DOE Office of Scientific and Technical Information (OSTI.GOV)
Kononenko, Oleksiy
2015-03-26
ACE3P is a 3D parallel simulation suite that is being developed at SLAC National Accelerator Laboratory. Effectively utilizing supercomputer resources, ACE3P has become a key tool for the coupled electromagnetic, thermal and mechanical research and design of particle accelerators. Based on the existing finite-element infrastructure, a massively parallel eigensolver is developed for modal analysis of mechanical structures. It complements a set of the multiphysics tools in ACE3P and, in particular, can be used for the comprehensive study of microphonics in accelerating cavities ensuring the operational reliability of a particle accelerator.
Finite Element Analysis of Adaptive-Stiffening and Shape-Control SMA Hybrid Composites
NASA Technical Reports Server (NTRS)
Gao, Xiu-Jie; Turner, Travis L.; Burton, Deborah; Brinson, L. Catherine
2005-01-01
The usage of shape memory materials has extended rapidly to many fields, including medical devices, actuators, composites, structures and MEMS devices. For these various applications, shape memory alloys (SMAs) are available in various forms: bulk, wire, ribbon, thin film, and porous. In this work, the focus is on SMA hybrid composites with adaptive-stiffening or morphing functions. These composites are created by using SMA ribbons or wires embedded in a polymeric based composite panel/beam. Adaptive stiffening or morphing is activated via selective resistance heating or uniform thermal loads. To simulate the thermomechanical behavior of these composites, a SMA model was implemented using ABAQUS user element interface and finite element simulations of the systems were studied. Several examples are presented which show that the implemented model can be a very useful design and simulation tool for SMA hybrid composites.
NASA Astrophysics Data System (ADS)
Abd El Baky, Hussien
This research work is devoted to theoretical and numerical studies on the flexural behaviour of FRP-strengthened concrete beams. The objectives of this research are to extend and generalize the results of simple experiments, to recommend new design guidelines based on accurate numerical tools, and to enhance our comprehension of the bond performance of such beams. These numerical tools can be exploited to bridge the existing gaps in the development of analysis and modelling approaches that can predict the behaviour of FRP-strengthened concrete beams. The research effort here begins with the formulation of a concrete model and development of FRP/concrete interface constitutive laws, followed by finite element simulations for beams strengthened in flexure. Finally, a statistical analysis is carried out taking the advantage of the aforesaid numerical tools to propose design guidelines. In this dissertation, an alternative incremental formulation of the M4 microplane model is proposed to overcome the computational complexities associated with the original formulation. Through a number of numerical applications, this incremental formulation is shown to be equivalent to the original M4 model. To assess the computational efficiency of the incremental formulation, the "arc-length" numerical technique is also considered and implemented in the original Bazant et al. [2000] M4 formulation. Finally, the M4 microplane concrete model is coded in FORTRAN and implemented as a user-defined subroutine into the commercial software package ADINA, Version 8.4. Then this subroutine is used with the finite element package to analyze various applications involving FRP strengthening. In the first application a nonlinear micromechanics-based finite element analysis is performed to investigate the interfacial behaviour of FRP/concrete joints subjected to direct shear loadings. The intention of this part is to develop a reliable bond--slip model for the FRP/concrete interface. The bond--slip relation is developed considering the interaction between the interfacial normal and shear stress components along the bonded length. A new approach is proposed to describe the entire tau-s relationship based on three separate models. The first model captures the shear response of an orthotropic FRP laminate. The second model simulates the shear characteristics of an adhesive layer, while the third model represents the shear nonlinearity of a thin layer inside the concrete, referred to as the interfacial layer. The proposed bond--slip model reflects the geometrical and material characteristics of the FRP, concrete, and adhesive layers. Two-dimensional and three-dimensional nonlinear displacement-controlled finite element (FE) models are then developed to investigate the flexural and FRP/concrete interfacial responses of FRP-strengthened reinforced concrete beams. The three-dimensional finite element model is created to accommodate cases of beams having FRP anchorage systems. Discrete interface elements are proposed and used to simulate the FRP/concrete interfacial behaviour before and after cracking. The FE models are capable of simulating the various failure modes, including debonding of the FRP either at the plate end or at intermediate cracks. Particular attention is focused on the effect of crack initiation and propagation on the interfacial behaviour. This study leads to an accurate and refined interpretation of the plate-end and intermediate crack debonding failure mechanisms for FRP-strengthened beams with and without FRP anchorage systems. Finally, the FE models are used to conduct a parametric study to generalize the findings of the FE analysis. The variables under investigation include two material characteristics; namely, the concrete compressive strength and axial stiffness of the FRP laminates as well as three geometric properties; namely, the steel reinforcement ratio, the beam span length and the beam depth. The parametric study is followed by a statistical analysis for 43 strengthened beams involving the five aforementioned variables. The response surface methodology (RSM) technique is employed to optimize the accuracy of the statistical models while minimizing the numbers of finite element runs. In particular, a face-centred design (FCD) is applied to evaluate the influence of the critical variables on the debonding load and debonding strain limits in the FRP laminates. Based on these statistical models, a nonlinear statistical regression analysis is used to propose design guidelines for the FRP flexural strengthening of reinforced concrete beams. (Abstract shortened by UMI.)
Enhanced Schapery Theory Software Development for Modeling Failure of Fiber-Reinforced Laminates
NASA Technical Reports Server (NTRS)
Pineda, Evan J.; Waas, Anthony M.
2013-01-01
Progressive damage and failure analysis (PDFA) tools are needed to predict the nonlinear response of advanced fiber-reinforced composite structures. Predictive tools should incorporate the underlying physics of the damage and failure mechanisms observed in the composite, and should utilize as few input parameters as possible. The purpose of the Enhanced Schapery Theory (EST) was to create a PDFA tool that operates in conjunction with a commercially available finite element (FE) code (Abaqus). The tool captures the physics of the damage and failure mechanisms that result in the nonlinear behavior of the material, and the failure methodology employed yields numerical results that are relatively insensitive to changes in the FE mesh. The EST code is written in Fortran and compiled into a static library that is linked to Abaqus. A Fortran Abaqus UMAT material subroutine is used to facilitate the communication between Abaqus and EST. A clear distinction between damage and failure is imposed. Damage mechanisms result in pre-peak nonlinearity in the stress strain curve. Four internal state variables (ISVs) are utilized to control the damage and failure degradation. All damage is said to result from matrix microdamage, and a single ISV marks the micro-damage evolution as it is used to degrade the transverse and shear moduli of the lamina using a set of experimentally obtainable matrix microdamage functions. Three separate failure ISVs are used to incorporate failure due to fiber breakage, mode I matrix cracking, and mode II matrix cracking. Failure initiation is determined using a failure criterion, and the evolution of these ISVs is controlled by a set of traction-separation laws. The traction separation laws are postulated such that the area under the curves is equal to the fracture toughness of the material associated with the corresponding failure mechanism. A characteristic finite element length is used to transform the traction-separation laws into stress-strain laws. The ISV evolution equations are derived in a thermodynamically consistent manner by invoking the stationary principle on the total work of the system with respect to each ISV. A novel feature is the inclusion of both pre-peak damage and appropriately scaled, post-peak strain softening failure. Also, the characteristic elements used in the failure degradation scheme are calculated using the element nodal coordinates, rather than simply the square root of the area of the element.
Building Stories about Sea Level Rise through Interactive Visualizations
NASA Astrophysics Data System (ADS)
Stephens, S. H.; DeLorme, D. E.; Hagen, S. C.
2013-12-01
Digital media provide storytellers with dynamic new tools for communicating about scientific issues via interactive narrative visualizations. While traditional storytelling uses plot, characterization, and point of view to engage audiences with underlying themes and messages, interactive visualizations can be described as 'narrative builders' that promote insight through the process of discovery (Dove, G. & Jones, S. 2012, Proc. IHCI 2012). Narrative visualizations are used in online journalism to tell complex stories that allow readers to select aspects of datasets to explore and construct alternative interpretations of information (Segel, E. & Heer, J. 2010, IEEE Trans. Vis. Comp. Graph.16, 1139), thus enabling them to participate in the story-building process. Nevertheless, narrative visualizations also incorporate author-selected narrative elements that help guide and constrain the overall themes and messaging of the visualization (Hullman, J. & Diakopoulos, N. 2011, IEEE Trans. Vis. Comp. Graph. 17, 2231). One specific type of interactive narrative visualization that is used for science communication is the sea level rise (SLR) viewer. SLR viewers generally consist of a base map, upon which projections of sea level rise scenarios can be layered, and various controls for changing the viewpoint and scenario parameters. They are used to communicate the results of scientific modeling and help readers visualize the potential impacts of SLR on the coastal zone. Readers can use SLR viewers to construct personal narratives of the effects of SLR under different scenarios in locations that are important to them, thus extending the potential reach and impact of scientific research. With careful selection of narrative elements that guide reader interpretation, the communicative aspects of these visualizations may be made more effective. This presentation reports the results of a content analysis of a subset of existing SLR viewers selected in order to comprehensively identify and characterize the narrative elements that contribute to this storytelling medium. The results describe four layers of narrative elements in these viewers: data, visual representations, annotations, and interactivity; and explain the ways in which these elements are used to communicate about SLR. Most existing SLR viewers have been designed with attention to technical usability; however, careful design of narrative elements could increase their overall effectiveness as story-building tools. The analysis concludes with recommendations for narrative elements that should be considered when designing new SLR viewers, and offers suggestions for integrating these components to balance author-driven and reader-driven design features for more effective messaging.
LIBS: a potential tool for industrial/agricultural waste water analysis
NASA Astrophysics Data System (ADS)
Karpate, Tanvi; K. M., Muhammed Shameem; Nayak, Rajesh; V. K., Unnikrishnan; Santhosh, C.
2016-04-01
Laser Induced Breakdown Spectroscopy (LIBS) is a multi-elemental analysis technique with various advantages and has the ability to detect any element in real time. This technique holds a potential for environmental monitoring and various such analysis has been done in soil, glass, paint, water, plastic etc confirms the robustness of this technique for such applications. Compared to the currently available water quality monitoring methods and techniques, LIBS has several advantages, viz. no need for sample preparation, fast and easy operation, and chemical free during the process. In LIBS, powerful pulsed laser generates plasma which is then analyzed to get quantitative and qualitative details of the elements present in the sample. Another main advantage of LIBS technique is that it can perform in standoff mode for real time analysis. Water samples from industries and agricultural strata tend to have a lot of pollutants making it harmful for consumption. The emphasis of this project is to determine such harmful pollutants present in trace amounts in industrial and agricultural wastewater. When high intensity laser is made incident on the sample, a plasma is generated which gives a multielemental emission spectra. LIBS analysis has shown outstanding success for solids samples. For liquid samples, the analysis is challenging as the liquid sample has the chances of splashing due to the high energy of laser and thus making it difficult to generate plasma. This project also deals with determining the most efficient method for testing of water sample for qualitative as well as quantitative analysis using LIBS.
NASA Astrophysics Data System (ADS)
Genberg, Victor L.; Michels, Gregory J.
2017-08-01
The ultimate design goal of an optical system subjected to dynamic loads is to minimize system level wavefront error (WFE). In random response analysis, system WFE is difficult to predict from finite element results due to the loss of phase information. In the past, the use of ystem WFE was limited by the difficulty of obtaining a linear optics model. In this paper, an automated method for determining system level WFE using a linear optics model is presented. An error estimate is included in the analysis output based on fitting errors of mode shapes. The technique is demonstrated by example with SigFit, a commercially available tool integrating mechanical analysis with optical analysis.
NASA Technical Reports Server (NTRS)
Townsend, John S.; Peck, Jeff; Ayala, Samuel
2000-01-01
NASA has funded several major programs (the Probabilistic Structural Analysis Methods Project is an example) to develop probabilistic structural analysis methods and tools for engineers to apply in the design and assessment of aerospace hardware. A probabilistic finite element software code, known as Numerical Evaluation of Stochastic Structures Under Stress, is used to determine the reliability of a critical weld of the Space Shuttle solid rocket booster aft skirt. An external bracket modification to the aft skirt provides a comparison basis for examining the details of the probabilistic analysis and its contributions to the design process. Also, analysis findings are compared with measured Space Shuttle flight data.
Front lighted optical tooling method and apparatus
Stone, W.J.
1983-06-30
An optical tooling method and apparatus uses a front lighted shadowgraphic technique to enhance visual contrast of reflected light. The apparatus includes an optical assembly including a fiducial mark, such as cross hairs, reflecting polarized light with a first polarization, a polarizing element backing the fiducial mark and a reflective surface backing the polarizing element for reflecting polarized light bypassing the fiducial mark and traveling through the polarizing element. The light reflected by the reflecting surface is directed through a second pass of the polarizing element toward the frontal direction with a polarization differing from the polarization of the light reflected by the fiducial mark. When used as a tooling target, the optical assembly may be mounted directly to a reference surface or may be secured in a mounting, such as a magnetic mounting. The optical assembly may also be mounted in a plane defining structure and used as a spherometer in conjunction with an optical depth measuring instrument.
Microscopic analysis of Hopper flow with ellipsoidal particles
NASA Astrophysics Data System (ADS)
Liu, Sida; Zhou, Zongyan; Zou, Ruiping; Pinson, David; Yu, Aibing
2013-06-01
Hoppers are widely used in process industries. With such widespread application, difficulties in achieving desired operational behaviors have led to extensive experimental and mathematical studies in the past decades. Particularly, the discrete element method has become one of the most important simulation tools for design and analysis. So far, most studies are on spherical particles for computational convenience. In this work, ellipsoidal particles are used as they can represent a large variation of particle shapes. Hopper flow with ellipsoidal particles is presented highlighting the effect of particle shape on the microscopic properties.
Multidisciplinary Tool for Systems Analysis of Planetary Entry, Descent, and Landing
NASA Technical Reports Server (NTRS)
Samareh, Jamshid A.
2011-01-01
Systems analysis of a planetary entry (SAPE), descent, and landing (EDL) is a multidisciplinary activity in nature. SAPE improves the performance of the systems analysis team by automating and streamlining the process, and this improvement can reduce the errors that stem from manual data transfer among discipline experts. SAPE is a multidisciplinary tool for systems analysis of planetary EDL for Venus, Earth, Mars, Jupiter, Saturn, Uranus, Neptune, and Titan. It performs EDL systems analysis for any planet, operates cross-platform (i.e., Windows, Mac, and Linux operating systems), uses existing software components and open-source software to avoid software licensing issues, performs low-fidelity systems analysis in one hour on a computer that is comparable to an average laptop, and keeps discipline experts in the analysis loop. SAPE uses Python, a platform-independent, open-source language, for integration and for the user interface. Development has relied heavily on the object-oriented programming capabilities that are available in Python. Modules are provided to interface with commercial and government off-the-shelf software components (e.g., thermal protection systems and finite-element analysis). SAPE currently includes the following analysis modules: geometry, trajectory, aerodynamics, aerothermal, thermal protection system, and interface for structural sizing.
Yazaki, Junshi; Kikuchi, Shoshi
2005-01-01
We now have the various genomics tools for monocot (Oryza sativa) and a dicot (Arabidopsis thaliana) plant. Plant is not only a very important agricultural resource but also a model organism for biological research. It is important that the interaction between ABA and GA is investigated for controlling the transition from embryogenesis to germination in seeds using genomics tools. These studies have investigated the relationship between dormancy and germination using genomics tools. Genomics tools identified genes that had never before been annotated as ABA- or GA-responsive genes in plant, detected new interactions between genes responsive to the two hormones, comprehensively characterized cis-elements of hormone-responsive genes, and characterized cis-elements of rice and Arabidopsis. In these research, ABA- and GA-regulated genes have been classified as functional proteins (proteins that probably function in stress or PR tolerance) and regulatory proteins (protein factors involved in further regulation of signal transduction). Comparison between ABA and/or GA-responsive genes in rice and those in Arabidopsis has shown that the cis-element has specificity in each species. cis-Elements for the dehydration-stress response have been specified in Arabidopsis but not in rice. cis-Elements for protein storage are remarkably richer in the upstream regions of the rice gene than in those of Arabidopsis.
CAS-viewer: web-based tool for splicing-guided integrative analysis of multi-omics cancer data.
Han, Seonggyun; Kim, Dongwook; Kim, Youngjun; Choi, Kanghoon; Miller, Jason E; Kim, Dokyoon; Lee, Younghee
2018-04-20
The Cancer Genome Atlas (TCGA) project is a public resource that provides transcriptomic, DNA sequence, methylation, and clinical data for 33 cancer types. Transforming the large size and high complexity of TCGA cancer genome data into integrated knowledge can be useful to promote cancer research. Alternative splicing (AS) is a key regulatory mechanism of genes in human cancer development and in the interaction with epigenetic factors. Therefore, AS-guided integration of existing TCGA data sets will make it easier to gain insight into the genetic architecture of cancer risk and related outcomes. There are already existing tools analyzing and visualizing alternative mRNA splicing patterns for large-scale RNA-seq experiments. However, these existing web-based tools are limited to the analysis of individual TCGA data sets at a time, such as only transcriptomic information. We implemented CAS-viewer (integrative analysis of Cancer genome data based on Alternative Splicing), a web-based tool leveraging multi-cancer omics data from TCGA. It illustrates alternative mRNA splicing patterns along with methylation, miRNAs, and SNPs, and then provides an analysis tool to link differential transcript expression ratio to methylation, miRNA, and splicing regulatory elements for 33 cancer types. Moreover, one can analyze AS patterns with clinical data to identify potential transcripts associated with different survival outcome for each cancer. CAS-viewer is a web-based application for transcript isoform-driven integration of multi-omics data in multiple cancer types and will aid in the visualization and possible discovery of biomarkers for cancer by integrating multi-omics data from TCGA.
GC/MS analysis of pesticides in the Ferrara area (Italy) surface water: a chemometric study.
Pasti, Luisa; Nava, Elisabetta; Morelli, Marco; Bignami, Silvia; Dondi, Francesco
2007-01-01
The development of a network to monitor surface waters is a critical element in the assessment, restoration and protection of water quality. In this study, concentrations of 42 pesticides--determined by GC-MS on samples from 11 points along the Ferrara area rivers--have been analyzed by chemometric tools. The data were collected over a three-year period (2002-2004). Principal component analysis of the detected pesticides was carried out in order to define the best spatial locations for the sampling points. The results obtained have been interpreted in view of agricultural land use. Time series data regarding pesticide contents in surface waters has been analyzed using the Autocorrelation function. This chemometric tool allows for seasonal trends and makes it possible to optimize sampling frequency in order to detect the effective maximum pesticide content.
Transmission Bearing Damage Detection Using Decision Fusion Analysis
NASA Technical Reports Server (NTRS)
Dempsey, Paula J.; Lewicki, David G.; Decker, Harry J.
2004-01-01
A diagnostic tool was developed for detecting fatigue damage to rolling element bearings in an OH-58 main rotor transmission. Two different monitoring technologies, oil debris analysis and vibration, were integrated using data fusion into a health monitoring system for detecting bearing surface fatigue pitting damage. This integrated system showed improved detection and decision-making capabilities as compared to using individual monitoring technologies. This diagnostic tool was evaluated by collecting vibration and oil debris data from tests performed in the NASA Glenn 500 hp Helicopter Transmission Test Stand. Data was collected during experiments performed in this test rig when two unanticipated bearing failures occurred. Results show that combining the vibration and oil debris measurement technologies improves the detection of pitting damage on spiral bevel gears duplex ball bearings and spiral bevel pinion triplex ball bearings in a main rotor transmission.
Modeling the Multi-Body System Dynamics of a Flexible Solar Sail Spacecraft
NASA Technical Reports Server (NTRS)
Kim, Young; Stough, Robert; Whorton, Mark
2005-01-01
Solar sail propulsion systems enable a wide range of space missions that are not feasible with current propulsion technology. Hardware concepts and analytical methods have matured through ground development to the point that a flight validation mission is now realizable. Much attention has been given to modeling the structural dynamics of the constituent elements, but to date an integrated system level dynamics analysis has been lacking. Using a multi-body dynamics and control analysis tool called TREETOPS, the coupled dynamics of the sailcraft bus, sail membranes, flexible booms, and control system sensors and actuators of a representative solar sail spacecraft are investigated to assess system level dynamics and control issues. With this tool, scaling issues and parametric trade studies can be performed to study achievable performance, control authority requirements, and control/structure interaction assessments.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Zhang, L.; Lanza, R.C.
1999-12-01
The authors have developed a near field coded aperture imaging system for use with fast neutron techniques as a tool for the detection of contraband and hidden explosives through nuclear elemental analysis. The technique relies on the prompt gamma rays produced by fast neutron interactions with the object being examined. The position of the nuclear elements is determined by the location of the gamma emitters. For existing fast neutron techniques, in Pulsed Fast Neutron Analysis (PFNA), neutrons are used with very low efficiency; in Fast Neutron Analysis (FNS), the sensitivity for detection of the signature gamma rays is very low.more » For the Coded Aperture Fast Neutron Analysis (CAFNA{reg{underscore}sign}) the authors have developed, the efficiency for both using the probing fast neutrons and detecting the prompt gamma rays is high. For a probed volume of n{sup 3} volume elements (voxels) in a cube of n resolution elements on a side, they can compare the sensitivity with other neutron probing techniques. As compared to PFNA, the improvement for neutron utilization is n{sup 2}, where the total number of voxels in the object being examined is n{sup 3}. Compared to FNA, the improvement for gamma-ray imaging is proportional to the total open area of the coded aperture plane; a typical value is n{sup 2}/2, where n{sup 2} is the number of total detector resolution elements or the number of pixels in an object layer. It should be noted that the actual signal to noise ratio of a system depends also on the nature and distribution of background events and this comparison may reduce somewhat the effective sensitivity of CAFNA. They have performed analysis, Monte Carlo simulations, and preliminary experiments using low and high energy gamma-ray sources. The results show that a high sensitivity 3-D contraband imaging and detection system can be realized by using CAFNA.« less
Pegg, Elise C; Gill, Harinderjit S
2016-09-06
A new software tool to assign the material properties of bone to an ABAQUS finite element mesh was created and compared with Bonemat, a similar tool originally designed to work with Ansys finite element models. Our software tool (py_bonemat_abaqus) was written in Python, which is the chosen scripting language for ABAQUS. The purpose of this study was to compare the software packages in terms of the material assignment calculation and processing speed. Three element types were compared (linear hexahedral (C3D8), linear tetrahedral (C3D4) and quadratic tetrahedral elements (C3D10)), both individually and as part of a mesh. Comparisons were made using a CT scan of a hemi-pelvis as a test case. A small difference, of -0.05kPa on average, was found between Bonemat version 3.1 (the current version) and our Python package. Errors were found in the previous release of Bonemat (version 3.0 downloaded from www.biomedtown.org) during calculation of the quadratic tetrahedron Jacobian, and conversion of the apparent density to modulus when integrating over the Young׳s modulus field. These issues caused up to 2GPa error in the modulus assignment. For these reasons, we recommend users upgrade to the most recent release of Bonemat. Processing speeds were assessed for the three different element types. Our Python package took significantly longer (110s on average) to perform the calculations compared with the Bonemat software (10s). Nevertheless, the workflow advantages of the package and added functionality makes 'py_bonemat_abaqus' a useful tool for ABAQUS users. Copyright © 2016 Elsevier Ltd. All rights reserved.
Telfer, Scott; Erdemir, Ahmet; Woodburn, James; Cavanagh, Peter R
2014-01-01
Over the past two decades finite element (FE) analysis has become a popular tool for researchers seeking to simulate the biomechanics of the healthy and diabetic foot. The primary aims of these simulations have been to improve our understanding of the foot's complicated mechanical loading in health and disease and to inform interventions designed to prevent plantar ulceration, a major complication of diabetes. This article provides a systematic review and summary of the findings from FE analysis-based computational simulations of the diabetic foot. A systematic literature search was carried out and 31 relevant articles were identified covering three primary themes: methodological aspects relevant to modelling the diabetic foot; investigations of the pathomechanics of the diabetic foot; and simulation-based design of interventions to reduce ulceration risk. Methodological studies illustrated appropriate use of FE analysis for simulation of foot mechanics, incorporating nonlinear tissue mechanics, contact and rigid body movements. FE studies of pathomechanics have provided estimates of internal soft tissue stresses, and suggest that such stresses may often be considerably larger than those measured at the plantar surface and are proportionally greater in the diabetic foot compared to controls. FE analysis allowed evaluation of insole performance and development of new insole designs, footwear and corrective surgery to effectively provide intervention strategies. The technique also presents the opportunity to simulate the effect of changes associated with the diabetic foot on non-mechanical factors such as blood supply to local tissues. While significant advancement in diabetic foot research has been made possible by the use of FE analysis, translational utility of this powerful tool for routine clinical care at the patient level requires adoption of cost-effective (both in terms of labour and computation) and reliable approaches with clear clinical validity for decision making.
Kruser, Jacqueline M; Taylor, Lauren J; Campbell, Toby C; Zelenski, Amy; Johnson, Sara K; Nabozny, Michael J; Steffens, Nicole M; Tucholka, Jennifer L; Kwekkeboom, Kris L; Schwarze, Margaret L
2017-04-01
Older adults often have surgery in the months preceding death, which can initiate postoperative treatments inconsistent with end-of-life values. "Best Case/Worst Case" (BC/WC) is a communication tool designed to promote goal-concordant care during discussions about high-risk surgery. The objective of this study was to evaluate a structured training program designed to teach surgeons how to use BC/WC. Twenty-five surgeons from one tertiary care hospital completed a two-hour training session followed by individual coaching. We audio-recorded surgeons using BC/WC with standardized patients and 20 hospitalized patients. Hospitalized patients and their families participated in an open-ended interview 30 to 120 days after enrollment. We used a checklist of 11 BC/WC elements to measure tool fidelity and surgeons completed the Practitioner Opinion Survey to measure acceptability of the tool. We used qualitative analysis to evaluate variability in tool content and to characterize patient and family perceptions of the tool. Surgeons completed a median of 10 of 11 BC/WC elements with both standardized and hospitalized patients (range 5-11). We found moderate variability in presentation of treatment options and description of outcomes. Three months after training, 79% of surgeons reported BC/WC is better than their usual approach and 71% endorsed active use of BC/WC in clinical practice. Patients and families found that BC/WC established expectations, provided clarity, and facilitated deliberation. Surgeons can learn to use BC/WC with older patients considering acute high-risk surgical interventions. Surgeons, patients, and family members endorse BC/WC as a strategy to support complex decision making. Copyright © 2017 American Academy of Hospice and Palliative Medicine. Published by Elsevier Inc. All rights reserved.
Buddhachat, Kittisak; Klinhom, Sarisa; Siengdee, Puntita; Brown, Janine L; Nomsiri, Raksiri; Kaewmong, Patcharaporn; Thitaram, Chatchote; Mahakkanukrauh, Pasuk; Nganvongpanit, Korakot
2016-01-01
Mineralized tissues accumulate elements that play crucial roles in animal health. Although elemental content of bone, blood and teeth of human and some animal species have been characterized, data for many others are lacking, as well as species comparisons. Here we describe the distribution of elements in horn (Bovidae), antler (Cervidae), teeth and bone (humerus) across a number of species determined by handheld X-ray fluorescence (XRF) to better understand differences and potential biological relevance. A difference in elemental profiles between horns and antlers was observed, possibly due to the outer layer of horns being comprised of keratin, whereas antlers are true bone. Species differences in tissue elemental content may be intrinsic, but also related to feeding habits that contribute to mineral accumulation, particularly for toxic heavy metals. One significant finding was a higher level of iron (Fe) in the humerus bone of elephants compared to other species. This may be an adaptation of the hematopoietic system by distributing Fe throughout the bone rather than the marrow, as elephant humerus lacks a marrow cavity. We also conducted discriminant analysis and found XRF was capable of distinguishing samples from different species, with humerus bone being the best source for species discrimination. For example, we found a 79.2% correct prediction and success rate of 80% for classification between human and non-human humerus bone. These findings show that handheld XRF can serve as an effective tool for the biological study of elemental composition in mineralized tissue samples and may have a forensic application.
Buddhachat, Kittisak; Klinhom, Sarisa; Siengdee, Puntita; Brown, Janine L.; Nomsiri, Raksiri; Kaewmong, Patcharaporn; Thitaram, Chatchote; Mahakkanukrauh, Pasuk; Nganvongpanit, Korakot
2016-01-01
Mineralized tissues accumulate elements that play crucial roles in animal health. Although elemental content of bone, blood and teeth of human and some animal species have been characterized, data for many others are lacking, as well as species comparisons. Here we describe the distribution of elements in horn (Bovidae), antler (Cervidae), teeth and bone (humerus) across a number of species determined by handheld X-ray fluorescence (XRF) to better understand differences and potential biological relevance. A difference in elemental profiles between horns and antlers was observed, possibly due to the outer layer of horns being comprised of keratin, whereas antlers are true bone. Species differences in tissue elemental content may be intrinsic, but also related to feeding habits that contribute to mineral accumulation, particularly for toxic heavy metals. One significant finding was a higher level of iron (Fe) in the humerus bone of elephants compared to other species. This may be an adaptation of the hematopoietic system by distributing Fe throughout the bone rather than the marrow, as elephant humerus lacks a marrow cavity. We also conducted discriminant analysis and found XRF was capable of distinguishing samples from different species, with humerus bone being the best source for species discrimination. For example, we found a 79.2% correct prediction and success rate of 80% for classification between human and non-human humerus bone. These findings show that handheld XRF can serve as an effective tool for the biological study of elemental composition in mineralized tissue samples and may have a forensic application. PMID:27196603
NASA Astrophysics Data System (ADS)
Kubala-Kukuś, A.; Banaś, D.; Braziewicz, J.; Góźdź, S.; Majewska, U.; Pajek, M.
2007-07-01
The total reflection X-ray fluorescence method was applied to study the trace element concentrations in human breast malignant and breast benign neoplasm tissues taken from the women who were patients of Holycross Cancer Centre in Kielce (Poland). These investigations were mainly focused on the development of new possibilities of cancer diagnosis and therapy monitoring. This systematic comparative study was based on relatively large (˜ 100) population studied, namely 26 samples of breast malignant and 68 samples of breast benign neoplasm tissues. The concentrations, being in the range from a few ppb to 0.1%, were determined for thirteen elements (from P to Pb). The results were carefully analysed to investigate the concentration distribution of trace elements in the studied samples. The measurements of concentration of trace elements by total reflection X-ray fluorescence were limited, however, by the detection limit of the method. It was observed that for more than 50% of elements determined, the concentrations were not measured in all samples. These incomplete measurements were treated within the statistical concept called left-random censoring and for the estimation of the mean value and median of censored concentration distributions, the Kaplan-Meier estimator was used. For comparison of concentrations in two populations, the log-rank test was applied, which allows to compare the censored total reflection X-ray fluorescence data. Found statistically significant differences are discussed in more details. It is noted that described data analysis procedures should be the standard tool to analyze the censored concentrations of trace elements analysed by X-ray fluorescence methods.
Parametric Modelling of As-Built Beam Framed Structure in Bim Environment
NASA Astrophysics Data System (ADS)
Yang, X.; Koehl, M.; Grussenmeyer, P.
2017-02-01
A complete documentation and conservation of a historic timber roof requires the integration of geometry modelling, attributional and dynamic information management and results of structural analysis. Recently developed as-built Building Information Modelling (BIM) technique has the potential to provide a uniform platform, which provides possibility to integrate the traditional geometry modelling, parametric elements management and structural analysis together. The main objective of the project presented in this paper is to develop a parametric modelling tool for a timber roof structure whose elements are leaning and crossing beam frame. Since Autodesk Revit, as the typical BIM software, provides the platform for parametric modelling and information management, an API plugin, able to automatically create the parametric beam elements and link them together with strict relationship, was developed. The plugin under development is introduced in the paper, which can obtain the parametric beam model via Autodesk Revit API from total station points and terrestrial laser scanning data. The results show the potential of automatizing the parametric modelling by interactive API development in BIM environment. It also integrates the separate data processing and different platforms into the uniform Revit software.
NASA Technical Reports Server (NTRS)
Pak, Chan-Gi
2013-01-01
Modern aircraft employ a significant fraction of their weight in composite materials to reduce weight and improve performance. Aircraft aeroservoelastic models are typically characterized by significant levels of model parameter uncertainty due to the composite manufacturing process. Small modeling errors in the finite element model will eventually induce errors in the structural flexibility and mass, thus propagating into unpredictable errors in the unsteady aerodynamics and the control law design. One of the primary objectives of Multi Utility Technology Test-bed (MUTT) aircraft is the flight demonstration of active flutter suppression, and therefore in this study, the identification of the primary and secondary modes for the structural model tuning based on the flutter analysis of MUTT aircraft. The ground vibration test-validated structural dynamic finite element model of the MUTT aircraft is created in this study. The structural dynamic finite element model of MUTT aircraft is improved using the in-house Multi-disciplinary Design, Analysis, and Optimization tool. In this study, two different weight configurations of MUTT aircraft have been improved simultaneously in a single model tuning procedure.
Integrated computer-aided design using minicomputers
NASA Technical Reports Server (NTRS)
Storaasli, O. O.
1980-01-01
Computer-Aided Design/Computer-Aided Manufacturing (CAD/CAM), a highly interactive software, has been implemented on minicomputers at the NASA Langley Research Center. CAD/CAM software integrates many formerly fragmented programs and procedures into one cohesive system; it also includes finite element modeling and analysis, and has been interfaced via a computer network to a relational data base management system and offline plotting devices on mainframe computers. The CAD/CAM software system requires interactive graphics terminals operating at a minimum of 4800 bits/sec transfer rate to a computer. The system is portable and introduces 'interactive graphics', which permits the creation and modification of models interactively. The CAD/CAM system has already produced designs for a large area space platform, a national transonic facility fan blade, and a laminar flow control wind tunnel model. Besides the design/drafting element analysis capability, CAD/CAM provides options to produce an automatic program tooling code to drive a numerically controlled (N/C) machine. Reductions in time for design, engineering, drawing, finite element modeling, and N/C machining will benefit productivity through reduced costs, fewer errors, and a wider range of configuration.
Proton Induced X-Ray Emission (PIXE): Determining the Concentration of Samples
NASA Astrophysics Data System (ADS)
McCarthy, Mallory; Rodriguez Manso, Alis; Pajouhafsar, Yasmin; J Yennello, Sherry
2017-09-01
We used Proton Induced X-ray Emission (PIXE) as an analysis technique to determine the composition of samples, in particular, the elemental constituents and the concentrations. Each of the samples are bombarded with protons, which in result displaces a lower level electron and causes a higher level electron to fall into its place. This displacement produces characteristic x-rays that are `fingerprints' for each element. The protons supplied for the bombardment are produced and accelerated by the K150 proton beam in the Cyclotron Institute at Texas A&M University. The products are detected by three x-ray detectors: XR-100CR Si-PIN, XR-100SDD, and XR-100T CdTe. The peaks of the spectrum are analyzed using a software analysis tool, GUPIXWIN, to determine the concentration of the known elements of each particular sample. The goals of this work are to test run the Proton Induced X-Ray Emission experimental set up at Texas A&M University (TAMU) and to determine the concentration of thin films containing KBr given by the TAMU Chemical Engineering Department.
Campbell, J Q; Petrella, A J
2016-09-06
Population-based modeling of the lumbar spine has the potential to be a powerful clinical tool. However, developing a fully parameterized model of the lumbar spine with accurate geometry has remained a challenge. The current study used automated methods for landmark identification to create a statistical shape model of the lumbar spine. The shape model was evaluated using compactness, generalization ability, and specificity. The primary shape modes were analyzed visually, quantitatively, and biomechanically. The biomechanical analysis was performed by using the statistical shape model with an automated method for finite element model generation to create a fully parameterized finite element model of the lumbar spine. Functional finite element models of the mean shape and the extreme shapes (±3 standard deviations) of all 17 shape modes were created demonstrating the robust nature of the methods. This study represents an advancement in finite element modeling of the lumbar spine and will allow population-based modeling in the future. Copyright © 2016 Elsevier Ltd. All rights reserved.
A self-report critical incident assessment tool for army night vision goggle helicopter operations.
Renshaw, Peter F; Wiggins, Mark W
2007-04-01
The present study sought to examine the utility of a self-report tool that was designed as a partial substitute for a face-to-face cognitive interview for critical incidents involving night vision goggles (NVGs). The use of NVGs remains problematic within the military environment, as these devices have been identified as a factor in a significant proportion of aircraft accidents and incidents. The self-report tool was structured to identify some of the cognitive features of human performance that were associated with critical incidents involving NVGs. The tool incorporated a number of different levels of analysis, ranging from specific behavioral responses to broader cognitive constructs. Reports were received from 30 active pilots within the Australian Army using the NVG Critical Incident Assessment Tool (NVGCIAT). The results revealed a correspondence between specific types of NVG-related errors and elements of the Human Factors Analysis and Classification System (HFACS). In addition, uncertainty emerged as a significant factor associated with the critical incidents that were recalled by operators. These results were broadly consistent with previous research and provide some support for the utility of subjective assessment tools as a means of extracting critical incident-related data when face-to-face cognitive interviews are not possible. In some circumstances, the NVGCIAT might be regarded as a substitute cognitive interview protocol with some level of diagnosticity.
Naville, Magali; Gautheret, Daniel
2010-01-01
Bacterial transcription attenuation occurs through a variety of cis-regulatory elements that control gene expression in response to a wide range of signals. The signal-sensing structures in attenuators are so diverse and rapidly evolving that only a small fraction have been properly annotated and characterized to date. Here we apply a broad-spectrum detection tool in order to achieve a more complete view of the transcriptional attenuation complement of key bacterial species. Our protocol seeks gene families with an unusual frequency of 5' terminators found across multiple species. Many of the detected attenuators are part of annotated elements, such as riboswitches or T-boxes, which often operate through transcriptional attenuation. However, a significant fraction of candidates were not previously characterized in spite of their unmistakable footprint. We further characterized some of these new elements using sequence and secondary structure analysis. We also present elements that may control the expression of several non-homologous genes, suggesting co-transcription and response to common signals. An important class of such elements, which we called mobile attenuators, is provided by 3' terminators of insertion sequences or prophages that may be exapted as 5' regulators when inserted directly upstream of a cellular gene. We show here that attenuators involve a complex landscape of signal-detection structures spanning the entire bacterial domain. We discuss possible scenarios through which these diverse 5' regulatory structures may arise or evolve.
Discrete elements for 3D microfluidics.
Bhargava, Krisna C; Thompson, Bryant; Malmstadt, Noah
2014-10-21
Microfluidic systems are rapidly becoming commonplace tools for high-precision materials synthesis, biochemical sample preparation, and biophysical analysis. Typically, microfluidic systems are constructed in monolithic form by means of microfabrication and, increasingly, by additive techniques. These methods restrict the design and assembly of truly complex systems by placing unnecessary emphasis on complete functional integration of operational elements in a planar environment. Here, we present a solution based on discrete elements that liberates designers to build large-scale microfluidic systems in three dimensions that are modular, diverse, and predictable by simple network analysis techniques. We develop a sample library of standardized components and connectors manufactured using stereolithography. We predict and validate the flow characteristics of these individual components to design and construct a tunable concentration gradient generator with a scalable number of parallel outputs. We show that these systems are rapidly reconfigurable by constructing three variations of a device for generating monodisperse microdroplets in two distinct size regimes and in a high-throughput mode by simple replacement of emulsifier subcircuits. Finally, we demonstrate the capability for active process monitoring by constructing an optical sensing element for detecting water droplets in a fluorocarbon stream and quantifying their size and frequency. By moving away from large-scale integration toward standardized discrete elements, we demonstrate the potential to reduce the practice of designing and assembling complex 3D microfluidic circuits to a methodology comparable to that found in the electronics industry.
Crystal Structure Prediction via Deep Learning.
Ryan, Kevin; Lengyel, Jeff; Shatruk, Michael
2018-06-06
We demonstrate the application of deep neural networks as a machine-learning tool for the analysis of a large collection of crystallographic data contained in the crystal structure repositories. Using input data in the form of multi-perspective atomic fingerprints, which describe coordination topology around unique crystallographic sites, we show that the neural-network model can be trained to effectively distinguish chemical elements based on the topology of their crystallographic environment. The model also identifies structurally similar atomic sites in the entire dataset of ~50000 crystal structures, essentially uncovering trends that reflect the periodic table of elements. The trained model was used to analyze templates derived from the known binary and ternary crystal structures in order to predict the likelihood to form new compounds that could be generated by placing elements into these structural templates in combinatorial fashion. Statistical analysis of predictive performance of the neural-network model, which was applied to a test set of structures never seen by the model during training, indicates its ability to predict known elemental compositions with a high likelihood of success. In ~30% of cases, the known compositions were found among top-10 most likely candidates proposed by the model. These results suggest that the approach developed in this work can be used to effectively guide the synthetic efforts in the discovery of new materials, especially in the case of systems composed of 3 or more chemical elements.
Design/Analysis of the JWST ISIM Bonded Joints for Survivability at Cryogenic Temperatures
NASA Technical Reports Server (NTRS)
Bartoszyk, Andrew; Johnston, John; Kaprielian, Charles; Kuhn, Jonathan; Kunt, Cengiz; Rodini,Benjamin; Young, Daniel
1990-01-01
A major design and analysis challenge for the JWST ISIM structure is thermal survivability of metal/composite bonded joints below the cryogenic temperature of 30K (-405 F). Current bonded joint concepts include internal invar plug fittings, external saddle titanium/invar fittings and composite gusset/clip joints all bonded to M55J/954-6 and T300/954-6 hybrid composite tubes (75mm square). Analytical experience and design work done on metal/composite bonded joints at temperatures below that of liquid nitrogen are limited and important analysis tools, material properties, and failure criteria for composites at cryogenic temperatures are sparse in the literature. Increasing this challenge is the difficulty in testing for these required tools and properties at cryogenic temperatures. To gain confidence in analyzing and designing the ISIM joints, a comprehensive joint development test program has been planned and is currently running. The test program is designed to produce required analytical tools and develop a composite failure criterion for bonded joint strengths at cryogenic temperatures. Finite element analysis is used to design simple test coupons that simulate anticipated stress states in the flight joints; subsequently the test results are used to correlate the analysis technique for the final design of the bonded joints. In this work, we present an overview of the analysis and test methodology, current results, and working joint designs based on developed techniques and properties.
Gage for micromachining system
Miller, Donald M.
1979-02-27
A gage for measuring the contour of the surface of an element of a micromachining tool system and of a work piece machined by the micromachining tool system. The gage comprises a glass plate containing two electrical contacts and supporting a steel ball resting against the contacts. As the element or workpiece is moved against the steel ball, the very slight contact pressure causes an extremely small movement of the steel ball which breaks the electrical circuit between the two contacts. The contour information is supplied to a dedicated computer controlling the micromachining tool so that the computer knows the contour of the element and the work piece to an accuracy of .+-. 25 nm. The micromachining tool system with X- and omega-axes is used to machine spherical, aspherical, and irregular surfaces with a maximum contour error of 100 nanometers (nm) and surface waviness of no more than 0.8 nm RMS.
Fujibuchi, Wataru; Anderson, John S. J.; Landsman, David
2001-01-01
Consensus pattern and matrix-based searches designed to predict cis-acting transcriptional regulatory sequences have historically been subject to large numbers of false positives. We sought to decrease false positives by incorporating expression profile data into a consensus pattern-based search method. We have systematically analyzed the expression phenotypes of over 6000 yeast genes, across 121 expression profile experiments, and correlated them with the distribution of 14 known regulatory elements over sequences upstream of the genes. Our method is based on a metric we term probabilistic element assessment (PEA), which is a ranking of potential sites based on sequence similarity in the upstream regions of genes with similar expression phenotypes. For eight of the 14 known elements that we examined, our method had a much higher selectivity than a naïve consensus pattern search. Based on our analysis, we have developed a web-based tool called PROSPECT, which allows consensus pattern-based searching of gene clusters obtained from microarray data. PMID:11574681
Lou, Yun-xiao; Fu, Xian-shu; Yu, Xiao-ping; Zhang, Ya-fen
2017-01-01
This paper focused on an effective method to discriminate the geographical origin of Wuyi-Rock tea by the stable isotope ratio (SIR) and metallic element profiling (MEP) combined with support vector machine (SVM) analysis. Wuyi-Rock tea (n = 99) collected from nine producing areas and non-Wuyi-Rock tea (n = 33) from eleven nonproducing areas were analysed for SIR and MEP by established methods. The SVM model based on coupled data produced the best prediction accuracy (0.9773). This prediction shows that instrumental methods combined with a classification model can provide an effective and stable tool for provenance discrimination. Moreover, every feature variable in stable isotope and metallic element data was ranked by its contribution to the model. The results show that δ2H, δ18O, Cs, Cu, Ca, and Rb contents are significant indications for provenance discrimination and not all of the metallic elements improve the prediction accuracy of the SVM model. PMID:28473941
Critical review of decision support tools for sustainability assessment of site remediation options.
Huysegoms, Lies; Cappuyns, Valérie
2017-07-01
In Europe alone, there are more than 2,5 million potentially contaminated sites of which 14% are expected to require remediation. Contaminated soil and groundwater can cause damage to human health as well as to valuable ecosystems. Globally more attention has been paid to this problem of soil contamination in the past decades. For example, more than 58 000 sites have been remediated in Europe between 2006 and 2011. Together with this increase in remediation projects there has been a surge in the development of new remediation technologies and decision support tools to be able to match every site and its specific characteristics to the best possible remediation alternative. In the past years the development of decision support tools (DST) has evolved in a more sustainable direction. Several DSTs added the claim not only to denote effective or technologically and economically feasible remediation alternatives but also to point out the more or most sustainable remediation alternatives. These trends in the evaluation of site remediation options left users with a confusing clew of possibly applicable tools to assist them in decision making for contaminated site remediation. This review provides a structured overview on the extent decision support tools for contaminated site remediation, that claim to assist in choosing the most sustainable remediation alternative, actually include the different elements of sustainability proposed in our assessment framework. The review contains an in-depth analysis of thirteen tools specifically developed to assess the sustainability of site remediation alternatives. This analysis is based on six criteria derived from the definition of sustainable development of the Brundtland report. The six criteria were concretized by using the three pillars of sustainability, applied to site remediation according to the SuRF-UK framework, two criteria derived from Life Cycle Assessment and Cost-Benefit Analysis, and an 'User friendly' criterion. These elements come together in a framework, drafted for this study, containing six criteria covering the environmental, economic, social, time, uncertainty aspects and user friendliness of a sustainable site remediation. The main remarks uncovered by this review are the imbalance of used indicators still expressing a strong preference for the environmental aspect at the expense of the economic and social aspects of sustainability, the lack of consistency in the terminology used within the field and the failure in adapting released tools to recent legislation or scientific advancements. Copyright © 2017 Elsevier Ltd. All rights reserved.
Cunningham, J C; Sinka, I C; Zavaliangos, A
2004-08-01
In this first of two articles on the modeling of tablet compaction, the experimental inputs related to the constitutive model of the powder and the powder/tooling friction are determined. The continuum-based analysis of tableting makes use of an elasto-plastic model, which incorporates the elements of yield, plastic flow potential, and hardening, to describe the mechanical behavior of microcrystalline cellulose over the range of densities experienced during tableting. Specifically, a modified Drucker-Prager/cap plasticity model, which includes material parameters such as cohesion, internal friction, and hydrostatic yield pressure that evolve with the internal state variable relative density, was applied. Linear elasticity is assumed with the elastic parameters, Young's modulus, and Poisson's ratio dependent on the relative density. The calibration techniques were developed based on a series of simple mechanical tests including diametrical compression, simple compression, and die compaction using an instrumented die. The friction behavior is measured using an instrumented die and the experimental data are analyzed using the method of differential slices. The constitutive model and frictional properties are essential experimental inputs to the finite element-based model described in the companion article. Copyright 2004 Wiley-Liss, Inc. and the American Pharmacists Association J Pharm Sci 93:2022-2039, 2004
ASTROS: A multidisciplinary automated structural design tool
NASA Technical Reports Server (NTRS)
Neill, D. J.
1989-01-01
ASTROS (Automated Structural Optimization System) is a finite-element-based multidisciplinary structural optimization procedure developed under Air Force sponsorship to perform automated preliminary structural design. The design task is the determination of the structural sizes that provide an optimal structure while satisfying numerous constraints from many disciplines. In addition to its automated design features, ASTROS provides a general transient and frequency response capability, as well as a special feature to perform a transient analysis of a vehicle subjected to a nuclear blast. The motivation for the development of a single multidisciplinary design tool is that such a tool can provide improved structural designs in less time than is currently needed. The role of such a tool is even more apparent as modern materials come into widespread use. Balancing conflicting requirements for the structure's strength and stiffness while exploiting the benefits of material anisotropy is perhaps an impossible task without assistance from an automated design tool. Finally, the use of a single tool can bring the design task into better focus among design team members, thereby improving their insight into the overall task.
NASA Astrophysics Data System (ADS)
Farina, Simone; Thepsonti, Thanongsak; Ceretti, Elisabetta; Özel, Tugrul
2011-05-01
Titanium alloys offer superb properties in strength, corrosion resistance and biocompatibility and are commonly utilized in medical devices and implants. Micro-end milling process is a direct and rapid fabrication method for manufacturing medical devices and implants in titanium alloys. Process performance and quality depend upon an understanding of the relationship between cutting parameters and forces and resultant tool deflections to avoid tool breakage. For this purpose, FE simulations of chip formation during micro-end milling of Ti-6Al-4V alloy with an ultra-fine grain solid carbide two-flute micro-end mill are investigated using DEFORM software. At first, specific forces in tangential and radial directions of cutting during micro-end milling for varying feed advance and rotational speeds have been determined using designed FE simulations for chip formation process. Later, these forces are applied to the micro-end mill geometry along the axial depth of cut in 3D analysis of ABAQUS. Consequently, 3D distributions for tool deflections & von Misses stress are determined. These analyses will yield in establishing integrated multi-physics process models for high performance micro-end milling and a leap-forward to process improvements.
An efficient framework for Java data processing systems in HPC environments
NASA Astrophysics Data System (ADS)
Fries, Aidan; Castañeda, Javier; Isasi, Yago; Taboada, Guillermo L.; Portell de Mora, Jordi; Sirvent, Raül
2011-11-01
Java is a commonly used programming language, although its use in High Performance Computing (HPC) remains relatively low. One of the reasons is a lack of libraries offering specific HPC functions to Java applications. In this paper we present a Java-based framework, called DpcbTools, designed to provide a set of functions that fill this gap. It includes a set of efficient data communication functions based on message-passing, thus providing, when a low latency network such as Myrinet is available, higher throughputs and lower latencies than standard solutions used by Java. DpcbTools also includes routines for the launching, monitoring and management of Java applications on several computing nodes by making use of JMX to communicate with remote Java VMs. The Gaia Data Processing and Analysis Consortium (DPAC) is a real case where scientific data from the ESA Gaia astrometric satellite will be entirely processed using Java. In this paper we describe the main elements of DPAC and its usage of the DpcbTools framework. We also assess the usefulness and performance of DpcbTools through its performance evaluation and the analysis of its impact on some DPAC systems deployed in the MareNostrum supercomputer (Barcelona Supercomputing Center).
Sakurai, Nozomu; Ara, Takeshi; Kanaya, Shigehiko; Nakamura, Yukiko; Iijima, Yoko; Enomoto, Mitsuo; Motegi, Takeshi; Aoki, Koh; Suzuki, Hideyuki; Shibata, Daisuke
2013-01-15
High-accuracy mass values detected by high-resolution mass spectrometry analysis enable prediction of elemental compositions, and thus are used for metabolite annotations in metabolomic studies. Here, we report an application of a relational database to significantly improve the rate of elemental composition predictions. By searching a database of pre-calculated elemental compositions with fixed kinds and numbers of atoms, the approach eliminates redundant evaluations of the same formula that occur in repeated calculations with other tools. When our approach is compared with HR2, which is one of the fastest tools available, our database search times were at least 109 times shorter than those of HR2. When a solid-state drive (SSD) was applied, the search time was 488 times shorter at 5 ppm mass tolerance and 1833 times at 0.1 ppm. Even if the search by HR2 was performed with 8 threads in a high-spec Windows 7 PC, the database search times were at least 26 and 115 times shorter without and with the SSD. These improvements were enhanced in a low spec Windows XP PC. We constructed a web service 'MFSearcher' to query the database in a RESTful manner. Available for free at http://webs2.kazusa.or.jp/mfsearcher. The web service is implemented in Java, MySQL, Apache and Tomcat, with all major browsers supported. sakurai@kazusa.or.jp Supplementary data are available at Bioinformatics online.
Study of archaeological coins of different dynasties using libs coupled with multivariate analysis
NASA Astrophysics Data System (ADS)
Awasthi, Shikha; Kumar, Rohit; Rai, G. K.; Rai, A. K.
2016-04-01
Laser Induced Breakdown Spectroscopy (LIBS) is an atomic emission spectroscopic technique having unique capability of an in-situ monitoring tool for detection and quantification of elements present in different artifacts. Archaeological coins collected form G.R. Sharma Memorial Museum; University of Allahabad, India has been analyzed using LIBS technique. These coins were obtained from excavation of Kausambi, Uttar Pradesh, India. LIBS system assembled in the laboratory (laser Nd:YAG 532 nm, 4 ns pulse width FWHM with Ocean Optics LIBS 2000+ spectrometer) is employed for spectral acquisition. The spectral lines of Ag, Cu, Ca, Sn, Si, Fe and Mg are identified in the LIBS spectra of different coins. LIBS along with Multivariate Analysis play an effective role for classification and contribution of spectral lines in different coins. The discrimination between five coins with Archaeological interest has been carried out using Principal Component Analysis (PCA). The results show the potential relevancy of the methodology used in the elemental identification and classification of artifacts with high accuracy and robustness.
Coupled Thermo-Electro-Magneto-Elastic Response of Smart Stiffened Panels
NASA Technical Reports Server (NTRS)
Bednarcyk, Brett A.; Yarrington, Phillip W.
2009-01-01
This report documents the procedures developed for incorporating smart laminate and panel analysis capabilities within the HyperSizer aerospace structural sizing software package. HyperSizer analyzes stiffened panels composed of arbitrary composite laminates through stiffener homogenization, or "smearing " techniques. The result is an effective constitutive equation for the stiffened panel that is suitable for use in a full vehicle-scale finite element analysis via MSC/NASTRAN. The existing thermo-elastic capabilities of HyperSizer have herein been extended to include coupled thermo-electro-magneto-elastic analysis capabilities. This represents a significant step toward realization of design tools capable of guiding the development of the next generation of smart aerospace structures. Verification results are presented that compare the developed smart HyperSizer capability with an ABAQUS piezoelectric finite element solution for a facesheet-flange combination. These results show good agreement between HyperSizer and ABAQUS, but highlight a limitation of the HyperSizer formulation in that constant electric field components are assumed.
Etzl, Stefan; Lindner, Robert; Nelson, Matthew D; Winkler, Andreas
2018-06-08
Genetically targeting biological systems to control cellular processes with light is the concept of optogenetics. Despite impressive developments in this field, underlying molecular mechanisms of signal transduction of the employed photoreceptor modules are frequently not sufficiently understood to rationally design new optogenetic tools. Here, we investigate the requirements for functional coupling of red light-sensing phytochromes with non-natural enzymatic effectors by creating a series of constructs featuring the Deinococcus radiodurans bacteriophytochrome linked to a Synechocystis guanylate/adenylate cyclase. Incorporating characteristic structural elements important for cyclase regulation in our designs, we identified several red light-regulated fusions with promising properties. We provide details of one light-activated construct with low dark-state activity and high dynamic range that outperforms previous optogenetic tools in vitro and expands our in vivo toolkit, as demonstrated by manipulation of Caenorhabditis elegans locomotor activity. The full-length crystal structure of this phytochrome-linked cyclase revealed molecular details of photoreceptor-effector coupling, highlighting the importance of the regulatory cyclase element. Analysis of conformational dynamics by hydrogen-deuterium exchange in different functional states enriched our understanding of phytochrome signaling and signal integration by effectors. We found that light-induced conformational changes in the phytochrome destabilize the coiled-coil sensor-effector linker, which releases the cyclase regulatory element from an inhibited conformation, increasing cyclase activity of this artificial system. Future designs of optogenetic functionalities may benefit from our work, indicating that rational considerations for the effector improve the rate of success of initial designs to obtain optogenetic tools with superior properties. © 2018 Etzl et al.
Recommendations for tool-handle material choice based on finite element analysis.
Harih, Gregor; Dolšak, Bojan
2014-05-01
Huge areas of work are still done manually and require the usages of different powered and non-powered hand tools. In order to increase the user performance, satisfaction, and lower the risk of acute and cumulative trauma disorders, several researchers have investigated the sizes and shapes of tool-handles. However, only a few authors have investigated tool-handles' materials for further optimising them. Therefore, as presented in this paper, we have utilised a finite-element method for simulating human fingertip whilst grasping tool-handles. We modelled and simulated steel and ethylene propylene diene monomer (EPDM) rubber as homogeneous tool-handle materials and two composites consisting of EPDM rubber and EPDM foam, and also EPDM rubber and PU foam. The simulated finger force was set to obtain characteristic contact pressures of 20 kPa, 40 kPa, 80 kPa, and 100 kPa. Numerical tests have shown that EPDM rubber lowers the contact pressure just slightly. On the other hand, both composites showed significant reduction in contact pressure that could lower the risks of acute and cumulative trauma disorders which are pressure-dependent. Based on the results, it is also evident that a composite containing PU foam with a more evident and flat plateau deformed less at lower strain rates and deformed more when the plateau was reached, in comparison to the composite with EPDM foam. It was shown that hyper-elastic foam materials, which take into account the non-linear behaviour of fingertip soft tissue, can lower the contact pressure whilst maintaining low deformation rate of the tool-handle material for maintaining sufficient rate of stability of the hand tool in the hands. Lower contact pressure also lowers the risk of acute and cumulative trauma disorders, and increases comfort whilst maintaining performance. Copyright © 2013 Elsevier Ltd and The Ergonomics Society. All rights reserved.
Gagliardi, Anna R; Brouwers, Melissa C; Bhattacharyya, Onil K
2012-04-02
Modifying the format and content of guidelines may facilitate their use and lead to improved quality of care. We reviewed the medical literature to identify features desired by different users and associated with guideline use to develop a framework of implementability and found that most guidelines do not contain these elements. Further research is needed to develop and evaluate implementability tools. We are launching the Guideline Implementability Research and Application Network (GIRAnet) to enable the development and testing of implementability tools in three domains: Resource Implications, Implementation, and Evaluation. Partners include the Guidelines International Network (G-I-N) and its member guideline developers, implementers, and researchers. In phase one, international guidelines will be examined to identify and describe exemplar tools. Indication-specific and generic tools will populate a searchable repository. In phase two, qualitative analysis of cognitive interviews will be used to understand how developers can best integrate implementability tools in guidelines and how health professionals use them for interpreting and applying guidelines. In phase three, a small-scale pilot test will assess the impact of implementability tools based on quantitative analysis of chart-based behavioural outcomes and qualitative analysis of interviews with participants. The findings will be used to plan a more comprehensive future evaluation of implementability tools. Infrastructure funding to establish GIRAnet will be leveraged with the in-kind contributions of collaborating national and international guideline developers to advance our knowledge of implementation practice and science. Needs assessment and evaluation of GIRAnet will provide a greater understanding of how to develop and sustain such knowledge-exchange networks. Ultimately, by facilitating use of guidelines, this research may lead to improved delivery and outcomes of patient care.
Application of an enhanced discrete element method to oil and gas drilling processes
NASA Astrophysics Data System (ADS)
Ubach, Pere Andreu; Arrufat, Ferran; Ring, Lev; Gandikota, Raju; Zárate, Francisco; Oñate, Eugenio
2016-03-01
The authors present results on the use of the discrete element method (DEM) for the simulation of drilling processes typical in the oil and gas exploration industry. The numerical method uses advanced DEM techniques using a local definition of the DEM parameters and combined FEM-DEM procedures. This paper presents a step-by-step procedure to build a DEM model for analysis of the soil region coupled to a FEM model for discretizing the drilling tool that reproduces the drilling mechanics of a particular drill bit. A parametric study has been performed to determine the model parameters in order to maintain accurate solutions with reduced computational cost.
Quantitative aspects of inductively coupled plasma mass spectrometry
NASA Astrophysics Data System (ADS)
Bulska, Ewa; Wagner, Barbara
2016-10-01
Accurate determination of elements in various kinds of samples is essential for many areas, including environmental science, medicine, as well as industry. Inductively coupled plasma mass spectrometry (ICP-MS) is a powerful tool enabling multi-elemental analysis of numerous matrices with high sensitivity and good precision. Various calibration approaches can be used to perform accurate quantitative measurements by ICP-MS. They include the use of pure standards, matrix-matched standards, or relevant certified reference materials, assuring traceability of the reported results. This review critically evaluates the advantages and limitations of different calibration approaches, which are used in quantitative analyses by ICP-MS. Examples of such analyses are provided. This article is part of the themed issue 'Quantitative mass spectrometry'.
NASA Astrophysics Data System (ADS)
Zirari, M.; Abdellah El-Hadj, A.; Bacha, N.
2010-03-01
A finite element method is used to simulate the deposition of the thermal spray coating process. A set of governing equations is solving by a volume of fluid method. For the solidification phenomenon, we use the specific heat method (SHM). We begin by comparing the present model with experimental and numerical model available in the literature. In this study, completely molten or semi-molten aluminum particle impacts a H13 tool steel substrate is considered. Next we investigate the effect of inclination of impact of a partially molten particle on flat substrate. It was found that the melting state of the particle has great effects on the morphologies of the splat.
JPL-IDEAS - ITERATIVE DESIGN OF ANTENNA STRUCTURES
NASA Technical Reports Server (NTRS)
Levy, R.
1994-01-01
The Iterative DEsign of Antenna Structures (IDEAS) program is a finite element analysis and design optimization program with special features for the analysis and design of microwave antennas and associated sub-structures. As the principal structure analysis and design tool for the Jet Propulsion Laboratory's Ground Antenna and Facilities Engineering section of NASA's Deep Space Network, IDEAS combines flexibility with easy use. The relatively small bending stiffness of the components of large, steerable reflector antennas allows IDEAS to use pinjointed (three translational degrees of freedom per joint) models for modeling the gross behavior of these antennas when subjected to static and dynamic loading. This facilitates the formulation of the redesign algorithm which has only one design variable per structural element. Input data deck preparation has been simplified by the use of NAMELIST inputs to promote clarity of data input for problem defining parameters, user selection of execution and design options and output requests, and by the use of many attractive and familiar features of the NASTRAN program (in many cases, NASTRAN and IDEAS formatted bulk data cards are interchangeable). Features such as simulation of a full symmetric structure based on analyses of only half the structure make IDEAS a handy and efficient analysis tool, with many features unavailable in any other finite element analysis program. IDEAS can choose design variables such as areas of rods and thicknesses of plates to minimize total structure weight, constrain the structure weight to a specified value while maximizing a natural frequency or minimizing compliance measures, and can use a stress ratio algorithm to size each structural member so that it is at maximum or minimum stress level for at least one of the applied loads. Calculations of total structure weight can be broken down according to material. Center of gravity weight balance, static first and second moments about the center of mass and optionally about a user-specified gridpoint, and lumped structure weight at grid points can also be calculated. Other analysis outputs include calculation of reactions, displacements, and element stresses due to specified gravity, thermal, and external applied loads; calculations of linear combinations of specific node displacements (e.g. to represent motions of rigid attachments not included in the structure model), natural frequency eigenvalues and eigenvectors, structure reactions and element stresses, and coordinates of effective modal masses. Cassegrain antenna boresight error analysis of a best fitting paraboloid and Cassegrain microwave antenna root mean square half-pathlength error analysis of a best fitting paraboloid are also performed. The IDEAS program is written in ATHENA FORTRAN and ASSEMBLER for an EXEC 8 operating system and was implemented on a UNIVAC 1100 series computer. The minimum memory requirement for the program is approximately 42,000 36-bit words. This program is available on a 9-track 1600 BPI magnetic tape in UNIVAC FURPUR format only; since JPL-IDEAS will not run on other platforms, COSMIC will not reformat the code to be readable on other platforms. The program was developed in 1988.
Kausar, Rubina; Ahmad, Zulfiqar
2009-10-01
The present study deals with the ground water quality assessment in Kahuta Industrial Triangle Islamabad, Pakistan. The objective of the study was to assess ground water quality against the drinking water standards for various toxic inorganic elements. Representative groundwater samples were collected and analyzed in the Water Quality Laboratory of Pakistan Council of Research in Water Resources (PCRWR) at Islamabad, Pakistan. The samples were run on ICP-MS (Inductively coupled plasma mass spectrometry), which has the capability to separate and quantify 70 elements at a time. One of the finding of study is that ICP-MS is a very good tool to analyze broad range of toxic inorganic elements to the level of parts per billion (ppb). World Health Organization drinking water standards shows that these toxic inorganic elements such as heavy metals even at this concentration level (ppb) are injurious to human health. This analysis indicated pollution of various toxic elements including Selenium. Vertical leachate through industrial waste septic tanks is identified as major cause of groundwater pollution in the Industrial Triangle. Monitoring of the septic tanks and groundwater quality in study area is suggested along with remedial measures.