Construction of dynamic stochastic simulation models using knowledge-based techniques
NASA Technical Reports Server (NTRS)
Williams, M. Douglas; Shiva, Sajjan G.
1990-01-01
Over the past three decades, computer-based simulation models have proven themselves to be cost-effective alternatives to the more structured deterministic methods of systems analysis. During this time, many techniques, tools and languages for constructing computer-based simulation models have been developed. More recently, advances in knowledge-based system technology have led many researchers to note the similarities between knowledge-based programming and simulation technologies and to investigate the potential application of knowledge-based programming techniques to simulation modeling. The integration of conventional simulation techniques with knowledge-based programming techniques is discussed to provide a development environment for constructing knowledge-based simulation models. A comparison of the techniques used in the construction of dynamic stochastic simulation models and those used in the construction of knowledge-based systems provides the requirements for the environment. This leads to the design and implementation of a knowledge-based simulation development environment. These techniques were used in the construction of several knowledge-based simulation models including the Advanced Launch System Model (ALSYM).
2017-11-01
ARL-TR-8225 ● NOV 2017 US Army Research Laboratory Methodology for Designing and Developing a New Ultra-Wideband Antenna Based...Research Laboratory Methodology for Designing and Developing a New Ultra-Wideband Antenna Based on Bio-Inspired Optimization Techniques by...SUBTITLE Methodology for Designing and Developing a New Ultra-Wideband Antenna Based on Bio-Inspired Optimization Techniques 5a. CONTRACT NUMBER
Artificial Intelligence Techniques: Applications for Courseware Development.
ERIC Educational Resources Information Center
Dear, Brian L.
1986-01-01
Introduces some general concepts and techniques of artificial intelligence (natural language interfaces, expert systems, knowledge bases and knowledge representation, heuristics, user-interface metaphors, and object-based environments) and investigates ways these techniques might be applied to analysis, design, development, implementation, and…
Techniques for Enhancing Web-Based Education.
ERIC Educational Resources Information Center
Barbieri, Kathy; Mehringer, Susan
The Virtual Workshop is a World Wide Web-based set of modules on high performance computing developed at the Cornell Theory Center (CTC) (New York). This approach reaches a large audience, leverages staff effort, and poses challenges for developing interesting presentation techniques. This paper describes the following techniques with their…
Development of a Web-based financial application System
NASA Astrophysics Data System (ADS)
Hasan, M. R.; Ibrahimy, M. I.; Motakabber, S. M. A.; Ferdaus, M. M.; Khan, M. N. H.; Mostafa, M. G.
2013-12-01
The paper describes a technique to develop a web based financial system, following latest technology and business needs. In the development of web based application, the user friendliness and technology both are very important. It is used ASP .NET MVC 4 platform and SQL 2008 server for development of web based financial system. It shows the technique for the entry system and report monitoring of the application is user friendly. This paper also highlights the critical situations of development, which will help to develop the quality product.
Cell culture-based biosensing techniques for detecting toxicity in water.
Tan, Lu; Schirmer, Kristin
2017-06-01
The significant increase of contaminants entering fresh water bodies calls for the development of rapid and reliable methods to monitor the aquatic environment and to detect water toxicity. Cell culture-based biosensing techniques utilise the overall cytotoxic response to external stimuli, mediated by a transduced signal, to specify the toxicity of aqueous samples. These biosensing techniques can effectively indicate water toxicity for human safety and aquatic organism health. In this review we account for the recent developments of the mainstream cell culture-based biosensing techniques for water quality evaluation, discuss their key features, potentials and limitations, and outline the future prospects of their development. Copyright © 2017 The Authors. Published by Elsevier Ltd.. All rights reserved.
Digital Signal Processing Based Biotelemetry Receivers
NASA Technical Reports Server (NTRS)
Singh, Avtar; Hines, John; Somps, Chris
1997-01-01
This is an attempt to develop a biotelemetry receiver using digital signal processing technology and techniques. The receiver developed in this work is based on recovering signals that have been encoded using either Pulse Position Modulation (PPM) or Pulse Code Modulation (PCM) technique. A prototype has been developed using state-of-the-art digital signal processing technology. A Printed Circuit Board (PCB) is being developed based on the technique and technology described here. This board is intended to be used in the UCSF Fetal Monitoring system developed at NASA. The board is capable of handling a variety of PPM and PCM signals encoding signals such as ECG, temperature, and pressure. A signal processing program has also been developed to analyze the received ECG signal to determine heart rate. This system provides a base for using digital signal processing in biotelemetry receivers and other similar applications.
Applying knowledge compilation techniques to model-based reasoning
NASA Technical Reports Server (NTRS)
Keller, Richard M.
1991-01-01
Researchers in the area of knowledge compilation are developing general purpose techniques for improving the efficiency of knowledge-based systems. In this article, an attempt is made to define knowledge compilation, to characterize several classes of knowledge compilation techniques, and to illustrate how some of these techniques can be applied to improve the performance of model-based reasoning systems.
Automated Video-Based Traffic Count Analysis.
DOT National Transportation Integrated Search
2016-01-01
The goal of this effort has been to develop techniques that could be applied to the : detection and tracking of vehicles in overhead footage of intersections. To that end we : have developed and published techniques for vehicle tracking based on dete...
NASA Astrophysics Data System (ADS)
Roushangar, Kiyoumars; Mehrabani, Fatemeh Vojoudi; Shiri, Jalal
2014-06-01
This study presents Artificial Intelligence (AI)-based modeling of total bed material load through developing the accuracy level of the predictions of traditional models. Gene expression programming (GEP) and adaptive neuro-fuzzy inference system (ANFIS)-based models were developed and validated for estimations. Sediment data from Qotur River (Northwestern Iran) were used for developing and validation of the applied techniques. In order to assess the applied techniques in relation to traditional models, stream power-based and shear stress-based physical models were also applied in the studied case. The obtained results reveal that developed AI-based models using minimum number of dominant factors, give more accurate results than the other applied models. Nonetheless, it was revealed that k-fold test is a practical but high-cost technique for complete scanning of applied data and avoiding the over-fitting.
Leveraging Existing Mission Tools in a Re-Usable, Component-Based Software Environment
NASA Technical Reports Server (NTRS)
Greene, Kevin; Grenander, Sven; Kurien, James; z,s (fshir. z[orttr); z,scer; O'Reilly, Taifun
2006-01-01
Emerging methods in component-based software development offer significant advantages but may seem incompatible with existing mission operations applications. In this paper we relate our positive experiences integrating existing mission applications into component-based tools we are delivering to three missions. In most operations environments, a number of software applications have been integrated together to form the mission operations software. In contrast, with component-based software development chunks of related functionality and data structures, referred to as components, can be individually delivered, integrated and re-used. With the advent of powerful tools for managing component-based development, complex software systems can potentially see significant benefits in ease of integration, testability and reusability from these techniques. These benefits motivate us to ask how component-based development techniques can be relevant in a mission operations environment, where there is significant investment in software tools that are not component-based and may not be written in languages for which component-based tools even exist. Trusted and complex software tools for sequencing, validation, navigation, and other vital functions cannot simply be re-written or abandoned in order to gain the advantages offered by emerging component-based software techniques. Thus some middle ground must be found. We have faced exactly this issue, and have found several solutions. Ensemble is an open platform for development, integration, and deployment of mission operations software that we are developing. Ensemble itself is an extension of an open source, component-based software development platform called Eclipse. Due to the advantages of component-based development, we have been able to vary rapidly develop mission operations tools for three surface missions by mixing and matching from a common set of mission operation components. We have also had to determine how to integrate existing mission applications for sequence development, sequence validation, and high level activity planning, and other functions into a component-based environment. For each of these, we used a somewhat different technique based upon the structure and usage of the existing application.
Screening and Biosensor-Based Approaches for Lung Cancer Detection
Wang, Lulu
2017-01-01
Early diagnosis of lung cancer helps to reduce the cancer death rate significantly. Over the years, investigators worldwide have extensively investigated many screening modalities for lung cancer detection, including computerized tomography, chest X-ray, positron emission tomography, sputum cytology, magnetic resonance imaging and biopsy. However, these techniques are not suitable for patients with other pathologies. Developing a rapid and sensitive technique for early diagnosis of lung cancer is urgently needed. Biosensor-based techniques have been recently recommended as a rapid and cost-effective tool for early diagnosis of lung tumor markers. This paper reviews the recent development in screening and biosensor-based techniques for early lung cancer detection. PMID:29065541
Change detection from remotely sensed images: From pixel-based to object-based approaches
NASA Astrophysics Data System (ADS)
Hussain, Masroor; Chen, Dongmei; Cheng, Angela; Wei, Hui; Stanley, David
2013-06-01
The appetite for up-to-date information about earth's surface is ever increasing, as such information provides a base for a large number of applications, including local, regional and global resources monitoring, land-cover and land-use change monitoring, and environmental studies. The data from remote sensing satellites provide opportunities to acquire information about land at varying resolutions and has been widely used for change detection studies. A large number of change detection methodologies and techniques, utilizing remotely sensed data, have been developed, and newer techniques are still emerging. This paper begins with a discussion of the traditionally pixel-based and (mostly) statistics-oriented change detection techniques which focus mainly on the spectral values and mostly ignore the spatial context. This is succeeded by a review of object-based change detection techniques. Finally there is a brief discussion of spatial data mining techniques in image processing and change detection from remote sensing data. The merits and issues of different techniques are compared. The importance of the exponential increase in the image data volume and multiple sensors and associated challenges on the development of change detection techniques are highlighted. With the wide use of very-high-resolution (VHR) remotely sensed images, object-based methods and data mining techniques may have more potential in change detection.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Chen, Kuangcai
The goal of this study is to help with future data analysis and experiment designs in rotational dynamics research using DIC-based SPORT technique. Most of the current studies using DIC-based SPORT techniques are technical demonstrations. Understanding the mechanisms behind the observed rotational behaviors of the imaging probes should be the focus of the future SPORT studies. More efforts are still needed in the development of new imaging probes, particle tracking methods, instrumentations, and advanced data analysis methods to further extend the potential of DIC-based SPORT technique.
NASA Astrophysics Data System (ADS)
Vidya Sagar, R.; Raghu Prasad, B. K.
2012-03-01
This article presents a review of recent developments in parametric based acoustic emission (AE) techniques applied to concrete structures. It recapitulates the significant milestones achieved by previous researchers including various methods and models developed in AE testing of concrete structures. The aim is to provide an overview of the specific features of parametric based AE techniques of concrete structures carried out over the years. Emphasis is given to traditional parameter-based AE techniques applied to concrete structures. A significant amount of research on AE techniques applied to concrete structures has already been published and considerable attention has been given to those publications. Some recent studies such as AE energy analysis and b-value analysis used to assess damage of concrete bridge beams have also been discussed. The formation of fracture process zone and the AE energy released during the fracture process in concrete beam specimens have been summarised. A large body of experimental data on AE characteristics of concrete has accumulated over the last three decades. This review of parametric based AE techniques applied to concrete structures may be helpful to the concerned researchers and engineers to better understand the failure mechanism of concrete and evolve more useful methods and approaches for diagnostic inspection of structural elements and failure prediction/prevention of concrete structures.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Onoufriou, T.; Simpson, R.J.; Protopapas, M.
This paper presents the development and application of reliability based inspection planning techniques for floaters. Based on previous experience from jacket structure applications optimized inspection planning (OIP) techniques for floaters are developed. The differences between floaters and jacket structures in relation to fatigue damage, redundancy levels and inspection practice are examined and reflected in the proposed methodology. The application and benefits of these techniques is demonstrated through representative analyses and important trends are highlighted through the results of a parametric sensitivity study.
Aerodynamic measurement techniques. [laser based diagnostic techniques
NASA Technical Reports Server (NTRS)
Hunter, W. W., Jr.
1976-01-01
Laser characteristics of intensity, monochromatic, spatial coherence, and temporal coherence were developed to advance laser based diagnostic techniques for aerodynamic related research. Two broad categories of visualization and optical measurements were considered, and three techniques received significant attention. These are holography, laser velocimetry, and Raman scattering. Examples of the quantitative laser velocimeter and Raman scattering measurements of velocity, temperature, and density indicated the potential of these nonintrusive techniques.
Macready, Anna L; Fallaize, Rosalind; Butler, Laurie T; Ellis, Judi A; Kuznesof, Sharron; Frewer, Lynn J; Celis-Morales, Carlos; Livingstone, Katherine M; Araújo-Soares, Vera; Fischer, Arnout RH; Stewart-Knox, Barbara J; Mathers, John C
2018-01-01
Background To determine the efficacy of behavior change techniques applied in dietary and physical activity intervention studies, it is first necessary to record and describe techniques that have been used during such interventions. Published frameworks used in dietary and smoking cessation interventions undergo continuous development, and most are not adapted for Web-based delivery. The Food4Me study (N=1607) provided the opportunity to use existing frameworks to describe standardized Web-based techniques employed in a large-scale, internet-based intervention to change dietary behavior and physical activity. Objective The aims of this study were (1) to describe techniques embedded in the Food4Me study design and explain the selection rationale and (2) to demonstrate the use of behavior change technique taxonomies, develop standard operating procedures for training, and identify strengths and limitations of the Food4Me framework that will inform its use in future studies. Methods The 6-month randomized controlled trial took place simultaneously in seven European countries, with participants receiving one of four levels of personalized advice (generalized, intake-based, intake+phenotype–based, and intake+phenotype+gene–based). A three-phase approach was taken: (1) existing taxonomies were reviewed and techniques were identified a priori for possible inclusion in the Food4Me study, (2) a standard operating procedure was developed to maintain consistency in the use of methods and techniques across research centers, and (3) the Food4Me behavior change technique framework was reviewed and updated post intervention. An analysis of excluded techniques was also conducted. Results Of 46 techniques identified a priori as being applicable to Food4Me, 17 were embedded in the intervention design; 11 were from a dietary taxonomy, and 6 from a smoking cessation taxonomy. In addition, the four-category smoking cessation framework structure was adopted for clarity of communication. Smoking cessation texts were adapted for dietary use where necessary. A posteriori, a further 9 techniques were included. Examination of excluded items highlighted the distinction between techniques considered appropriate for face-to-face versus internet-based delivery. Conclusions The use of existing taxonomies facilitated the description and standardization of techniques used in Food4Me. We recommend that for complex studies of this nature, technique analysis should be conducted a priori to develop standardized procedures and training and reviewed a posteriori to audit the techniques actually adopted. The present framework description makes a valuable contribution to future systematic reviews and meta-analyses that explore technique efficacy and underlying psychological constructs. This was a novel application of the behavior change taxonomies and was the first internet-based personalized nutrition intervention to use such a framework remotely. Trial Registration ClinicalTrials.gov NCT01530139; https://clinicaltrials.gov/ct2/show/NCT01530139 (Archived by WebCite at http://www.webcitation.org/6y8XYUft1) PMID:29631993
1982-02-01
For these data elements, Initial Milestone 11 values were established as the Flanning Estimate (PE) with the Development Estimate ( DE ) to he based ...development of improved forensic collection techniques for Naval Investigative Agents on ships and overseas bases . As this is a continuing program, the above...overseas bases ), and continue development of improved forensic collection techniques for Naval Investigative Agents on ships and overseas baszs. 4. (U) FY
A methodology for developing remediation goals for sites with contaminated sediments is provided. The remediation goals are based upon the concentrations of chemicals in the sediment interstitial water measured using the passive sampling technique. The passive sampling technique ...
A demonstrative model of a lunar base simulation on a personal computer
NASA Technical Reports Server (NTRS)
1985-01-01
The initial demonstration model of a lunar base simulation is described. This initial model was developed on the personal computer level to demonstrate feasibility and technique before proceeding to a larger computer-based model. Lotus Symphony Version 1.1 software was used to base the demonstration model on an personal computer with an MS-DOS operating system. The personal computer-based model determined the applicability of lunar base modeling techniques developed at an LSPI/NASA workshop. In addition, the personnal computer-based demonstration model defined a modeling structure that could be employed on a larger, more comprehensive VAX-based lunar base simulation. Refinement of this personal computer model and the development of a VAX-based model is planned in the near future.
Wind Tunnel Test Technique and Instrumentation Development at LaRC
NASA Technical Reports Server (NTRS)
Putnam, Lawrence E.
1999-01-01
LaRC has an aggressive test technique development program underway. This program has been developed using 3rd Generation R&D management techniques and is a closely coordinated program between suppliers and wind tunnel operators- wind tunnel customers' informal input relative to their needs has been an essential ingredient in developing the research portfolio. An attempt has been made to balance this portfolio to meet near term and long term test technique needs. Major efforts are underway to develop techniques for determining model wing twist and location of boundary layer transition in the NTF (National Transonic Facility). The foundation of all new instrumentation developments, procurements, and upgrades will be based on uncertainty analysis.
The Evolvement of Automobile Steering System Based on TRIZ
NASA Astrophysics Data System (ADS)
Zhao, Xinjun; Zhang, Shuang
Products and techniques pass through a process of birth, growth, maturity, death and quit the stage like biological evolution process. The developments of products and techniques conform to some evolvement rules. If people know and hold these rules, they can design new kind of products and forecast the develop trends of the products. Thereby, enterprises can grasp the future technique directions of products, and make product and technique innovation. Below, based on TRIZ theory, the mechanism evolvement, the function evolvement and the appearance evolvement of automobile steering system had been analyzed and put forward some new ideas about future automobile steering system.
Monitoring Knowledge Base (MKB)
The Monitoring Knowledge Base (MKB) is a compilation of emissions measurement and monitoring techniques associated with air pollution control devices, industrial process descriptions, and permitting techniques, including flexible permit development. Using MKB, one can gain a comprehensive understanding of emissions sources, control devices, and monitoring techniques, enabling one to determine appropriate permit terms and conditions.
Silicon-based optoelectronics: Monolithic integration for WDM
NASA Astrophysics Data System (ADS)
Pearson, Matthew Richard T.
2000-10-01
This thesis details the development of enabling technologies required for inexpensive, monolithic integration of Si-based wavelength division multiplexing (WDM) components and photodetectors. The work involves the design and fabrication of arrayed waveguide grating demultiplexers in silicon-on-insulator (SOI), the development of advanced SiGe photodetectors capable of photodetection at 1.55 mum wavelengths, and the development of a low cost fabrication technique that enables the high volume production of Si-based photonic components. Arrayed waveguide grating (AWG) demultiplexers were designed and fabricated in SOI. The fabrication of AWGs in SOI has been reported in the literature, however there are a number of design issues specific to the SOI material system that can have a large effect on device performance and design, and have not been theoretically examined in earlier work. The SOI AWGs presented in this thesis are the smallest devices of this type reported, and they exhibit performance acceptable for commercial applications. The SiGe photodetectors reported in the literature exhibit extremely low responsivities at wavelengths near 1.55 mum. We present the first use of three dimensional growth modes to enhance the photoresponse of SiGe at 1.55 mum wavelengths. Metal semiconductor-metal (MSM) photodetectors were fabricated using this undulating quantum well structure, and demonstrate the highest responsivities yet reported for a SiGe-based photodetector at 1.55 mum. These detectors were monolithically integrated with low-loss SOI waveguides, enabling integration with nearly any Si-based passive WDM component. The pursuit of inexpensive Si-based photonic components also requires the development of new manufacturing techniques that are more suitable for high volume production. This thesis presents the development of a low cost fabrication technique based on the local oxidation of silicon (LOCOS), a standard processing technique used for Si integrated circuits. This process is developed for both SiGe and SOI waveguides, but is shown to be commercially suitable only for SOI waveguide devices. The technique allows nearly any Si microelectronics fabrication facility to begin manufacturing optical components with minimal change in processing equipment or techniques. These enabling technologies provide the critical elements for inexpensive, monolithic integration in a Si-based system.
Recent Progress in Optical Biosensors Based on Smartphone Platforms
Geng, Zhaoxin; Zhang, Xiong; Fan, Zhiyuan; Lv, Xiaoqing; Su, Yue; Chen, Hongda
2017-01-01
With a rapid improvement of smartphone hardware and software, especially complementary metal oxide semiconductor (CMOS) cameras, many optical biosensors based on smartphone platforms have been presented, which have pushed the development of the point-of-care testing (POCT). Imaging-based and spectrometry-based detection techniques have been widely explored via different approaches. Combined with the smartphone, imaging-based and spectrometry-based methods are currently used to investigate a wide range of molecular properties in chemical and biological science for biosensing and diagnostics. Imaging techniques based on smartphone-based microscopes are utilized to capture microscale analysts, while spectrometry-based techniques are used to probe reactions or changes of molecules. Here, we critically review the most recent progress in imaging-based and spectrometry-based smartphone-integrated platforms that have been developed for chemical experiments and biological diagnosis. We focus on the analytical performance and the complexity for implementation of the platforms. PMID:29068375
Recent Progress in Optical Biosensors Based on Smartphone Platforms.
Geng, Zhaoxin; Zhang, Xiong; Fan, Zhiyuan; Lv, Xiaoqing; Su, Yue; Chen, Hongda
2017-10-25
With a rapid improvement of smartphone hardware and software, especially complementary metal oxide semiconductor (CMOS) cameras, many optical biosensors based on smartphone platforms have been presented, which have pushed the development of the point-of-care testing (POCT). Imaging-based and spectrometry-based detection techniques have been widely explored via different approaches. Combined with the smartphone, imaging-based and spectrometry-based methods are currently used to investigate a wide range of molecular properties in chemical and biological science for biosensing and diagnostics. Imaging techniques based on smartphone-based microscopes are utilized to capture microscale analysts, while spectrometry-based techniques are used to probe reactions or changes of molecules. Here, we critically review the most recent progress in imaging-based and spectrometry-based smartphone-integrated platforms that have been developed for chemical experiments and biological diagnosis. We focus on the analytical performance and the complexity for implementation of the platforms.
Constraint-based integration of planning and scheduling for space-based observatory management
NASA Technical Reports Server (NTRS)
Muscettola, Nicola; Smith, Steven F.
1994-01-01
Progress toward the development of effective, practical solutions to space-based observatory scheduling problems within the HSTS scheduling framework is reported. HSTS was developed and originally applied in the context of the Hubble Space Telescope (HST) short-term observation scheduling problem. The work was motivated by the limitations of the current solution and, more generally, by the insufficiency of classical planning and scheduling approaches in this problem context. HSTS has subsequently been used to develop improved heuristic solution techniques in related scheduling domains and is currently being applied to develop a scheduling tool for the upcoming Submillimeter Wave Astronomy Satellite (SWAS) mission. The salient architectural characteristics of HSTS and their relationship to previous scheduling and AI planning research are summarized. Then, some key problem decomposition techniques underlying the integrated planning and scheduling approach to the HST problem are described; research results indicate that these techniques provide leverage in solving space-based observatory scheduling problems. Finally, more recently developed constraint-posting scheduling procedures and the current SWAS application focus are summarized.
El-Said, Waleed A; Yoon, Jinho; Choi, Jeong-Woo
2018-01-01
Discovering new anticancer drugs and screening their efficacy requires a huge amount of resources and time-consuming processes. The development of fast, sensitive, and nondestructive methods for the in vitro and in vivo detection of anticancer drugs' effects and action mechanisms have been done to reduce the time and resources required to discover new anticancer drugs. For the in vitro and in vivo detection of the efficiency, distribution, and action mechanism of anticancer drugs, the applications of electrochemical techniques such as electrochemical cell chips and optical techniques such as surface-enhanced Raman spectroscopy (SERS) have been developed based on the nanostructured surface. Research focused on electrochemical cell chips and the SERS technique have been reviewed here; electrochemical cell chips based on nanostructured surfaces have been developed for the in vitro detection of cell viability and the evaluation of the effects of anticancer drugs, which showed the high capability to evaluate the cytotoxic effects of several chemicals at low concentrations. SERS technique based on the nanostructured surface have been used as label-free, simple, and nondestructive techniques for the in vitro and in vivo monitoring of the distribution, mechanism, and metabolism of different anticancer drugs at the cellular level. The use of electrochemical cell chips and the SERS technique based on the nanostructured surface should be good tools to detect the effects and action mechanisms of anticancer drugs.
NASA Astrophysics Data System (ADS)
El-Said, Waleed A.; Yoon, Jinho; Choi, Jeong-Woo
2018-04-01
Discovering new anticancer drugs and screening their efficacy requires a huge amount of resources and time-consuming processes. The development of fast, sensitive, and nondestructive methods for the in vitro and in vivo detection of anticancer drugs' effects and action mechanisms have been done to reduce the time and resources required to discover new anticancer drugs. For the in vitro and in vivo detection of the efficiency, distribution, and action mechanism of anticancer drugs, the applications of electrochemical techniques such as electrochemical cell chips and optical techniques such as surface-enhanced Raman spectroscopy (SERS) have been developed based on the nanostructured surface. Research focused on electrochemical cell chips and the SERS technique have been reviewed here; electrochemical cell chips based on nanostructured surfaces have been developed for the in vitro detection of cell viability and the evaluation of the effects of anticancer drugs, which showed the high capability to evaluate the cytotoxic effects of several chemicals at low concentrations. SERS technique based on the nanostructured surface have been used as label-free, simple, and nondestructive techniques for the in vitro and in vivo monitoring of the distribution, mechanism, and metabolism of different anticancer drugs at the cellular level. The use of electrochemical cell chips and the SERS technique based on the nanostructured surface should be good tools to detect the effects and action mechanisms of anticancer drugs.
Progress of new label-free techniques for biosensors: a review.
Sang, Shengbo; Wang, Yajun; Feng, Qiliang; Wei, Ye; Ji, Jianlong; Zhang, Wendong
2016-01-01
The detection techniques used in biosensors can be broadly classified into label-based and label-free. Label-based detection relies on the specific properties of labels for detecting a particular target. In contrast, label-free detection is suitable for the target molecules that are not labeled or the screening of analytes which are not easy to tag. Also, more types of label-free biosensors have emerged with developments in biotechnology. The latest developed techniques in label-free biosensors, such as field-effect transistors-based biosensors including carbon nanotube field-effect transistor biosensors, graphene field-effect transistor biosensors and silicon nanowire field-effect transistor biosensors, magnetoelastic biosensors, optical-based biosensors, surface stress-based biosensors and other type of biosensors based on the nanotechnology are discussed. The sensing principles, configurations, sensing performance, applications, advantages and restriction of different label-free based biosensors are considered and discussed in this review. Most concepts included in this survey could certainly be applied to the development of this kind of biosensor in the future.
Development Context Driven Change Awareness and Analysis Framework
NASA Technical Reports Server (NTRS)
Sarma, Anita; Branchaud, Josh; Dwyer, Matthew B.; Person, Suzette; Rungta, Neha
2014-01-01
Recent work on workspace monitoring allows conflict prediction early in the development process, however, these approaches mostly use syntactic differencing techniques to compare different program versions. In contrast, traditional change-impact analysis techniques analyze related versions of the program only after the code has been checked into the master repository. We propose a novel approach, De- CAF (Development Context Analysis Framework), that leverages the development context to scope a change impact analysis technique. The goal is to characterize the impact of each developer on other developers in the team. There are various client applications such as task prioritization, early conflict detection, and providing advice on testing that can benefit from such a characterization. The DeCAF framework leverages information from the development context to bound the iDiSE change impact analysis technique to analyze only the parts of the code base that are of interest. Bounding the analysis can enable DeCAF to efficiently compute the impact of changes using a combination of program dependence and symbolic execution based approaches.
Development Context Driven Change Awareness and Analysis Framework
NASA Technical Reports Server (NTRS)
Sarma, Anita; Branchaud, Josh; Dwyer, Matthew B.; Person, Suzette; Rungta, Neha; Wang, Yurong; Elbaum, Sebastian
2014-01-01
Recent work on workspace monitoring allows conflict prediction early in the development process, however, these approaches mostly use syntactic differencing techniques to compare different program versions. In contrast, traditional change-impact analysis techniques analyze related versions of the program only after the code has been checked into the master repository. We propose a novel approach, DeCAF (Development Context Analysis Framework), that leverages the development context to scope a change impact analysis technique. The goal is to characterize the impact of each developer on other developers in the team. There are various client applications such as task prioritization, early conflict detection, and providing advice on testing that can benefit from such a characterization. The DeCAF framework leverages information from the development context to bound the iDiSE change impact analysis technique to analyze only the parts of the code base that are of interest. Bounding the analysis can enable DeCAF to efficiently compute the impact of changes using a combination of program dependence and symbolic execution based approaches.
Modeling and managing risk early in software development
NASA Technical Reports Server (NTRS)
Briand, Lionel C.; Thomas, William M.; Hetmanski, Christopher J.
1993-01-01
In order to improve the quality of the software development process, we need to be able to build empirical multivariate models based on data collectable early in the software process. These models need to be both useful for prediction and easy to interpret, so that remedial actions may be taken in order to control and optimize the development process. We present an automated modeling technique which can be used as an alternative to regression techniques. We show how it can be used to facilitate the identification and aid the interpretation of the significant trends which characterize 'high risk' components in several Ada systems. Finally, we evaluate the effectiveness of our technique based on a comparison with logistic regression based models.
Tucker, Melanie T; Lewis, Dwight W; Payne Foster, Pamela; Lucky, Felecia; Yerby, Lea G; Hites, Lisle; Higginbotham, John C
2016-11-01
Developing meaningful community-based participatory relationships between researchers and the community can be challenging. The overall success of a community-based participatory relationship should be predicated on commitment and respect from empowered stakeholders. Prior to developing the technique discussed in this article, we hypothesized that the process of fostering relationships between researchers and the community was much like a social relationship: It has to develop organically and cannot be forced. To address this challenge, we developed a community-based participatory research-speed dating technique to foster relationships based on common interests, which we call CBPR-SD. This article describes the logistics of implementing CBPR-SD to foster scholarly collaborations. As part of a federally funded community-based research project, the speed dating technique was implemented for 10 researchers and 11 community leaders with a goal of developing scholarly collaborative groups who will submit applications for community-based research grants. In the end, four collaborative groups developed through CBPR-SD, three (75%) successfully submitted grant applications to fund pilot studies addressing obesity-related disparities in rural communities. Our preliminary findings suggest that CBPR-SD is a successful tool for promoting productive scholarly relationships between researchers and community leaders. © 2016 Society for Public Health Education.
Constraint-based component-modeling for knowledge-based design
NASA Technical Reports Server (NTRS)
Kolb, Mark A.
1992-01-01
The paper describes the application of various advanced programming techniques derived from artificial intelligence research to the development of flexible design tools for conceptual design. Special attention is given to two techniques which appear to be readily applicable to such design tools: the constraint propagation technique and the object-oriented programming. The implementation of these techniques in a prototype computer tool, Rubber Airplane, is described.
Kranz, Christine
2014-01-21
In recent years, major developments in scanning electrochemical microscopy (SECM) have significantly broadened the application range of this electroanalytical technique from high-resolution electrochemical imaging via nanoscale probes to large scale mapping using arrays of microelectrodes. A major driving force in advancing the SECM methodology is based on developing more sophisticated probes beyond conventional micro-disc electrodes usually based on noble metals or carbon microwires. This critical review focuses on the design and development of advanced electrochemical probes particularly enabling combinations of SECM with other analytical measurement techniques to provide information beyond exclusively measuring electrochemical sample properties. Consequently, this critical review will focus on recent progress and new developments towards multifunctional imaging.
Writing with Basals: A Sentence Combining Approach to Comprehension.
ERIC Educational Resources Information Center
Reutzel, D. Ray; Merrill, Jimmie D.
Sentence combining techniques can be used with basal readers to help students develop writing skills. The first technique is addition, characterized by using the connecting word "and" to join two or more base sentences together. The second technique is called "embedding," and is characterized by putting parts of two or more base sentences together…
NASA Astrophysics Data System (ADS)
Ageev, O. A.; Il'in, O. I.; Rubashkina, M. V.; Smirnov, V. A.; Fedotov, A. A.; Tsukanova, O. G.
2015-07-01
Techniques are developed to determine the resistance per unit length and the electrical resistivity of vertically aligned carbon nanotubes (VA CNTs) using atomic force microscopy (AFM) and scanning tunneling microscopy (STM). These techniques are used to study the resistance of VA CNTs. The resistance of an individual VA CNT calculated with the AFM-based technique is shown to be higher than the resistance of VA CNTs determined by the STM-based technique by a factor of 200, which is related to the influence of the resistance of the contact of an AFM probe to VA CNTs. The resistance per unit length and the electrical resistivity of an individual VA CNT 118 ± 39 nm in diameter and 2.23 ± 0.37 μm in height that are determined by the STM-based technique are 19.28 ± 3.08 kΩ/μm and 8.32 ± 3.18 × 10-4 Ω m, respectively. The STM-based technique developed to determine the resistance per unit length and the electrical resistivity of VA CNTs can be used to diagnose the electrical parameters of VA CNTs and to create VA CNT-based nanoelectronic elements.
Sen, Novonil; Kundu, Tribikram
2018-07-01
Estimating the location of an acoustic source in a structure is an important step towards passive structural health monitoring. Techniques for localizing an acoustic source in isotropic structures are well developed in the literature. Development of similar techniques for anisotropic structures, however, has gained attention only in the recent years and has a scope of further improvement. Most of the existing techniques for anisotropic structures either assume a straight line wave propagation path between the source and an ultrasonic sensor or require the material properties to be known. This study considers different shapes of the wave front generated during an acoustic event and develops a methodology to localize the acoustic source in an anisotropic plate from those wave front shapes. An elliptical wave front shape-based technique was developed first, followed by the development of a parametric curve-based technique for non-elliptical wave front shapes. The source coordinates are obtained by minimizing an objective function. The proposed methodology does not assume a straight line wave propagation path and can predict the source location without any knowledge of the elastic properties of the material. A numerical study presented here illustrates how the proposed methodology can accurately estimate the source coordinates. Copyright © 2018 Elsevier B.V. All rights reserved.
DEVELOPMENT OF ULTRATRACE LASER SPECTROMETRY TECHNIQUES FOR MEASUREMENTS OF ARSENIC
Development of Arsenic Speciation Techniques Based on High Performance Liquid Chromatography and Atomic Fluorescence Spectrometry
J.B. Simeonsson, H.D. Beach and D.J. Thomas
US EPA, Office of Research and Development, National Health and Environmental Effects Resear...
The Effects of Practice-Based Training on Graduate Teaching Assistants’ Classroom Practices
Becker, Erin A.; Easlon, Erin J.; Potter, Sarah C.; Guzman-Alvarez, Alberto; Spear, Jensen M.; Facciotti, Marc T.; Igo, Michele M.; Singer, Mitchell; Pagliarulo, Christopher
2017-01-01
Evidence-based teaching is a highly complex skill, requiring repeated cycles of deliberate practice and feedback to master. Despite existing well-characterized frameworks for practice-based training in K–12 teacher education, the major principles of these frameworks have not yet been transferred to instructor development in higher educational contexts, including training of graduate teaching assistants (GTAs). We sought to determine whether a practice-based training program could help GTAs learn and use evidence-based teaching methods in their classrooms. We implemented a weekly training program for introductory biology GTAs that included structured drills of techniques selected to enhance student practice, logic development, and accountability and reduce apprehension. These elements were selected based on their previous characterization as dimensions of active learning. GTAs received regular performance feedback based on classroom observations. To quantify use of target techniques and levels of student participation, we collected and coded 160 h of video footage. We investigated the relationship between frequency of GTA implementation of target techniques and student exam scores; however, we observed no significant relationship. Although GTAs adopted and used many of the target techniques with high frequency, techniques that enforced student participation were not stably adopted, and their use was unresponsive to formal feedback. We also found that techniques discussed in training, but not practiced, were not used at quantifiable frequencies, further supporting the importance of practice-based training for influencing instructional practices. PMID:29146664
Novel diagnostic techniques for celiac disease.
Kurppa, Kalle; Taavela, Juha; Saavalainen, Päivi; Kaukinen, Katri; Lindfors, Katri
2016-07-01
The diagnosis of celiac disease has long been based on the demonstration of gluten-induced small-bowel mucosal damage. However, due to the constantly increasing disease prevalence and limitations in the histology-based criteria there is a pressure towards more serology-based diagnostics. The serological tools are being improved and new non-invasive methods are being developed, but the constantly refined endoscopic and histologic techniques may still prove helpful. Moreover, growing understanding of the disease pathogenesis has led researchers to suggest completely novel approaches to celiac disease diagnostics regardless of disease activity. In this review, we will elucidate the most recent development and possible future innovations in the diagnostic techniques for celiac disease.
Inversion of particle-size distribution from angular light-scattering data with genetic algorithms.
Ye, M; Wang, S; Lu, Y; Hu, T; Zhu, Z; Xu, Y
1999-04-20
A stochastic inverse technique based on a genetic algorithm (GA) to invert particle-size distribution from angular light-scattering data is developed. This inverse technique is independent of any given a priori information of particle-size distribution. Numerical tests show that this technique can be successfully applied to inverse problems with high stability in the presence of random noise and low susceptibility to the shape of distributions. It has also been shown that the GA-based inverse technique is more efficient in use of computing time than the inverse Monte Carlo method recently developed by Ligon et al. [Appl. Opt. 35, 4297 (1996)].
Study of fault tolerant software technology for dynamic systems
NASA Technical Reports Server (NTRS)
Caglayan, A. K.; Zacharias, G. L.
1985-01-01
The major aim of this study is to investigate the feasibility of using systems-based failure detection isolation and compensation (FDIC) techniques in building fault-tolerant software and extending them, whenever possible, to the domain of software fault tolerance. First, it is shown that systems-based FDIC methods can be extended to develop software error detection techniques by using system models for software modules. In particular, it is demonstrated that systems-based FDIC techniques can yield consistency checks that are easier to implement than acceptance tests based on software specifications. Next, it is shown that systems-based failure compensation techniques can be generalized to the domain of software fault tolerance in developing software error recovery procedures. Finally, the feasibility of using fault-tolerant software in flight software is investigated. In particular, possible system and version instabilities, and functional performance degradation that may occur in N-Version programming applications to flight software are illustrated. Finally, a comparative analysis of N-Version and recovery block techniques in the context of generic blocks in flight software is presented.
ERIC Educational Resources Information Center
Neiva de Figueiredo, Joao; Mauri, Alfredo J.
2013-01-01
This article describes the "Cross-Cultural Assignment," an experiential learning technique for students of business that deepens self-awareness of their own attitudes toward different cultures and develops international managerial skills. The technique consists of pairing up small teams of U.S.-based business students with small teams of…
Hill, Ryan T.
2015-01-01
The unique optical properties of plasmon resonant nanostructures enable exploration of nanoscale environments using relatively simple optical characterization techniques. For this reason, the field of plasmonics continues to garner the attention of the biosensing community. Biosensors based on propagating surface plasmon resonances (SPRs) in films are the most well-recognized plasmonic biosensors, but there is great potential for the new, developing technologies to surpass the robustness and popularity of film-based SPR sensing. This review surveys the current plasmonic biosensor landscape with emphasis on the basic operating principles of each plasmonic sensing technique and the practical considerations when developing a sensing platform with the various techniques. The “gold standard” film SPR technique is reviewed briefly, but special emphasis is devoted to the up-and-coming LSPR-based and plasmonically coupled sensor technology. PMID:25377594
Mirapeix, J; Cobo, A; González, D A; López-Higuera, J M
2007-02-19
A new plasma spectroscopy analysis technique based on the generation of synthetic spectra by means of optimization processes is presented in this paper. The technique has been developed for its application in arc-welding quality assurance. The new approach has been checked through several experimental tests, yielding results in reasonably good agreement with the ones offered by the traditional spectroscopic analysis technique.
Using cognitive task analysis to develop simulation-based training for medical tasks.
Cannon-Bowers, Jan; Bowers, Clint; Stout, Renee; Ricci, Katrina; Hildabrand, Annette
2013-10-01
Pressures to increase the efficacy and effectiveness of medical training are causing the Department of Defense to investigate the use of simulation technologies. This article describes a comprehensive cognitive task analysis technique that can be used to simultaneously generate training requirements, performance metrics, scenario requirements, and simulator/simulation requirements for medical tasks. On the basis of a variety of existing techniques, we developed a scenario-based approach that asks experts to perform the targeted task multiple times, with each pass probing a different dimension of the training development process. In contrast to many cognitive task analysis approaches, we argue that our technique can be highly cost effective because it is designed to accomplish multiple goals. The technique was pilot tested with expert instructors from a large military medical training command. These instructors were employed to generate requirements for two selected combat casualty care tasks-cricothyroidotomy and hemorrhage control. Results indicated that the technique is feasible to use and generates usable data to inform simulation-based training system design. Reprint & Copyright © 2013 Association of Military Surgeons of the U.S.
Nonlinear relaxation algorithms for circuit simulation
DOE Office of Scientific and Technical Information (OSTI.GOV)
Saleh, R.A.
Circuit simulation is an important Computer-Aided Design (CAD) tool in the design of Integrated Circuits (IC). However, the standard techniques used in programs such as SPICE result in very long computer-run times when applied to large problems. In order to reduce the overall run time, a number of new approaches to circuit simulation were developed and are described. These methods are based on nonlinear relaxation techniques and exploit the relative inactivity of large circuits. Simple waveform-processing techniques are described to determine the maximum possible speed improvement that can be obtained by exploiting this property of large circuits. Three simulation algorithmsmore » are described, two of which are based on the Iterated Timing Analysis (ITA) method and a third based on the Waveform-Relaxation Newton (WRN) method. New programs that incorporate these techniques were developed and used to simulate a variety of industrial circuits. The results from these simulations are provided. The techniques are shown to be much faster than the standard approach. In addition, a number of parallel aspects of these algorithms are described, and a general space-time model of parallel-task scheduling is developed.« less
Dynamic Domains in Data Production Planning
NASA Technical Reports Server (NTRS)
Golden, Keith; Pang, Wanlin
2005-01-01
This paper discusses a planner-based approach to automating data production tasks, such as producing fire forecasts from satellite imagery and weather station data. Since the set of available data products is large, dynamic and mostly unknown, planning techniques developed for closed worlds are unsuitable. We discuss a number of techniques we have developed to cope with data production domains, including a novel constraint propagation algorithm based on planning graphs and a constraint-based approach to interleaved planning, sensing and execution.
NASA Astrophysics Data System (ADS)
Juanle, Wang; Shuang, Li; Yunqiang, Zhu
2005-10-01
According to the requirements of China National Scientific Data Sharing Program (NSDSP), the research and development of web oriented RS Image Publication System (RSIPS) is based on Java Servlet technique. The designing of RSIPS framework is composed of 3 tiers, which is Presentation Tier, Application Service Tier and Data Resource Tier. Presentation Tier provides user interface for data query, review and download. For the convenience of users, visual spatial query interface is included. Served as a middle tier, Application Service Tier controls all actions between users and databases. Data Resources Tier stores RS images in file and relationship databases. RSIPS is developed with cross platform programming based on Java Servlet tools, which is one of advanced techniques in J2EE architecture. RSIPS's prototype has been developed and applied in the geosciences clearinghouse practice which is among the experiment units of NSDSP in China.
Macready, Anna L; Fallaize, Rosalind; Butler, Laurie T; Ellis, Judi A; Kuznesof, Sharron; Frewer, Lynn J; Celis-Morales, Carlos; Livingstone, Katherine M; Araújo-Soares, Vera; Fischer, Arnout Rh; Stewart-Knox, Barbara J; Mathers, John C; Lovegrove, Julie A
2018-04-09
To determine the efficacy of behavior change techniques applied in dietary and physical activity intervention studies, it is first necessary to record and describe techniques that have been used during such interventions. Published frameworks used in dietary and smoking cessation interventions undergo continuous development, and most are not adapted for Web-based delivery. The Food4Me study (N=1607) provided the opportunity to use existing frameworks to describe standardized Web-based techniques employed in a large-scale, internet-based intervention to change dietary behavior and physical activity. The aims of this study were (1) to describe techniques embedded in the Food4Me study design and explain the selection rationale and (2) to demonstrate the use of behavior change technique taxonomies, develop standard operating procedures for training, and identify strengths and limitations of the Food4Me framework that will inform its use in future studies. The 6-month randomized controlled trial took place simultaneously in seven European countries, with participants receiving one of four levels of personalized advice (generalized, intake-based, intake+phenotype-based, and intake+phenotype+gene-based). A three-phase approach was taken: (1) existing taxonomies were reviewed and techniques were identified a priori for possible inclusion in the Food4Me study, (2) a standard operating procedure was developed to maintain consistency in the use of methods and techniques across research centers, and (3) the Food4Me behavior change technique framework was reviewed and updated post intervention. An analysis of excluded techniques was also conducted. Of 46 techniques identified a priori as being applicable to Food4Me, 17 were embedded in the intervention design; 11 were from a dietary taxonomy, and 6 from a smoking cessation taxonomy. In addition, the four-category smoking cessation framework structure was adopted for clarity of communication. Smoking cessation texts were adapted for dietary use where necessary. A posteriori, a further 9 techniques were included. Examination of excluded items highlighted the distinction between techniques considered appropriate for face-to-face versus internet-based delivery. The use of existing taxonomies facilitated the description and standardization of techniques used in Food4Me. We recommend that for complex studies of this nature, technique analysis should be conducted a priori to develop standardized procedures and training and reviewed a posteriori to audit the techniques actually adopted. The present framework description makes a valuable contribution to future systematic reviews and meta-analyses that explore technique efficacy and underlying psychological constructs. This was a novel application of the behavior change taxonomies and was the first internet-based personalized nutrition intervention to use such a framework remotely. ClinicalTrials.gov NCT01530139; https://clinicaltrials.gov/ct2/show/NCT01530139 (Archived by WebCite at http://www.webcitation.org/6y8XYUft1). ©Anna L Macready, Rosalind Fallaize, Laurie T Butler, Judi A Ellis, Sharron Kuznesof, Lynn J Frewer, Carlos Celis-Morales, Katherine M Livingstone, Vera Araújo-Soares, Arnout RH Fischer, Barbara J Stewart-Knox, John C Mathers, Julie A Lovegrove. Originally published in JMIR Research Protocols (http://www.researchprotocols.org), 09.04.2018.
Diagnostics and Active Control of Aircraft Interior Noise
NASA Technical Reports Server (NTRS)
Fuller, C. R.
1998-01-01
This project deals with developing advanced methods for investigating and controlling interior noise in aircraft. The work concentrates on developing and applying the techniques of Near Field Acoustic Holography (NAH) and Principal Component Analysis (PCA) to the aircraft interior noise dynamic problem. This involves investigating the current state of the art, developing new techniques and then applying them to the particular problem being studied. The knowledge gained under the first part of the project was then used to develop and apply new, advanced noise control techniques for reducing interior noise. A new fully active control approach based on the PCA was developed and implemented on a test cylinder. Finally an active-passive approach based on tunable vibration absorbers was to be developed and analytically applied to a range of test structures from simple plates to aircraft fuselages.
NASA Astrophysics Data System (ADS)
Sokkar, T. Z. N.; El-Farahaty, K. A.; El-Bakary, M. A.; Raslan, M. I.; Omar, E. Z.; Hamza, A. A.
2018-03-01
The optical setup of the transport intensity equation (TIE) technique is developed to be valid for measuring the optical properties of the highly-oriented anisotropic fibres. This development is based on the microstructure models of the highly-oriented anisotropic fibres and the principle of anisotropy. We provide the setup of TIE technique with polarizer which is controlled via stepper motor. This developed technique is used to investigate the refractive indices in the parallel and perpendicular polarization directions of light for the highly-oriented poly (ethylene terephthalate) (PET) fibres and hence its birefringence. The obtained results through the developed TIE technique for PET fibre are compared with that determined experimentally using the Mach-Zehnder interferometer under the same conditions. The comparison shows a good agreement between the obtained results from the developed technique and that obtained from the Mach-Zehnder interferometer technique.
Using Every-Pupil Response Techniques to Develop Decoding Ability.
ERIC Educational Resources Information Center
Whisler, Nancy G.
Noting that "every pupil response" (EPR) techniques allow for all students in a group to respond to each question asked by a teacher, this paper explains how EPR techniques may be incorporated in a reading lesson to help students develop decoding ability. The paper offers a justification for using these directed teaching strategies based on…
Leveraging Experiential Learning Techniques for Transfer
ERIC Educational Resources Information Center
Furman, Nate; Sibthorp, Jim
2013-01-01
Experiential learning techniques can be helpful in fostering learning transfer. Techniques such as project-based learning, reflective learning, and cooperative learning provide authentic platforms for developing rich learning experiences. In contrast to more didactic forms of instruction, experiential learning techniques foster a depth of learning…
Using Psychodrama Techniques to Promote Counselor Identity Development in Group Supervision
ERIC Educational Resources Information Center
Scholl, Mark B.; Smith-Adcock, Sondra
2007-01-01
The authors briefly introduce the concepts, techniques, and theory of identity development associated with J. L. Moreno's (1946, 1969, 1993) Psychodrama. Based upon Loganbill, Hardy, and Delworth's (1982) model, counselor identity development is conceptualized as consisting of seven developmental themes or vectors (e.g., issues of awareness and…
Deductive Glue Code Synthesis for Embedded Software Systems Based on Code Patterns
NASA Technical Reports Server (NTRS)
Liu, Jian; Fu, Jicheng; Zhang, Yansheng; Bastani, Farokh; Yen, I-Ling; Tai, Ann; Chau, Savio N.
2006-01-01
Automated code synthesis is a constructive process that can be used to generate programs from specifications. It can, thus, greatly reduce the software development cost and time. The use of formal code synthesis approach for software generation further increases the dependability of the system. Though code synthesis has many potential benefits, the synthesis techniques are still limited. Meanwhile, components are widely used in embedded system development. Applying code synthesis to component based software development (CBSD) process can greatly enhance the capability of code synthesis while reducing the component composition efforts. In this paper, we discuss the issues and techniques for applying deductive code synthesis techniques to CBSD. For deductive synthesis in CBSD, a rule base is the key for inferring appropriate component composition. We use the code patterns to guide the development of rules. Code patterns have been proposed to capture the typical usages of the components. Several general composition operations have been identified to facilitate systematic composition. We present the technique for rule development and automated generation of new patterns from existing code patterns. A case study of using this method in building a real-time control system is also presented.
Integer Linear Programming in Computational Biology
NASA Astrophysics Data System (ADS)
Althaus, Ernst; Klau, Gunnar W.; Kohlbacher, Oliver; Lenhof, Hans-Peter; Reinert, Knut
Computational molecular biology (bioinformatics) is a young research field that is rich in NP-hard optimization problems. The problem instances encountered are often huge and comprise thousands of variables. Since their introduction into the field of bioinformatics in 1997, integer linear programming (ILP) techniques have been successfully applied to many optimization problems. These approaches have added much momentum to development and progress in related areas. In particular, ILP-based approaches have become a standard optimization technique in bioinformatics. In this review, we present applications of ILP-based techniques developed by members and former members of Kurt Mehlhorn’s group. These techniques were introduced to bioinformatics in a series of papers and popularized by demonstration of their effectiveness and potential.
ERIC Educational Resources Information Center
Schneider, Marlene; Robin, Arthur
This manual describes the Turtle Technique which was developed to help children with behavior problems control their own disruptive behavior. The technique differs from other behavior modification techniques in that it is based upon self-control rather than external control of disruptive behavior. The Turtle Technique first teaches the child how…
A Review of Recent Developments in X-Ray Diagnostics for Turbulent and Optically Dense Rocket Sprays
NASA Technical Reports Server (NTRS)
Radke, Christopher; Halls, Benjamin; Kastengren, Alan; Meyer, Terrence
2017-01-01
Highly efficient mixing and atomization of fuel and oxidizers is an important factor in many propulsion and power generating applications. To better quantify breakup and mixing in atomizing sprays, several diagnostic techniques have been developed to collect droplet information and spray statistics. Several optical based techniques, such as Ballistic Imaging and SLIPI have previously demonstrated qualitative measurements in optically dense sprays, however these techniques have produced limited quantitative information in the near injector region. To complement to these advances, a recent wave of developments utilizing synchrotron based x-rays have been successful been implemented facilitating the collection of quantitative measurements in optically dense sprays.
Distributed acoustic sensing technique and its field trial in SAGD well
NASA Astrophysics Data System (ADS)
Han, Li; He, Xiangge; Pan, Yong; Liu, Fei; Yi, Duo; Hu, Chengjun; Zhang, Min; Gu, Lijuan
2017-10-01
Steam assisted gravity drainage (SAGD) is a very promising way for the development of heavy oil, extra heavy oil and tight oil reservoirs. Proper monitoring of the SAGD operations is essential to avoid operational issues and improve efficiency. Among all the monitoring techniques, micro-seismic monitoring and related interpretation method can give useful information about the steam chamber development and has been extensively studied. Distributed acoustic sensor (DAS) based on Rayleigh backscattering is a newly developed technique that can measure acoustic signal at all points along the sensing fiber. In this paper, we demonstrate a DAS system based on dual-pulse heterodyne demodulation technique and did field trial in SAGD well located in Xinjiang Oilfield, China. The field trail results validated the performance of the DAS system and indicated its applicability in steam-chamber monitoring and hydraulic monitoring.
NASA Technical Reports Server (NTRS)
Woodward, W. A.; Gray, H. L.
1983-01-01
Efforts in support of the development of multicrop production monitoring capability are reported. In particular, segment level proportion estimation techniques based upon a mixture model were investigated. Efforts have dealt primarily with evaluation of current techniques and development of alternative ones. A comparison of techniques is provided on both simulated and LANDSAT data along with an analysis of the quality of profile variables obtained from LANDSAT data.
NASA Astrophysics Data System (ADS)
McCann, C.; Repasky, K. S.; Morin, M.; Lawrence, R. L.; Powell, S. L.
2016-12-01
Compact, cost-effective, flight-based hyperspectral imaging systems can provide scientifically relevant data over large areas for a variety of applications such as ecosystem studies, precision agriculture, and land management. To fully realize this capability, unsupervised classification techniques based on radiometrically-calibrated data that cluster based on biophysical similarity rather than simply spectral similarity are needed. An automated technique to produce high-resolution, large-area, radiometrically-calibrated hyperspectral data sets based on the Landsat surface reflectance data product as a calibration target was developed and applied to three subsequent years of data covering approximately 1850 hectares. The radiometrically-calibrated data allows inter-comparison of the temporal series. Advantages of the radiometric calibration technique include the need for minimal site access, no ancillary instrumentation, and automated processing. Fitting the reflectance spectra of each pixel using a set of biophysically relevant basis functions reduces the data from 80 spectral bands to 9 parameters providing noise reduction and data compression. Examination of histograms of these parameters allows for determination of natural splitting into biophysical similar clusters. This method creates clusters that are similar in terms of biophysical parameters, not simply spectral proximity. Furthermore, this method can be applied to other data sets, such as urban scenes, by developing other physically meaningful basis functions. The ability to use hyperspectral imaging for a variety of important applications requires the development of data processing techniques that can be automated. The radiometric-calibration combined with the histogram based unsupervised classification technique presented here provide one potential avenue for managing big-data associated with hyperspectral imaging.
Additive manufacturing techniques for the production of tissue engineering constructs.
Mota, Carlos; Puppi, Dario; Chiellini, Federica; Chiellini, Emo
2015-03-01
'Additive manufacturing' (AM) refers to a class of manufacturing processes based on the building of a solid object from three-dimensional (3D) model data by joining materials, usually layer upon layer. Among the vast array of techniques developed for the production of tissue-engineering (TE) scaffolds, AM techniques are gaining great interest for their suitability in achieving complex shapes and microstructures with a high degree of automation, good accuracy and reproducibility. In addition, the possibility of rapidly producing tissue-engineered constructs meeting patient's specific requirements, in terms of tissue defect size and geometry as well as autologous biological features, makes them a powerful way of enhancing clinical routine procedures. This paper gives an extensive overview of different AM techniques classes (i.e. stereolithography, selective laser sintering, 3D printing, melt-extrusion-based techniques, solution/slurry extrusion-based techniques, and tissue and organ printing) employed for the development of tissue-engineered constructs made of different materials (i.e. polymeric, ceramic and composite, alone or in combination with bioactive agents), by highlighting their principles and technological solutions. Copyright © 2012 John Wiley & Sons, Ltd.
The design and implementation of hydrographical information management system (HIMS)
NASA Astrophysics Data System (ADS)
Sui, Haigang; Hua, Li; Wang, Qi; Zhang, Anming
2005-10-01
With the development of hydrographical work and information techniques, the large variety of hydrographical information including electronic charts, documents and other materials are widely used, and the traditional management mode and techniques are unsuitable for the development of the Chinese Marine Safety Administration Bureau (CMSAB). How to manage all kinds of hydrographical information has become an important and urgent problem. A lot of advanced techniques including GIS, RS, spatial database management and VR techniques are introduced for solving these problems. Some design principles and key techniques of the HIMS including the mixed mode base on B/S, C/S and stand-alone computer mode, multi-source & multi-scale data organization and management, multi-source data integration and diverse visualization of digital chart, efficient security control strategies are illustrated in detail. Based on the above ideas and strategies, an integrated system named Hydrographical Information Management System (HIMS) was developed. And the HIMS has been applied in the Shanghai Marine Safety Administration Bureau and obtained good evaluation.
DOT National Transportation Integrated Search
2011-01-01
This study develops an enhanced transportation planning framework by augmenting the sequential four-step : planning process with post-processing techniques. The post-processing techniques are incorporated through a feedback : mechanism and aim to imp...
NASA Technical Reports Server (NTRS)
Coen, Peter G.
1991-01-01
A new computer technique for the analysis of transport aircraft sonic boom signature characteristics was developed. This new technique, based on linear theory methods, combines the previously separate equivalent area and F function development with a signature propagation method using a single geometry description. The new technique was implemented in a stand-alone computer program and was incorporated into an aircraft performance analysis program. Through these implementations, both configuration designers and performance analysts are given new capabilities to rapidly analyze an aircraft's sonic boom characteristics throughout the flight envelope.
Machine learning modelling for predicting soil liquefaction susceptibility
NASA Astrophysics Data System (ADS)
Samui, P.; Sitharam, T. G.
2011-01-01
This study describes two machine learning techniques applied to predict liquefaction susceptibility of soil based on the standard penetration test (SPT) data from the 1999 Chi-Chi, Taiwan earthquake. The first machine learning technique which uses Artificial Neural Network (ANN) based on multi-layer perceptions (MLP) that are trained with Levenberg-Marquardt backpropagation algorithm. The second machine learning technique uses the Support Vector machine (SVM) that is firmly based on the theory of statistical learning theory, uses classification technique. ANN and SVM have been developed to predict liquefaction susceptibility using corrected SPT [(N1)60] and cyclic stress ratio (CSR). Further, an attempt has been made to simplify the models, requiring only the two parameters [(N1)60 and peck ground acceleration (amax/g)], for the prediction of liquefaction susceptibility. The developed ANN and SVM models have also been applied to different case histories available globally. The paper also highlights the capability of the SVM over the ANN models.
Classification of the Regional Ionospheric Disturbance Based on Machine Learning Techniques
NASA Astrophysics Data System (ADS)
Terzi, Merve Begum; Arikan, Orhan; Karatay, Secil; Arikan, Feza; Gulyaeva, Tamara
2016-08-01
In this study, Total Electron Content (TEC) estimated from GPS receivers is used to model the regional and local variability that differs from global activity along with solar and geomagnetic indices. For the automated classification of regional disturbances, a classification technique based on a robust machine learning technique that have found wide spread use, Support Vector Machine (SVM) is proposed. Performance of developed classification technique is demonstrated for midlatitude ionosphere over Anatolia using TEC estimates generated from GPS data provided by Turkish National Permanent GPS Network (TNPGN-Active) for solar maximum year of 2011. As a result of implementing developed classification technique to Global Ionospheric Map (GIM) TEC data, which is provided by the NASA Jet Propulsion Laboratory (JPL), it is shown that SVM can be a suitable learning method to detect anomalies in TEC variations.
Low cost MATLAB-based pulse oximeter for deployment in research and development applications.
Shokouhian, M; Morling, R C S; Kale, I
2013-01-01
Problems such as motion artifact and effects of ambient lights have forced developers to design different signal processing techniques and algorithms to increase the reliability and accuracy of the conventional pulse oximeter device. To evaluate the robustness of these techniques, they are applied either to recorded data or are implemented on chip to be applied to real-time data. Recorded data is the most common method of evaluating however it is not as reliable as real-time measurements. On the other hand, hardware implementation can be both expensive and time consuming. This paper presents a low cost MATLAB-based pulse oximeter that can be used for rapid evaluation of newly developed signal processing techniques and algorithms. Flexibility to apply different signal processing techniques, providing both processed and unprocessed data along with low implementation cost are the important features of this design which makes it ideal for research and development purposes, as well as commercial, hospital and healthcare application.
NASA Astrophysics Data System (ADS)
Kleshnin, Mikhail; Orlova, Anna; Kirillin, Mikhail; Golubiatnikov, German; Turchin, Ilya
2017-07-01
A new approach to optical measuring blood oxygen saturation was developed and implemented. This technique is based on an original three-stage algorithm for reconstructing the relative concentration of biological chromophores (hemoglobin, water, lipids) from the measured spectra of diffusely scattered light at different distances from the probing radiation source. The numerical experiments and approbation of the proposed technique on a biological phantom have shown the high reconstruction accuracy and the possibility of correct calculation of hemoglobin oxygenation in the presence of additive noise and calibration errors. The obtained results of animal studies have agreed with the previously published results of other research groups and demonstrated the possibility to apply the developed technique to monitor oxygen saturation in tumor tissue.
Standoff laser-based spectroscopy for explosives detection
NASA Astrophysics Data System (ADS)
Gaft, M.; Nagli, L.
2007-10-01
Real time detection and identification of explosives at a standoff distance is a major issue in efforts to develop defense against so-called Improvised Explosive Devices (IED). It is recognized that the only technique, which is potentially capable to standoff detection of minimal amounts of explosives is laser-based spectroscopy. LDS activity is based on a combination of laser-based spectroscopic methods with orthogonal capabilities. Our technique belongs to trace detection, namely to its micro-particles variety. It is based on commonly held belief that surface contamination was very difficult to avoid and could be exploited for standoff detection. We has applied optical techniques including gated Raman and time-resolved luminescence spectroscopy for detection of main explosive materials, both factory and homemade. We developed and tested a Raman system for the field remote detection and identification of minimal amounts of explosives on relevant surfaces at a distance of up to 30 meters.
NASA Technical Reports Server (NTRS)
Djorgovski, George
1993-01-01
The existing and forthcoming data bases from NASA missions contain an abundance of information whose complexity cannot be efficiently tapped with simple statistical techniques. Powerful multivariate statistical methods already exist which can be used to harness much of the richness of these data. Automatic classification techniques have been developed to solve the problem of identifying known types of objects in multiparameter data sets, in addition to leading to the discovery of new physical phenomena and classes of objects. We propose an exploratory study and integration of promising techniques in the development of a general and modular classification/analysis system for very large data bases, which would enhance and optimize data management and the use of human research resource.
NASA Technical Reports Server (NTRS)
Djorgovski, Stanislav
1992-01-01
The existing and forthcoming data bases from NASA missions contain an abundance of information whose complexity cannot be efficiently tapped with simple statistical techniques. Powerful multivariate statistical methods already exist which can be used to harness much of the richness of these data. Automatic classification techniques have been developed to solve the problem of identifying known types of objects in multi parameter data sets, in addition to leading to the discovery of new physical phenomena and classes of objects. We propose an exploratory study and integration of promising techniques in the development of a general and modular classification/analysis system for very large data bases, which would enhance and optimize data management and the use of human research resources.
Warmerdam, G; Vullings, R; Van Pul, C; Andriessen, P; Oei, S G; Wijn, P
2013-01-01
Non-invasive fetal electrocardiography (ECG) can be used for prolonged monitoring of the fetal heart rate (FHR). However, the signal-to-noise-ratio (SNR) of non-invasive ECG recordings is often insufficient for reliable detection of the FHR. To overcome this problem, source separation techniques can be used to enhance the fetal ECG. This study uses a physiology-based source separation (PBSS) technique that has already been demonstrated to outperform widely used blind source separation techniques. Despite the relatively good performance of PBSS in enhancing the fetal ECG, PBSS is still susceptible to artifacts. In this study an augmented PBSS technique is developed to reduce the influence of artifacts. The performance of the developed method is compared to PBSS on multi-channel non-invasive fetal ECG recordings. Based on this comparison, the developed method is shown to outperform PBSS for the enhancement of the fetal ECG.
Theoretical Bound of CRLB for Energy Efficient Technique of RSS-Based Factor Graph Geolocation
NASA Astrophysics Data System (ADS)
Kahar Aziz, Muhammad Reza; Heriansyah; Saputra, EfaMaydhona; Musa, Ardiansyah
2018-03-01
To support the increase of wireless geolocation development as the key of the technology in the future, this paper proposes theoretical bound derivation, i.e., Cramer Rao lower bound (CRLB) for energy efficient of received signal strength (RSS)-based factor graph wireless geolocation technique. The theoretical bound derivation is crucially important to evaluate whether the energy efficient technique of RSS-based factor graph wireless geolocation is effective as well as to open the opportunity to further innovation of the technique. The CRLB is derived in this paper by using the Fisher information matrix (FIM) of the main formula of the RSS-based factor graph geolocation technique, which is lied on the Jacobian matrix. The simulation result shows that the derived CRLB has the highest accuracy as a bound shown by its lowest root mean squared error (RMSE) curve compared to the RMSE curve of the RSS-based factor graph geolocation technique. Hence, the derived CRLB becomes the lower bound for the efficient technique of RSS-based factor graph wireless geolocation.
NASA Technical Reports Server (NTRS)
Otugen, M. Volkan
1997-01-01
Non-intrusive techniques for the dynamic measurement of gas flow properties such as density, temperature and velocity, are needed in the research leading to the development of new generation high-speed aircraft. Accurate velocity, temperature and density data obtained in ground testing and in-flight measurements can help understand the flow physics leading to transition and turbulence in supersonic, high-altitude flight. Such non-intrusive measurement techniques can also be used to study combustion processes of hydrocarbon fuels in aircraft engines. Reliable, time and space resolved temperature measurements in various combustor configurations can lead to a better understanding of high temperature chemical reaction dynamics thus leading to improved modeling and better prediction of such flows. In view of this, a research program was initiated at Polytechnic University's Aerodynamics Laboratory with support from NASA Lewis Research Center through grants NAG3-1301 and NAG3-1690. The overall objective of this program has been to develop laser-based, non-contact, space- and time-resolved temperature and velocity measurement techniques. In the initial phase of the program a ND:YAG laser-based dual-line Rayleigh scattering technique was developed and tested for the accurate measurement of gas temperature in the presence of background laser glare. Effort was next directed towards the development of a filtered, spectrally-resolved Rayleigh/Mie scattering technique with the objective of developing an interferometric method for time-frozen velocity measurements in high-speed flows utilizing the uv line of an ND:YAG laser and an appropriate molecular absorption filter. This effort included both a search for an appropriate filter material for the 266 nm laser line and the development and testing of several image processing techniques for the fast processing of Fabry-Perot images for velocity and temperature information. Finally, work was also carried out for the development of a new laser-based strain-rate and vorticity technique for the time-resolved measurement of vorticity and strain-rates in turbulent flows.
The Role of a Physical Analysis Laboratory in a 300 mm IC Development and Manufacturing Centre
NASA Astrophysics Data System (ADS)
Kwakman, L. F. Tz.; Bicais-Lepinay, N.; Courtas, S.; Delille, D.; Juhel, M.; Trouiller, C.; Wyon, C.; de la Bardonnie, M.; Lorut, F.; Ross, R.
2005-09-01
To remain competitive IC manufacturers have to accelerate the development of most advanced (CMOS) technology and to deliver high yielding products with best cycle times and at a competitive pricing. With the increase of technology complexity, also the need for physical characterization support increases, however many of the existing techniques are no longer adequate to effectively support the 65-45 nm technology node developments. New and improved techniques are definitely needed to better characterize the often marginal processes, but these should not significantly impact fabrication costs or cycle time. Hence, characterization and metrology challenges in state-of-the-art IC manufacturing are both of technical and economical nature. TEM microscopy is needed for high quality, high volume analytical support but several physical and practical hurdles have to be taken. The success rate of FIB-SEM based failure analysis drops as defects often are too small to be detected and fault isolation becomes more difficult in the nano-scale device structures. To remain effective and efficient, SEM and OBIRCH techniques have to be improved or complemented with other more effective methods. Chemical analysis of novel materials and critical interfaces requires improvements in the field of e.g. SIMS, ToF-SIMS. Techniques that previously were only used sporadically, like EBSD and XRD, have become a `must' to properly support backend process development. At the bright side, thanks to major technical advances, techniques that previously were practiced at laboratory level only now can be used effectively for at-line fab metrology: Voltage Contrast based defectivity control, XPS based gate dielectric metrology and XRD based control of copper metallization processes are practical examples. In this paper capabilities and shortcomings of several techniques and corresponding equipment are presented with practical illustrations of use in our Crolles facilities.
Model-Based GN and C Simulation and Flight Software Development for Orion Missions beyond LEO
NASA Technical Reports Server (NTRS)
Odegard, Ryan; Milenkovic, Zoran; Henry, Joel; Buttacoli, Michael
2014-01-01
For Orion missions beyond low Earth orbit (LEO), the Guidance, Navigation, and Control (GN&C) system is being developed using a model-based approach for simulation and flight software. Lessons learned from the development of GN&C algorithms and flight software for the Orion Exploration Flight Test One (EFT-1) vehicle have been applied to the development of further capabilities for Orion GN&C beyond EFT-1. Continuing the use of a Model-Based Development (MBD) approach with the Matlab®/Simulink® tool suite, the process for GN&C development and analysis has been largely improved. Furthermore, a model-based simulation environment in Simulink, rather than an external C-based simulation, greatly eases the process for development of flight algorithms. The benefits seen by employing lessons learned from EFT-1 are described, as well as the approach for implementing additional MBD techniques. Also detailed are the key enablers for improvements to the MBD process, including enhanced configuration management techniques for model-based software systems, automated code and artifact generation, and automated testing and integration.
Ground Vibration Test Planning and Pre-Test Analysis for the X-33 Vehicle
NASA Technical Reports Server (NTRS)
Bedrossian, Herand; Tinker, Michael L.; Hidalgo, Homero
2000-01-01
This paper describes the results of the modal test planning and the pre-test analysis for the X-33 vehicle. The pre-test analysis included the selection of the target modes, selection of the sensor and shaker locations and the development of an accurate Test Analysis Model (TAM). For target mode selection, four techniques were considered, one based on the Modal Cost technique, one based on Balanced Singular Value technique, a technique known as the Root Sum Squared (RSS) method, and a Modal Kinetic Energy (MKE) approach. For selecting sensor locations, four techniques were also considered; one based on the Weighted Average Kinetic Energy (WAKE), one based on Guyan Reduction (GR), one emphasizing engineering judgment, and one based on an optimum sensor selection technique using Genetic Algorithm (GA) search technique combined with a criteria based on Hankel Singular Values (HSV's). For selecting shaker locations, four techniques were also considered; one based on the Weighted Average Driving Point Residue (WADPR), one based on engineering judgment and accessibility considerations, a frequency response method, and an optimum shaker location selection based on a GA search technique combined with a criteria based on HSV's. To evaluate the effectiveness of the proposed sensor and shaker locations for exciting the target modes, extensive numerical simulations were performed. Multivariate Mode Indicator Function (MMIF) was used to evaluate the effectiveness of each sensor & shaker set with respect to modal parameter identification. Several TAM reduction techniques were considered including, Guyan, IRS, Modal, and Hybrid. Based on a pre-test cross-orthogonality checks using various reduction techniques, a Hybrid TAM reduction technique was selected and was used for all three vehicle fuel level configurations.
Application of light and ultrasound for medical diagnostics and treatment
NASA Astrophysics Data System (ADS)
Esenaliev, Rinat O.
2002-07-01
We develop novel optical and ultrasound techniques for medical noninvasive diagnostics and treatment. In this review, we present our results on the development of: (1) optoacoustic technique for detection of small tumors; (2) optoacoustic monitoring of blood oxygenation; (3) optoacoustic monitoring during thermotherapy; (4) optical coherence tomography for monitoring of blood glucose concentration; and (5) laser- and ultrasound-based anti- cancer drug delivery technique. Motivation, experimental methods, results obtained in vitro and in vivo with the use of these techniques are presented.
The anatomy of floating shock fitting. [shock waves computation for flow field
NASA Technical Reports Server (NTRS)
Salas, M. D.
1975-01-01
The floating shock fitting technique is examined. Second-order difference formulas are developed for the computation of discontinuities. A procedure is developed to compute mesh points that are crossed by discontinuities. The technique is applied to the calculation of internal two-dimensional flows with arbitrary number of shock waves and contact surfaces. A new procedure, based on the coalescence of characteristics, is developed to detect the formation of shock waves. Results are presented to validate and demonstrate the versatility of the technique.
Terminology model discovery using natural language processing and visualization techniques.
Zhou, Li; Tao, Ying; Cimino, James J; Chen, Elizabeth S; Liu, Hongfang; Lussier, Yves A; Hripcsak, George; Friedman, Carol
2006-12-01
Medical terminologies are important for unambiguous encoding and exchange of clinical information. The traditional manual method of developing terminology models is time-consuming and limited in the number of phrases that a human developer can examine. In this paper, we present an automated method for developing medical terminology models based on natural language processing (NLP) and information visualization techniques. Surgical pathology reports were selected as the testing corpus for developing a pathology procedure terminology model. The use of a general NLP processor for the medical domain, MedLEE, provides an automated method for acquiring semantic structures from a free text corpus and sheds light on a new high-throughput method of medical terminology model development. The use of an information visualization technique supports the summarization and visualization of the large quantity of semantic structures generated from medical documents. We believe that a general method based on NLP and information visualization will facilitate the modeling of medical terminologies.
Develop real-time dosimetry concepts and instrumentation for long term missions
NASA Technical Reports Server (NTRS)
Braby, L. A.
1981-01-01
The development of a rugged portable dosimetry system, based on microdosimetry techniques, which will measure dose and evaluate dose equivalent in a mixed radiation field is described. Progress in the desired dosimetry system can be divided into three distinct areas: development of the radiation detector, and electron system are presented. The mathematical techniques required are investigated.
Reliable quantification techniques for carbon nanotubes (CNTs) are limited. In this study, a new procedure was developed for quantifying multi-walled carbon nanotubes (MWNTs) in earthworms (Eisenia fetida) based on freeze drying and microwave-induced heating. Specifically, earthw...
Bilek, Maciej; Namieśnik, Jacek
2016-01-01
For a long time, chromatographic techniques and techniques related to them have stimulated the development of new procedures in the field of pharmaceutical analysis. The newly developed methods, characterized by improved metrological parameters, allow for more accurate testing of, among others, the composition of raw materials, intermediates and final products. The chromatographic techniques also enable studies on waste generated in research laboratories and factories producing pharmaceuticals and parapharmaceuticals. Based on the review of reports published in Polish pharmaceutical journals, we assessed the impact of chromatographic techniques on the development of pharmaceutical analysis. The first chromatographic technique used in pharmaceutical analysis was a so-called capillary analysis. It was applied in the 1930s to control the identity of pharmaceutical formulations. In the 1940s and 1950s, the chromatographic techniques were mostly a subject of review publications, while their use in experimental work was rare. Paper chromatography and thin layer chromatography were introduced in the 1960s and 1970s, respectively. These new analytical tools have contributed to the intensive development of research in the field of phytochemistry and the analysis of herbal medicines. The development of colunm chromatography-based techniques, i.e., gas chromatography and high performance liquid chromatography took place in the end of 20th century. Both aforementioned techniques were widely applied in pharmaceutical analysis, for example, to assess the stability of drugs, test for impurities and degradation products as well as in pharmacokinetics studies. The first decade of 21" century was the time of new detection methods in gas and liquid chromatography. The information sources used to write this article were Polish pharmaceutical journals, both professional and scientific, originating from the interwar and post-war period, i.e., "Kronika Farmaceutyczna", "Farmacja Współczesna", "Wiadomości Farmaceutyczne", "Acta Poloniae Pharmaceutica", "Farmacja Polska", "Dissertationes Pharmaceuticae", "Annales UMCS sectio DDD Phamacia". The number of published works using various chromatography techniques was assessed based on the content description of individual issues of the journal "Acta Poloniae Pharmaceutica".
An adaptive technique to maximize lossless image data compression of satellite images
NASA Technical Reports Server (NTRS)
Stewart, Robert J.; Lure, Y. M. Fleming; Liou, C. S. Joe
1994-01-01
Data compression will pay an increasingly important role in the storage and transmission of image data within NASA science programs as the Earth Observing System comes into operation. It is important that the science data be preserved at the fidelity the instrument and the satellite communication systems were designed to produce. Lossless compression must therefore be applied, at least, to archive the processed instrument data. In this paper, we present an analysis of the performance of lossless compression techniques and develop an adaptive approach which applied image remapping, feature-based image segmentation to determine regions of similar entropy and high-order arithmetic coding to obtain significant improvements over the use of conventional compression techniques alone. Image remapping is used to transform the original image into a lower entropy state. Several techniques were tested on satellite images including differential pulse code modulation, bi-linear interpolation, and block-based linear predictive coding. The results of these experiments are discussed and trade-offs between computation requirements and entropy reductions are used to identify the optimum approach for a variety of satellite images. Further entropy reduction can be achieved by segmenting the image based on local entropy properties then applying a coding technique which maximizes compression for the region. Experimental results are presented showing the effect of different coding techniques for regions of different entropy. A rule-base is developed through which the technique giving the best compression is selected. The paper concludes that maximum compression can be achieved cost effectively and at acceptable performance rates with a combination of techniques which are selected based on image contextual information.
Single-molecule detection: applications to ultrasensitive biochemical analysis
NASA Astrophysics Data System (ADS)
Castro, Alonso; Shera, E. Brooks
1995-06-01
Recent developments in laser-based detection of fluorescent molecules have made possible the implementation of very sensitive techniques for biochemical analysis. We present and discuss our experiments on the applications of our recently developed technique of single-molecule detection to the analysis of molecules of biological interest. These newly developed methods are capable of detecting and identifying biomolecules at the single-molecule level of sensitivity. In one case, identification is based on measuring fluorescence brightness from single molecules. In another, molecules are classified by determining their electrophoretic velocities.
Note: development of high speed confocal 3D profilometer.
Ang, Kar Tien; Fang, Zhong Ping; Tay, Arthur
2014-11-01
A high-speed confocal 3D profilometer based on the chromatic confocal technology and spinning Nipkow disk technique has been developed and tested. It can measure a whole surface topography by taking only one image that requires less than 0.3 s. Surface height information is retrieved based on the ratios of red, green, and blue color information. A new vector projection technique has developed to enhance the vertical resolution of the measurement. The measurement accuracy of the prototype system has been verified via different test samples.
Berton, Paula; Lana, Nerina B; Ríos, Juan M; García-Reyes, Juan F; Altamirano, Jorgelina C
2016-01-28
Green chemistry principles for developing methodologies have gained attention in analytical chemistry in recent decades. A growing number of analytical techniques have been proposed for determination of organic persistent pollutants in environmental and biological samples. In this light, the current review aims to present state-of-the-art sample preparation approaches based on green analytical principles proposed for the determination of polybrominated diphenyl ethers (PBDEs) and metabolites (OH-PBDEs and MeO-PBDEs) in environmental and biological samples. Approaches to lower the solvent consumption and accelerate the extraction, such as pressurized liquid extraction, microwave-assisted extraction, and ultrasound-assisted extraction, are discussed in this review. Special attention is paid to miniaturized sample preparation methodologies and strategies proposed to reduce organic solvent consumption. Additionally, extraction techniques based on alternative solvents (surfactants, supercritical fluids, or ionic liquids) are also commented in this work, even though these are scarcely used for determination of PBDEs. In addition to liquid-based extraction techniques, solid-based analytical techniques are also addressed. The development of greener, faster and simpler sample preparation approaches has increased in recent years (2003-2013). Among green extraction techniques, those based on the liquid phase predominate over those based on the solid phase (71% vs. 29%, respectively). For solid samples, solvent assisted extraction techniques are preferred for leaching of PBDEs, and liquid phase microextraction techniques are mostly used for liquid samples. Likewise, green characteristics of the instrumental analysis used after the extraction and clean-up steps are briefly discussed. Copyright © 2015 Elsevier B.V. All rights reserved.
Kelay, Tanika; Chan, Kah Leong; Ako, Emmanuel; Yasin, Mohammad; Costopoulos, Charis; Gold, Matthew; Kneebone, Roger K; Malik, Iqbal S; Bello, Fernando
2017-01-01
Distributed Simulation is the concept of portable, high-fidelity immersive simulation. Here, it is used for the development of a simulation-based training programme for cardiovascular specialities. We present an evidence base for how accessible, portable and self-contained simulated environments can be effectively utilised for the modelling, development and testing of a complex training framework and assessment methodology. Iterative user feedback through mixed-methods evaluation techniques resulted in the implementation of the training programme. Four phases were involved in the development of our immersive simulation-based training programme: ( 1) initial conceptual stage for mapping structural criteria and parameters of the simulation training framework and scenario development ( n = 16), (2) training facility design using Distributed Simulation , (3) test cases with clinicians ( n = 8) and collaborative design, where evaluation and user feedback involved a mixed-methods approach featuring (a) quantitative surveys to evaluate the realism and perceived educational relevance of the simulation format and framework for training and (b) qualitative semi-structured interviews to capture detailed feedback including changes and scope for development. Refinements were made iteratively to the simulation framework based on user feedback, resulting in (4) transition towards implementation of the simulation training framework, involving consistent quantitative evaluation techniques for clinicians ( n = 62). For comparative purposes, clinicians' initial quantitative mean evaluation scores for realism of the simulation training framework, realism of the training facility and relevance for training ( n = 8) are presented longitudinally, alongside feedback throughout the development stages from concept to delivery, including the implementation stage ( n = 62). Initially, mean evaluation scores fluctuated from low to average, rising incrementally. This corresponded with the qualitative component, which augmented the quantitative findings; trainees' user feedback was used to perform iterative refinements to the simulation design and components (collaborative design), resulting in higher mean evaluation scores leading up to the implementation phase. Through application of innovative Distributed Simulation techniques, collaborative design, and consistent evaluation techniques from conceptual, development, and implementation stages, fully immersive simulation techniques for cardiovascular specialities are achievable and have the potential to be implemented more broadly.
NASA Astrophysics Data System (ADS)
Saetchnikov, Vladimir A.; Tcherniavskaia, Elina A.; Saetchnikov, Anton V.; Schweiger, Gustav; Ostendorf, Andreas
2014-05-01
Experimental data on detection and identification of variety of biochemical agents, such as proteins, microelements, antibiotic of different generation etc. in both single and multi component solutions under varied in wide range concentration analyzed on the light scattering parameters of whispering gallery mode optical resonance based sensor are represented. Multiplexing on parameters and components has been realized using developed fluidic sensor cell with fixed in adhesive layer dielectric microspheres and data processing. Biochemical component identification has been performed by developed network analysis techniques. Developed approach is demonstrated to be applicable both for single agent and for multi component biochemical analysis. Novel technique based on optical resonance on microring structures, plasmon resonance and identification tools has been developed. To improve a sensitivity of microring structures microspheres fixed by adhesive had been treated previously by gold nanoparticle solution. Another technique used thin film gold layers deposited on the substrate below adhesive. Both biomolecule and nanoparticle injections caused considerable changes of optical resonance spectra. Plasmonic gold layers under optimized thickness also improve parameters of optical resonance spectra. Biochemical component identification has been also performed by developed network analysis techniques both for single and for multi component solution. So advantages of plasmon enhancing optical microcavity resonance with multiparameter identification tools is used for development of a new platform for ultra sensitive label-free biomedical sensor.
Vision-based obstacle recognition system for automated lawn mower robot development
NASA Astrophysics Data System (ADS)
Mohd Zin, Zalhan; Ibrahim, Ratnawati
2011-06-01
Digital image processing techniques (DIP) have been widely used in various types of application recently. Classification and recognition of a specific object using vision system require some challenging tasks in the field of image processing and artificial intelligence. The ability and efficiency of vision system to capture and process the images is very important for any intelligent system such as autonomous robot. This paper gives attention to the development of a vision system that could contribute to the development of an automated vision based lawn mower robot. The works involve on the implementation of DIP techniques to detect and recognize three different types of obstacles that usually exist on a football field. The focus was given on the study on different types and sizes of obstacles, the development of vision based obstacle recognition system and the evaluation of the system's performance. Image processing techniques such as image filtering, segmentation, enhancement and edge detection have been applied in the system. The results have shown that the developed system is able to detect and recognize various types of obstacles on a football field with recognition rate of more 80%.
Prediction of drug synergy in cancer using ensemble-based machine learning techniques
NASA Astrophysics Data System (ADS)
Singh, Harpreet; Rana, Prashant Singh; Singh, Urvinder
2018-04-01
Drug synergy prediction plays a significant role in the medical field for inhibiting specific cancer agents. It can be developed as a pre-processing tool for therapeutic successes. Examination of different drug-drug interaction can be done by drug synergy score. It needs efficient regression-based machine learning approaches to minimize the prediction errors. Numerous machine learning techniques such as neural networks, support vector machines, random forests, LASSO, Elastic Nets, etc., have been used in the past to realize requirement as mentioned above. However, these techniques individually do not provide significant accuracy in drug synergy score. Therefore, the primary objective of this paper is to design a neuro-fuzzy-based ensembling approach. To achieve this, nine well-known machine learning techniques have been implemented by considering the drug synergy data. Based on the accuracy of each model, four techniques with high accuracy are selected to develop ensemble-based machine learning model. These models are Random forest, Fuzzy Rules Using Genetic Cooperative-Competitive Learning method (GFS.GCCL), Adaptive-Network-Based Fuzzy Inference System (ANFIS) and Dynamic Evolving Neural-Fuzzy Inference System method (DENFIS). Ensembling is achieved by evaluating the biased weighted aggregation (i.e. adding more weights to the model with a higher prediction score) of predicted data by selected models. The proposed and existing machine learning techniques have been evaluated on drug synergy score data. The comparative analysis reveals that the proposed method outperforms others in terms of accuracy, root mean square error and coefficient of correlation.
Comparison of global cloud liquid water path derived from microwave measurements with CERES-MODIS
NASA Astrophysics Data System (ADS)
Yi, Y.; Minnis, P.; Huang, J.; Lin, B.; Ayers, K.; Sun-Mack, S.; Fan, A.
Cloud liquid water path LWP is a crucial parameter for climate studies due to the link that it provides between the atmospheric hydrological and radiative budgets Satellite-based visible infrared techniques such as the Visible Infrared Solar Split-Window Technique VISST can retrieve LWP for water clouds assumes single-layer over a variety of surfaces If the water clouds are overlapped by ice clouds the LWP of the underlying clouds can not be retrieved by such techniques However microwave techniques may be used to retrieve the LWP underneath ice clouds due to the microwave s insensitivity to cloud ice particles LWP is typically retrieved from satellite-observed microwave radiances only over ocean due to variations of land surface temperature and emissivity Recently Deeter and Vivekanandan 2006 developed a new technique for retrieving LWP over land In order to overcome the sensitivity to land surface temperature and emissivity their technique is based on a parameterization of microwave polarization-difference signals In this study a similar regression-based technique for retrieving LWP over land and ocean using Advanced Microwave Scanning Radiometer - EOS AMSR-E measurements is developed Furthermore the microwave surface emissivities are also derived using clear-sky fields of view based on the Clouds and Earth s Radiant Energy System Moderate-resolution Imaging Spectroradiometer CERES-MODIS cloud mask These emissivities are used in an alternate form of the technique The results are evaluated using independent measurements such
Developing a hybrid dictionary-based bio-entity recognition technique.
Song, Min; Yu, Hwanjo; Han, Wook-Shin
2015-01-01
Bio-entity extraction is a pivotal component for information extraction from biomedical literature. The dictionary-based bio-entity extraction is the first generation of Named Entity Recognition (NER) techniques. This paper presents a hybrid dictionary-based bio-entity extraction technique. The approach expands the bio-entity dictionary by combining different data sources and improves the recall rate through the shortest path edit distance algorithm. In addition, the proposed technique adopts text mining techniques in the merging stage of similar entities such as Part of Speech (POS) expansion, stemming, and the exploitation of the contextual cues to further improve the performance. The experimental results show that the proposed technique achieves the best or at least equivalent performance among compared techniques, GENIA, MESH, UMLS, and combinations of these three resources in F-measure. The results imply that the performance of dictionary-based extraction techniques is largely influenced by information resources used to build the dictionary. In addition, the edit distance algorithm shows steady performance with three different dictionaries in precision whereas the context-only technique achieves a high-end performance with three difference dictionaries in recall.
Developing a hybrid dictionary-based bio-entity recognition technique
2015-01-01
Background Bio-entity extraction is a pivotal component for information extraction from biomedical literature. The dictionary-based bio-entity extraction is the first generation of Named Entity Recognition (NER) techniques. Methods This paper presents a hybrid dictionary-based bio-entity extraction technique. The approach expands the bio-entity dictionary by combining different data sources and improves the recall rate through the shortest path edit distance algorithm. In addition, the proposed technique adopts text mining techniques in the merging stage of similar entities such as Part of Speech (POS) expansion, stemming, and the exploitation of the contextual cues to further improve the performance. Results The experimental results show that the proposed technique achieves the best or at least equivalent performance among compared techniques, GENIA, MESH, UMLS, and combinations of these three resources in F-measure. Conclusions The results imply that the performance of dictionary-based extraction techniques is largely influenced by information resources used to build the dictionary. In addition, the edit distance algorithm shows steady performance with three different dictionaries in precision whereas the context-only technique achieves a high-end performance with three difference dictionaries in recall. PMID:26043907
Laser applications and system considerations in ocular imaging
Elsner, Ann E.; Muller, Matthew S.
2009-01-01
We review laser applications for primarily in vivo ocular imaging techniques, describing their constraints based on biological tissue properties, safety, and the performance of the imaging system. We discuss the need for cost effective sources with practical wavelength tuning capabilities for spectral studies. Techniques to probe the pathological changes of layers beneath the highly scattering retina and diagnose the onset of various eye diseases are described. The recent development of several optical coherence tomography based systems for functional ocular imaging is reviewed, as well as linear and nonlinear ocular imaging techniques performed with ultrafast lasers, emphasizing recent source developments and methods to enhance imaging contrast. PMID:21052482
Development and evaluation of vision rehabilitation devices.
Luo, Gang; Peli, Eli
2011-01-01
We have developed a range of vision rehabilitation devices and techniques for people with impaired vision due to either central vision loss or severely restricted peripheral visual field. We have conducted evaluation studies with patients to test the utilities of these techniques in an effort to document their advantages as well as their limitations. Here we describe our work on a visual field expander based on a head mounted display (HMD) for tunnel vision, a vision enhancement device for central vision loss, and a frequency domain JPEG/MPEG based image enhancement technique. All the evaluation studies included visual search paradigms that are suitable for conducting indoor controllable experiments.
Transient Faults in Computer Systems
NASA Technical Reports Server (NTRS)
Masson, Gerald M.
1993-01-01
A powerful technique particularly appropriate for the detection of errors caused by transient faults in computer systems was developed. The technique can be implemented in either software or hardware; the research conducted thus far primarily considered software implementations. The error detection technique developed has the distinct advantage of having provably complete coverage of all errors caused by transient faults that affect the output produced by the execution of a program. In other words, the technique does not have to be tuned to a particular error model to enhance error coverage. Also, the correctness of the technique can be formally verified. The technique uses time and software redundancy. The foundation for an effective, low-overhead, software-based certification trail approach to real-time error detection resulting from transient fault phenomena was developed.
Modeling and prototyping of biometric systems using dataflow programming
NASA Astrophysics Data System (ADS)
Minakova, N.; Petrov, I.
2018-01-01
The development of biometric systems is one of the labor-intensive processes. Therefore, the creation and analysis of approaches and techniques is an urgent task at present. This article presents a technique of modeling and prototyping biometric systems based on dataflow programming. The technique includes three main stages: the development of functional blocks, the creation of a dataflow graph and the generation of a prototype. A specially developed software modeling environment that implements this technique is described. As an example of the use of this technique, an example of the implementation of the iris localization subsystem is demonstrated. A variant of modification of dataflow programming is suggested to solve the problem related to the undefined order of block activation. The main advantage of the presented technique is the ability to visually display and design the model of the biometric system, the rapid creation of a working prototype and the reuse of the previously developed functional blocks.
Signal analysis techniques for incipient failure detection in turbomachinery
NASA Technical Reports Server (NTRS)
Coffin, T.
1985-01-01
Signal analysis techniques for the detection and classification of incipient mechanical failures in turbomachinery were developed, implemented and evaluated. Signal analysis techniques available to describe dynamic measurement characteristics are reviewed. Time domain and spectral methods are described, and statistical classification in terms of moments is discussed. Several of these waveform analysis techniques were implemented on a computer and applied to dynamic signals. A laboratory evaluation of the methods with respect to signal detection capability is described. Plans for further technique evaluation and data base development to characterize turbopump incipient failure modes from Space Shuttle main engine (SSME) hot firing measurements are outlined.
In this study, a new analytical technique was developed for the identification and quantification of multi-functional compounds containing simultaneously at least one hydroxyl or one carboxylic group, or both. This technique is based on derivatizing first the carboxylic group(s) ...
Developing Coping Skills in Early Childhood: Theory and Techniques.
ERIC Educational Resources Information Center
Forquer, Sandra L.
1982-01-01
Discusses techniques that child care workers can utilize to foster the development of coping skills in young children. Emphasizes the difference between psychological immunity to stress based on problem-solving abilities and pseudo-immunity created by overprotectiveness. Holds that challenges build children's competence and self-esteem.…
ERIC Educational Resources Information Center
Henry, Gary T.; And Others
1992-01-01
A statistical technique is presented for developing performance standards based on benchmark groups. The benchmark groups are selected using a multivariate technique that relies on a squared Euclidean distance method. For each observation unit (a school district in the example), a unique comparison group is selected. (SLD)
The GenTechnique Project: Developing an Open Environment for Learning Molecular Genetics.
ERIC Educational Resources Information Center
Calza, R. E.; Meade, J. T.
1998-01-01
The GenTechnique project at Washington State University uses a networked learning environment for molecular genetics learning. The project is developing courseware featuring animation, hyper-link controls, and interactive self-assessment exercises focusing on fundamental concepts. The first pilot course featured a Web-based module on DNA…
Knowledge-based system verification and validation
NASA Technical Reports Server (NTRS)
Johnson, Sally C.
1990-01-01
The objective of this task is to develop and evaluate a methodology for verification and validation (V&V) of knowledge-based systems (KBS) for space station applications with high reliability requirements. The approach consists of three interrelated tasks. The first task is to evaluate the effectiveness of various validation methods for space station applications. The second task is to recommend requirements for KBS V&V for Space Station Freedom (SSF). The third task is to recommend modifications to the SSF to support the development of KBS using effectiveness software engineering and validation techniques. To accomplish the first task, three complementary techniques will be evaluated: (1) Sensitivity Analysis (Worchester Polytechnic Institute); (2) Formal Verification of Safety Properties (SRI International); and (3) Consistency and Completeness Checking (Lockheed AI Center). During FY89 and FY90, each contractor will independently demonstrate the user of his technique on the fault detection, isolation, and reconfiguration (FDIR) KBS or the manned maneuvering unit (MMU), a rule-based system implemented in LISP. During FY91, the application of each of the techniques to other knowledge representations and KBS architectures will be addressed. After evaluation of the results of the first task and examination of Space Station Freedom V&V requirements for conventional software, a comprehensive KBS V&V methodology will be developed and documented. Development of highly reliable KBS's cannot be accomplished without effective software engineering methods. Using the results of current in-house research to develop and assess software engineering methods for KBS's as well as assessment of techniques being developed elsewhere, an effective software engineering methodology for space station KBS's will be developed, and modification of the SSF to support these tools and methods will be addressed.
NASA Technical Reports Server (NTRS)
1994-01-01
This manual presents a series of recommended techniques that can increase overall operational effectiveness of both flight and ground based NASA systems. It provides a set of tools that minimizes risk associated with: (1) restoring failed functions (both ground and flight based); (2) conducting complex and highly visible maintenance operations; and (3) sustaining a technical capability to support the NASA mission using aging equipment or facilities. It considers (1) program management - key elements of an effective maintainability effort; (2) design and development - techniques that have benefited previous programs; (3) analysis and test - quantitative and qualitative analysis processes and testing techniques; and (4) operations and operational design techniques that address NASA field experience. This document is a valuable resource for continuous improvement ideas in executing the systems development process in accordance with the NASA 'better, faster, smaller, and cheaper' goal without compromising safety.
Finite volume model for two-dimensional shallow environmental flow
Simoes, F.J.M.
2011-01-01
This paper presents the development of a two-dimensional, depth integrated, unsteady, free-surface model based on the shallow water equations. The development was motivated by the desire of balancing computational efficiency and accuracy by selective and conjunctive use of different numerical techniques. The base framework of the discrete model uses Godunov methods on unstructured triangular grids, but the solution technique emphasizes the use of a high-resolution Riemann solver where needed, switching to a simpler and computationally more efficient upwind finite volume technique in the smooth regions of the flow. Explicit time marching is accomplished with strong stability preserving Runge-Kutta methods, with additional acceleration techniques for steady-state computations. A simplified mass-preserving algorithm is used to deal with wet/dry fronts. Application of the model is made to several benchmark cases that show the interplay of the diverse solution techniques.
Optical fiber sensors measurement system and special fibers improvement
NASA Astrophysics Data System (ADS)
Jelinek, Michal; Hrabina, Jan; Hola, Miroslava; Hucl, Vaclav; Cizek, Martin; Rerucha, Simon; Lazar, Josef; Mikel, Bretislav
2017-06-01
We present method for the improvement of the measurement accuracy in the optical frequency spectra measurements based on tunable optical filters. The optical filter was used during the design and realization of the measurement system for the inspection of the fiber Bragg gratings. The system incorporates a reference block for the compensation of environmental influences, an interferometric verification subsystem and a PC - based control software implemented in LabView. The preliminary experimental verification of the measurement principle and the measurement system functionality were carried out on a testing rig with a specially prepared concrete console in the UJV Řež. The presented system is the laboratory version of the special nuclear power plant containment shape deformation measurement system which was installed in the power plant Temelin during last year. On the base of this research we started with preparation other optical fiber sensors to nuclear power plants measurement. These sensors will be based on the microstructured and polarization maintaining optical fibers. We started with development of new methods and techniques of the splicing and shaping optical fibers. We are able to made optical tapers from ultra-short called adiabatic with length around 400 um up to long tapers with length up to 6 millimeters. We developed new techniques of splicing standard Single Mode (SM) and Multimode (MM) optical fibers and splicing of optical fibers with different diameters in the wavelength range from 532 to 1550 nm. Together with development these techniques we prepared other techniques to splicing and shaping special optical fibers like as Polarization-Maintaining (PM) or hollow core Photonic Crystal Fiber (PCF) and theirs cross splicing methods with focus to minimalize backreflection and attenuation. The splicing special optical fibers especially PCF fibers with standard telecommunication and other SM fibers can be done by our developed techniques. Adjustment of the splicing process has to be prepared for any new optical fibers and new fibers combinations. The splicing of the same types of fibers from different manufacturers can be adjusted by several tested changes in the splicing process. We are able to splice PCF with standard telecommunication fiber with attenuation up to 2 dB. The method is also presented. Development of these new techniques and methods of the optical fibers splicing are made with respect to using these fibers to another research and development in the field of optical fibers sensors, laser frequency stabilization and laser interferometry based on optical fibers. Especially for the field of laser frequency stabilization we developed and present new techniques to closing microstructured fibers with gases inside.
Wavelet Transform Based Filter to Remove the Notches from Signal Under Harmonic Polluted Environment
NASA Astrophysics Data System (ADS)
Das, Sukanta; Ranjan, Vikash
2017-12-01
The work proposes to annihilate the notches present in the synchronizing signal required for converter operation appearing due to switching of semiconductor devices connected to the system in the harmonic polluted environment. The disturbances in the signal are suppressed by wavelet based novel filtering technique. In the proposed technique, the notches in the signal are determined and eliminated by the wavelet based multi-rate filter using `Daubechies4' (db4) as mother wavelet. The computational complexity of the adapted technique is very less as compared to any other conventional notch filtering techniques. The proposed technique is developed in MATLAB/Simulink and finally validated with dSPACE-1103 interface. The recovered signal, thus obtained, is almost free of the notches.
Hill, Ryan T
2015-01-01
The unique optical properties of plasmon resonant nanostructures enable exploration of nanoscale environments using relatively simple optical characterization techniques. For this reason, the field of plasmonics continues to garner the attention of the biosensing community. Biosensors based on propagating surface plasmon resonances (SPRs) in films are the most well-recognized plasmonic biosensors, but there is great potential for the new, developing technologies to surpass the robustness and popularity of film-based SPR sensing. This review surveys the current plasmonic biosensor landscape with emphasis on the basic operating principles of each plasmonic sensing technique and the practical considerations when developing a sensing platform with the various techniques. The 'gold standard' film SPR technique is reviewed briefly, but special emphasis is devoted to the up-and-coming localized surface plasmon resonance and plasmonically coupled sensor technology. © 2014 Wiley Periodicals, Inc.
A study on laser-based ultrasonic technique by the use of guided wave tomographic imaging
DOE Office of Scientific and Technical Information (OSTI.GOV)
Park, Junpil, E-mail: jpp@pusan.ac.kr; Lim, Juyoung, E-mail: jpp@pusan.ac.kr; Cho, Younho
2015-03-31
Guided wave tests are impractical for investigating specimens with limited accessibility and coarse surfaces or geometrically complicated features. A non-contact setup with a laser ultrasonic transmitter and receiver is the classic attractive for guided wave inspection. The present work was done to develop a non-contact guided-wave tomography technique by laser ultrasonic technique in a plate-like structure. A method for Lam wave generation and detection in an aluminum plate with a pulse laser ultrasonic transmitter and a Michelson interferometer receiver has been developed. In the images obtained by laser scanning, the defect shape and area showed good agreement with the actualmore » defect. The proposed approach can be used as a non-contact-based online inspection and monitoring technique.« less
A New Femtosecond Laser-Based Three-Dimensional Tomography Technique
NASA Astrophysics Data System (ADS)
Echlin, McLean P.
2011-12-01
Tomographic imaging has dramatically changed science, most notably in the fields of medicine and biology, by producing 3D views of structures which are too complex to understand in any other way. Current tomographic techniques require extensive time both for post-processing and data collection. Femtosecond laser based tomographic techniques have been developed in both standard atmosphere (femtosecond laser-based serial sectioning technique - FSLSS) and in vacuum (Tri-Beam System) for the fast collection (10 5mum3/s) of mm3 sized 3D datasets. Both techniques use femtosecond laser pulses to selectively remove layer-by-layer areas of material with low collateral damage and a negligible heat affected zone. To the authors knowledge, femtosecond lasers have never been used to serial section and these techniques have been entirely and uniquely developed by the author and his collaborators at the University of Michigan and University of California Santa Barbara. The FSLSS was applied to measure the 3D distribution of TiN particles in a 4330 steel. Single pulse ablation morphologies and rates were measured and collected from literature. Simultaneous two-phase ablation of TiN and steel matrix was shown to occur at fluences of 0.9-2 J/cm2. Laser scanning protocols were developed minimizing surface roughness to 0.1-0.4 mum for laser-based sectioning. The FSLSS technique was used to section and 3D reconstruct titanium nitride (TiN) containing 4330 steel. Statistical analysis of 3D TiN particle sizes, distribution parameters, and particle density were measured. A methodology was developed to use the 3D datasets to produce statistical volume elements (SVEs) for toughness modeling. Six FSLSS TiN datasets were sub-sampled into 48 SVEs for statistical analysis and toughness modeling using the Rice-Tracey and Garrison-Moody models. A two-parameter Weibull analysis was performed and variability in the toughness data agreed well with Ruggieri et al. bulk toughness measurements. The Tri-Beam system combines the benefits of laser based material removal (speed, low-damage, automated) with detectors that collect chemical, structural, and topological information. Multi-modal sectioning information was collected after many laser scanning passes demonstrating the capability of the Tri-Beam system.
Potential for Imaging Engineered Tissues with X-Ray Phase Contrast
Appel, Alyssa; Anastasio, Mark A.
2011-01-01
As the field of tissue engineering advances, it is crucial to develop imaging methods capable of providing detailed three-dimensional information on tissue structure. X-ray imaging techniques based on phase-contrast (PC) have great potential for a number of biomedical applications due to their ability to provide information about soft tissue structure without exogenous contrast agents. X-ray PC techniques retain the excellent spatial resolution, tissue penetration, and calcified tissue contrast of conventional X-ray techniques while providing drastically improved imaging of soft tissue and biomaterials. This suggests that X-ray PC techniques are very promising for evaluation of engineered tissues. In this review, four different implementations of X-ray PC imaging are described and applications to tissues of relevance to tissue engineering reviewed. In addition, recent applications of X-ray PC to the evaluation of biomaterial scaffolds and engineered tissues are presented and areas for further development and application of these techniques are discussed. Imaging techniques based on X-ray PC have significant potential for improving our ability to image and characterize engineered tissues, and their continued development and optimization could have significant impact on the field of tissue engineering. PMID:21682604
Knowledge-based geographic information systems (KBGIS): New analytic and data management tools
Albert, T.M.
1988-01-01
In its simplest form, a geographic information system (GIS) may be viewed as a data base management system in which most of the data are spatially indexed, and upon which sets of procedures operate to answer queries about spatial entities represented in the data base. Utilization of artificial intelligence (AI) techniques can enhance greatly the capabilities of a GIS, particularly in handling very large, diverse data bases involved in the earth sciences. A KBGIS has been developed by the U.S. Geological Survey which incorporates AI techniques such as learning, expert systems, new data representation, and more. The system, which will be developed further and applied, is a prototype of the next generation of GIS's, an intelligent GIS, as well as an example of a general-purpose intelligent data handling system. The paper provides a description of KBGIS and its application, as well as the AI techniques involved. ?? 1988 International Association for Mathematical Geology.
Statistical approach for selection of biologically informative genes.
Das, Samarendra; Rai, Anil; Mishra, D C; Rai, Shesh N
2018-05-20
Selection of informative genes from high dimensional gene expression data has emerged as an important research area in genomics. Many gene selection techniques have been proposed so far are either based on relevancy or redundancy measure. Further, the performance of these techniques has been adjudged through post selection classification accuracy computed through a classifier using the selected genes. This performance metric may be statistically sound but may not be biologically relevant. A statistical approach, i.e. Boot-MRMR, was proposed based on a composite measure of maximum relevance and minimum redundancy, which is both statistically sound and biologically relevant for informative gene selection. For comparative evaluation of the proposed approach, we developed two biological sufficient criteria, i.e. Gene Set Enrichment with QTL (GSEQ) and biological similarity score based on Gene Ontology (GO). Further, a systematic and rigorous evaluation of the proposed technique with 12 existing gene selection techniques was carried out using five gene expression datasets. This evaluation was based on a broad spectrum of statistically sound (e.g. subject classification) and biological relevant (based on QTL and GO) criteria under a multiple criteria decision-making framework. The performance analysis showed that the proposed technique selects informative genes which are more biologically relevant. The proposed technique is also found to be quite competitive with the existing techniques with respect to subject classification and computational time. Our results also showed that under the multiple criteria decision-making setup, the proposed technique is best for informative gene selection over the available alternatives. Based on the proposed approach, an R Package, i.e. BootMRMR has been developed and available at https://cran.r-project.org/web/packages/BootMRMR. This study will provide a practical guide to select statistical techniques for selecting informative genes from high dimensional expression data for breeding and system biology studies. Published by Elsevier B.V.
Moisture determination in composite materials using positron lifetime techniques
NASA Technical Reports Server (NTRS)
Singh, J. J.; Holt, W. R.; Mock, W., Jr.
1980-01-01
A technique was developed which has the potential of providing information on the moisture content as well as its depth in the specimen. This technique was based on the dependence of positron lifetime on the moisture content of the composite specimen. The positron lifetime technique of moisture determination and the results of the initial studies are described.
The Effects of Practice-Based Training on Graduate Teaching Assistants' Classroom Practices.
Becker, Erin A; Easlon, Erin J; Potter, Sarah C; Guzman-Alvarez, Alberto; Spear, Jensen M; Facciotti, Marc T; Igo, Michele M; Singer, Mitchell; Pagliarulo, Christopher
2017-01-01
Evidence-based teaching is a highly complex skill, requiring repeated cycles of deliberate practice and feedback to master. Despite existing well-characterized frameworks for practice-based training in K-12 teacher education, the major principles of these frameworks have not yet been transferred to instructor development in higher educational contexts, including training of graduate teaching assistants (GTAs). We sought to determine whether a practice-based training program could help GTAs learn and use evidence-based teaching methods in their classrooms. We implemented a weekly training program for introductory biology GTAs that included structured drills of techniques selected to enhance student practice, logic development, and accountability and reduce apprehension. These elements were selected based on their previous characterization as dimensions of active learning. GTAs received regular performance feedback based on classroom observations. To quantify use of target techniques and levels of student participation, we collected and coded 160 h of video footage. We investigated the relationship between frequency of GTA implementation of target techniques and student exam scores; however, we observed no significant relationship. Although GTAs adopted and used many of the target techniques with high frequency, techniques that enforced student participation were not stably adopted, and their use was unresponsive to formal feedback. We also found that techniques discussed in training, but not practiced, were not used at quantifiable frequencies, further supporting the importance of practice-based training for influencing instructional practices. © 2017 E. A. Becker et al. CBE—Life Sciences Education © 2017 The American Society for Cell Biology. This article is distributed by The American Society for Cell Biology under license from the author(s). It is available to the public under an Attribution–Noncommercial–Share Alike 3.0 Unported Creative Commons License (http://creativecommons.org/licenses/by-nc-sa/3.0).
Test techniques for evaluating flight displays
NASA Technical Reports Server (NTRS)
Haworth, Loran A.; Newman, Richard L.
1993-01-01
The rapid development of graphics technology allows for greater flexibility in aircraft displays, but display evaluation techniques have not kept pace. Historically, display evaluation has been based on subjective opinion and not on the actual aircraft/pilot performance. Existing electronic display specifications and evaluation techniques are reviewed. A display rating technique analogous to handling qualities ratings was developed and is recommended for future evaluations. The choice of evaluation pilots is also discussed and the use of a limited number of trained evaluators is recommended over the use of a large number of operational pilots.
Development and evaluation of an automatic labeling technique for spring small grains
NASA Technical Reports Server (NTRS)
Crist, E. P.; Malila, W. A. (Principal Investigator)
1981-01-01
A labeling technique is described which seeks to associate a sampling entity with a particular crop or crop group based on similarity of growing season and temporal-spectral patterns of development. Human analyst provide contextual information, after which labeling decisions are made automatically. Results of a test of the technique on a large, multi-year data set are reported. Grain labeling accuracies are similar to those achieved by human analysis techniques, while non-grain accuracies are lower. Recommendations for improvments and implications of the test results are discussed.
NASA Astrophysics Data System (ADS)
Chowdhury, D. P.; Pal, Sujit; Parthasarathy, R.; Mathur, P. K.; Kohli, A. K.; Limaye, P. K.
1998-09-01
Thin layer activation (TLA) technique has been developed in Zr based alloy materials, e.g., zircaloy II, using 40 MeV α-particles from Variable Energy Cyclotron Centre at Calcutta. A brief description of the methodology of TLA technique is presented to determine the surface wear. The sensitivity of the measurement of surface wear in zircaloy material is found to be 0.22±0.05 μm. The surface wear is determined by TLA technique in zircaloy material which is used in pressurised heavy water reactor and the values have been compared with that obtained by conventional technique for the analytical validation of the TLA technique.
An overview of the Office of Space Flight satellite servicing program plan
NASA Technical Reports Server (NTRS)
Levin, George M.; Erwin, Harry O., Jr.
1987-01-01
A comprehensive program for the development of satellite servicing tools and techniques is being currently carried out by the Office of Space Flight. The program is based on a satellite servicing infrastructure formulated by analyzing satellite servicing requirements; the program is Shuttle-based and compatible with the Orbital Maneuvering Vehicle and Space Station. The content of the satellite servicing program is reviewed with reference to the tools, techniques, and procedures being developed for refueling (or consumables resupply), repairing, and retrieving.
MESA: An Interactive Modeling and Simulation Environment for Intelligent Systems Automation
NASA Technical Reports Server (NTRS)
Charest, Leonard
1994-01-01
This report describes MESA, a software environment for creating applications that automate NASA mission opterations. MESA enables intelligent automation by utilizing model-based reasoning techniques developed in the field of Artificial Intelligence. Model-based reasoning techniques are realized in Mesa through native support of causal modeling and discrete event simulation.
Molecular-Based Optical Measurement Techniques for Transition and Turbulence in High-Speed Flow
NASA Technical Reports Server (NTRS)
Bathel, Brett F.; Danehy, Paul M.; Cutler, Andrew D.
2013-01-01
High-speed laminar-to-turbulent transition and turbulence affect the control of flight vehicles, the heat transfer rate to a flight vehicle's surface, the material selected to protect such vehicles from high heating loads, the ultimate weight of a flight vehicle due to the presence of thermal protection systems, the efficiency of fuel-air mixing processes in high-speed combustion applications, etc. Gaining a fundamental understanding of the physical mechanisms involved in the transition process will lead to the development of predictive capabilities that can identify transition location and its impact on parameters like surface heating. Currently, there is no general theory that can completely describe the transition-to-turbulence process. However, transition research has led to the identification of the predominant pathways by which this process occurs. For a truly physics-based model of transition to be developed, the individual stages in the paths leading to the onset of fully turbulent flow must be well understood. This requires that each pathway be computationally modeled and experimentally characterized and validated. This may also lead to the discovery of new physical pathways. This document is intended to describe molecular based measurement techniques that have been developed, addressing the needs of the high-speed transition-to-turbulence and high-speed turbulence research fields. In particular, we focus on techniques that have either been used to study high speed transition and turbulence or techniques that show promise for studying these flows. This review is not exhaustive. In addition to the probe-based techniques described in the previous paragraph, several other classes of measurement techniques that are, or could be, used to study high speed transition and turbulence are excluded from this manuscript. For example, surface measurement techniques such as pressure and temperature paint, phosphor thermography, skin friction measurements and photogrammetry (for model attitude and deformation measurement) are excluded to limit the scope of this report. Other physical probes such as heat flux gauges, total temperature probes are also excluded. We further exclude measurement techniques that require particle seeding though particle based methods may still be useful in many high speed flow applications. This manuscript details some of the more widely used molecular-based measurement techniques for studying transition and turbulence: laser-induced fluorescence (LIF), Rayleigh and Raman Scattering and coherent anti-Stokes Raman scattering (CARS). These techniques are emphasized, in part, because of the prior experience of the authors. Additional molecular based techniques are described, albeit in less detail. Where possible, an effort is made to compare the relative advantages and disadvantages of the various measurement techniques, although these comparisons can be subjective views of the authors. Finally, the manuscript concludes by evaluating the different measurement techniques in view of the precision requirements described in this chapter. Additional requirements and considerations are discussed to assist with choosing an optical measurement technique for a given application.
The Pixon Method for Data Compression Image Classification, and Image Reconstruction
NASA Technical Reports Server (NTRS)
Puetter, Richard; Yahil, Amos
2002-01-01
As initially proposed, this program had three goals: (1) continue to develop the highly successful Pixon method for image reconstruction and support other scientist in implementing this technique for their applications; (2) develop image compression techniques based on the Pixon method; and (3) develop artificial intelligence algorithms for image classification based on the Pixon approach for simplifying neural networks. Subsequent to proposal review the scope of the program was greatly reduced and it was decided to investigate the ability of the Pixon method to provide superior restorations of images compressed with standard image compression schemes, specifically JPEG-compressed images.
Aircraft propeller induced structure-borne noise
NASA Technical Reports Server (NTRS)
Unruh, James F.
1989-01-01
A laboratory-based test apparatus employing components typical of aircraft construction was developed that would allow the study of structure-borne noise transmission due to propeller induced wake/vortex excitation of in-wake structural appendages. The test apparatus was employed to evaluate several aircraft installation effects (power plant placement, engine/nacelle mass loading, and wing/fuselage attachment methods) and several structural response modifications for structure-borne noise control (the use of wing blocking mass/fuel, wing damping treaments, and tuned mechanical dampers). Most important was the development of in-flight structure-borne noise transmission detection techniques using a combination of ground-based frequency response function testing and in-flight structural response measurement. Propeller wake/vortex excitation simulation techniques for improved ground-based testing were also developed to support the in-flight structure-borne noise transmission detection development.
Ghorpade, Uma; Suryawanshi, Mahesh; Shin, Seung Wook; Gurav, Kishor; Patil, Pramod; Pawar, Sambhaji; Hong, Chang Woo; Kim, Jin Hyeok; Kolekar, Sanjay
2014-10-07
With the earth's abundance of kesterite, recent progress in chalcogenide based Cu2ZnSn(Sx,Se1-x)4 (CZTSSe) thin films has drawn prime attention in thin film solar cells (TFSCs) research and development. This review is focused on the current developments in the synthesis of CZTS nanocrystals (NCs) using a hot injection (HI) technique and provides comprehensive discussions on the current status of CZTSSe TFSCs. This article begins with a description of the advantages of nanoparticulate based thin films, and then introduces the basics of this technique and the corresponding growth mechanism is also discussed. A brief overview further addresses a series of investigations on the developments in the HI based CZTSSe NCs using different solvents in terms of their high toxicity to environmentally benign materials. A variety of recipes and techniques for the NCs ink formulation and thereby the preparation of absorber layers using NC inks are outlined, respectively. The deposition of precursor thin films, post-deposition processes such as sulfurization or selenization treatments and the fabrication of CZTSSe NCs based solar cells and their performances are discussed. Finally, we discussed concluding remarks and the perspectives for further developments in the existing research on CZTSSe based nanoparticulate (NP) TFSCs towards future green technology.
Guidelines for using the Delphi Technique to develop habitat suitability index curves
Crance, Johnie H.
1987-01-01
Habitat Suitability Index (SI) curves are one method of presenting species habitat suitability criteria. The curves are often used with the Habitat Evaluation Procedures (HEP) and are necessary components of the Instream Flow Incremental Methodology (IFIM) (Armour et al. 1984). Bovee (1986) described three categories of SI curves or habitat suitability criteria based on the procedures and data used to develop the criteria. Category I curves are based on professional judgment, with 1ittle or no empirical data. Both Category II (utilization criteria) and Category III (preference criteria) curves have as their source data collected at locations where target species are observed or collected. Having Category II and Category III curves for all species of concern would be ideal. In reality, no SI curves are available for many species, and SI curves that require intensive field sampling often cannot be developed under prevailing constraints on time and costs. One alternative under these circumstances is the development and interim use of SI curves based on expert opinion. The Delphi technique (Pill 1971; Delbecq et al. 1975; Linstone and Turoff 1975) is one method used for combining the knowledge and opinions of a group of experts. The purpose of this report is to describe how the Delphi technique may be used to develop expert-opinion-based SI curves.
Focus Group Meets Nominal Group Technique: An Effective Combination for Student Evaluation?
ERIC Educational Resources Information Center
Varga-Atkins, Tünde; McIsaac, Jaye; Willis, Ian
2017-01-01
In Higher Education Focus Groups and Nominal Group Technique are two well-established methods for obtaining student feedback about their learning experience. These methods are regularly used for the enhancement and quality assurance. Based on small-scale research of educational developers' practice in curriculum development, this study presents…
Qu, Yongzhi; He, David; Yoon, Jae; Van Hecke, Brandon; Bechhoefer, Eric; Zhu, Junda
2014-01-01
In recent years, acoustic emission (AE) sensors and AE-based techniques have been developed and tested for gearbox fault diagnosis. In general, AE-based techniques require much higher sampling rates than vibration analysis-based techniques for gearbox fault diagnosis. Therefore, it is questionable whether an AE-based technique would give a better or at least the same performance as the vibration analysis-based techniques using the same sampling rate. To answer the question, this paper presents a comparative study for gearbox tooth damage level diagnostics using AE and vibration measurements, the first known attempt to compare the gearbox fault diagnostic performance of AE- and vibration analysis-based approaches using the same sampling rate. Partial tooth cut faults are seeded in a gearbox test rig and experimentally tested in a laboratory. Results have shown that the AE-based approach has the potential to differentiate gear tooth damage levels in comparison with the vibration-based approach. While vibration signals are easily affected by mechanical resonance, the AE signals show more stable performance. PMID:24424467
Development and Evaluation of the Method with an Affective Interface for Promoting Employees' Morale
NASA Astrophysics Data System (ADS)
Fujino, Hidenori; Ishii, Hirotake; Shimoda, Hiroshi; Yoshikawa, Hidekazu
For the sustainable society, organization management not based on the mass production and mass consumption but having the flexibility to meet to various social needs precisely is required. For realizing such management, the emploees' work morale is required. Recently, however, the emploees' work morale is tend to decrease. Therefore, in this study, the authors developed the model of the method for promoting and keeping employees' work morale effectively and efficiently. Especially the authors thought “work morale” of “attitude to the work”. Based on this idea, it could be considered that the theory of the persuasion psychology and various persuasion techniques. Therefore, the model of the method applying the character agent was developed based on the forced compliance which is one of persuasion techniques based on the theory of the cognitive dissonance. By the evaluation experiment using human subjects, it was confirmed that developed method could improve workers' work morle effectively.
NASA Technical Reports Server (NTRS)
Colwell, R. N. (Principal Investigator); Hay, C. M.; Thomas, R. W.; Benson, A. S.
1976-01-01
The progress of research conducted in support of the Large Area Crop Inventory Experiment (LACIE) is documented. Specific tasks include (1) evaluation of the static stratification procedure and modification of that procedure if warranted, and (2) the development of alternative photointerpretative techniques to the present LACIE procedures for the identification and selection of training fields (areas).
NASA Technical Reports Server (NTRS)
Colwell, R. N. (Principal Investigator); Hay, C. M.; Thomas, R. W.; Benson, A. S.
1977-01-01
Progress in the evaluation of the static stratification procedure and the development of alternative photointerpretive techniques to the present LACIE procedure for the identification of training fields is reported. Statistically significant signature controlling variables were defined for use in refining the stratification procedure. A subset of the 1973-74 Kansas LACIE segments for wheat was analyzed.
NASA Technical Reports Server (NTRS)
Bedewi, Nabih E.; Yang, Jackson C. S.
1987-01-01
Identification of the system parameters of a randomly excited structure may be treated using a variety of statistical techniques. Of all these techniques, the Random Decrement is unique in that it provides the homogeneous component of the system response. Using this quality, a system identification technique was developed based on a least-squares fit of the signatures to estimate the mass, damping, and stiffness matrices of a linear randomly excited system. The results of an experiment conducted on an offshore platform scale model to verify the validity of the technique and to demonstrate its application in damage detection are presented.
Noninvasive in vivo glucose sensing using an iris based technique
NASA Astrophysics Data System (ADS)
Webb, Anthony J.; Cameron, Brent D.
2011-03-01
Physiological glucose monitoring is important aspect in the treatment of individuals afflicted with diabetes mellitus. Although invasive techniques for glucose monitoring are widely available, it would be very beneficial to make such measurements in a noninvasive manner. In this study, a New Zealand White (NZW) rabbit animal model was utilized to evaluate a developed iris-based imaging technique for the in vivo measurement of physiological glucose concentration. The animals were anesthetized with isoflurane and an insulin/dextrose protocol was used to control blood glucose concentration. To further help restrict eye movement, a developed ocular fixation device was used. During the experimental time frame, near infrared illuminated iris images were acquired along with corresponding discrete blood glucose measurements taken with a handheld glucometer. Calibration was performed using an image based Partial Least Squares (PLS) technique. Independent validation was also performed to assess model performance along with Clarke Error Grid Analysis (CEGA). Initial validation results were promising and show that a high percentage of the predicted glucose concentrations are within 20% of the reference values.
Visualization of delamination in composite materials utilizing advanced X-ray imaging techniques
NASA Astrophysics Data System (ADS)
Vavrik, D.; Jakubek, J.; Jandejsek, I.; Krejci, F.; Kumpova, I.; Zemlicka, J.
2015-04-01
This work is focused on the development of instrumental radiographic methods for detection of delaminations in layered carbon fibre reinforced plastic composites used in the aerospace industry. The main limitation of current visualisation techniques is a very limited possibility to image so-called closed delaminations in which delaminated layers are in contact practically with no physical gap. In this contribution we report the development of innovative methods for closed delamination detection using an X-ray phase contrast technique for which the distance between delamination surfaces is not relevant. The approach is based on the energetic sensitivity of phase-enhanced radiography. Based on the applied methodology, we can distinguish both closed and open delamination. Further we have demonstrated the possibility to visualise open delaminations characterised by a physical gap between delaminated layers. This delamination type was successfully identified and visualized utilizing a high resolution and computed tomography table-top technique based on proper beam-hardening effect correction.
An Object-Based Requirements Modeling Method.
ERIC Educational Resources Information Center
Cordes, David W.; Carver, Doris L.
1992-01-01
Discusses system modeling and specification as it relates to object-based information systems development and software development. An automated system model based on the objects in the initial requirements document is described, the requirements document translator is explained, and a sample application of the technique is provided. (12…
NASA Astrophysics Data System (ADS)
Schooneveld, E. M.; Pietropaolo, A.; Andreani, C.; Perelli Cippo, E.; Rhodes, N. J.; Senesi, R.; Tardocchi, M.; Gorini, G.
2016-09-01
Neutron scattering techniques are attracting an increasing interest from scientists in various research fields, ranging from physics and chemistry to biology and archaeometry. The success of these neutron scattering applications is stimulated by the development of higher performance instrumentation. The development of new techniques and concepts, including radiative capture based neutron detection, is therefore a key issue to be addressed. Radiative capture based neutron detectors utilize the emission of prompt gamma rays after neutron absorption in a suitable isotope and the detection of those gammas by a photon counter. They can be used as simple counters in the thermal region and (simultaneously) as energy selector and counters for neutrons in the eV energy region. Several years of extensive development have made eV neutron spectrometers operating in the so-called resonance detector spectrometer (RDS) configuration outperform their conventional counterparts. In fact, the VESUVIO spectrometer, a flagship instrument at ISIS serving a continuous user programme for eV inelastic neutron spectroscopy measurements, is operating in the RDS configuration since 2007. In this review, we discuss the physical mechanism underlying the RDS configuration and the development of associated instrumentation. A few successful neutron scattering experiments that utilize the radiative capture counting techniques will be presented together with the potential of this technique for thermal neutron diffraction measurements. We also outline possible improvements and future perspectives for radiative capture based neutron detectors in neutron scattering application at pulsed neutron sources.
Spectroscopic techniques to study the immune response in human saliva
NASA Astrophysics Data System (ADS)
Nepomnyashchaya, E.; Savchenko, E.; Velichko, E.; Bogomaz, T.; Aksenov, E.
2018-01-01
Studies of the immune response dynamics by means of spectroscopic techniques, i.e., laser correlation spectroscopy and fluorescence spectroscopy, are described. The laser correlation spectroscopy is aimed at measuring sizes of particles in biological fluids. The fluorescence spectroscopy allows studying of the conformational and other structural changings in immune complex. We have developed a new scheme of a laser correlation spectrometer and an original signal processing algorithm. We have suggested a new fluorescence detection scheme based on a prism and an integrating pin diode. The developed system based on the spectroscopic techniques allows studies of complex process in human saliva and opens some prospects for an individual treatment of immune diseases.
Zhang, Ao; Yan, Xing-Ke; Liu, An-Guo
2016-12-25
In the present paper, the authors introduce a newly-developed "Acupuncture Needle Manipulation Training-evaluation System" based on optical motion capture technique. It is composed of two parts, sensor and software, and overcomes some shortages of mechanical motion capture technique. This device is able to analyze the data of operations of the pressing-hand and needle-insertion hand during acupuncture performance and its software contains personal computer (PC) version, Android version, and Internetwork Operating System (IOS) Apple version. It is competent in recording and analyzing information of any ope-rator's needling manipulations, and is quite helpful for teachers in teaching, training and examining students in clinical practice.
Pattern recognition of satellite cloud imagery for improved weather prediction
NASA Technical Reports Server (NTRS)
Gautier, Catherine; Somerville, Richard C. J.; Volfson, Leonid B.
1986-01-01
The major accomplishment was the successful development of a method for extracting time derivative information from geostationary meteorological satellite imagery. This research is a proof-of-concept study which demonstrates the feasibility of using pattern recognition techniques and a statistical cloud classification method to estimate time rate of change of large-scale meteorological fields from remote sensing data. The cloud classification methodology is based on typical shape function analysis of parameter sets characterizing the cloud fields. The three specific technical objectives, all of which were successfully achieved, are as follows: develop and test a cloud classification technique based on pattern recognition methods, suitable for the analysis of visible and infrared geostationary satellite VISSR imagery; develop and test a methodology for intercomparing successive images using the cloud classification technique, so as to obtain estimates of the time rate of change of meteorological fields; and implement this technique in a testbed system incorporating an interactive graphics terminal to determine the feasibility of extracting time derivative information suitable for comparison with numerical weather prediction products.
State-of-the-art characterization techniques for advanced lithium-ion batteries
NASA Astrophysics Data System (ADS)
Lu, Jun; Wu, Tianpin; Amine, Khalil
2017-03-01
To meet future needs for industries from personal devices to automobiles, state-of-the-art rechargeable lithium-ion batteries will require both improved durability and lowered costs. To enhance battery performance and lifetime, understanding electrode degradation mechanisms is of critical importance. Various advanced in situ and operando characterization tools developed during the past few years have proven indispensable for optimizing battery materials, understanding cell degradation mechanisms, and ultimately improving the overall battery performance. Here we review recent progress in the development and application of advanced characterization techniques such as in situ transmission electron microscopy for high-performance lithium-ion batteries. Using three representative electrode systems—layered metal oxides, Li-rich layered oxides and Si-based or Sn-based alloys—we discuss how these tools help researchers understand the battery process and design better battery systems. We also summarize the application of the characterization techniques to lithium-sulfur and lithium-air batteries and highlight the importance of those techniques in the development of next-generation batteries.
Development of a sensitivity analysis technique for multiloop flight control systems
NASA Technical Reports Server (NTRS)
Vaillard, A. H.; Paduano, J.; Downing, D. R.
1985-01-01
This report presents the development and application of a sensitivity analysis technique for multiloop flight control systems. This analysis yields very useful information on the sensitivity of the relative-stability criteria of the control system, with variations or uncertainties in the system and controller elements. The sensitivity analysis technique developed is based on the computation of the singular values and singular-value gradients of a feedback-control system. The method is applicable to single-input/single-output as well as multiloop continuous-control systems. Application to sampled-data systems is also explored. The sensitivity analysis technique was applied to a continuous yaw/roll damper stability augmentation system of a typical business jet, and the results show that the analysis is very useful in determining the system elements which have the largest effect on the relative stability of the closed-loop system. As a secondary product of the research reported here, the relative stability criteria based on the concept of singular values were explored.
Preliminary Evaluation of BIM-based Approaches for Schedule Delay Analysis
NASA Astrophysics Data System (ADS)
Chou, Hui-Yu; Yang, Jyh-Bin
2017-10-01
The problem of schedule delay commonly occurs in construction projects. The quality of delay analysis depends on the availability of schedule-related information and delay evidence. More information used in delay analysis usually produces more accurate and fair analytical results. How to use innovative techniques to improve the quality of schedule delay analysis results have received much attention recently. As Building Information Modeling (BIM) technique has been quickly developed, using BIM and 4D simulation techniques have been proposed and implemented. Obvious benefits have been achieved especially in identifying and solving construction consequence problems in advance of construction. This study preforms an intensive literature review to discuss the problems encountered in schedule delay analysis and the possibility of using BIM as a tool in developing a BIM-based approach for schedule delay analysis. This study believes that most of the identified problems can be dealt with by BIM technique. Research results could be a fundamental of developing new approaches for resolving schedule delay disputes.
Web image retrieval using an effective topic and content-based technique
NASA Astrophysics Data System (ADS)
Lee, Ching-Cheng; Prabhakara, Rashmi
2005-03-01
There has been an exponential growth in the amount of image data that is available on the World Wide Web since the early development of Internet. With such a large amount of information and image available and its usefulness, an effective image retrieval system is thus greatly needed. In this paper, we present an effective approach with both image matching and indexing techniques that improvise on existing integrated image retrieval methods. This technique follows a two-phase approach, integrating query by topic and query by example specification methods. In the first phase, The topic-based image retrieval is performed by using an improved text information retrieval (IR) technique that makes use of the structured format of HTML documents. This technique consists of a focused crawler that not only provides for the user to enter the keyword for the topic-based search but also, the scope in which the user wants to find the images. In the second phase, we use query by example specification to perform a low-level content-based image match in order to retrieve smaller and relatively closer results of the example image. From this, information related to the image feature is automatically extracted from the query image. The main objective of our approach is to develop a functional image search and indexing technique and to demonstrate that better retrieval results can be achieved.
Hocalar, A; Türker, M; Karakuzu, C; Yüzgeç, U
2011-04-01
In this study, previously developed five different state estimation methods are examined and compared for estimation of biomass concentrations at a production scale fed-batch bioprocess. These methods are i. estimation based on kinetic model of overflow metabolism; ii. estimation based on metabolic black-box model; iii. estimation based on observer; iv. estimation based on artificial neural network; v. estimation based on differential evaluation. Biomass concentrations are estimated from available measurements and compared with experimental data obtained from large scale fermentations. The advantages and disadvantages of the presented techniques are discussed with regard to accuracy, reproducibility, number of primary measurements required and adaptation to different working conditions. Among the various techniques, the metabolic black-box method seems to have advantages although the number of measurements required is more than that for the other methods. However, the required extra measurements are based on commonly employed instruments in an industrial environment. This method is used for developing a model based control of fed-batch yeast fermentations. Copyright © 2010 ISA. Published by Elsevier Ltd. All rights reserved.
The detection of bulk explosives using nuclear-based techniques
DOE Office of Scientific and Technical Information (OSTI.GOV)
Morgado, R.E.; Gozani, T.; Seher, C.C.
1988-01-01
In 1986 we presented a rationale for the detection of bulk explosives based on nuclear techniques that addressed the requirements of civil aviation security in the airport environment. Since then, efforts have intensified to implement a system based on thermal neutron activation (TNA), with new work developing in fast neutron and energetic photon reactions. In this paper we will describe these techniques and present new results from laboratory and airport testing. Based on preliminary results, we contended in our earlier paper that nuclear-based techniques did provide sufficiently penetrating probes and distinguishable detectable reaction products to achieve the FAA operational goals;more » new data have supported this contention. The status of nuclear-based techniques for the detection of bulk explosives presently under investigation by the US Federal Aviation Administration (FAA) is reviewed. These include thermal neutron activation (TNA), fast neutron activation (FNA), the associated particle technique, nuclear resonance absorption, and photoneutron activation. The results of comprehensive airport testing of the TNA system performed during 1987-88 are summarized. From a technical point of view, nuclear-based techniques now represent the most comprehensive and feasible approach for meeting the operational criteria of detection, false alarms, and throughput. 9 refs., 5 figs., 2 tabs.« less
Unsupervised, Robust Estimation-based Clustering for Multispectral Images
NASA Technical Reports Server (NTRS)
Netanyahu, Nathan S.
1997-01-01
To prepare for the challenge of handling the archiving and querying of terabyte-sized scientific spatial databases, the NASA Goddard Space Flight Center's Applied Information Sciences Branch (AISB, Code 935) developed a number of characterization algorithms that rely on supervised clustering techniques. The research reported upon here has been aimed at continuing the evolution of some of these supervised techniques, namely the neural network and decision tree-based classifiers, plus extending the approach to incorporating unsupervised clustering algorithms, such as those based on robust estimation (RE) techniques. The algorithms developed under this task should be suited for use by the Intelligent Information Fusion System (IIFS) metadata extraction modules, and as such these algorithms must be fast, robust, and anytime in nature. Finally, so that the planner/schedule module of the IlFS can oversee the use and execution of these algorithms, all information required by the planner/scheduler must be provided to the IIFS development team to ensure the timely integration of these algorithms into the overall system.
Measurements of Cuspal Slope Inclination Angles in Palaeoanthropological Applications
NASA Astrophysics Data System (ADS)
Gaboutchian, A. V.; Knyaz, V. A.; Leybova, N. A.
2017-05-01
Tooth crown morphological features, studied in palaeoanthropology, provide valuable information about human evolution and development of civilization. Tooth crown morphology represents biological and historical data of high taxonomical value as it characterizes genetically conditioned tooth relief features averse to substantial changes under environmental factors during lifetime. Palaeoanthropological studies are still based mainly on descriptive techniques and manual measurements of limited number of morphological parameters. Feature evaluation and measurement result analysis are expert-based. Development of new methods and techniques in 3D imaging creates a background provides for better value of palaeoanthropological data processing, analysis and distribution. The goals of the presented research are to propose new features for automated odontometry and to explore their applicability to paleoanthropological studies. A technique for automated measuring of given morphological tooth parameters needed for anthropological study is developed. It is based on using original photogrammetric system as a teeth 3D models acquisition device and on a set of algorithms for given tooth parameters estimation.
A study of trends and techniques for space base electronics
NASA Technical Reports Server (NTRS)
Trotter, J. D.; Wade, T. E.; Gassaway, J. D.; Mahmood, Q.
1978-01-01
A sputtering system was developed to deposit aluminum and aluminum alloys by the dc sputtering technique. This system is designed for a high level of cleanliness and for monitoring the deposition parameters during film preparation. This system is now ready for studying the deposition and annealing parameters upon double-level metal preparation. A technique recently applied for semiconductor analysis, the finite element method, was studied for use in the computer modeling of two dimensional MOS transistor structures. It was concluded that the method has not been sufficiently well developed for confident use at this time. An algorithm was developed for confident use at this time. An algorithm was developed for implementing a computer study which is based upon the finite difference method. The program which was developed was modified and used to calculate redistribution data for boron and phosphorous which had been predeposited by ion implantation with range and straggle conditions. Data were generated for 111 oriented SOS films with redistribution in N2, dry O2 and steam ambients.
Hybrid approach for robust diagnostics of cutting tools
NASA Astrophysics Data System (ADS)
Ramamurthi, K.; Hough, C. L., Jr.
1994-03-01
A new multisensor based hybrid technique has been developed for robust diagnosis of cutting tools. The technique combines the concepts of pattern classification and real-time knowledge based systems (RTKBS) and draws upon their strengths; learning facility in the case of pattern classification and a higher level of reasoning in the case of RTKBS. It eliminates some of their major drawbacks: false alarms or delayed/lack of diagnosis in case of pattern classification and tedious knowledge base generation in case of RTKBS. It utilizes a dynamic distance classifier, developed upon a new separability criterion and a new definition of robust diagnosis for achieving these benefits. The promise of this technique has been proven concretely through an on-line diagnosis of drill wear. Its suitability for practical implementation is substantiated by the use of practical, inexpensive, machine-mounted sensors and low-cost delivery systems.
New Researches and Application Progress of Commonly Used Optical Molecular Imaging Technology
Chen, Zhi-Yi; Yang, Feng; Lin, Yan; Zhou, Qiu-Lan; Liao, Yang-Ying
2014-01-01
Optical molecular imaging, a new medical imaging technique, is developed based on genomics, proteomics and modern optical imaging technique, characterized by non-invasiveness, non-radiativity, high cost-effectiveness, high resolution, high sensitivity and simple operation in comparison with conventional imaging modalities. Currently, it has become one of the most widely used molecular imaging techniques and has been applied in gene expression regulation and activity detection, biological development and cytological detection, drug research and development, pathogenesis research, pharmaceutical effect evaluation and therapeutic effect evaluation, and so forth, This paper will review the latest researches and application progresses of commonly used optical molecular imaging techniques such as bioluminescence imaging and fluorescence molecular imaging. PMID:24696850
Development of a Transportable Gravity Gradiometer Based on Atom Interferometry
NASA Astrophysics Data System (ADS)
Yu, N.; Kohel, J. M.; Aveline, D. C.; Kellogg, J. R.; Thompson, R. J.; Maleki, L.
2007-12-01
JPL is developing a transportable gravity gradiometer based on light-pulse atom interferometers for NASA's Earth Science Technology Office's Instrument Incubator Program. The inertial sensors in this instrument employ a quantum interference measurement technique, analogous to the precise phase measurements in atomic clocks, which offers increased sensitivity and improved long-term stability over traditional mechanical devices. We report on the implementation of this technique in JPL's gravity gradiometer, and on the current performance of the mobile instrument. We also discuss the prospects for satellite-based gravity field mapping, including high-resolution monitoring of time-varying fields from a single satellite platform and multi-component measurements of the gravitational gradient tensor, using atom interferometer-based instruments.
Vision based techniques for rotorcraft low altitude flight
NASA Technical Reports Server (NTRS)
Sridhar, Banavar; Suorsa, Ray; Smith, Philip
1991-01-01
An overview of research in obstacle detection at NASA Ames Research Center is presented. The research applies techniques from computer vision to automation of rotorcraft navigation. The development of a methodology for detecting the range to obstacles based on the maximum utilization of passive sensors is emphasized. The development of a flight and image data base for verification of vision-based algorithms, and a passive ranging methodology tailored to the needs of helicopter flight are discussed. Preliminary results indicate that it is possible to obtain adequate range estimates except at regions close to the FOE. Closer to the FOE, the error in range increases since the magnitude of the disparity gets smaller, resulting in a low SNR.
Cost and schedule analytical techniques development
NASA Technical Reports Server (NTRS)
1994-01-01
This contract provided technical services and products to the Marshall Space Flight Center's Engineering Cost Office (PP03) and the Program Plans and Requirements Office (PP02) for the period of 3 Aug. 1991 - 30 Nov. 1994. Accomplishments summarized cover the REDSTAR data base, NASCOM hard copy data base, NASCOM automated data base, NASCOM cost model, complexity generators, program planning, schedules, NASA computer connectivity, other analytical techniques, and special project support.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Solaimani, Mohiuddin; Iftekhar, Mohammed; Khan, Latifur
Anomaly detection refers to the identi cation of an irregular or unusual pat- tern which deviates from what is standard, normal, or expected. Such deviated patterns typically correspond to samples of interest and are assigned different labels in different domains, such as outliers, anomalies, exceptions, or malware. Detecting anomalies in fast, voluminous streams of data is a formidable chal- lenge. This paper presents a novel, generic, real-time distributed anomaly detection framework for heterogeneous streaming data where anomalies appear as a group. We have developed a distributed statistical approach to build a model and later use it to detect anomaly. Asmore » a case study, we investigate group anomaly de- tection for a VMware-based cloud data center, which maintains a large number of virtual machines (VMs). We have built our framework using Apache Spark to get higher throughput and lower data processing time on streaming data. We have developed a window-based statistical anomaly detection technique to detect anomalies that appear sporadically. We then relaxed this constraint with higher accuracy by implementing a cluster-based technique to detect sporadic and continuous anomalies. We conclude that our cluster-based technique out- performs other statistical techniques with higher accuracy and lower processing time.« less
Sun, Tie Gang; Xiao, Rong Bo; Cai, Yun Nan; Wang, Yao Wu; Wu, Chang Guang
2016-08-01
Quantitative assessment of urban thermal environment has become a focus for urban climate and environmental science since the concept of urban heat island has been proposed. With the continual development of space information and computer simulation technology, substantial progresses have been made on quantitative assessment techniques and methods of urban thermal environment. The quantitative assessment techniques have been developed to dynamics simulation and forecast of thermal environment at various scales based on statistical analysis of thermal environment on urban-scale using the historical data of weather stations. This study reviewed the development progress of ground meteorological observation, thermal infrared remote sensing and numerical simulation. Moreover, the potential advantages and disadvantages, applicability and the development trends of these techniques were also summarized, aiming to add fundamental knowledge of understanding the urban thermal environment assessment and optimization.
New techniques for test development for tactical auto-pilots using microprocessors
NASA Astrophysics Data System (ADS)
Shemeta, E. H.
1980-07-01
This paper reports on a demonstration of the application of the method to generate system level tests for a typical tactical missile autopilot. The test algorithms are based on the autopilot control law. When loaded on the tester with appropriate control information, the complete autopilot is tested to establish if the specified control law requirements are met. Thus, the test procedure not only checks to see if the hardware is functional, but also checks the operational software. The technique also uses a 'learning' mode to allow minor timing or functional deviations from the expected responses to be incorporated in the test procedures. A potential application of this test development technique is the extraction of production test data for the various subassemblies. The technique will 'learn' the input-output patterns forming the basis for developement and production tests. If successful, these new techniques should allow the test development process to keep pace with semiconductor progress.
EB and EUV lithography using inedible cellulose-based biomass resist material
NASA Astrophysics Data System (ADS)
Takei, Satoshi; Hanabata, Makoto; Oshima, Akihiro; Kashiwakura, Miki; Kozawa, Takahiro; Tagawa, Seiichi
2016-03-01
The validity of our approach of inedible cellulose-based resist material derived from woody biomass has been confirmed experimentally for the use of pure water in organic solvent-free water spin-coating and tetramethylammonium hydroxide(TMAH)-free water-developable techniques of eco-conscious electron beam (EB) and extreme-ultraviolet (EUV) lithography. The water developable, non-chemically amplified, high sensitive, and negative tone resist material in EB and EUV lithography was developed for environmental affair, safety, easiness of handling, and health of the working people. The inedible cellulose-based biomass resist material was developed by replacing the hydroxyl groups in the beta-linked disaccharides with EB and EUV sensitive groups. The 50-100 nm line and space width, and little footing profiles of cellulose-based biomass resist material on hardmask and layer were resolved at the doses of 10-30 μC/cm2. The eco-conscious lithography techniques was referred to as green EB and EUV lithography using inedible cellulose-based biomass resist material.
Coater/developer based techniques to improve high-resolution EUV patterning defectivity
NASA Astrophysics Data System (ADS)
Hontake, Koichi; Huli, Lior; Lemley, Corey; Hetzer, Dave; Liu, Eric; Ko, Akiteru; Kawakami, Shinichiro; Shimoaoki, Takeshi; Hashimoto, Yusaku; Tanaka, Koichiro; Petrillo, Karen; Meli, Luciana; De Silva, Anuja; Xu, Yongan; Felix, Nelson; Johnson, Richard; Murray, Cody; Hubbard, Alex
2017-10-01
Extreme ultraviolet lithography (EUVL) technology is one of the leading candidates under consideration for enabling the next generation of devices, for 7nm node and beyond. As the focus shifts to driving down the 'effective' k1 factor and enabling the full scaling entitlement of EUV patterning, new techniques and methods must be developed to reduce the overall defectivity, mitigate pattern collapse, and eliminate film-related defects. In addition, CD uniformity and LWR/LER must be improved in terms of patterning performance. Tokyo Electron Limited (TEL™) and IBM Corporation are continuously developing manufacturing quality processes for EUV. In this paper, we review the ongoing progress in coater/developer based processes (coating, developing, baking) that are required to enable EUV patterning.
Imputatoin and Model-Based Updating Technique for Annual Forest Inventories
Ronald E. McRoberts
2001-01-01
The USDA Forest Service is developing an annual inventory system to establish the capability of producing annual estimates of timber volume and related variables. The inventory system features measurement of an annual sample of field plots with options for updating data for plots measured in previous years. One imputation and two model-based updating techniques are...
DOE Office of Scientific and Technical Information (OSTI.GOV)
Harrison, Richard Karl; Martin, Jeffrey B.; Wiemann, Dora K.
We developed new detector technologies to identify the presence of radioactive materials for nuclear forensics applications. First, we investigated an optical radiation detection technique based on imaging nitrogen fluorescence excited by ionizing radiation. We demonstrated optical detection in air under indoor and outdoor conditions for alpha particles and gamma radiation at distances up to 75 meters. We also contributed to the development of next generation systems and concepts that could enable remote detection at distances greater than 1 km, and originated a concept that could enable daytime operation of the technique. A second area of research was the development ofmore » room-temperature graphene-based sensors for radiation detection and measurement. In this project, we observed tunable optical and charged particle detection, and developed improved devices. With further development, the advancements described in this report could enable new capabilities for nuclear forensics applications.« less
DOE Office of Scientific and Technical Information (OSTI.GOV)
Jin, J; Kong, V; Zhang, H
Purpose: Three dimensional (3D) Grid Therapy using MLC-based inverse-planning has been proposed to achieve the features of both conformal radiotherapy and spatially fractionated radiotherapy, which may deliver very high dose in a single fraction to portions of a large tumor with relatively low normal tissue dose. However, the technique requires relatively long delivery time. This study aims to develop a collimator-based 3D grid therapy technique. Here we report the development of the technique in a small animal radiation research platform. Methods: Similar as in the MLC-based technique, 9 non-coplanar beams in special channeling directions were used for the 3D gridmore » therapy technique. Two specially designed grid collimators were fabricated, and one of them was selectively used to match the corresponding gantry/couch angles so that the grid opening of all 9 beams are met in the 3D space in the target. A stack of EBT3 films were used as 3D dosimetry to demonstrate the 3D grid-like dose distribution in the target. Three 1-mm beams were delivered to the stack of films in the area outside the target for alignment when all the films were scanned to reconstruct the 3D dosimtric image. Results: 3D film dosimetry showed a lattice-like dose distribution in the 3D target as well as in the axial, sagittal and coronal planes. The dose outside the target also showed a grid like dose distribution, and the average dose gradually decreased with the distance to the target. The peak to valley ratio was approximately 5:1. The delivery time was 7 minutes for 18 Gy peak dose, comparing to 6 minutes to deliver a 18-Gy 3D conformal plan. Conclusion: We have demonstrated the feasibility of the collimator-based 3D grid therapy technique which can significantly reduce delivery time comparing to MLC-based inverse planning technique.« less
NASA Astrophysics Data System (ADS)
Rajshekhar, G.; Gorthi, Sai Siva; Rastogi, Pramod
2010-04-01
For phase estimation in digital holographic interferometry, a high-order instantaneous moments (HIM) based method was recently developed which relies on piecewise polynomial approximation of phase and subsequent evaluation of the polynomial coefficients using the HIM operator. A crucial step in the method is mapping the polynomial coefficient estimation to single-tone frequency determination for which various techniques exist. The paper presents a comparative analysis of the performance of the HIM operator based method in using different single-tone frequency estimation techniques for phase estimation. The analysis is supplemented by simulation results.
NASA Technical Reports Server (NTRS)
Ray, Ronald J.
1994-01-01
New flight test maneuvers and analysis techniques for evaluating the dynamic response of in-flight thrust models during throttle transients have been developed and validated. The approach is based on the aircraft and engine performance relationship between thrust and drag. Two flight test maneuvers, a throttle step and a throttle frequency sweep, were developed and used in the study. Graphical analysis techniques, including a frequency domain analysis method, were also developed and evaluated. They provide quantitative and qualitative results. Four thrust calculation methods were used to demonstrate and validate the test technique. Flight test applications on two high-performance aircraft confirmed the test methods as valid and accurate. These maneuvers and analysis techniques were easy to implement and use. Flight test results indicate the analysis techniques can identify the combined effects of model error and instrumentation response limitations on the calculated thrust value. The methods developed in this report provide an accurate approach for evaluating, validating, or comparing thrust calculation methods for dynamic flight applications.
The Empirical Investigation of Perspective-Based Reading
NASA Technical Reports Server (NTRS)
Basili, Victor R.; Green, Scott; Laitenberger, Oliver; Shull, Forrest; Sorumgard, Sivert; Zelkowitz, Marvin V.
1996-01-01
We consider reading techniques a fundamental means of achieving high quality software. Due to the lack of research in this area, we are experimenting with the application and comparison of various reading techniques. This paper deals with our experiences with Perspective-Based Reading (PBR), a particular reading technique for requirements documents. The goal of PBR is to provide operational scenarios where members of a review team read a document from a particular perspective (e.g., tester, developer, user). Our assumption is that the combination of different perspectives provides better coverage of the document than the same number of readers using their usual technique.
Advances in fragment-based drug discovery platforms.
Orita, Masaya; Warizaya, Masaichi; Amano, Yasushi; Ohno, Kazuki; Niimi, Tatsuya
2009-11-01
Fragment-based drug discovery (FBDD) has been established as a powerful alternative and complement to traditional high-throughput screening techniques for identifying drug leads. At present, this technique is widely used among academic groups as well as small biotech and large pharmaceutical companies. In recent years, > 10 new compounds developed with FBDD have entered clinical development, and more and more attention in the drug discovery field is being focused on this technique. Under the FBDD approach, a fragment library of relatively small compounds (molecular mass = 100 - 300 Da) is screened by various methods and the identified fragment hits which normally weakly bind to the target are used as starting points to generate more potent drug leads. Because FBDD is still a relatively new drug discovery technology, further developments and optimizations in screening platforms and fragment exploitation can be expected. This review summarizes recent advances in FBDD platforms and discusses the factors important for the successful application of this technique. Under the FBDD approach, both identifying the starting fragment hit to be developed and generating the drug lead from that starting fragment hit are important. Integration of various techniques, such as computational technology, X-ray crystallography, NMR, surface plasmon resonance, isothermal titration calorimetry, mass spectrometry and high-concentration screening, must be applied in a situation-appropriate manner.
NASA Technical Reports Server (NTRS)
Harrison, P. Ann
1993-01-01
All the NASA VEGetation Workbench (VEG) goals except the Learning System provide the scientist with several different techniques. When VEG is run, rules assist the scientist in selecting the best of the available techniques to apply to the sample of cover type data being studied. The techniques are stored in the VEG knowledge base. The design and implementation of an interface that allows the scientist to add new techniques to VEG without assistance from the developer were completed. A new interface that enables the scientist to add techniques to VEG without assistance from the developer was designed and implemented. This interface does not require the scientist to have a thorough knowledge of Knowledge Engineering Environment (KEE) by Intellicorp or a detailed knowledge of the structure of VEG. The interface prompts the scientist to enter the required information about the new technique. It prompts the scientist to enter the required Common Lisp functions for executing the technique and the left hand side of the rule that causes the technique to be selected. A template for each function and rule and detailed instructions about the arguments of the functions, the values they should return, and the format of the rule are displayed. Checks are made to ensure that the required data were entered, the functions compiled correctly, and the rule parsed correctly before the new technique is stored. The additional techniques are stored separately from the VEG knowledge base. When the VEG knowledge base is loaded, the additional techniques are not normally loaded. The interface allows the scientist the option of adding all the previously defined new techniques before running VEG. When the techniques are added, the required units to store the additional techniques are created automatically in the correct places in the VEG knowledge base. The methods file containing the functions required by the additional techniques is loaded. New rule units are created to store the new rules. The interface that allow the scientist to select which techniques to use is updated automatically to include the new techniques. Task H was completed. The interface that allows the scientist to add techniques to VEG was implemented and comprehensively tested. The Common Lisp code for the Add Techniques system is listed in Appendix A.
Development of a HIV-1 Virus Detection System Based on Nanotechnology.
Lee, Jin-Ho; Oh, Byung-Keun; Choi, Jeong-Woo
2015-04-27
Development of a sensitive and selective detection system for pathogenic viral agents is essential for medical healthcare from diagnostics to therapeutics. However, conventional detection systems are time consuming, resource-intensive and tedious to perform. Hence, the demand for sensitive and selective detection system for virus are highly increasing. To attain this aim, different aspects and techniques have been applied to develop virus sensor with improved sensitivity and selectivity. Here, among those aspects and techniques, this article reviews HIV virus particle detection systems incorporated with nanotechnology to enhance the sensitivity. This review mainly focused on four different detection system including vertically configured electrical detection based on scanning tunneling microscopy (STM), electrochemical detection based on direct electron transfer in virus, optical detection system based on localized surface plasmon resonance (LSPR) and surface enhanced Raman spectroscopy (SERS) using plasmonic nanoparticle.
NASA Technical Reports Server (NTRS)
Coleman, R. A.; Cofer, W. R., III; Edahl, R. A., Jr.
1985-01-01
An analytical technique for the determination of trace (sub-ppbv) quantities of volatile organic compounds in air was developed. A liquid nitrogen-cooled trap operated at reduced pressures in series with a Dupont Nafion-based drying tube and a gas chromatograph was utilized. The technique is capable of analyzing a variety of organic compounds, from simple alkanes to alcohols, while offering a high level of precision, peak sharpness, and sensitivity.
Model-based RSA of a femoral hip stem using surface and geometrical shape models.
Kaptein, Bart L; Valstar, Edward R; Spoor, Cees W; Stoel, Berend C; Rozing, Piet M
2006-07-01
Roentgen stereophotogrammetry (RSA) is a highly accurate three-dimensional measuring technique for assessing micromotion of orthopaedic implants. A drawback is that markers have to be attached to the implant. Model-based techniques have been developed to prevent using special marked implants. We compared two model-based RSA methods with standard marker-based RSA techniques. The first model-based RSA method used surface models, and the second method used elementary geometrical shape (EGS) models. We used a commercially available stem to perform experiments with a phantom as well as reanalysis of patient RSA radiographs. The data from the phantom experiment indicated the accuracy and precision of the elementary geometrical shape model-based RSA method is equal to marker-based RSA. For model-based RSA using surface models, the accuracy is equal to the accuracy of marker-based RSA, but its precision is worse. We found no difference in accuracy and precision between the two model-based RSA techniques in clinical data. For this particular hip stem, EGS model-based RSA is a good alternative for marker-based RSA.
Interventional radiology of the thyroid gland: critical review and state of the art
Quarchioni, Simone; Bruno, Federico; Ierardi, Anna Maria; Arrigoni, Francesco; Giordano, Aldo Victor; Carducci, Sergio; Varrassi, Marco; Carrafiello, Giampaolo; Caranci, Ferdinando; Splendiani, Alessandra; Di Cesare, Ernesto; Masciocchi, Carlo
2018-01-01
Thyroid nodules are a common incidental finding during a routinely ultrasound (US) exam unrelated to the thyroid gland in the healthy adult population with a prevalence of 20–76%. As treated before with surgery, in the last years new minimally invasive techniques have been developed as an alternative to surgery. The aim of this review, based on newly revised guidelines, is to provide some information regarding the basic principles, indications, materials, techniques, and results of mini-invasive procedures or treatments for thyroid nodules. We performed a narrative review including both newest and representative papers and guidelines based on the different procedures of ablation techniques developed in the last years for the diagnosis and the treatment of thyroid nodules. All examined papers referred very good results in term of volume nodule reduction, improvement in related symptoms and cosmetic problems, with a very low rate of complications and side effects for all the minimally invasive technique analyzed. Obviously, some differents between technique based on different kind of thyroid nodules and different indication were found. In conclusion, many thyroid nodules nowadays could be treated thanks to the advent of new mini-invasive technique that are less expensive and present a lower risk of major complications and side effects compared to surgery. PMID:29770309
Development and comparison of projection and image space 3D nodule insertion techniques
NASA Astrophysics Data System (ADS)
Robins, Marthony; Solomon, Justin; Sahbaee, Pooyan; Samei, Ehsan
2016-04-01
This study aimed to develop and compare two methods of inserting computerized virtual lesions into CT datasets. 24 physical (synthetic) nodules of three sizes and four morphologies were inserted into an anthropomorphic chest phantom (LUNGMAN, KYOTO KAGAKU). The phantom was scanned (Somatom Definition Flash, Siemens Healthcare) with and without nodules present, and images were reconstructed with filtered back projection and iterative reconstruction (SAFIRE) at 0.6 mm slice thickness using a standard thoracic CT protocol at multiple dose settings. Virtual 3D CAD models based on the physical nodules were virtually inserted (accounting for the system MTF) into the nodule-free CT data using two techniques. These techniques include projection-based and image-based insertion. Nodule volumes were estimated using a commercial segmentation tool (iNtuition, TeraRecon, Inc.). Differences were tested using paired t-tests and R2 goodness of fit between the virtually and physically inserted nodules. Both insertion techniques resulted in nodule volumes very similar to the real nodules (<3% difference) and in most cases the differences were not statistically significant. Also, R2 values were all <0.97 for both insertion techniques. These data imply that these techniques can confidently be used as a means of inserting virtual nodules in CT datasets. These techniques can be instrumental in building hybrid CT datasets composed of patient images with virtually inserted nodules.
Finite Element Modelling and Analysis of Conventional Pultrusion Processes
NASA Astrophysics Data System (ADS)
Akishin, P.; Barkanov, E.; Bondarchuk, A.
2015-11-01
Pultrusion is one of many composite manufacturing techniques and one of the most efficient methods for producing fiber reinforced polymer composite parts with a constant cross-section. Numerical simulation is helpful for understanding the manufacturing process and developing scientific means for the pultrusion tooling design. Numerical technique based on the finite element method has been developed for the simulation of pultrusion processes. It uses the general purpose finite element software ANSYS Mechanical. It is shown that the developed technique predicts the temperature and cure profiles, which are in good agreement with those published in the open literature.
Simulation verification techniques study
NASA Technical Reports Server (NTRS)
Schoonmaker, P. B.; Wenglinski, T. H.
1975-01-01
Results are summarized of the simulation verification techniques study which consisted of two tasks: to develop techniques for simulator hardware checkout and to develop techniques for simulation performance verification (validation). The hardware verification task involved definition of simulation hardware (hardware units and integrated simulator configurations), survey of current hardware self-test techniques, and definition of hardware and software techniques for checkout of simulator subsystems. The performance verification task included definition of simulation performance parameters (and critical performance parameters), definition of methods for establishing standards of performance (sources of reference data or validation), and definition of methods for validating performance. Both major tasks included definition of verification software and assessment of verification data base impact. An annotated bibliography of all documents generated during this study is provided.
Space station advanced automation
NASA Technical Reports Server (NTRS)
Woods, Donald
1990-01-01
In the development of a safe, productive and maintainable space station, Automation and Robotics (A and R) has been identified as an enabling technology which will allow efficient operation at a reasonable cost. The Space Station Freedom's (SSF) systems are very complex, and interdependent. The usage of Advanced Automation (AA) will help restructure, and integrate system status so that station and ground personnel can operate more efficiently. To use AA technology for the augmentation of system management functions requires a development model which consists of well defined phases of: evaluation, development, integration, and maintenance. The evaluation phase will consider system management functions against traditional solutions, implementation techniques and requirements; the end result of this phase should be a well developed concept along with a feasibility analysis. In the development phase the AA system will be developed in accordance with a traditional Life Cycle Model (LCM) modified for Knowledge Based System (KBS) applications. A way by which both knowledge bases and reasoning techniques can be reused to control costs is explained. During the integration phase the KBS software must be integrated with conventional software, and verified and validated. The Verification and Validation (V and V) techniques applicable to these KBS are based on the ideas of consistency, minimal competency, and graph theory. The maintenance phase will be aided by having well designed and documented KBS software.
The Population of Small Comets: Optimum Techniques for Detection
NASA Technical Reports Server (NTRS)
Brandt, John C.
1997-01-01
The goals of this project were: (1) to present evidence to the scientific community for the importance of the small comet population and (2) to develop techniques for optimum detection in order to characterize the population. Our work on techniques has been to develop algorithms for searching images for SCs based on the distinctive properties of comets; (1) motion with respect to background stars; (2) extended source with most light coming from the coma rather than the nucleus; and characteristic spectral signature.
Propulsion Health Monitoring for Enhanced Safety
NASA Technical Reports Server (NTRS)
Butz, Mark G.; Rodriguez, Hector M.
2003-01-01
This report presents the results of the NASA contract Propulsion System Health Management for Enhanced Safety performed by General Electric Aircraft Engines (GE AE), General Electric Global Research (GE GR), and Pennsylvania State University Applied Research Laboratory (PSU ARL) under the NASA Aviation Safety Program. This activity supports the overall goal of enhanced civil aviation safety through a reduction in the occurrence of safety-significant propulsion system malfunctions. Specific objectives are to develop and demonstrate vibration diagnostics techniques for the on-line detection of turbine rotor disk cracks, and model-based fault tolerant control techniques for the prevention and mitigation of in-flight engine shutdown, surge/stall, and flameout events. The disk crack detection work was performed by GE GR which focused on a radial-mode vibration monitoring technique, and PSU ARL which focused on a torsional-mode vibration monitoring technique. GE AE performed the Model-Based Fault Tolerant Control work which focused on the development of analytical techniques for detecting, isolating, and accommodating gas-path faults.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Aziz, H. M. Abdul; Ukkusuri, Satish V.
We present that EPA-MOVES (Motor Vehicle Emission Simulator) is often integrated with traffic simulators to assess emission levels of large-scale urban networks with signalized intersections. High variations in speed profiles exist in the context of congested urban networks with signalized intersections. The traditional average-speed-based emission estimation technique with EPA-MOVES provides faster execution while underestimates the emissions in most cases because of ignoring the speed variation at congested networks with signalized intersections. In contrast, the atomic second-by-second speed profile (i.e., the trajectory of each vehicle)-based technique provides accurate emissions at the cost of excessive computational power and time. We addressed thismore » issue by developing a novel method to determine the link-driving-schedules (LDSs) for the EPA-MOVES tool. Our research developed a hierarchical clustering technique with dynamic time warping similarity measures (HC-DTW) to find the LDS for EPA-MOVES that is capable of producing emission estimates better than the average-speed-based technique with execution time faster than the atomic speed profile approach. We applied the HC-DTW on a sample data from a signalized corridor and found that HC-DTW can significantly reduce computational time without compromising the accuracy. The developed technique in this research can substantially contribute to the EPA-MOVES-based emission estimation process for large-scale urban transportation network by reducing the computational time with reasonably accurate estimates. This method is highly appropriate for transportation networks with higher variation in speed such as signalized intersections. Lastly, experimental results show error difference ranging from 2% to 8% for most pollutants except PM 10.« less
Aziz, H. M. Abdul; Ukkusuri, Satish V.
2017-06-29
We present that EPA-MOVES (Motor Vehicle Emission Simulator) is often integrated with traffic simulators to assess emission levels of large-scale urban networks with signalized intersections. High variations in speed profiles exist in the context of congested urban networks with signalized intersections. The traditional average-speed-based emission estimation technique with EPA-MOVES provides faster execution while underestimates the emissions in most cases because of ignoring the speed variation at congested networks with signalized intersections. In contrast, the atomic second-by-second speed profile (i.e., the trajectory of each vehicle)-based technique provides accurate emissions at the cost of excessive computational power and time. We addressed thismore » issue by developing a novel method to determine the link-driving-schedules (LDSs) for the EPA-MOVES tool. Our research developed a hierarchical clustering technique with dynamic time warping similarity measures (HC-DTW) to find the LDS for EPA-MOVES that is capable of producing emission estimates better than the average-speed-based technique with execution time faster than the atomic speed profile approach. We applied the HC-DTW on a sample data from a signalized corridor and found that HC-DTW can significantly reduce computational time without compromising the accuracy. The developed technique in this research can substantially contribute to the EPA-MOVES-based emission estimation process for large-scale urban transportation network by reducing the computational time with reasonably accurate estimates. This method is highly appropriate for transportation networks with higher variation in speed such as signalized intersections. Lastly, experimental results show error difference ranging from 2% to 8% for most pollutants except PM 10.« less
Oliver, Kelly; Manton, David John
2015-01-01
Effective behavior management guides children through the complex social context of dentistry utilizing techniques based on a current understanding of the social, emotional, and cognitive development of children. Behavior management techniques facilitate effective communication and establish social and behavioral guidelines for the dental environment. Contemporary parenting styles, expectations, and attitudes of modern parents and society have influenced the use of behavior management techniques with a prevailing emphasis on communicative techniques and pharmacological management over aversive techniques.
Harnessing Solid-State Ionic Transport for Nanomanufacturing and Nanodevices
ERIC Educational Resources Information Center
Hsu, Keng Hao
2009-01-01
Through this work a new all-solid, ambient processing condition direct metal patterning technique has been developed and characterized. This ionic-transport-based patterning technique is capable of sub-50nm feature resolution under ambient conditions. It generates features with a rate that is comparable to conventional dry-etching techniques. A…
Biosensor-based microRNA detection: techniques, design, performance, and challenges.
Johnson, Blake N; Mutharasan, Raj
2014-04-07
The current state of biosensor-based techniques for amplification-free microRNA (miRNA) detection is critically reviewed. Comparison with non-sensor and amplification-based molecular techniques (MTs), such as polymerase-based methods, is made in terms of transduction mechanism, associated protocol, and sensitivity. Challenges associated with miRNA hybridization thermodynamics which affect assay selectivity and amplification bias are briefly discussed. Electrochemical, electromechanical, and optical classes of miRNA biosensors are reviewed in terms of transduction mechanism, limit of detection (LOD), time-to-results (TTR), multiplexing potential, and measurement robustness. Current trends suggest that biosensor-based techniques (BTs) for miRNA assay will complement MTs due to the advantages of amplification-free detection, LOD being femtomolar (fM)-attomolar (aM), short TTR, multiplexing capability, and minimal sample preparation requirement. Areas of future importance in miRNA BT development are presented which include focus on achieving high measurement confidence and multiplexing capabilities.
Review of online coupling of sample preparation techniques with liquid chromatography.
Pan, Jialiang; Zhang, Chengjiang; Zhang, Zhuomin; Li, Gongke
2014-03-07
Sample preparation is still considered as the bottleneck of the whole analytical procedure, and efforts has been conducted towards the automation, improvement of sensitivity and accuracy, and low comsuption of organic solvents. Development of online sample preparation techniques (SP) coupled with liquid chromatography (LC) is a promising way to achieve these goals, which has attracted great attention. This article reviews the recent advances on the online SP-LC techniques. Various online SP techniques have been described and summarized, including solid-phase-based extraction, liquid-phase-based extraction assisted with membrane, microwave assisted extraction, ultrasonic assisted extraction, accelerated solvent extraction and supercritical fluids extraction. Specially, the coupling approaches of online SP-LC systems and the corresponding interfaces have been discussed and reviewed in detail, such as online injector, autosampler combined with transport unit, desorption chamber and column switching. Typical applications of the online SP-LC techniques have been summarized. Then the problems and expected trends in this field are attempted to be discussed and proposed in order to encourage the further development of online SP-LC techniques. Copyright © 2014 Elsevier B.V. All rights reserved.
Digital Mapping Techniques '11–12 workshop proceedings
Soller, David R.
2014-01-01
At these meetings, oral and poster presentations and special discussion sessions emphasized: (1) methods for creating and publishing map products (here, "publishing" includes Web-based release); (2) field data capture software and techniques, including the use of LiDAR; (3) digital cartographic techniques; (4) migration of digital maps into ArcGIS Geodatabase formats; (5) analytical GIS techniques; and (6) continued development of the National Geologic Map Database.
A solution to the Navier-Stokes equations based upon the Newton Kantorovich method
NASA Technical Reports Server (NTRS)
Davis, J. E.; Gabrielsen, R. E.; Mehta, U. B.
1977-01-01
An implicit finite difference scheme based on the Newton-Kantorovich technique was developed for the numerical solution of the nonsteady, incompressible, two-dimensional Navier-Stokes equations in conservation-law form. The algorithm was second-order-time accurate, noniterative with regard to the nonlinear terms in the vorticity transport equation except at the earliest few time steps, and spatially factored. Numerical results were obtained with the technique for a circular cylinder at Reynolds number 15. Results indicate that the technique is in excellent agreement with other numerical techniques for all geometries and Reynolds numbers investigated, and indicates a potential for significant reduction in computation time over current iterative techniques.
NASA Technical Reports Server (NTRS)
Bedewi, Nabih E.; Yang, Jackson C. S.
1987-01-01
Identification of the system parameters of a randomly excited structure may be treated using a variety of statistical techniques. Of all these techniques, the Random Decrement is unique in that it provides the homogeneous component of the system response. Using this quality, a system identification technique was developed based on a least-squares fit of the signatures to estimate the mass, damping, and stiffness matrices of a linear randomly excited system. The mathematics of the technique is presented in addition to the results of computer simulations conducted to demonstrate the prediction of the response of the system and the random forcing function initially introduced to excite the system.
Optical coherence tomography angiography in glaucoma care.
Chansangpetch, Sunee; Lin, Shan C
2018-05-14
Rapid improvements in optical coherence tomography (OCT) technology have allowed for enhancement of both image resolution and scanning speed, and the development of vascular assessment modality. Optical coherence tomography angiography (OCTA) is the non-invasive in vivo imaging of the vasculature located within the retina and optic nerve head area. The principle of OCTA is to use the variations in OCT signals caused by moving particles as the contrast mechanism for imaging of flow. Several algorithms which aim to maximize the contrast signal and minimize the noise have been developed including the phase-based techniques, intensity-based techniques (e.g., split-spectrum amplitude decorrelation angiography (SSADA)), and complex-based techniques (e.g., optical microangiography (OMAG)). With its reliable technique, high image resolution, and current availability, OCTA has been widely used in the assessment of posterior segment diseases including glaucoma in which ocular perfusion dysfunction has been proposed as a pathophysiological mechanism. This review will provide the reader with information on the principle techniques of OCTA; the current literature on OCTA reproducibility; its applications to glaucoma detection and monitoring of progression; and the role of OCTA in the assessment of the vascular component in glaucoma pathogenesis.
A brief review of extrusion-based tissue scaffold bio-printing.
Ning, Liqun; Chen, Xiongbiao
2017-08-01
Extrusion-based bio-printing has great potential as a technique for manipulating biomaterials and living cells to create three-dimensional (3D) scaffolds for damaged tissue repair and function restoration. Over the last two decades, advances in both engineering techniques and life sciences have evolved extrusion-based bio-printing from a simple technique to one able to create diverse tissue scaffolds from a wide range of biomaterials and cell types. However, the complexities associated with synthesis of materials for bio-printing and manipulation of multiple materials and cells in bio-printing pose many challenges for scaffold fabrication. This paper presents an overview of extrusion-based bio-printing for scaffold fabrication, focusing on the prior-printing considerations (such as scaffold design and materials/cell synthesis), working principles, comparison to other techniques, and to-date achievements. This paper also briefly reviews the recent development of strategies with regard to hydrogel synthesis, multi-materials/cells manipulation, and process-induced cell damage in extrusion-based bio-printing. The key issue and challenges for extrusion-based bio-printing are also identified and discussed along with recommendations for future, aimed at developing novel biomaterials and bio-printing systems, creating patterned vascular networks within scaffolds, and preserving the cell viability and functions in scaffold bio-printing. The address of these challenges will significantly enhance the capability of extrusion-based bio-printing. Copyright © 2017 WILEY-VCH Verlag GmbH & Co. KGaA, Weinheim.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Keepin, G.R.
Over the years the Los Alamos safeguards program has developed, tested, and implemented a broad range of passive and active nondestructive analysis (NDA) instruments (based on gamma and x-ray detection and neutron counting) that are now widely employed in safeguarding nuclear materials of all forms. Here very briefly, the major categories of gamma ray and neutron based NDA techniques, give some representative examples of NDA instruments currently in use, and cite a few notable instances of state-of-the-art NDA technique development. Historical aspects and a broad overview of the safeguards program are also presented.
Surface plasmon resonance sensing: from purified biomolecules to intact cells.
Su, Yu-Wen; Wang, Wei
2018-04-12
Surface plasmon resonance (SPR) has become a well-recognized label-free technique for measuring the binding kinetics between biomolecules since the invention of the first SPR-based immunosensor in 1980s. The most popular and traditional format for SPR analysis is to monitor the real-time optical signals when a solution containing ligand molecules is flowing over a sensor substrate functionalized with purified receptor molecules. In recent years, rapid development of several kinds of SPR imaging techniques have allowed for mapping the dynamic distribution of local mass density within single living cells with high spatial and temporal resolutions and reliable sensitivity. Such capability immediately enabled one to investigate the interaction between important biomolecules and intact cells in a label-free, quantitative, and single cell manner, leading to an exciting new trend of cell-based SPR bioanalysis. In this Trend Article, we first describe the principle and technical features of two types of SPR imaging techniques based on prism and objective, respectively. Then we survey the intact cell-based applications in both fundamental cell biology and drug discovery. We conclude the article with comments and perspectives on the future developments. Graphical abstract Recent developments in surface plasmon resonance (SPR) imaging techniques allow for label-free mapping the mass-distribution within single living cells, leading to great expansions in biomolecular interactions studies from homogeneous substrates functionalized with purified biomolecules to heterogeneous substrates containing individual living cells.
NASA Technical Reports Server (NTRS)
Hinton, David A.
1989-01-01
Numerous air carrier accidents and incidents result from encounters with the atmospheric wind shear associated with microburst phenomena, in some cases resulting in heavy loss of life. An important issue in current wind shear research is how to best manage aircraft performance during an inadvertent wind shear encounter. The goals of this study were to: (1) develop techniques and guidance for maximizing an aircraft's ability to recover from microburst encounters following takeoff, (2) develop an understanding of how theoretical predictions of wind shear recovery performance might be achieved in actual use, and (3) gain insight into the piloting factors associated with recovery from microburst encounters. Three recovery strategies were implemented and tested in piloted simulation. Results show that a recovery strategy based on flying a flight path angle schedule produces improved performance over constant pitch attitude or acceleration-based recovery techniques. The best recovery technique was initially counterintuitive to the pilots who participated in the study. Evidence was found to indicate that the techniques required for flight through the turbulent vortex of a microburst may differ from the techniques being developed using classical, nonturbulent microburst models.
[Three-dimensional endoscopic endonasal study of skull base anatomy].
Abarca-Olivas, Javier; Monjas-Cánovas, Irene; López-Álvarez, Beatriz; Lloret-García, Jaime; Sanchez-del Campo, Jose; Gras-Albert, Juan Ramon; Moreno-López, Pedro
2014-01-01
Training in dissection of the paranasal sinuses and the skull base is essential for anatomical understanding and correct surgical techniques. Three-dimensional (3D) visualisation of endoscopic skull base anatomy increases spatial orientation and allows depth perception. To show endoscopic skull base anatomy based on the 3D technique. We performed endoscopic dissection in cadaveric specimens fixed with formalin and with the Thiel technique, both prepared using intravascular injection of coloured material. Endonasal approaches were performed with conventional 2D endoscopes. Then we applied the 3D anaglyph technique to illustrate the pictures in 3D. The most important anatomical structures and landmarks of the sellar region under endonasal endoscopic vision are illustrated in 3D images. The skull base consists of complex bony and neurovascular structures. Experience with cadaver dissection is essential to understand complex anatomy and develop surgical skills. A 3D view constitutes a useful tool for understanding skull base anatomy. Copyright © 2012 Sociedad Española de Neurocirugía. Published by Elsevier España. All rights reserved.
Flight test derived heating math models for critical locations on the orbiter during reentry
NASA Technical Reports Server (NTRS)
Hertzler, E. K.; Phillips, P. W.
1983-01-01
An analysis technique was developed for expanding the aerothermodynamic envelope of the Space Shuttle without subjecting the vehicle to sustained flight at more stressing heating conditions. A transient analysis program was developed to take advantage of the transient maneuvers that were flown as part of this analysis technique. Heat rates were derived from flight test data for various locations on the orbiter. The flight derived heat rates were used to update heating models based on predicted data. Future missions were then analyzed based on these flight adjusted models. A technique for comparing flight and predicted heating rate data and the extrapolation of the data to predict the aerothermodynamic environment of future missions is presented.
Modified Redundancy based Technique—a New Approach to Combat Error Propagation Effect of AES
NASA Astrophysics Data System (ADS)
Sarkar, B.; Bhunia, C. T.; Maulik, U.
2012-06-01
Advanced encryption standard (AES) is a great research challenge. It has been developed to replace the data encryption standard (DES). AES suffers from a major limitation of error propagation effect. To tackle this limitation, two methods are available. One is redundancy based technique and the other one is bite based parity technique. The first one has a significant advantage of correcting any error on definite term over the second one but at the cost of higher level of overhead and hence lowering the processing speed. In this paper, a new approach based on the redundancy based technique is proposed that would certainly speed up the process of reliable encryption and hence the secured communication.
[Organization development of the public health system].
Pfaff, Holger; Klein, Jürgen
2002-05-15
Changes in the German health care system require changes in health care institutions. Organizational development (OD) techniques can help them to cope successfully with their changing environment. OD is defined as a collective process of learning aiming to induce intended organizational change. OD is based on social science methods and conducted by process-oriented consultants. In contrast to techniques of organizational design, OD is characterized by employee participation. One of the most important elements of OD is the so-called "survey-feedback-technique". Five examples illustrate how the survey-feedback-technique can be used to facilitate organisational learning. OD technique supports necessary change in health care organizations. It should be used more frequently.
Implementation plans included in World Health Organisation guidelines.
Wang, Zhicheng; Norris, Susan L; Bero, Lisa
2016-05-20
The implementation of high-quality guidelines is essential to improve clinical practice and public health. The World Health Organisation (WHO) develops evidence-based public health and other guidelines that are used or adapted by countries around the world. Detailed implementation plans are often necessary for local policymakers to properly use the guidelines developed by WHO. This paper describes the plans for guideline implementation reported in WHO guidelines and indicates which of these plans are evidence-based. We conducted a content analysis of the implementation sections of WHO guidelines approved by the WHO guideline review committee between December 2007 and May 2015. The implementation techniques reported in each guideline were coded according to the Cochrane Collaboration's Effective Practice and Organisation of Care (EPOC) taxonomy and classified as passive, active or policy strategies. The frequencies of implementation techniques are reported. The WHO guidelines (n = 123) analysed mentioned implementation techniques 800 times, although most mentioned implementation techniques very briefly, if at all. Passive strategies (21 %, 167/800) and general policy strategies (62 %, 496/800) occurred most often. Evidence-based active implementation methods were generally neglected with no guideline mentioning reminders (computerised or paper) and only one mentioning a multifaceted approach. Many guidelines contained implementation sections that were identical to those used in older guidelines produced by the same WHO technical unit. The prevalence of passive and policy-based implementation techniques as opposed to evidence-based active techniques suggests that WHO guidelines should contain stronger guidance for implementation. This could include structured and increased detail on implementation considerations, accompanying or linked documents that provide information on what is needed to contextualise or adapt a guideline and specific options from among evidence-based implementation strategies.
Microneedle Coating Techniques for Transdermal Drug Delivery
Haj-Ahmad, Rita; Khan, Hashim; Arshad, Muhammad Sohail; Rasekh, Manoochehr; Hussain, Amjad; Walsh, Susannah; Li, Xiang; Chang, Ming-Wei; Ahmad, Zeeshan
2015-01-01
Drug administration via the transdermal route is an evolving field that provides an alternative to oral and parenteral routes of therapy. Several microneedle (MN) based approaches have been developed. Among these, coated MNs (typically where drug is deposited on MN tips) are a minimally invasive method to deliver drugs and vaccines through the skin. In this review, we describe several processes to coat MNs. These include dip coating, gas jet drying, spray coating, electrohydrodynamic atomisation (EHDA) based processes and piezoelectric inkjet printing. Examples of process mechanisms, conditions and tested formulations are provided. As these processes are independent techniques, modifications to facilitate MN coatings are elucidated. In summary, the outcomes and potential value for each technique provides opportunities to overcome formulation or dosage form limitations. While there are significant developments in solid degradable MNs, coated MNs (through the various techniques described) have potential to be utilized in personalized drug delivery via controlled deposition onto MN templates. PMID:26556364
NASA Astrophysics Data System (ADS)
Natali, Marco; Reggente, Melania; Passeri, Daniele; Rossi, Marco
2016-06-01
The development of polymer-based nanocomposites to be used in critical thermal environments requires the characterization of their mechanical properties, which are related to their chemical composition, size, morphology and operating temperature. Atomic force microscopy (AFM) has been proven to be a useful tool to develop techniques for the mechanical characterization of these materials, thanks to its nanometer lateral resolution and to the capability of exerting ultra-low loads, down to the piconewton range. In this work, we demonstrate two techniques, one quasi-static, i.e., AFM-based indentation (I-AFM), and one dynamic, i.e., contact resonance AFM (CR-AFM), for the mechanical characterization of compliant materials at variable temperature. A cross-validation of I-AFM and CR-AFM has been performed by comparing the results obtained on two reference materials, i.e., low-density polyethylene (LDPE) and polycarbonate (PC), which demonstrated the accuracy of the techniques.
Cryogenic Cathode Cooling Techniques for Improved SABRE Extraction Ion Diode Li Beam Generation
NASA Astrophysics Data System (ADS)
Hanson, D. L.; Johnston, R. R.; Cuneo, M. E.; Menge, P. R.; Fowler, W. E.; Armijo, J.; Nielsen, D. S.; Petmecky, D.
1997-11-01
We are developing techniques for cryogenic cooling of the SABRE extraction ion diode cathode that, combined with source cleaning, should improve the purity and brightness of Li beams for ICF light ion fusion. By liquid helium (LHe) cathode cooling, we have been able to maintain A-K gap base pressures in the range of 5 - 7x10-8 Torr for about 45 minutes. These base pressures extend the monolayer formation time for the worst beam contaminants (H2 and water vapor) to 10 - 100 sec or longer, which should allow the accelerator to be fired without significant Li source recontamination. This technique is compatible with He glow discharge cleaning, laser cleaning, and in situ Li deposition. We are also developing techniques for Ti-gettering of H2 and for cryogenic cooling of cathode electrodes to delay cathode plasma expansion.
NASA Astrophysics Data System (ADS)
Henderson, Charles; Yerushalmi, Edit; Kuo, Vince H.; Heller, Kenneth; Heller, Patricia
2007-12-01
To identify and describe the basis upon which instructors make curricular and pedagogical decisions, we have developed an artifact-based interview and an analysis technique based on multilayered concept maps. The policy capturing technique used in the interview asks instructors to make judgments about concrete instructional artifacts similar to those they likely encounter in their teaching environment. The analysis procedure alternatively employs both an a priori systems view analysis and an emergent categorization to construct a multilayered concept map, which is a hierarchically arranged set of concept maps where child maps include more details than parent maps. Although our goal was to develop a model of physics faculty beliefs about the teaching and learning of problem solving in the context of an introductory calculus-based physics course, the techniques described here are applicable to a variety of situations in which instructors make decisions that influence teaching and learning.
NASA Astrophysics Data System (ADS)
Patsariya, Ajay; Rai, Shiwani; Kumar, Yogendra, Dr.; Kirar, Mukesh, Dr.
2017-08-01
The energy crisis particularly with developing GDPs, has bring up to a new panorama of sustainable power source like solar energy, which has encountered huge development. Progressively high infiltration level of photovoltaic (PV) era emerges in keen matrix. Sunlight based power is irregular and variable, as the sun based source at the ground level is exceedingly subject to overcast cover inconstancy, environmental vaporized levels, and other climate parameters. The inalienable inconstancy of substantial scale sun based era acquaints huge difficulties with keen lattice vitality administration. Exact determining of sun powered power/irradiance is basic to secure financial operation of the shrewd framework. In this paper a noble TLBO-MPPT technique has been proposed to address the vitality of solar energy. A comparative analysis has been presented between conventional PO, IC and the proposed MPPT technique. The research has been done on Matlab Simulink software version 2013.
DOT National Transportation Integrated Search
1976-04-01
The development and testing of incident detection algorithms was based on Los Angeles and Minneapolis freeway surveillance data. Algorithms considered were based on times series and pattern recognition techniques. Attention was given to the effects o...
The Empirical Investigation of Perspective-Based Reading
NASA Technical Reports Server (NTRS)
Basili, Victor R.; Green, Scott; Laitenberger, Oliver; Shull, Forrest; Sorumgard, Sivert; Zelkowitz, Marvin V.
1995-01-01
We consider reading techniques a fundamental means of achieving high quality software. Due to lack of research in this area, we are experimenting with the application and comparison of various reading techniques. This paper deals with our experiences with Perspective Based Reading (PBR) a particular reading technique for requirement documents. The goal of PBR is to provide operation scenarios where members of a review team read a document from a particular perspective (eg., tester, developer, user). Our assumption is that the combination of different perspective provides better coverage of the document than the same number of readers using their usual technique. To test the efficacy of PBR, we conducted two runs of a controlled experiment in the environment of NASA GSFC Software Engineering Laboratory (SEL), using developers from the environment. The subjects read two types of documents, one generic in nature and the other from the NASA Domain, using two reading techniques, PBR and their usual technique. The results from these experiment as well as the experimental design, are presented and analyzed. When there is a statistically significant distinction, PBR performs better than the subjects' usual technique. However, PBR appears to be more effective on the generic documents than on the NASA documents.
ERIC Educational Resources Information Center
Small, Jason W.; Lee, Jon; Frey, Andy J.; Seeley, John R.; Walker, Hill M.
2014-01-01
As specialized instructional support personnel begin learning and using motivational interviewing (MI) techniques in school-based settings, there is growing need for context-specific measures to assess initial MI skill development. In this article, we describe the iterative development and preliminary evaluation of two measures of MI skill adapted…
A Weighted Least Squares Approach To Robustify Least Squares Estimates.
ERIC Educational Resources Information Center
Lin, Chowhong; Davenport, Ernest C., Jr.
This study developed a robust linear regression technique based on the idea of weighted least squares. In this technique, a subsample of the full data of interest is drawn, based on a measure of distance, and an initial set of regression coefficients is calculated. The rest of the data points are then taken into the subsample, one after another,…
Hu, Yunzi; Daoud, Walid A.; Cheuk, Kevin Ka Leung; Lin, Carol Sze Ki
2016-01-01
Polycondensation and ring-opening polymerization are two important polymer synthesis methods. Poly(lactic acid), the most typical biodegradable polymer, has been researched extensively from 1900s. It is of significant importance to have an up-to-date review on the recent improvement in techniques for biodegradable polymers. This review takes poly(lactic acid) as the example to present newly developed polymer synthesis techniques on polycondensation and ring-opening polymerization reported in the recent decade (2005–2015) on the basis of industrial technique modifications and advanced laboratory research. Different polymerization methods, including various solvents, heating programs, reaction apparatus and catalyst systems, are summarized and compared with the current industrial production situation. Newly developed modification techniques for polymer properties improvement are also discussed based on the case of poly(lactic acid). PMID:28773260
Defogging of road images using gain coefficient-based trilateral filter
NASA Astrophysics Data System (ADS)
Singh, Dilbag; Kumar, Vijay
2018-01-01
Poor weather conditions are responsible for most of the road accidents year in and year out. Poor weather conditions, such as fog, degrade the visibility of objects. Thus, it becomes difficult for drivers to identify the vehicles in a foggy environment. The dark channel prior (DCP)-based defogging techniques have been found to be an efficient way to remove fog from road images. However, it produces poor results when image objects are inherently similar to airlight and no shadow is cast on them. To eliminate this problem, a modified restoration model-based DCP is developed to remove the fog from road images. The transmission map is also refined by developing a gain coefficient-based trilateral filter. Thus, the proposed technique has an ability to remove fog from road images in an effective manner. The proposed technique is compared with seven well-known defogging techniques on two benchmark foggy images datasets and five real-time foggy images. The experimental results demonstrate that the proposed approach is able to remove the different types of fog from roadside images as well as significantly improve the image's visibility. It also reveals that the restored image has little or no artifacts.
Next Generation LOCAD-PTS Cartridge Development
NASA Technical Reports Server (NTRS)
Morris, H.; Nutter, D.; Weite, E.; Wells, M.; Maule, J.; Damon, M.; Monaco, L.; Steele, A.; Wainwright, N.
2008-01-01
Future astrobiology exploration missions will require rapid, point-of-use techniques for surface science experiments and contamination monitoring. The Lab-On-a-Chip Application Development (LOCAD) team is developing operational instruments that advance spaceflight technologies to molecular-based methods. Currently, LOCAD-Portable Test System (PTS) is quantifying levels of the bacterial molecule endotoxin onboard the Internatioal Space Station. Future research and development will focus on more sensitive molecular techniques that expand the number of compounds detected to include beta-glucan from fungal cell walls.
Plasticity models of material variability based on uncertainty quantification techniques
DOE Office of Scientific and Technical Information (OSTI.GOV)
Jones, Reese E.; Rizzi, Francesco; Boyce, Brad
The advent of fabrication techniques like additive manufacturing has focused attention on the considerable variability of material response due to defects and other micro-structural aspects. This variability motivates the development of an enhanced design methodology that incorporates inherent material variability to provide robust predictions of performance. In this work, we develop plasticity models capable of representing the distribution of mechanical responses observed in experiments using traditional plasticity models of the mean response and recently developed uncertainty quantification (UQ) techniques. Lastly, we demonstrate that the new method provides predictive realizations that are superior to more traditional ones, and how these UQmore » techniques can be used in model selection and assessing the quality of calibrated physical parameters.« less
NASA Astrophysics Data System (ADS)
Vanvyve, E.; Magontier, P.; Vandenberghe, F. C.; Delle Monache, L.; Dickinson, K.
2012-12-01
Wind energy is amongst the fastest growing sources of renewable energy in the U.S. and could supply up to 20 % of the U.S power production by 2030. An accurate and reliable wind resource assessment for prospective wind farm sites is a challenging task, yet is crucial for evaluating the long-term profitability and feasibility of a potential development. We have developed an accurate and computationally efficient wind resource assessment technique for prospective wind farm sites, which incorporates innovative statistical techniques and the new NASA Earth science dataset MERRA. This technique produces a wind resource estimate that is more accurate than that obtained by the wind energy industry's standard technique, while providing a reliable quantification of its uncertainty. The focus now is on evaluating the socio-economic value of this new technique upon using the industry's standard technique. Would it yield lower financing costs? Could it result in lower electricity prices? Are there further down-the-line positive consequences, e.g. job creation, time saved, greenhouse gas decrease? Ultimately, we expect our results will inform efforts to refine and disseminate the new technique to support the development of the U.S. renewable energy infrastructure. In order to address the above questions, we are carrying out a cost-benefit analysis based on the net present worth of the technique. We will describe this approach, including the cash-flow process of wind farm financing, how the wind resource assessment factors in, and will present current results for various hypothetical candidate wind farm sites.
Tandon, Nikhil; Kalra, Sanjay; Balhara, Yatan Pal Singh; Baruah, Manash P.; Chadha, Manoj; Chandalia, Hemraj B.; Chowdhury, Subhankar; Jothydev, Kesavadev; Kumar, Prasanna K. M.; V., Madhu S.; Mithal, Ambrish; Modi, Sonal; Pitale, Shailesh; Sahay, Rakesh; Shukla, Rishi; Sundaram, Annamalai; Unnikrishnan, Ambika G.; Wangnoo, Subhash K.
2015-01-01
As injectable therapies such as human insulin, insulin analogs, and glucagon-like peptide-1 receptor agonists are used to manage diabetes, correct injection technique is vital for the achievement of glycemic control. The forum for injection technique India acknowledged this need for the first time in India and worked to develop evidence-based recommendations on insulin injection technique, to assist healthcare practitioners in their clinical practice. PMID:25932385
NASA Technical Reports Server (NTRS)
Lih, Shyh-Shiuh; Bar-Cohen, Yoseph; Lee, Hyeong Jae; Takano, Nobuyuki; Bao, Xiaoqi
2013-01-01
An advanced signal processing methodology is being developed to monitor the height of condensed water thru the wall of a steel pipe while operating at temperatures as high as 250deg. Using existing techniques, previous study indicated that, when the water height is low or there is disturbance in the environment, the predicted water height may not be accurate. In recent years, the use of the autocorrelation and envelope techniques in the signal processing has been demonstrated to be a very useful tool for practical applications. In this paper, various signal processing techniques including the auto correlation, Hilbert transform, and the Shannon Energy Envelope methods were studied and implemented to determine the water height in the steam pipe. The results have shown that the developed method provides a good capability for monitoring the height in the regular conditions. An alternative solution for shallow water or no water conditions based on a developed hybrid method based on Hilbert transform (HT) with a high pass filter and using the optimized windowing technique is suggested. Further development of the reported methods would provide a powerful tool for the identification of the disturbances of water height inside the pipe.
Robotic Anterior and Midline Skull Base Surgery: Preclinical Investigations
DOE Office of Scientific and Technical Information (OSTI.GOV)
O'Malley, Bert W.; Weinstein, Gregory S.
Purpose: To develop a minimally invasive surgical technique to access the midline and anterior skull base using the optical and technical advantages of robotic surgical instrumentation. Methods and Materials: Ten experimental procedures focusing on approaches to the nasopharynx, clivus, sphenoid, pituitary sella, and suprasellar regions were performed on one cadaver and one live mongrel dog. Both the cadaver and canine procedures were performed in an approved training facility using the da Vinci Surgical Robot. For the canine experiments, a transoral robotic surgery (TORS) approach was used, and for the cadaver a newly developed combined cervical-transoral robotic surgery (C-TORS) approach wasmore » investigated and compared with standard TORS. The ability to access and dissect tissues within the various areas of the midline and anterior skull base were evaluated, and techniques to enhance visualization and instrumentation were developed. Results: Standard TORS approaches did not provide adequate access to the midline and anterior skull base; however, the newly developed C-TORS approach was successful in providing the surgical access to these regions of the skull base. Conclusion: Robotic surgery is an exciting minimally invasive approach to the skull base that warrants continued preclinical investigation and development.« less
Passive Optical Locking Techniques for Diode Lasers
NASA Astrophysics Data System (ADS)
Zhang, Quan
1995-01-01
Most current diode-based nonlinear frequency converters utilize electronic frequency locking techniques. However, this type of locking technique typically involves very complex electronics, and suffers the 'power-drop' problem. This dissertation is devoted to the development of an all-optical passive locking technique that locks the diode laser frequency to the external cavity resonance stably without using any kind of electronic servo. The amplitude noise problem associated with the strong optical locking has been studied. Single-mode operation of a passively locked single-stripe diode with an amplitude stability better than 1% has been achieved. This passive optical locking technique applies to broad-area diodes as well as single-stripe diodes, and can be easily used to generate blue light. A schematic of a milliwatt level blue laser based on the single-stripe diode locking technique has been proposed. A 120 mW 467 nm blue laser has been built using the tapered amplifier locking technique. In addition to diode-based blue lasers, this passive locking technique has applications in nonlinear frequency conversions, resonant spectroscopy, particle counter devices, telecommunications, and medical devices.
Compiler-Assisted Multiple Instruction Rollback Recovery Using a Read Buffer. Ph.D. Thesis
NASA Technical Reports Server (NTRS)
Alewine, Neal Jon
1993-01-01
Multiple instruction rollback (MIR) is a technique to provide rapid recovery from transient processor failures and was implemented in hardware by researchers and slow in mainframe computers. Hardware-based MIR designs eliminate rollback data hazards by providing data redundancy implemented in hardware. Compiler-based MIR designs were also developed which remove rollback data hazards directly with data flow manipulations, thus eliminating the need for most data redundancy hardware. Compiler-assisted techniques to achieve multiple instruction rollback recovery are addressed. It is observed that data some hazards resulting from instruction rollback can be resolved more efficiently by providing hardware redundancy while others are resolved more efficiently with compiler transformations. A compiler-assisted multiple instruction rollback scheme is developed which combines hardware-implemented data redundancy with compiler-driven hazard removal transformations. Experimental performance evaluations were conducted which indicate improved efficiency over previous hardware-based and compiler-based schemes. Various enhancements to the compiler transformations and to the data redundancy hardware developed for the compiler-assisted MIR scheme are described and evaluated. The final topic deals with the application of compiler-assisted MIR techniques to aid in exception repair and branch repair in a speculative execution architecture.
Teaching Techniques in Clinical Chemistry.
ERIC Educational Resources Information Center
Wilson, Diane
This master's thesis presents several instructional methods and techniques developed for each of eleven topics or subject areas in clinical chemistry: carbohydrate metabolism, lipid metabolism, diagnostic enzymology, endocrinology, toxicology, quality control, electrolytes, acid base balance, hepatic function, nonprotein nitrogenous compounds, and…
Improved Photoresist Coating for Making CNT Field Emitters
NASA Technical Reports Server (NTRS)
Toda, Risaku; Manohara, Harish
2009-01-01
An improved photoresist-coating technique has been developed for use in the fabrication of carbon-nanotube- (CNT) based field emitters is described. The improved photoresist coating technique overcomes what, heretofore, has been a major difficulty in the fabrication process.
Advances in Testing Techniques for Digital Microfluidic Biochips
Shukla, Vineeta; Hussin, Fawnizu Azmadi; Hamid, Nor Hisham; Zain Ali, Noohul Basheer
2017-01-01
With the advancement of digital microfluidics technology, applications such as on-chip DNA analysis, point of care diagnosis and automated drug discovery are common nowadays. The use of Digital Microfluidics Biochips (DMFBs) in disease assessment and recognition of target molecules had become popular during the past few years. The reliability of these DMFBs is crucial when they are used in various medical applications. Errors found in these biochips are mainly due to the defects developed during droplet manipulation, chip degradation and inaccuracies in the bio-assay experiments. The recently proposed Micro-electrode-dot Array (MEDA)-based DMFBs involve both fluidic and electronic domains in the micro-electrode cell. Thus, the testing techniques for these biochips should be revised in order to ensure proper functionality. This paper describes recent advances in the testing technologies for digital microfluidics biochips, which would serve as a useful platform for developing revised/new testing techniques for MEDA-based biochips. Therefore, the relevancy of these techniques with respect to testing of MEDA-based biochips is analyzed in order to exploit the full potential of these biochips. PMID:28749411
Speckle-based at-wavelength metrology of X-ray mirrors with super accuracy
DOE Office of Scientific and Technical Information (OSTI.GOV)
Kashyap, Yogesh; Wang, Hongchang; Sawhney, Kawal, E-mail: kawal.sawhney@diamond.ac.uk
2016-05-15
X-ray active mirrors, such as bimorph and mechanically bendable mirrors, are increasingly being used on beamlines at modern synchrotron source facilities to generate either focused or “tophat” beams. As well as optical tests in the metrology lab, it is becoming increasingly important to optimise and characterise active optics under actual beamline operating conditions. Recently developed X-ray speckle-based at-wavelength metrology technique has shown great potential. The technique has been established and further developed at the Diamond Light Source and is increasingly being used to optimise active mirrors. Details of the X-ray speckle-based at-wavelength metrology technique and an example of its applicabilitymore » in characterising and optimising a micro-focusing bimorph X-ray mirror are presented. Importantly, an unprecedented angular sensitivity in the range of two nanoradians for measuring the slope error of an optical surface has been demonstrated. Such a super precision metrology technique will be beneficial to the manufacturers of polished mirrors and also in optimization of beam shaping during experiments.« less
Advances in Testing Techniques for Digital Microfluidic Biochips.
Shukla, Vineeta; Hussin, Fawnizu Azmadi; Hamid, Nor Hisham; Zain Ali, Noohul Basheer
2017-07-27
With the advancement of digital microfluidics technology, applications such as on-chip DNA analysis, point of care diagnosis and automated drug discovery are common nowadays. The use of Digital Microfluidics Biochips (DMFBs) in disease assessment and recognition of target molecules had become popular during the past few years. The reliability of these DMFBs is crucial when they are used in various medical applications. Errors found in these biochips are mainly due to the defects developed during droplet manipulation, chip degradation and inaccuracies in the bio-assay experiments. The recently proposed Micro-electrode-dot Array (MEDA)-based DMFBs involve both fluidic and electronic domains in the micro-electrode cell. Thus, the testing techniques for these biochips should be revised in order to ensure proper functionality. This paper describes recent advances in the testing technologies for digital microfluidics biochips, which would serve as a useful platform for developing revised/new testing techniques for MEDA-based biochips. Therefore, the relevancy of these techniques with respect to testing of MEDA-based biochips is analyzed in order to exploit the full potential of these biochips.
NASA Astrophysics Data System (ADS)
Cerwin, Steve; Barnes, Julie; Kell, Scott; Walters, Mark
2003-09-01
This paper describes development and application of a novel method to accomplish real-time solid angle acoustic direction finding using two 8-element orthogonal microphone arrays. The developed prototype system was intended for localization and signature recognition of ground-based sounds from a small UAV. Recent advances in computer speeds have enabled the implementation of microphone arrays in many audio applications. Still, the real-time presentation of a two-dimensional sound field for the purpose of audio target localization is computationally challenging. In order to overcome this challenge, a crosspower spectrum phase1 (CSP) technique was applied to each 8-element arm of a 16-element cross array to provide audio target localization. In this paper, we describe the technique and compare it with two other commonly used techniques; Cross-Spectral Matrix2 and MUSIC3. The results show that the CSP technique applied to two 8-element orthogonal arrays provides a computationally efficient solution with reasonable accuracy and tolerable artifacts, sufficient for real-time applications. Additional topics include development of a synchronized 16-channel transmitter and receiver to relay the airborne data to the ground-based processor and presentation of test data demonstrating both ground-mounted operation and airborne localization of ground-based gunshots and loud engine sounds.
Remote sensing of high-latitude ionization profiles by ground-based and spaceborne instrumentation
NASA Technical Reports Server (NTRS)
Vondrak, R. R.
1981-01-01
Ionospheric specification and modeling are now largely based on data provided by active remote sensing with radiowave techniques (ionosondes, incoherent-scatter radars, and satellite beacons). More recently, passive remote sensing techniques have been developed that can be used to monitor quantitatively the spatial distribution of high-latitude E-region ionization. These passive methods depend on the measurement, or inference, of the energy distribution of precipitating kilovolt electrons, the principal source of the nighttime E-region at high latitudes. To validate these techniques, coordinated measurements of the auroral ionosphere have been made with the Chatanika incoherent-scatter radar and a variety of ground-based and spaceborne sensors
Obstacle Avoidance On Roadways Using Range Data
NASA Astrophysics Data System (ADS)
Dunlay, R. Terry; Morgenthaler, David G.
1987-02-01
This report describes range data based obstacle avoidance techniques developed for use on an autonomous road-following robot vehicle. The purpose of these techniques is to detect and locate obstacles present in a road environment for navigation of a robot vehicle equipped with an active laser-based range sensor. Techniques are presented for obstacle detection, obstacle location, and coordinate transformations needed in the construction of Scene Models (symbolic structures representing the 3-D obstacle boundaries used by the vehicle's Navigator for path planning). These techniques have been successfully tested on an outdoor robotic vehicle, the Autonomous Land Vehicle (ALV), at speeds up to 3.5 km/hour.
Field results of antifouling techniques for optical instruments
Strahle, W.J.; Hotchkiss, F.S.; Martini, Marinna A.
1998-01-01
An anti-fouling technique is developed for the protection of optical instruments from biofouling which leaches a bromide compound into a sample chamber and pumps new water into the chamber prior to measurement. The primary advantage of using bromide is that it is less toxic than the metal-based antifoulants. The drawback of the bromide technique is also discussed.
ERIC Educational Resources Information Center
Lamichhane, Kamal
2011-01-01
There are many techniques for communicating with individuals who are deaf-blind. In Japan, another technique, known as fingerbraille, is also used. Fingerbraille is a tactile method of communication that is based on Japanese braille script. This article introduces the reader to the history of the development of this communication technique and to…
Joining of Silicon Carbide-Based Ceramics by Reaction Forming Method
NASA Technical Reports Server (NTRS)
Singh, M.; Kiser, J. D.
1997-01-01
Recently, there has been a surge of interest in the development and testing of silicon-based ceramics and composite components for a number of aerospace and ground based systems. The designs often require fabrication of complex shaped parts which can be quite expensive. One attractive way of achieving this goal is to build up complex shapes by joining together geometrically simple shapes. However, the joints should have good mechanical strength and environmental stability comparable to the bulk materials. These joints should also be able to maintain their structural integrity at high temperatures. In addition, the joining technique should be practical, reliable, and affordable. Thus, joining has been recognized as one of the enabling technologies for the successful utilization of silicon carbide based ceramic components in high temperature applications. Overviews of various joining techniques, i.e., mechanical fastening, adhesive bonding, welding, brazing, and soldering have been provided in recent publications. The majority of the techniques used today are based on the joining of monolithic ceramics with metals either by diffusion bonding, metal brazing, brazing with oxides and oxynitrides, or diffusion welding. These techniques need either very high temperatures for processing or hot pressing (high pressures). The joints produced by these techniques have different thermal expansion coefficients than the ceramic materials, which creates a stress concentration in the joint area. The use temperatures for these joints are around 700 C. Ceramic joint interlayers have been developed as a means of obtaining high temperature joints. These joint interlayers have been produced via pre-ceramic polymers, in-situ displacement reactions, and reaction bonding techniques. Joints produced by the pre-ceramic polymer approach exhibit a large amounts of porosity and poor mechanical properties. On the other hand, hot pressing or high pressures are needed for in-situ displacement reactions and reaction bonding techniques. Due to the equipment required, these techniques are impractical for joining large or complex shaped components.
Nano-Al Based Energetics: Rapid Heating Studies and a New Preparation Technique
NASA Astrophysics Data System (ADS)
Sullivan, Kyle; Kuntz, Josh; Gash, Alex; Zachariah, Michael
2011-06-01
Nano-Al based thermites have become an attractive alternative to traditional energetic formulations due to their increased energy density and high reactivity. Understanding the intrinsic reaction mechanism has been a difficult task, largely due to the lack of experimental techniques capable of rapidly and uniform heating a sample (~104- 108 K/s). The current work presents several studies on nano-Al based thermites, using rapid heating techniques. A new mechanism termed a Reactive Sintering Mechanism is proposed for nano-Al based thermites. In addition, new experimental techniques for nanocomposite thermite deposition onto thin Pt electrodes will be discussed. This combined technique will offer more precise control of the deposition, and will serve to further our understanding of the intrinsic reaction mechanism of rapidly heated energetic systems. An improved mechanistic understanding will lead to the development of optimized formulations and architectures. This work performed under the auspices of the U.S. Department of Energy by Lawrence Livermore National Laboratory under Contract DE-AC52-07NA27344.
A spline-based parameter estimation technique for static models of elastic structures
NASA Technical Reports Server (NTRS)
Dutt, P.; Taasan, S.
1986-01-01
The problem of identifying the spatially varying coefficient of elasticity using an observed solution to the forward problem is considered. Under appropriate conditions this problem can be treated as a first order hyperbolic equation in the unknown coefficient. Some continuous dependence results are developed for this problem and a spline-based technique is proposed for approximating the unknown coefficient, based on these results. The convergence of the numerical scheme is established and error estimates obtained.
A Novel Flame Thermometer Based on the Doppler Width of Ro-Vibrational Transitions
1993-05-01
flame thermometric technique based on the infrared spectroscopy of flame species is reported. It involves the use of a narrow linewidth (ɘ.001 cm-1...study. Output radiation from the laser diode is incident onto a parabolic aluminum mirror which collimates the beam and directs it into a monochromator...infrared beam. 5. SUMMARY AND CONCLUSIONS A novel flame thermometric technique has been developed which is based on the infrared spectroscopy of flame
The Development of Neutron Radiography and Tomography on a SLOWPOKE-2 Reactor
NASA Astrophysics Data System (ADS)
Bennett, L. G. I.; Lewis, W. J.; Hungler, P. C.
Development of neutron radiography at the Royal Military College of Canada (RMC) started by trying to interest the Royal Canadian Air Force (RCAF) in this new non-destructive testing (NDT) technique. A Californium-252 based device was ordered and then installed at RMC for development of applicable techniques for aircraft by the first author. A second and transportable device was then designed, modified and used in trials at RCAF Bases and other locations for one year. This activity was the only foreign loan of the U.S. Californium Loan Program. Around this time, SLOWPOKE-2 reactors were being installed at four Canadian universities, while a new science and engineering building was being built at RMC. A reactor pool was incorporated and efforts to procure a reactor succeeded a decade later with a SLOWPOKE-2 reactor being installed at RMC. The only modification by the vendor for RMC was a thermal column replacing an irradiation site inside the reactor container for a later installation of a neutron beam tube (NBT). Development of a working NBT took several years, starting with the second author. A demonstration of the actual worth of neutron radiography took place with a CF-18 Hornet aircraft being neutron and X-radiographed at McClellan Air Force Base, Sacramento, CA. This inspection was followed by one of the rudders that had indications of water ingress being radiographed successfully at RMC just after the NBT became functional. The next step was to develop a neutron radioscopy system (NRS), initially employing film and then digital imaging, and is in use today for all flight control surfaces (FCS). With the third author, a technique capable of removing water from affected FCS was developed at RMC. Heating equipment and a vacuum system were utilized to carefully remove the water. This technique was proven using a sequence of near real time neutron images obtained during the drying process. The results of the drying process were correlated with a relative humidity gauge and an NDT technique that could be performed at Canadian Forces (CF) Bases was developed. In order to determine the structural integrity of the component having undergone this water removal, further research was required into the effect of water inside composite honeycomb structures. This need has led to the present development of neutron tomography on the reactor at RMC, which is capable of determining the exact location of water ingress inside composite components. This technique has been successfully applied to coupons as well as to complete rudders.
Digital Mapping Techniques '09-Workshop Proceedings, Morgantown, West Virginia, May 10-13, 2009
Soller, David R.
2011-01-01
As in the previous years' meetings, the objective was to foster informal discussion and exchange of technical information, principally in order to develop more efficient methods for digital mapping, cartography, GIS analysis, and information management. At this meeting, oral and poster presentations and special discussion sessions emphasized (1) methods for creating and publishing map products (here, "publishing" includes Web-based release); (2) field data capture software and techniques, including the use of LiDAR; (3) digital cartographic techniques; (4) migration of digital maps into ArcGIS Geodatabase format; (5) analytical GIS techniques; and (6) continued development of the National Geologic Map Database.
NASA Technical Reports Server (NTRS)
Garai, Anirban; Diosady, Laslo T.; Murman, Scott M.; Madavan, Nateri K.
2016-01-01
The perfectly matched layer (PML) technique is developed in the context of a high- order spectral-element Discontinuous-Galerkin (DG) method. The technique is applied to a range of test cases and is shown to be superior compared to other approaches, such as those based on using characteristic boundary conditions and sponge layers, for treating the inflow and outflow boundaries of computational domains. In general, the PML technique improves the quality of the numerical results for simulations of practical flow configurations, but it also exhibits some instabilities for large perturbations. A preliminary analysis that attempts to understand the source of these instabilities is discussed.
[Progress in industrial bioprocess engineering in China].
Zhuang, Yingping; Chen, Hongzhang; Xia, Jianye; Tang, Wenjun; Zhao, Zhimin
2015-06-01
The advances of industrial biotechnology highly depend on the development of industrial bioprocess researches. In China, we are facing several challenges because of a huge national industrial fermentation capacity. The industrial bioprocess development experienced several main stages. This work mainly reviews the development of the industrial bioprocess in China during the past 30 or 40 years: including the early stage kinetics model study derived from classical chemical engineering, researching method based on control theory, multiple-parameter analysis techniques of on-line measuring instruments and techniques, and multi-scale analysis theory, and also solid state fermentation techniques and fermenters. In addition, the cutting edge of bioprocess engineering was also addressed.
Tip-Based Nanofabrication for Scalable Manufacturing
Hu, Huan; Kim, Hoe; Somnath, Suhas
2017-03-16
Tip-based nanofabrication (TBN) is a family of emerging nanofabrication techniques that use a nanometer scale tip to fabricate nanostructures. Here in this review, we first introduce the history of the TBN and the technology development. We then briefly review various TBN techniques that use different physical or chemical mechanisms to fabricate features and discuss some of the state-of-the-art techniques. Subsequently, we focus on those TBN methods that have demonstrated potential to scale up the manufacturing throughput. Finally, we discuss several research directions that are essential for making TBN a scalable nano-manufacturing technology.
Tip-Based Nanofabrication for Scalable Manufacturing
DOE Office of Scientific and Technical Information (OSTI.GOV)
Hu, Huan; Kim, Hoe; Somnath, Suhas
Tip-based nanofabrication (TBN) is a family of emerging nanofabrication techniques that use a nanometer scale tip to fabricate nanostructures. Here in this review, we first introduce the history of the TBN and the technology development. We then briefly review various TBN techniques that use different physical or chemical mechanisms to fabricate features and discuss some of the state-of-the-art techniques. Subsequently, we focus on those TBN methods that have demonstrated potential to scale up the manufacturing throughput. Finally, we discuss several research directions that are essential for making TBN a scalable nano-manufacturing technology.
Head ballistocardiogram based on wireless multi-location sensors.
Onizuka, Kohei; Sodini, Charles G
2015-08-01
Recently a wearable BCG monitoring technique based on an accelerometer worn at the ear was demonstrated to replace a conventional bulky BCG acquisition system. In this work, a multi-location wireless vital signs monitor was developed, and at least two common acceleration vectors correlating to sitting-BCG were found in the supine position by using head PPG signal as a reference for eight healthy human subjects. The head side amplitude in the supine position is roughly proportional to the sitting amplitude that is in turn proportional to the stroke volume. Signal processing techniques to identify J-waves in a subject having small amplitude was also developed based on the two common vectors at the head side and top.
NASA Astrophysics Data System (ADS)
Azrina Talik, Noor; Boon Kar, Yap; Noradhlia Mohamad Tukijan, Siti; Wong, Chuan Ling
2017-10-01
To date, the state of art organic semiconductor distributed feedback (DFB) lasers gains tremendous interest in the organic device industry. This paper presents a short reviews on the fabrication techniques of DFB based laser by focusing on the fabrication method of DFB corrugated structure and the deposition of organic gain on the nano-patterned DFB resonator. The fabrication techniques such as Laser Direct Writing (LDW), ultrafast photo excitation dynamics, Laser Interference Lithography (LIL) and Nanoimprint Lithography (NIL) for DFB patterning are presented. In addition to that, the method for gain medium deposition method is also discussed. The technical procedures of the stated fabrication techniques are summarized together with their benefits and comparisons to the traditional fabrication techniques.
The development of selected data base applications for the crustal dynamics data information system
NASA Technical Reports Server (NTRS)
Noll, C. E.
1981-01-01
The development of a data base and its accompanying software for the data information system of crustal dynamics project is described. Background information concerning this project, and a definition of the techniques used in the implementation of an operational data base, are presented. Examples of key applications are included and interpreted.
Practical Team-Based Learning from Planning to Implementation
Bell, Edward; Eng, Marty; Fuentes, David G.; Helms, Kristen L.; Maki, Erik D.; Vyas, Deepti
2015-01-01
Team-based learning (TBL) helps instructors develop an active teaching approach for the classroom through group work. The TBL infrastructure engages students in the learning process through the Readiness Assessment Process, problem-solving through team discussions, and peer feedback to ensure accountability. This manuscript describes the benefits and barriers of TBL, and the tools necessary for developing, implementing, and critically evaluating the technique within coursework in a user-friendly method. Specifically, the manuscript describes the processes underpinning effective TBL development, preparation, implementation, assessment, and evaluation, as well as practical techniques and advice from authors’ classroom experiences. The paper also highlights published articles in the area of TBL in education, with a focus on pharmacy education. PMID:26889061
Modelling Technique for Demonstrating Gravity Collapse Structures in Jointed Rock.
ERIC Educational Resources Information Center
Stimpson, B.
1979-01-01
Described is a base-friction modeling technique for studying the development of collapse structures in jointed rocks. A moving belt beneath weak material is designed to simulate gravity. A description is given of the model frame construction. (Author/SA)
Feathering effect detection and artifact agglomeration index-based video deinterlacing technique
NASA Astrophysics Data System (ADS)
Martins, André Luis; Rodrigues, Evandro Luis Linhari; de Paiva, Maria Stela Veludo
2018-03-01
Several video deinterlacing techniques have been developed, and each one presents a better performance in certain conditions. Occasionally, even the most modern deinterlacing techniques create frames with worse quality than primitive deinterlacing processes. This paper validates that the final image quality can be improved by combining different types of deinterlacing techniques. The proposed strategy is able to select between two types of deinterlaced frames and, if necessary, make the local correction of the defects. This decision is based on an artifact agglomeration index obtained from a feathering effect detection map. Starting from a deinterlaced frame produced by the "interfield average" method, the defective areas are identified, and, if deemed appropriate, these areas are replaced by pixels generated through the "edge-based line average" method. Test results have proven that the proposed technique is able to produce video frames with higher quality than applying a single deinterlacing technique through getting what is good from intra- and interfield methods.
A Real-Time Earthquake Precursor Detection Technique Using TEC from a GPS Network
NASA Astrophysics Data System (ADS)
Alp Akyol, Ali; Arikan, Feza; Arikan, Orhan
2016-07-01
Anomalies have been observed in the ionospheric electron density distribution prior to strong earthquakes. However, most of the reported results are obtained by earthquake analysis. Therefore, their implementation in practice is highly problematic. Recently, a novel earthquake precursor detection technique based on spatio-temporal analysis of Total Electron Content (TEC) data obtained from Turkish National Permanent GPS Network (TNPGN) is developed by IONOLAB group (www.ionolab.org). In the present study, the developed detection technique is implemented in a causal setup over the available data set in test phase that enables the real time implementation. The performance of the developed earthquake prediction technique is evaluated by using 10 fold cross validation over the data obtained in 2011. Among the 23 earthquakes that have magnitudes higher than 5, the developed technique can detect precursors of 14 earthquakes while producing 8 false alarms. This study is supported by TUBITAK 115E915 and Joint TUBITAK 114E092 and AS CR 14/001 projects.
X-ray micro-beam techniques and phase contrast tomography applied to biomaterials
NASA Astrophysics Data System (ADS)
Fratini, Michela; Campi, Gaetano; Bukreeva, Inna; Pelliccia, Daniele; Burghammer, Manfred; Tromba, Giuliana; Cancedda, Ranieri; Mastrogiacomo, Maddalena; Cedola, Alessia
2015-12-01
A deeper comprehension of the biomineralization (BM) process is at the basis of tissue engineering and regenerative medicine developments. Several in-vivo and in-vitro studies were dedicated to this purpose via the application of 2D and 3D diagnostic techniques. Here, we develop a new methodology, based on different complementary experimental techniques (X-ray phase contrast tomography, micro-X-ray diffraction and micro-X-ray fluorescence scanning technique) coupled to new analytical tools. A qualitative and quantitative structural investigation, from the atomic to the micrometric length scale, is obtained for engineered bone tissues. The high spatial resolution achieved by X-ray scanning techniques allows us to monitor the bone formation at the first-formed mineral deposit at the organic-mineral interface within a porous scaffold. This work aims at providing a full comprehension of the morphology and functionality of the biomineralization process, which is of key importance for developing new drugs for preventing and healing bone diseases and for the development of bio-inspired materials.
DOT National Transportation Integrated Search
2016-04-01
In this study, we developed an adaptive signal control (ASC) framework for connected vehicles (CVs) using agent-based modeling technique. : The proposed framework consists of two types of agents: 1) vehicle agents (VAs); and 2) signal controller agen...
NASA Technical Reports Server (NTRS)
Butera, M. K.
1981-01-01
An automatic technique has been developed to measure marsh plant production by inference from a species classification derived from Landsat MSS data. A separate computer technique has been developed to calculate the transport path length of detritus and nutrients from their point of origin in the marsh to the shoreline from Landsat data. A nutrient availability indicator, the ratio of production to transport path length, was derived for each marsh-identified Landsat cell. The use of a data base compatible with the Landsat format facilitated data handling and computations.
NASA Technical Reports Server (NTRS)
Smith, Phillip N.
1990-01-01
The automation of low-altitude rotorcraft flight depends on the ability to detect, locate, and navigate around obstacles lying in the rotorcraft's intended flightpath. Computer vision techniques provide a passive method of obstacle detection and range estimation, for obstacle avoidance. Several algorithms based on computer vision methods have been developed for this purpose using laboratory data; however, further development and validation of candidate algorithms require data collected from rotorcraft flight. A data base containing low-altitude imagery augmented with the rotorcraft and sensor parameters required for passive range estimation is not readily available. Here, the emphasis is on the methodology used to develop such a data base from flight-test data consisting of imagery, rotorcraft and sensor parameters, and ground-truth range measurements. As part of the data preparation, a technique for obtaining the sensor calibration parameters is described. The data base will enable the further development of algorithms for computer vision-based obstacle detection and passive range estimation, as well as provide a benchmark for verification of range estimates against ground-truth measurements.
Robust Approach for Nonuniformity Correction in Infrared Focal Plane Array.
Boutemedjet, Ayoub; Deng, Chenwei; Zhao, Baojun
2016-11-10
In this paper, we propose a new scene-based nonuniformity correction technique for infrared focal plane arrays. Our work is based on the use of two well-known scene-based methods, namely, adaptive and interframe registration-based exploiting pure translation motion model between frames. The two approaches have their benefits and drawbacks, which make them extremely effective in certain conditions and not adapted for others. Following on that, we developed a method robust to various conditions, which may slow or affect the correction process by elaborating a decision criterion that adapts the process to the most effective technique to ensure fast and reliable correction. In addition to that, problems such as bad pixels and ghosting artifacts are also dealt with to enhance the overall quality of the correction. The performance of the proposed technique is investigated and compared to the two state-of-the-art techniques cited above.
Robust Approach for Nonuniformity Correction in Infrared Focal Plane Array
Boutemedjet, Ayoub; Deng, Chenwei; Zhao, Baojun
2016-01-01
In this paper, we propose a new scene-based nonuniformity correction technique for infrared focal plane arrays. Our work is based on the use of two well-known scene-based methods, namely, adaptive and interframe registration-based exploiting pure translation motion model between frames. The two approaches have their benefits and drawbacks, which make them extremely effective in certain conditions and not adapted for others. Following on that, we developed a method robust to various conditions, which may slow or affect the correction process by elaborating a decision criterion that adapts the process to the most effective technique to ensure fast and reliable correction. In addition to that, problems such as bad pixels and ghosting artifacts are also dealt with to enhance the overall quality of the correction. The performance of the proposed technique is investigated and compared to the two state-of-the-art techniques cited above. PMID:27834893
Lightweight and Statistical Techniques for Petascale PetaScale Debugging
DOE Office of Scientific and Technical Information (OSTI.GOV)
Miller, Barton
2014-06-30
This project investigated novel techniques for debugging scientific applications on petascale architectures. In particular, we developed lightweight tools that narrow the problem space when bugs are encountered. We also developed techniques that either limit the number of tasks and the code regions to which a developer must apply a traditional debugger or that apply statistical techniques to provide direct suggestions of the location and type of error. We extend previous work on the Stack Trace Analysis Tool (STAT), that has already demonstrated scalability to over one hundred thousand MPI tasks. We also extended statistical techniques developed to isolate programming errorsmore » in widely used sequential or threaded applications in the Cooperative Bug Isolation (CBI) project to large scale parallel applications. Overall, our research substantially improved productivity on petascale platforms through a tool set for debugging that complements existing commercial tools. Previously, Office Of Science application developers relied either on primitive manual debugging techniques based on printf or they use tools, such as TotalView, that do not scale beyond a few thousand processors. However, bugs often arise at scale and substantial effort and computation cycles are wasted in either reproducing the problem in a smaller run that can be analyzed with the traditional tools or in repeated runs at scale that use the primitive techniques. New techniques that work at scale and automate the process of identifying the root cause of errors were needed. These techniques significantly reduced the time spent debugging petascale applications, thus leading to a greater overall amount of time for application scientists to pursue the scientific objectives for which the systems are purchased. We developed a new paradigm for debugging at scale: techniques that reduced the debugging scenario to a scale suitable for traditional debuggers, e.g., by narrowing the search for the root-cause analysis to a small set of nodes or by identifying equivalence classes of nodes and sampling our debug targets from them. We implemented these techniques as lightweight tools that efficiently work on the full scale of the target machine. We explored four lightweight debugging refinements: generic classification parameters, such as stack traces, application-specific classification parameters, such as global variables, statistical data acquisition techniques and machine learning based approaches to perform root cause analysis. Work done under this project can be divided into two categories, new algorithms and techniques for scalable debugging, and foundation infrastructure work on our MRNet multicast-reduction framework for scalability, and Dyninst binary analysis and instrumentation toolkits.« less
Olyaeemanesh, Alireza; Bavandpour, Elahe; Mobinizadeh, Mohammadreza; Ashrafinia, Mansoor; Bavandpour, Maryam; Nouhi, Mojtaba
2017-01-01
Background: Caesarean section (C-section) is the most common surgery among women worldwide, and the global rate of this surgical procedure has been continuously rising. Hence, it is significantly crucial to develop and apply highly effective and safe caesarean section techniques. In this review study, we aimed at assessing the safety and effectiveness of the Joel-Cohen-based technique and comparing the results with the transverse Pfannenstiel incision for C-section. Methods: In this study, various reliable databases such as the PubMed Central, COCHRANE, DARE, and Ovid MEDLINE were targeted. Reviews, systematic reviews, and randomized clinical trial studies comparing the Joel-Cohen-based technique and the transverse Pfannenstiel incision were selected based on the inclusion criteria. Selected studies were checked by 2 independent reviewers based on the inclusion criteria, and the quality of these studies was assessed. Then, their data were extracted and analyzed. Results: Five randomized clinical trial studies met the inclusion criteria. According to the exiting evidence, statistical results of the Joel-Cohen-based technique showed that this technique is more effective compared to the transverse Pfannenstiel incision. Metaanalysis results of the 3 outcomes were as follow: operation time (5 trials, 764 women; WMD -9.78; 95% CI:-14.49-5.07 minutes, p<0.001), blood loss (3 trials, 309 women; WMD -53.23ml; 95% -CI: 90.20-16.26 ml, p= 0.004), and post-operative hospital stay (3 trials, 453 women; WMD -.69 day; 95% CI: 1.4-0.03 day, p<0.001). Statistical results revealed a significant difference between the 2 techniques. Conclusion: According to the literature, despite having a number of side effects, the Joel-Cohen-based technique is generally more effective than the Pfannenstiel incision technique. In addition, it was recommended that the Joel-Cohen-based technique be used as a replacement for the Pfannenstiel incision technique according to the surgeons' preferences and the patients' conditions.
Wei, Xueyong
2010-11-01
Since it was invented two decades ago, Nanosphere Lithography (NSL) has been widely studied as a low cost and flexible technique to fabricate nanostructures. Based on the registered patents and some selected papers, this review will discuss recent developments of different NSL strategies for the fabrication of ordered nanostructure arrays. The mechanism of self-assembly process and the techniques for preparing the self-assembled nanosphere template are first briefly introduced. The nanosphere templates are used either as shadow masks or as moulds for pattern transfer. Much more work now combines NSL with other lithographic techniques and material growth methods to form novel nanostructures of complex shape or various materials. Hence, this review finally gives a discussion on some future directions in NSL study.
NASA Astrophysics Data System (ADS)
Li, Qing; Lin, Haibo; Xiu, Yu-Feng; Wang, Ruixue; Yi, Chuijie
The test platform of wheat precision seeding based on image processing techniques is designed to develop the wheat precision seed metering device with high efficiency and precision. Using image processing techniques, this platform gathers images of seeds (wheat) on the conveyer belt which are falling from seed metering device. Then these data are processed and analyzed to calculate the qualified rate, reseeding rate and leakage sowing rate, etc. This paper introduces the whole structure, design parameters of the platform and hardware & software of the image acquisition system were introduced, as well as the method of seed identification and seed-space measurement using image's threshold and counting the seed's center. By analyzing the experimental result, the measurement error is less than ± 1mm.
Xiping Wang; James P. Wacker; Robert J. Ross; Brian K. Brashaw; Robert Vatalaro
2005-01-01
This paper describes an effort to develop a global dynamic testing technique for evaluating the overall stiffness of timber bridge superstructures. A forced vibration method was used to measure the natural frequency of single-span timber bridges in the laboratory and field. An analytical model based on simple beam theory was proposed to represent the relationship...
[Development of the automatic dental X-ray film processor].
Bai, J; Chen, H
1999-07-01
This paper introduces a multiple-point detecting technique of the density of dental X-ray films. With the infrared ray multiple-point detecting technique, a single-chip microcomputer control system is used to analyze the effectiveness of the film-developing in real time in order to achieve a good image. Based on the new technology, We designed the intelligent automatic dental X-ray film processing.
ERIC Educational Resources Information Center
Charconnet, Marie-George
This study describes various patterns of peer tutoring and is based on the use of cultural traditions and endogenous methods, on techniques and equipment acquired from other cultures, on problems presented by the adoption of educational technologies, and on methods needing little sophisticated equipment. A dozen peer tutoring systems are…
DOE Office of Scientific and Technical Information (OSTI.GOV)
Yinfa, Ma.
Thin-layer chromatography (TLC) is a broadly applicable separation technique. It offers many advantages over high performance liquid chromatography (HPLC), such as easily adapted for two-dimensional separation, for whole-column'' detection and for handling multiple samples, etc. However, due to its draggy development of detection techniques comparing with HPLC, TLC has not received the attention it deserves. Therefore, exploring new detection techniques is very important to the development of TLC. It is the principal of this dissertation to present a new detection method for TLC -- indirect fluorometric detection method. This detection technique is universal sensitive, nondestructive, and simple. This will bemore » described in detail from Sections 1 through Section 5. Section 1 and 3 describe the indirect fluorometric detection of anions and nonelectrolytes in TLC. In Section 2, a detection method for cations based on fluorescence quenching of ethidium bromide is presented. In Section 4, a simple and interesting TLC experiment is designed, three different fluorescence detection principles are used for the determination of caffeine, saccharin and sodium benzoate in beverages. A laser-based indirect fluorometric detection technique in TLC is developed in Section 5. Section 6 is totally different from Sections 1 through 5. An ultrasonic effect on the separation of DNA fragments in agarose gel electrophoresis is investigated. 262 refs.« less
Zhang, Xingwang; Xing, Huijie; Zhao, Yue; Ma, Zhiguo
2018-06-23
Over the past decades, a large number of drugs as well as drug candidates with poor dissolution characteristics have been witnessed, which invokes great interest in enabling formulation of these active ingredients. Poorly water-soluble drugs, especially biopharmaceutical classification system (BCS) II ones, are preferably designed as oral dosage forms if the dissolution limit can be broken through. Minimizing a drug’s size is an effective means to increase its dissolution and hence the bioavailability, which can be achieved by specialized dispersion techniques. This article reviews the most commonly used dispersion techniques for pharmaceutical processing that can practically enhance the dissolution and bioavailability of poorly water-soluble drugs. Major interests focus on solid dispersion, lipid-based dispersion (nanoencapsulation), and liquisolid dispersion (drug solubilized in a non-volatile solvent and dispersed in suitable solid excipients for tableting or capsulizing), covering the formulation development, preparative technique and potential applications for oral drug delivery. Otherwise, some other techniques that can increase the dispersibility of a drug such as co-precipitation, concomitant crystallization and inclusion complexation are also discussed. Various dispersion techniques provide a productive platform for addressing the formulation challenge of poorly water-soluble drugs. Solid dispersion and liquisolid dispersion are most likely to be successful in developing oral dosage forms. Lipid-based dispersion represents a promising approach to surmounting the bioavailability of low-permeable drugs, though the technique needs to traverse the obstacle from liquid to solid transformation. Novel dispersion techniques are highly encouraged to develop for formulation of poorly water-soluble drugs.
Bhat, Riyaz A; Lahaye, Thomas; Panstruga, Ralph
2006-01-01
Non-invasive fluorophore-based protein interaction assays like fluorescence resonance energy transfer (FRET) and bimolecular fluorescence complementation (BiFC, also referred to as "split YFP") have been proven invaluable tools to study protein-protein interactions in living cells. Both methods are now frequently used in the plant sciences and are likely to develop into standard techniques for the identification, verification and in-depth analysis of polypeptide interactions. In this review, we address the individual strengths and weaknesses of both approaches and provide an outlook about new directions and possible future developments for both techniques. PMID:16800872
Nanomaterials-Based Optical Techniques for the Detection of Acetylcholinesterase and Pesticides
Xia, Ning; Wang, Qinglong; Liu, Lin
2015-01-01
The large amount of pesticide residues in the environment is a threat to global health by inhibition of acetylcholinesterase (AChE). Biosensors for inhibition of AChE have been thus developed for the detection of pesticides. In line with the rapid development of nanotechnology, nanomaterials have attracted great attention and have been intensively studied in biological analysis due to their unique chemical, physical and size properties. The aim of this review is to provide insight into nanomaterial-based optical techniques for the determination of AChE and pesticides, including colorimetric and fluorescent assays and surface plasmon resonance. PMID:25558991
Lunar Contour Crafting: A Novel Technique for ISRU-Based Habitat Development
NASA Technical Reports Server (NTRS)
Khoshnevis, Behrokh; Bodiford, Melanie P.; Burks, Kevin H.; Ethridge, Ed; Tucker, Dennis; Kim, Won; Toutanji, Houssam; Fiske, Michael R.
2004-01-01
As the nation prepares to return to the Moon, it is apparent that the viability of long duration visits with appropriate radiation shielding/crew protection, hinges on the development of Lunar structures, preferably in advance of a manned landing, and preferably utilizing in-situ resources. Contour Crafting is a USC-patented technique for automated development of terrestrial concrete-based structures. The process is relatively fast, completely automated, and supports the incorporation of various infrastructure elements such as plumbing and electrical wiring. This paper will present a conceptual design of a Lunar Contour Crafting system designed to autonomously fabricate integrated structures on the Lunar surface using high-strength concrete based on Lunar regolith, including glass reinforcement rods or fibers fabricated from melted regolith. Design concepts will be presented, as well as results of initial tests aimed at concrete and glass production using Lunar regolith simulant. Key issues and concerns will be presented, along with design concepts for an LCC testbed to be developed at MSFC's Prototype Development Laboratory (PDL).
New developments of a knowledge based system (VEG) for inferring vegetation characteristics
NASA Technical Reports Server (NTRS)
Kimes, D. S.; Harrison, P. A.; Harrison, P. R.
1992-01-01
An extraction technique for inferring physical and biological surface properties of vegetation using nadir and/or directional reflectance data as input has been developed. A knowledge-based system (VEG) accepts spectral data of an unknown target as input, determines the best strategy for inferring the desired vegetation characteristic, applies the strategy to the target data, and provides a rigorous estimate of the accuracy of the inference. Progress in developing the system is presented. VEG combines methods from remote sensing and artificial intelligence, and integrates input spectral measurements with diverse knowledge bases. VEG has been developed to (1) infer spectral hemispherical reflectance from any combination of nadir and/or off-nadir view angles; (2) test and develop new extraction techniques on an internal spectral database; (3) browse, plot, or analyze directional reflectance data in the system's spectral database; (4) discriminate between user-defined vegetation classes using spectral and directional reflectance relationships; and (5) infer unknown view angles from known view angles (known as view angle extension).
NASA Astrophysics Data System (ADS)
Nitta, Noriko; Taniwaki, Masafumi
2006-04-01
The present authors proposed a novel nano-fabrication technique that is able to arrange the fine cells orderly, based on their finding in GaSb implanted at a low temperature. In this article, first the experimental results that anomalous cellular structure was formed in GaSb by ion implantation is introduced and the self-organizational formation mechanism of the structure is described. Next a nano-fabrication technique that utilizes focused ion beam is described. This technique consists of two procedures, i.e. the formation process of the voids array and the development of the initial array to ordered cellular structure. Finally, the nano-fabrication is actually performed by this technique and their results are reported. Fabrication succeeded in structures where the dot (cell) interval was 100 nm or larger. The minimum ion dose for initial voids which develops to the ordered cellular structure is evaluated. It is also shown that the substrate temperature during implantation is an essential parameter for this technique.
NASA Astrophysics Data System (ADS)
Setiyoko, A.; Dharma, I. G. W. S.; Haryanto, T.
2017-01-01
Multispectral data and hyperspectral data acquired from satellite sensor have the ability in detecting various objects on the earth ranging from low scale to high scale modeling. These data are increasingly being used to produce geospatial information for rapid analysis by running feature extraction or classification process. Applying the most suited model for this data mining is still challenging because there are issues regarding accuracy and computational cost. This research aim is to develop a better understanding regarding object feature extraction and classification applied for satellite image by systematically reviewing related recent research projects. A method used in this research is based on PRISMA statement. After deriving important points from trusted sources, pixel based and texture-based feature extraction techniques are promising technique to be analyzed more in recent development of feature extraction and classification.
NASA Technical Reports Server (NTRS)
Williams, B. F.
1976-01-01
Manufacturing techniques are evaluated using expenses based on experience and studying basic cost factors for each step to evaluate expenses from a first-principles point of view. A formal cost accounting procedure is developed which is used throughout the study for cost comparisons. The first test of this procedure is a comparison of its predicted costs for array module manufacturing with costs from a study which is based on experience factors. A manufacturing cost estimate for array modules of $10/W is based on present-day manufacturing techniques, expenses, and materials costs.
Use of activity theory-based need finding for biomedical device development.
Rismani, Shalaleh; Ratto, Matt; Machiel Van der Loos, H F
2016-08-01
Identifying the appropriate needs for biomedical device design is challenging, especially for less structured environments. The paper proposes an alternate need-finding method based on Cultural Historical Activity Theory and expanded to explicitly examine the role of devices within a socioeconomic system. This is compared to a conventional need-finding technique in a preliminary study with engineering student teams. The initial results show that the Activity Theory-based technique allows teams to gain deeper insights into their needs space.
NASA Astrophysics Data System (ADS)
Kougioumtzoglou, Ioannis A.; dos Santos, Ketson R. M.; Comerford, Liam
2017-09-01
Various system identification techniques exist in the literature that can handle non-stationary measured time-histories, or cases of incomplete data, or address systems following a fractional calculus modeling. However, there are not many (if any) techniques that can address all three aforementioned challenges simultaneously in a consistent manner. In this paper, a novel multiple-input/single-output (MISO) system identification technique is developed for parameter identification of nonlinear and time-variant oscillators with fractional derivative terms subject to incomplete non-stationary data. The technique utilizes a representation of the nonlinear restoring forces as a set of parallel linear sub-systems. In this regard, the oscillator is transformed into an equivalent MISO system in the wavelet domain. Next, a recently developed L1-norm minimization procedure based on compressive sensing theory is applied for determining the wavelet coefficients of the available incomplete non-stationary input-output (excitation-response) data. Finally, these wavelet coefficients are utilized to determine appropriately defined time- and frequency-dependent wavelet based frequency response functions and related oscillator parameters. Several linear and nonlinear time-variant systems with fractional derivative elements are used as numerical examples to demonstrate the reliability of the technique even in cases of noise corrupted and incomplete data.
Using mediation techniques to manage conflict and create healthy work environments.
Gerardi, Debra
2004-01-01
Healthcare organizations must find ways for managing conflict and developing effective working relationships to create healthy work environments. The effects of unresolved conflict on clinical outcomes, staff retention, and the financial health of the organization lead to many unnecessary costs that divert resources from clinical care. The complexity of delivering critical care services makes conflict resolution difficult. Developing collaborative working relationships helps to manage conflict in complex environments. Working relationships are based on the ability to deal with differences. Dealing with differences requires skill development and techniques for balancing interests and communicating effectively. Techniques used by mediators are effective for resolving disputes and developing working relationships. With practice, these techniques are easily transferable to the clinical setting. Listening for understanding, reframing, elevating the definition of the problem, and forming clear agreements can foster working relationships, decrease the level of conflict, and create healthy work environments that benefit patients and professionals.
NASA Technical Reports Server (NTRS)
Thomas, J. B.; Fanselow, J. L.; Macdoran, P. F.; Skjerve, L. J.; Spitzmesser, D. J.; Fliegel, H. F.
1976-01-01
Radio interferometry promises eventually to measure directly, with accuracies of a few centimeters, both whole earth motions and relative crustal motions with respect to an 'inertial' reference frame. Interferometry measurements of arbitrarily long base lines require, however, the development of new techniques for independent-station observation. In connection with the development of such techniques, a series of short base line demonstration experiments has been conducted between two antennas. The experiments were related to a program involving the design of independent-station instrumentation capable of making three-dimensional earth-fixed base line measurements with an accuracy of a few centimeters. Attention is given to the instrumentation used in the experiments, aspects of data analysis, and the experimental results.
MEMS-based platforms for mechanical manipulation and characterization of cells
NASA Astrophysics Data System (ADS)
Pan, Peng; Wang, Wenhui; Ru, Changhai; Sun, Yu; Liu, Xinyu
2017-12-01
Mechanical manipulation and characterization of single cells are important experimental techniques in biological and medical research. Because of the microscale sizes and highly fragile structures of cells, conventional cell manipulation and characterization techniques are not accurate and/or efficient enough or even cannot meet the more and more demanding needs in different types of cell-based studies. To this end, novel microelectromechanical systems (MEMS)-based technologies have been developed to improve the accuracy, efficiency, and consistency of various cell manipulation and characterization tasks, and enable new types of cell research. This article summarizes existing MEMS-based platforms developed for cell mechanical manipulation and characterization, highlights their specific design considerations making them suitable for their designated tasks, and discuss their advantages and limitations. In closing, an outlook into future trends is also provided.
Compiler-assisted multiple instruction rollback recovery using a read buffer
NASA Technical Reports Server (NTRS)
Alewine, N. J.; Chen, S.-K.; Fuchs, W. K.; Hwu, W.-M.
1993-01-01
Multiple instruction rollback (MIR) is a technique that has been implemented in mainframe computers to provide rapid recovery from transient processor failures. Hardware-based MIR designs eliminate rollback data hazards by providing data redundancy implemented in hardware. Compiler-based MIR designs have also been developed which remove rollback data hazards directly with data-flow transformations. This paper focuses on compiler-assisted techniques to achieve multiple instruction rollback recovery. We observe that some data hazards resulting from instruction rollback can be resolved efficiently by providing an operand read buffer while others are resolved more efficiently with compiler transformations. A compiler-assisted multiple instruction rollback scheme is developed which combines hardware-implemented data redundancy with compiler-driven hazard removal transformations. Experimental performance evaluations indicate improved efficiency over previous hardware-based and compiler-based schemes.
Adair, P M; Burnside, G; Pine, C M
2013-01-01
To improve oral health in children, the key behaviours (tooth brushing and sugar control) responsible for development of dental caries need to be better understood, as well as how to promote these behaviours effectively so they become habitual; and, the specific, optimal techniques to use in interventions. The aim of this paper is to describe and analyse the behaviour change techniques that have been used in primary school-based interventions to prevent dental caries (utilizing a Cochrane systematic review that we have undertaken) and to identify opportunities for improving future interventions by incorporating a comprehensive range of behaviour change techniques. Papers of five interventions were reviewed and data were independently extracted. Results indicate that behaviour change techniques were limited to information-behaviour links, information on consequences, instruction and demonstration of behaviours. None of the interventions were based on behaviour change theory. We conclude that behaviour change techniques used in school interventions to reduce dental caries were limited and focused around providing information about how behaviour impacts on health and the consequences of not developing the correct health behaviours as well as providing oral hygiene instruction. Establishing which techniques are effective is difficult due to poor reporting of interventions in studies. Future design of oral health promotion interventions using behaviour change theory for development and evaluation (and reporting results in academic journals) could strengthen the potential for efficacy and provide a framework to use a much wider range of behaviour change techniques. Future studies should include development and publication of intervention manuals which is becoming standard practice in other health promoting programmes. © 2013 S. Karger AG, Basel.
Model-based Clustering of High-Dimensional Data in Astrophysics
NASA Astrophysics Data System (ADS)
Bouveyron, C.
2016-05-01
The nature of data in Astrophysics has changed, as in other scientific fields, in the past decades due to the increase of the measurement capabilities. As a consequence, data are nowadays frequently of high dimensionality and available in mass or stream. Model-based techniques for clustering are popular tools which are renowned for their probabilistic foundations and their flexibility. However, classical model-based techniques show a disappointing behavior in high-dimensional spaces which is mainly due to their dramatical over-parametrization. The recent developments in model-based classification overcome these drawbacks and allow to efficiently classify high-dimensional data, even in the "small n / large p" situation. This work presents a comprehensive review of these recent approaches, including regularization-based techniques, parsimonious modeling, subspace classification methods and classification methods based on variable selection. The use of these model-based methods is also illustrated on real-world classification problems in Astrophysics using R packages.
Priority Techniques for High Occupancy Vehicles : State-of-the-Art Overview
DOT National Transportation Integrated Search
1975-11-01
The report, part of a series of publications based on research and development efforts is a concise state-of-the-art overview of priority techniques for high occupancy vehicles (buses, carpools, and vanpools). The report identifies and summarizes sel...
Fu, Yu; Pedrini, Giancarlo
2014-01-01
In recent years, optical interferometry-based techniques have been widely used to perform noncontact measurement of dynamic deformation in different industrial areas. In these applications, various physical quantities need to be measured in any instant and the Nyquist sampling theorem has to be satisfied along the time axis on each measurement point. Two types of techniques were developed for such measurements: one is based on high-speed cameras and the other uses a single photodetector. The limitation of the measurement range along the time axis in camera-based technology is mainly due to the low capturing rate, while the photodetector-based technology can only do the measurement on a single point. In this paper, several aspects of these two technologies are discussed. For the camera-based interferometry, the discussion includes the introduction of the carrier, the processing of the recorded images, the phase extraction algorithms in various domains, and how to increase the temporal measurement range by using multiwavelength techniques. For the detector-based interferometry, the discussion mainly focuses on the single-point and multipoint laser Doppler vibrometers and their applications for measurement under extreme conditions. The results show the effort done by researchers for the improvement of the measurement capabilities using interferometry-based techniques to cover the requirements needed for the industrial applications. PMID:24963503
Total Reconstruction of the Auricle: Our Experiences on Indications and Recent Techniques
Storck, K.; Staudenmaier, R.; Buchberger, M.; Strenger, T.; Kreutzer, K.; von Bomhard, A.; Stark, T.
2014-01-01
Introduction. Auricular reconstruction is a great challenge in facial plastic surgery. With the advances in surgical techniques and biotechnology, different options are available for consideration. The aim of this paper is to review the knowledge about the various techniques for total auricular reconstruction based on the literature and our experience. Methods. Approximately 179 articles published from 1980 to 2013 were identified, and 59 articles were included. We have focused on the current status of total auricular reconstruction based on our personal experience and on papers of particular interest, published within the period of review. We have also included a prospective view on the tissue engineering of cartilage. Results. Most surgeons still practice total auricular reconstruction by employing techniques developed by Brent, Nagata, and Firmin with autologous rib cartilage. Within the last years, alloplastic frameworks for reconstruction have become well established. Choosing the reconstruction techniques depends mainly on the surgeon's preference and experience. Prosthetic reconstruction is still reserved for special conditions, even though the material is constantly improving. Tissue engineering has a growing potential for clinical applicability. Conclusion. Auricular reconstruction still receives attention of plastic/maxillofacial surgeons and otolaryngologists. Even though clinical applicability lags behind initial expectations, the development of tissue-engineered constructs continues its potential development. PMID:24822198
MO-G-BRD-01: Point/Counterpoint Debate: Arc Based Techniques Will Make Conventional IMRT Obsolete
DOE Office of Scientific and Technical Information (OSTI.GOV)
Shepard, D; Popple, R; Balter, P
2014-06-15
A variety of intensity modulated radiation therapy (IMRT) delivery techniques have been developed that have provided clinicians with the ability to deliver highly conformal dose distributions. The delivery techniques include compensators, step-and-shoot IMRT, sliding window IMRT, volumetric modulated arc therapy (VMAT), and tomotherapy. A key development in the field of IMRT was the introduction of new planning algorithms and delivery control systems in 2007 that made it possible to coordinate the gantry rotation speed, dose rate, and multileaf collimator leaf positions during the delivery of arc therapy. With these developments, VMAT became a routine clinical tool. The use of VMATmore » has continued to grow in recent years and some would argue that this will soon make conventional IMRT obsolete, and this is the premise of this debate. To introduce the debate, David Shepard, Ph.D. will provide an overview of IMRT delivery techniques including historical context and how they are being used today. The debate will follow with Richard Popple, Ph.D. arguing FOR the Proposition and Peter Balter, Ph.D. arguing AGAINST it. Learning Objectives: Understand the different delivery techniques for IMRT. Understand the potential benefits of conventional IMRT. Understand the potential benefits of arc-based IMRT delivery.« less
Towards an Intelligent Planning Knowledge Base Development Environment
NASA Technical Reports Server (NTRS)
Chien, S.
1994-01-01
ract describes work in developing knowledge base editing and debugging tools for the Multimission VICAR Planner (MVP) system. MVP uses artificial intelligence planning techniques to automatically construct executable complex image processing procedures (using models of the smaller constituent image processing requests made to the JPL Multimission Image Processing Laboratory.
Predicting Plywood Properties with Wood-based Composite Models
Christopher Adam Senalik; Robert J. Ross
2015-01-01
Previous research revealed that stress wave nondestructive testing techniques could be used to evaluate the tensile and flexural properties of wood-based composite materials. Regression models were developed that related stress wave transmission characteristics (velocity and attenuation) to modulus of elasticity and strength. The developed regression models accounted...
Jäger, Jessica
2013-07-01
This article reports on a follow-up study exploring the use of play-based evaluation methods to facilitate children's views of therapy. The development and piloting of these techniques, with 12 children in the author's own practice, was previously reported in this journal. It was argued that play-based evaluation methods reduce the power imbalance inherent in adult researcher/interviewer-child relationships and provide children with meaningful ways to share their views. In this article, follow-up research into play-based evaluations with 20 children and 7 different play therapists is drawn upon to explore in greater depth the strengths and weaknesses of these techniques. The study shows that play-based evaluation techniques are important and flexible methods for facilitating children's views of child therapy. It is argued that those play therapists who incorporate their therapeutic skills effectively, maintain flexibility and sensitively attune to the child during the evaluation session, enable the child to explore their views most fully.
Damage of composite structures: Detection technique, dynamic response and residual strength
NASA Astrophysics Data System (ADS)
Lestari, Wahyu
2001-10-01
Reliable and accurate health monitoring techniques can prevent catastrophic failures of structures. Conventional damage detection methods are based on visual or localized experimental methods and very often require prior information concerning the vicinity of the damage or defect. The structure must also be readily accessible for inspections. The techniques are also labor intensive. In comparison to these methods, health-monitoring techniques that are based on the structural dynamic response offers unique information on failure of structures. However, systematic relations between the experimental data and the defect are not available and frequently, the number of vibration modes needed for an accurate identification of defects is much higher than the number of modes that can be readily identified in the experiment. These motivated us to develop an experimental data based detection method with systematic relationships between the experimentally identified information and the analytical or mathematical model representing the defective structures. The developed technique use changes in vibrational curvature modes and natural frequencies. To avoid misinterpretation of the identified information, we also need to understand the effects of defects on the structural dynamic response prior to developing health-monitoring techniques. In this thesis work we focus on two type of defects in composite structures, namely delamination and edge notch like defect. Effects of nonlinearity due to the presence of defect and due to the axial stretching are studied for beams with delamination. Once defects are detected in a structure, next concern is determining the effects of the defects on the strength of the structure and its residual stiffness under dynamic loading. In this thesis, energy release rate due to dynamic loading in a delaminated structure is studied, which will be a foundation toward determining the residual strength of the structure.
Development of a Pulsed Pressure-Based Technique for Cavitation Damage Study
DOE Office of Scientific and Technical Information (OSTI.GOV)
Ren, Fei; Wang, Jy-An John; Liu, Yun
2012-01-01
Cavitation occurs in many fluid systems and can lead to severe material damage. To assist the study of cavitation damage, a novel testing method utilizing pulsed pressure was developed. In this talk, the scientific background and the technical approach of this development are present and preliminary testing results are discussed. It is expected that this technique can be used to evaluate cavitation damage under various testing conditions including harsh environments such as those relevant to geothermal power generation.
NASA Technical Reports Server (NTRS)
Colwell, R. N. (Principal Investigator)
1977-01-01
The results and progress of work conducted in support of the Large Area Crop Inventory Experiment (LACIE) are documented. Research was conducted for two tasks. These tasks include: (1) evaluation of the UCB static stratification procedure and modification of that procedure if warranted; and (2) the development of alternative photointerpretive techniques to the present LACIE procedure for the identification and selection of training areas for machine-processing of LACIE segments.
Development of Optimized Combustors and Thermoelectric Generators for Palm Power Generation
2004-10-26
manufacturing techniques and microfabrication, on the chemical kinetics of JP-8 surrogates and on the development of advanced laser diagnostics for JP-8...takes the shape of a cone from the tip of which a thin liquid thread emerges, in the so-called cone-jet mode [1]. This microjet breaks into a stream of...combustion systems. 2. The development of a diagnostic technique based on two-color laser induced fluorescence from fluorescence tags added to the fuel
Short Duration Base Heating Test Improvements
NASA Technical Reports Server (NTRS)
Bender, Robert L.; Dagostino, Mark G.; Engel, Bradley A.; Engel, Carl D.
1999-01-01
Significant improvements have been made to a short duration space launch vehicle base heating test technique. This technique was first developed during the 1960's to investigate launch vehicle plume induced convective environments. Recent improvements include the use of coiled nitrogen buffer gas lines upstream of the hydrogen / oxygen propellant charge tubes, fast acting solenoid valves, stand alone gas delivery and data acquisition systems, and an integrated model design code. Technique improvements were successfully demonstrated during a 2.25% scale X-33 base heating test conducted in the NASA/MSFC Nozzle Test Facility in early 1999. Cost savings of approximately an order of magnitude over previous tests were realized due in large part to these improvements.
Zeiler, Frederick A; Donnelly, Joseph; Calviello, Leanne; Menon, David K; Smielewski, Peter; Czosnyka, Marek
2017-12-01
The purpose of this study was to perform a systematic, scoping review of commonly described intermittent/semi-intermittent autoregulation measurement techniques in adult traumatic brain injury (TBI). Nine separate systematic reviews were conducted for each intermittent technique: computed tomographic perfusion (CTP)/Xenon-CT (Xe-CT), positron emission tomography (PET), magnetic resonance imaging (MRI), arteriovenous difference in oxygen (AVDO 2 ) technique, thigh cuff deflation technique (TCDT), transient hyperemic response test (THRT), orthostatic hypotension test (OHT), mean flow index (Mx), and transfer function autoregulation index (TF-ARI). MEDLINE ® , BIOSIS, EMBASE, Global Health, Scopus, Cochrane Library (inception to December 2016), and reference lists of relevant articles were searched. A two tier filter of references was conducted. The total number of articles utilizing each of the nine searched techniques for intermittent/semi-intermittent autoregulation techniques in adult TBI were: CTP/Xe-CT (10), PET (6), MRI (0), AVDO 2 (10), ARI-based TCDT (9), THRT (6), OHT (3), Mx (17), and TF-ARI (6). The premise behind all of the intermittent techniques is manipulation of systemic blood pressure/blood volume via either chemical (such as vasopressors) or mechanical (such as thigh cuffs or carotid compression) means. Exceptionally, Mx and TF-ARI are based on spontaneous fluctuations of cerebral perfusion pressure (CPP) or mean arterial pressure (MAP). The method for assessing the cerebral circulation during these manipulations varies, with both imaging-based techniques and TCD utilized. Despite the limited literature for intermittent/semi-intermittent techniques in adult TBI (minus Mx), it is important to acknowledge the availability of such tests. They have provided fundamental insight into human autoregulatory capacity, leading to the development of continuous and more commonly applied techniques in the intensive care unit (ICU). Numerous methods of intermittent/semi-intermittent pressure autoregulation assessment in adult TBI exist, including: CTP/Xe-CT, PET, AVDO 2 technique, TCDT-based ARI, THRT, OHT, Mx, and TF-ARI. MRI-based techniques in adult TBI are yet to be described, with the main focus of MRI techniques on metabolic-based cerebrovascular reactivity (CVR) and not pressure-based autoregulation.
An improved sample loading technique for cellular metabolic response monitoring under pressure
NASA Astrophysics Data System (ADS)
Gikunda, Millicent Nkirote
To monitor cellular metabolism under pressure, a pressure chamber designed around a simple-to-construct capillary-based spectroscopic chamber coupled to a microliter-flow perfusion system is used in the laboratory. Although cyanide-induced metabolic responses from Saccharomyces cerevisiae (baker's yeast) could be controllably induced and monitored under pressure, previously used sample loading technique was not well controlled. An improved cell-loading technique which is based on use of a secondary inner capillary into which the sample is loaded then inserted into the capillary pressure chamber, has been developed. As validation, we demonstrate the ability to measure the chemically-induced metabolic responses at pressures of up to 500 bars. This technique is shown to be less prone to sample loss due to perfusive flow than the previous techniques used.
NASA Technical Reports Server (NTRS)
Morelli, Eugene A.; Cunningham, Kevin; Hill, Melissa A.
2013-01-01
Flight test and modeling techniques were developed for efficiently identifying global aerodynamic models that can be used to accurately simulate stall, upset, and recovery on large transport airplanes. The techniques were developed and validated in a high-fidelity fixed-base flight simulator using a wind-tunnel aerodynamic database, realistic sensor characteristics, and a realistic flight deck representative of a large transport aircraft. Results demonstrated that aerodynamic models for stall, upset, and recovery can be identified rapidly and accurately using relatively simple piloted flight test maneuvers. Stall maneuver predictions and comparisons of identified aerodynamic models with data from the underlying simulation aerodynamic database were used to validate the techniques.
NASA Technical Reports Server (NTRS)
Schmid, F.; Khattak, C. P.; Smith, M. B.; Lynch, L. D.
1982-01-01
Slicing is an important processing step for all technologies based on the use of ingots. A comparison of the economics of three slicing techniques shows that the fixed abrasive slicing technique (FAST) is superior to the internal diameter (ID) and the multiblade slurry (MBS) techniques. Factors affecting contact length are discussed, taking into account kerf width, rocking angle, ingot size, and surface speed. Aspects of blade development are also considered. A high concentration of diamonds on wire has been obtained in wire packs usd for FAST slicing. The material removal rate was found to be directly proportional to the pressure at the diamond tips.
NASA Astrophysics Data System (ADS)
Antolín-Urbaneja, J. C.; Eguizabal, I.; Briz, N.; Dominguez, A.; Estensoro, P.; Secchi, A.; Varriale, A.; Di Giovanni, S.; D'Auria, S.
2013-05-01
Several techniques for detecting chemical drug precursors have been developed in the last decade. Most of them are able to identify molecules at very low concentration under lab conditions. Other commercial devices are able to detect a fixed number and type of target substances based on a single detection technique providing an absence of flexibility with respect to target compounds. The construction of compact and easy to use detection systems providing screening for a large number of compounds being able to discriminate them with low false alarm rate and high probability of detection is still an open concern. Under CUSTOM project, funded by the European Commission within the FP7, a stand-alone portable sensing device based on multiple techniques is being developed. One of these techniques is based on the LED induced fluorescence polarization to detect Ephedrine and Benzyl Methyl Keton (BMK) as a first approach. This technique is highly selective with respect to the target compounds due to the generation of properly engineered fluorescent proteins which are able to bind the target analytes, as it happens in an "immune-type reaction". This paper deals with the advances in the design, construction and validation of the LED induced fluorescence sensor to detect BMK analytes. This sensor includes an analysis module based on high performance LED and PMT detector, a fluidic system to dose suitable quantities of reagents and some printed circuit boards, all of them fixed in a small structure (167mm × 193mm × 228mm) with the capability of working as a stand-alone application.
Optimization of Turbine Blade Design for Reusable Launch Vehicles
NASA Technical Reports Server (NTRS)
Shyy, Wei
1998-01-01
To facilitate design optimization of turbine blade shape for reusable launching vehicles, appropriate techniques need to be developed to process and estimate the characteristics of the design variables and the response of the output with respect to the variations of the design variables. The purpose of this report is to offer insight into developing appropriate techniques for supporting such design and optimization needs. Neural network and polynomial-based techniques are applied to process aerodynamic data obtained from computational simulations for flows around a two-dimensional airfoil and a generic three- dimensional wing/blade. For the two-dimensional airfoil, a two-layered radial-basis network is designed and trained. The performances of two different design functions for radial-basis networks, one based on the accuracy requirement, whereas the other one based on the limit on the network size. While the number of neurons needed to satisfactorily reproduce the information depends on the size of the data, the neural network technique is shown to be more accurate for large data set (up to 765 simulations have been used) than the polynomial-based response surface method. For the three-dimensional wing/blade case, smaller aerodynamic data sets (between 9 to 25 simulations) are considered, and both the neural network and the polynomial-based response surface techniques improve their performance as the data size increases. It is found while the relative performance of two different network types, a radial-basis network and a back-propagation network, depends on the number of input data, the number of iterations required for radial-basis network is less than that for the back-propagation network.
Development and validation of a sensor-based health monitoring model for the Parkview Bridge deck.
DOT National Transportation Integrated Search
2012-01-31
Accelerated bridge construction (ABC) using full-depth precast deck panels is an innovative technique that brings all : the benefits listed under ABC to full fruition. However, this technique needs to be evaluated and the performance of : the bridge ...
APPLICATION OF JET REMPI AND LIBS TO AIR TOXIC MONITORING
The paper discusses three advanced, laser-based monitoring techniques that the EPA is assisting in developing for real time measurement of toxic aerosol compounds. One of the three techniques is jet resonance enhanced multiphoton ionization (Jet REMPI) coupled with a time-of-flig...
Underwater Acoustic Propagation and Communications: A Coupled Research Program
2015-06-15
coding technique suitable for both SIMO and MIMO systems. 4. an adaptive OFDM modulation technique, whereby the transmitter acts in response to...timate based adaptation for SIMO and MIMO systems in a interactive turbo-equalization framework were developed and analyzed. MIMO and SISO
Evidence-based medicine in metastatic spine disease.
Dea, Nicolas; Fisher, Charles G
2014-06-01
Treatment modalities for metastatic spine disease have significantly expanded over the last two decades. This expansion occurred in many different fields. Improvement in surgical techniques and instrumentation now allow the oncologic spine surgeons to effectively circumferentially decompress the neural elements without compromising stability. Percutaneous techniques, both vertebral augmentation and pre-operative endovascular embolization procedures, also greatly benefit patients suffering from spinal column metastasis. Imaging technology advances has contributed to better pre-operative planning and the development of highly conformational radiation techniques, thus permitting the delivery of high-dose radiation to tumors, while avoiding radiotoxicity to the spinal cord and other vital structures. These new developments, combined with evidence-based stability and disease-specific quality of life scores now allow not only better treatment, but also a solid foundation for high-quality research. Spine oncology literature currently suffers from a lack of high-quality evidence due to low prevalence of the disease and complex methodological issues. However, when following evidence-based medicine principles, which incorporate best available evidence, clinical expertise and patient preference, sound, evidence-based recommendations can be made regarding the abovementioned treatment modalities.
López, Lissett; Venteo, Angel; Aguirre, Enara; García, Marga; Rodríguez, Majosé; Amusátegui, Inmaculada; Tesouro, Miguel A; Vela, Carmen; Sainz, Angel; Rueda, Paloma
2007-11-01
An indirect enzyme-linked immunosorbent assay (ELISA) based on baculovirus recombinant P30 protein of Ehrlichia canis and the 1BH4 anticanine IgG monoclonal antibody was developed and evaluated by examining a panel of 98 positive and 157 negative sera using the indirect fluorescent antibody (IFA) test as the reference technique. The P30-based ELISA appeared to be sensitive and specific (77.55% and 95.54%, respectively) when qualitative results (positive/negative) were compared with those of the IFA test; the coefficient of correlation (R) between the 2 tests was 0.833. Furthermore, it was possible to establish a mathematical formula for use in comparing the results of both techniques. These results indicate that recombinant P30 antigen-based ELISA is a suitable alternative of the IFA test for simple, consistent, and rapid serodiagnosis of canine ehrlichiosis. Moreover, the use of this recombinant protein as antigen offers a great advantage for antigen preparation in comparison with other techniques in which the whole E. canis organism is used as antigen.
NASA Technical Reports Server (NTRS)
Gasiewski, Albin J.
1992-01-01
This technique for electronically rotating the polarization basis of an orthogonal-linear polarization radiometer is based on the measurement of the first three feedhorn Stokes parameters, along with the subsequent transformation of this measured Stokes vector into a rotated coordinate frame. The technique requires an accurate measurement of the cross-correlation between the two orthogonal feedhorn modes, for which an innovative polarized calibration load was developed. The experimental portion of this investigation consisted of a proof of concept demonstration of the technique of electronic polarization basis rotation (EPBR) using a ground based 90-GHz dual orthogonal-linear polarization radiometer. Practical calibration algorithms for ground-, aircraft-, and space-based instruments were identified and tested. The theoretical effort consisted of radiative transfer modeling using the planar-stratified numerical model described in Gasiewski and Staelin (1990).
A data compression technique for synthetic aperture radar images
NASA Technical Reports Server (NTRS)
Frost, V. S.; Minden, G. J.
1986-01-01
A data compression technique is developed for synthetic aperture radar (SAR) imagery. The technique is based on an SAR image model and is designed to preserve the local statistics in the image by an adaptive variable rate modification of block truncation coding (BTC). A data rate of approximately 1.6 bit/pixel is achieved with the technique while maintaining the image quality and cultural (pointlike) targets. The algorithm requires no large data storage and is computationally simple.
NASA Astrophysics Data System (ADS)
Shao, Xupeng
2017-04-01
Glutenite bodies are widely developed in northern Minfeng zone of Dongying Sag. Their litho-electric relationship is not clear. In addition, as the conventional sequence stratigraphic research method drawbacks of involving too many subjective human factors, it has limited deepening of the regional sequence stratigraphic research. The wavelet transform technique based on logging data and the time-frequency analysis technique based on seismic data have advantages of dividing sequence stratigraphy quantitatively comparing with the conventional methods. Under the basis of the conventional sequence research method, this paper used the above techniques to divide the fourth-order sequence of the upper Es4 in northern Minfeng zone of Dongying Sag. The research shows that the wavelet transform technique based on logging data and the time-frequency analysis technique based on seismic data are essentially consistent, both of which divide sequence stratigraphy quantitatively in the frequency domain; wavelet transform technique has high resolutions. It is suitable for areas with wells. The seismic time-frequency analysis technique has wide applicability, but a low resolution. Both of the techniques should be combined; the upper Es4 in northern Minfeng zone of Dongying Sag is a complete set of third-order sequence, which can be further subdivided into 5 fourth-order sequences that has the depositional characteristics of fine-upward sequence in granularity. Key words: Dongying sag, northern Minfeng zone, wavelet transform technique, time-frequency analysis technique ,the upper Es4, sequence stratigraphy
Progress in EEG-Based Brain Robot Interaction Systems
Li, Mengfan; Niu, Linwei; Xian, Bin; Zeng, Ming; Chen, Genshe
2017-01-01
The most popular noninvasive Brain Robot Interaction (BRI) technology uses the electroencephalogram- (EEG-) based Brain Computer Interface (BCI), to serve as an additional communication channel, for robot control via brainwaves. This technology is promising for elderly or disabled patient assistance with daily life. The key issue of a BRI system is to identify human mental activities, by decoding brainwaves, acquired with an EEG device. Compared with other BCI applications, such as word speller, the development of these applications may be more challenging since control of robot systems via brainwaves must consider surrounding environment feedback in real-time, robot mechanical kinematics, and dynamics, as well as robot control architecture and behavior. This article reviews the major techniques needed for developing BRI systems. In this review article, we first briefly introduce the background and development of mind-controlled robot technologies. Second, we discuss the EEG-based brain signal models with respect to generating principles, evoking mechanisms, and experimental paradigms. Subsequently, we review in detail commonly used methods for decoding brain signals, namely, preprocessing, feature extraction, and feature classification, and summarize several typical application examples. Next, we describe a few BRI applications, including wheelchairs, manipulators, drones, and humanoid robots with respect to synchronous and asynchronous BCI-based techniques. Finally, we address some existing problems and challenges with future BRI techniques. PMID:28484488
Development of Al2O3 fiber-reinforced Al2O3-based ceramics.
Tanimoto, Yasuhiro; Nemoto, Kimiya
2004-09-01
The purpose of this study was to use a tape casting technique to develop an Al2O3 fiber-reinforced Al2O3-based ceramic material (Al2O3-fiber/Al2O3 composite) into a new type of dental ceramic. The Al2O3-based ceramic used a matrix consisting of 60 wt% Al2O3 powder and 40 wt% SiO2-B2O3 powder. The prepreg sheets of Al2O3-fiber/Al2O3 composite (in which uniaxially aligned Al2O3 fibers were infiltrated with the Al2O3-based matrix) were fabricated continuously using tape casting technique with a doctor blade system. Multilayer preforms of Al2O3-fiber/Al2O3 composite sheets were then sintered at a maximum temperature of 1000 degrees C under an atmospheric pressure in a furnace. The results showed that the shrinkage and bending properties of Al2O3-fiber/Al2O3 composite exceeded those of unreinforced Al2O3--hence demonstrating the positive effects of fiber reinforcement. In conclusion, the tape casting technique has been utilized to successfully develop a new type of dental ceramic material.
DOE Office of Scientific and Technical Information (OSTI.GOV)
C. Priniski, T. Dodson, M. Duco, S. Raftopoulos, R. Ellis, and A. Brooks
In support of the National Compact Stellerator Experiment (NCSX), stellerator assembly activities continued this past year at the Princeton Plasma Physics Laboratory (PPPL) in partnership with the Oak Ridge National Laboratory (ORNL). The construction program saw the completion of the first two Half Field-Period Assemblies (HPA), each consisting of three modular coils. The full machine includes six such sub-assemblies. A single HPA consists of three of the NCSX modular coils wound and assembled at PPPL. These geometrically-complex threedimensional coils were wound using computer-aided metrology and CAD models to tolerances within +/- 0.5mm. The assembly of these coils required similar accuracymore » on a larger scale with the added complexity of more individual parts and fewer degrees of freedom for correction. Several new potential positioning issues developed for which measurement and control techniques were developed. To accomplish this, CAD coordinate-based computer metrology equipment and software similar to the solutions employed for winding the modular coils was used. Given the size of the assemblies, the primary tools were both interferometeraided and Absolute Distance Measurement (ADM)-only based laser trackers. In addition, portable Coordinate Measurement Machine (CMM) arms and some novel indirect measurement techniques were employed. This paper will detail both the use of CAD coordinate-based metrology technology and the techniques developed and employed for dimensional control of NSCX subassemblies. The results achieved and possible improvements to techniques will be discussed.« less
Development of tritium permeation barriers on Al base in Europe
NASA Astrophysics Data System (ADS)
Benamati, G.; Chabrol, C.; Perujo, A.; Rigal, E.; Glasbrenner, H.
The development of the water cooled lithium lead (WCLL) DEMO fusion reactor requires the production of a material capable of acting as a tritium permeation barrier (TPB). In the DEMO blanket reactor permeation barriers on the structural material are required to reduce the tritium permeation from the Pb-17Li or the plasma into the cooling water to acceptable levels (<1 g/d). Because of experimental work previously performed, one of the most promising TPB candidates is A1 base coatings. Within the EU a large R&D programme is in progress to develop a TPB fabrication technique, compatible with the structural materials requirements and capable of producing coatings with acceptable performances. The research is focused on chemical vapour deposition (CVD), hot dipping, hot isostatic pressing (HIP) technology and spray (this one developed also for repair) deposition techniques. The final goal is to select a reference technique to be used in the blanket of the DEMO reactor and in the ITER test module fabrication. The activities performed in four European laboratories are summarised here.
NASA Technical Reports Server (NTRS)
Coon, Craig R.; Cardullo, Frank M.; Zaychik, Kirill B.
2014-01-01
The ability to develop highly advanced simulators is a critical need that has the ability to significantly impact the aerospace industry. The aerospace industry is advancing at an ever increasing pace and flight simulators must match this development with ever increasing urgency. In order to address both current problems and potential advancements with flight simulator techniques, several aspects of current control law technology of the National Aeronautics and Space Administration (NASA) Langley Research Center's Cockpit Motion Facility (CMF) motion base simulator were examined. Preliminary investigation of linear models based upon hardware data were examined to ensure that the most accurate models are used. This research identified both system improvements in the bandwidth and more reliable linear models. Advancements in the compensator design were developed and verified through multiple techniques. The position error rate feedback, the acceleration feedback and the force feedback were all analyzed in the heave direction using the nonlinear model of the hardware. Improvements were made using the position error rate feedback technique. The acceleration feedback compensator also provided noteworthy improvement, while attempts at implementing a force feedback compensator proved unsuccessful.
Internal corrosion monitoring of subsea oil and gas production equipment
DOE Office of Scientific and Technical Information (OSTI.GOV)
Joosten, M.W.; Fischer, K.P.; Strommen, R.
1995-04-01
Nonintrusive techniques will dominate subsea corrosion monitoring compared with the intrusive methods because such methods do not interfere with pipeline operations. The long-term reliability of the nonintrusive techniques in general is considered to be much better than that of intrusive-type probes. The nonintrusive techniques based on radioactive tracers (TLA, NA) and FSM and UT are expected to be the main types of subsea corrosion monitoring equipment in the coming years. Available techniques that could be developed specifically for subsea applications are: electrochemical noise, corrosion potentials (using new types of reference electrodes), multiprobe system for electrochemical measurements, and video camera inspectionmore » (mini-video camera with light source). The following innovative techniques have potential but need further development: ion selective electrodes, radioactive tracers, and Raman spectroscopy.« less
NASA Astrophysics Data System (ADS)
Demigha, Souâd.
2016-03-01
The paper presents a Case-Based Reasoning Tool for Breast Cancer Knowledge Management to improve breast cancer screening. To develop this tool, we combine both concepts and techniques of Case-Based Reasoning (CBR) and Data Mining (DM). Physicians and radiologists ground their diagnosis on their expertise (past experience) based on clinical cases. Case-Based Reasoning is the process of solving new problems based on the solutions of similar past problems and structured as cases. CBR is suitable for medical use. On the other hand, existing traditional hospital information systems (HIS), Radiological Information Systems (RIS) and Picture Archiving Information Systems (PACS) don't allow managing efficiently medical information because of its complexity and heterogeneity. Data Mining is the process of mining information from a data set and transform it into an understandable structure for further use. Combining CBR to Data Mining techniques will facilitate diagnosis and decision-making of medical experts.
Qualitative and quantitative detection of T7 bacteriophages using paper based sandwich ELISA.
Khan, Mohidus Samad; Pande, Tripti; van de Ven, Theo G M
2015-08-01
Viruses cause many infectious diseases and consequently epidemic health threats. Paper based diagnostics and filters can offer attractive options for detecting and deactivating pathogens. However, due to their infectious characteristics, virus detection using paper diagnostics is more challenging compared to the detection of bacteria, enzymes, DNA or antigens. The major objective of this study was to prepare reliable, degradable and low cost paper diagnostics to detect viruses, without using sophisticated optical or microfluidic analytical instruments. T7 bacteriophage was used as a model virus. A paper based sandwich ELISA technique was developed to detect and quantify the T7 phages in solution. The paper based sandwich ELISA detected T7 phage concentrations as low as 100 pfu/mL to as high as 10(9) pfu/mL. The compatibility of paper based sandwich ELISA with the conventional titre count was tested using T7 phage solutions of unknown concentrations. The paper based sandwich ELISA technique is faster and economical compared to the traditional detection techniques. Therefore, with proper calibration and right reagents, and by following the biosafety regulations, the paper based technique can be said to be compatible and economical to the sophisticated laboratory diagnostic techniques applied to detect pathogenic viruses and other microorganisms. Copyright © 2015 Elsevier B.V. All rights reserved.
Human age estimation combining third molar and skeletal development.
Thevissen, P W; Kaur, J; Willems, G
2012-03-01
The wide prediction intervals obtained with age estimation methods based on third molar development could be reduced by combining these dental observations with age-related skeletal information. Therefore, on cephalometric radiographs, the most accurate age-estimating skeletal variable and related registration method were searched and added to a regression model, with age as response and third molar stages as explanatory variable. In a pilot set up on a dataset of 496 (283 M; 213 F) cephalometric radiographs, the techniques of Baccetti et al. (2005) (BA), Seedat et al. (2005) (SE), Caldas et al. (2007) and Rai et al. (2008) (RA) were verified. In the main study, data from 460 (208 F, 224 M) individuals in an age range between 3 and 26 years, for which at the same day an orthopantogram and a cephalogram were taken, were collected. On the orthopantomograms, the left third molar development was registered using the scoring system described by Gleiser and Hunt (1955) and modified by Köhler (1994) (GH). On the cephalograms, cervical vertebrae development was registered according to the BA and SE techniques. A regression model, with age as response and the GH scores as explanatory variable, was fitted to the data. Next, information of BA, SE and BA + SE was, respectively, added to this model. From all obtained models, the determination coefficients and the root mean squared errors were calculated. Inclusion of information from cephalograms based on the BA, as well as the SE, technique improved the amount of explained variance in age acquired from panoramic radiographs using the GH technique with 48%. Inclusion of cephalometric BA + SE information marginally improved the previous result (+1%). The RMSE decreased with 1.93, 1.85 and 2.03 years by adding, respectively, BA, SE and BA + SE information to the GH model. The SE technique allows clinically the fastest and easiest registration of the degree of development of the cervical vertebrae. Therefore, the choice of technique to classify cervical vertebrae development in addition to third molar development is preferably the SE technique.
NASA Technical Reports Server (NTRS)
Sellers, William L., III; Dwoyer, Douglas L.
1992-01-01
The design of a hypersonic aircraft poses unique challenges to the engineering community. Problems with duplicating flight conditions in ground based facilities have made performance predictions risky. Computational fluid dynamics (CFD) has been proposed as an additional means of providing design data. At the present time, CFD codes are being validated based on sparse experimental data and then used to predict performance at flight conditions with generally unknown levels of uncertainty. This paper will discuss the facility and measurement techniques that are required to support CFD development for the design of hypersonic aircraft. Illustrations are given of recent success in combining experimental and direct numerical simulation in CFD model development and validation for hypersonic perfect gas flows.
Yu, Peiqiang
2007-01-01
Synchrotron-based Fourier transform infrared microspectroscopy (S-FTIR) has been developed as a rapid, direct, non-destructive, bioanalytical technique. This technique takes advantage of synchrotron light brightness and small effective source size and is capable of exploring the molecular chemical features and make-up within microstructures of a biological tissue without destruction of inherent structures at ultra-spatial resolutions within cellular dimension. To date there has been very little application of this advanced synchrotron technique to the study of plant and animal tissues' inherent structure at a cellular or subcellular level. In this article, a novel approach was introduced to show the potential of themore » newly developed, advanced synchrotron-based analytical technology, which can be used to reveal molecular structural-chemical features of various plant and animal tissues.« less
Roeber, Florian; Kahn, Lewis
2014-10-15
The specific diagnosis of gastrointestinal nematode infections in ruminants is routinely based on larval culture technique and on the morphological identification of developed third-stage larvae. However, research on the ecology and developmental requirements of different species suggests that environmental conditions (e.g., temperature and humidity) for optimal development to occur vary between the different species. Thus, employing a common culture protocol for all species will favour the development of certain species over others and can cause a biased result in particular when species proportions in a mixed infection are to be determined. Furthermore, the morphological identification of L3 larvae is complicated by a lack of distinctive, obvious features that would allow the identification of all key species. In the present paper we review in detail the potential limitations of larval culture technique and morphological identification and provide account to some modern molecular alternatives to the specific diagnosis of gastrointestinal nematode infection in ruminants. Copyright © 2014 Elsevier B.V. All rights reserved.
NASA Technical Reports Server (NTRS)
Turso, James; Lawrence, Charles; Litt, Jonathan
2004-01-01
The development of a wavelet-based feature extraction technique specifically targeting FOD-event induced vibration signal changes in gas turbine engines is described. The technique performs wavelet analysis of accelerometer signals from specified locations on the engine and is shown to be robust in the presence of significant process and sensor noise. It is envisioned that the technique will be combined with Kalman filter thermal/health parameter estimation for FOD-event detection via information fusion from these (and perhaps other) sources. Due to the lack of high-frequency FOD-event test data in the open literature, a reduced-order turbofan structural model (ROM) was synthesized from a finite element model modal analysis to support the investigation. In addition to providing test data for algorithm development, the ROM is used to determine the optimal sensor location for FOD-event detection. In the presence of significant noise, precise location of the FOD event in time was obtained using the developed wavelet-based feature.
NASA Technical Reports Server (NTRS)
Turso, James A.; Lawrence, Charles; Litt, Jonathan S.
2007-01-01
The development of a wavelet-based feature extraction technique specifically targeting FOD-event induced vibration signal changes in gas turbine engines is described. The technique performs wavelet analysis of accelerometer signals from specified locations on the engine and is shown to be robust in the presence of significant process and sensor noise. It is envisioned that the technique will be combined with Kalman filter thermal/ health parameter estimation for FOD-event detection via information fusion from these (and perhaps other) sources. Due to the lack of high-frequency FOD-event test data in the open literature, a reduced-order turbofan structural model (ROM) was synthesized from a finite-element model modal analysis to support the investigation. In addition to providing test data for algorithm development, the ROM is used to determine the optimal sensor location for FOD-event detection. In the presence of significant noise, precise location of the FOD event in time was obtained using the developed wavelet-based feature.
Large Terrain Modeling and Visualization for Planets
NASA Technical Reports Server (NTRS)
Myint, Steven; Jain, Abhinandan; Cameron, Jonathan; Lim, Christopher
2011-01-01
Physics-based simulations are actively used in the design, testing, and operations phases of surface and near-surface planetary space missions. One of the challenges in realtime simulations is the ability to handle large multi-resolution terrain data sets within models as well as for visualization. In this paper, we describe special techniques that we have developed for visualization, paging, and data storage for dealing with these large data sets. The visualization technique uses a real-time GPU-based continuous level-of-detail technique that delivers multiple frames a second performance even for planetary scale terrain model sizes.
Accurate low-cost methods for performance evaluation of cache memory systems
NASA Technical Reports Server (NTRS)
Laha, Subhasis; Patel, Janak H.; Iyer, Ravishankar K.
1988-01-01
Methods of simulation based on statistical techniques are proposed to decrease the need for large trace measurements and for predicting true program behavior. Sampling techniques are applied while the address trace is collected from a workload. This drastically reduces the space and time needed to collect the trace. Simulation techniques are developed to use the sampled data not only to predict the mean miss rate of the cache, but also to provide an empirical estimate of its actual distribution. Finally, a concept of primed cache is introduced to simulate large caches by the sampling-based method.
NASA Technical Reports Server (NTRS)
Sheffner, E. J.; Hlavka, C. A.; Bauer, E. M.
1984-01-01
Two techniques have been developed for the mapping and area estimation of small grains in California from Landsat digital data. The two techniques are Band Ratio Thresholding, a semi-automated version of a manual procedure, and LCLS, a layered classification technique which can be fully automated and is based on established clustering and classification technology. Preliminary evaluation results indicate that the two techniques have potential for providing map products which can be incorporated into existing inventory procedures and automated alternatives to traditional inventory techniques and those which currently employ Landsat imagery.
Shock and vibration technology with applications to electrical systems
NASA Technical Reports Server (NTRS)
Eshleman, R. L.
1972-01-01
A survey is presented of shock and vibration technology for electrical systems developed by the aerospace programs. The shock environment is surveyed along with new techniques for modeling, computer simulation, damping, and response analysis. Design techniques based on the use of analog computers, shock spectra, optimization, and nonlinear isolation are discussed. Shock mounting of rotors for performance and survival, and vibration isolation techniques are reviewed.
Interventional MR: vascular applications.
Smits, H F; Bos, C; van der Weide, R; Bakker, C J
1999-01-01
Three strategies for visualisation of MR-dedicated guidewires and catheters have been proposed, namely active tracking, the technique of locally induced field inhomogeneity and passive susceptibility-based tracking. In this article the pros and cons of these techniques are discussed, including the development of MR-dedicated guidewires and catheters, scan techniques, post-processing tools, and display facilities for MR tracking. Finally, some of the results obtained with MR tracking are discussed.
Foodomics: MS-based strategies in modern food science and nutrition.
Herrero, Miguel; Simó, Carolina; García-Cañas, Virginia; Ibáñez, Elena; Cifuentes, Alejandro
2012-01-01
Modern research in food science and nutrition is moving from classical methodologies to advanced analytical strategies in which MS-based techniques play a crucial role. In this context, Foodomics has been recently defined as a new discipline that studies food and nutrition domains through the application of advanced omics technologies in which MS techniques are considered indispensable. Applications of Foodomics include the genomic, transcriptomic, proteomic, and/or metabolomic study of foods for compound profiling, authenticity, and/or biomarker-detection related to food quality or safety; the development of new transgenic foods, food contaminants, and whole toxicity studies; new investigations on food bioactivity, food effects on human health, etc. This review work does not intend to provide an exhaustive revision of the many works published so far on food analysis using MS techniques. The aim of the present work is to provide an overview of the different MS-based strategies that have been (or can be) applied in the new field of Foodomics, discussing their advantages and drawbacks. Besides, some ideas about the foreseen development and applications of MS-techniques in this new discipline are also provided. Copyright © 2011 Wiley Periodicals, Inc.
Wei, Xuelei; Dong, Fuhui
2011-12-01
To review recent advance in the research and application of computer aided forming techniques for constructing bone tissue engineering scaffolds. The literature concerning computer aided forming techniques for constructing bone tissue engineering scaffolds in recent years was reviewed extensively and summarized. Several studies over last decade have focused on computer aided forming techniques for bone scaffold construction using various scaffold materials, which is based on computer aided design (CAD) and bone scaffold rapid prototyping (RP). CAD include medical CAD, STL, and reverse design. Reverse design can fully simulate normal bone tissue and could be very useful for the CAD. RP techniques include fused deposition modeling, three dimensional printing, selected laser sintering, three dimensional bioplotting, and low-temperature deposition manufacturing. These techniques provide a new way to construct bone tissue engineering scaffolds with complex internal structures. With rapid development of molding and forming techniques, computer aided forming techniques are expected to provide ideal bone tissue engineering scaffolds.
NASA Technical Reports Server (NTRS)
Grubbs, Guy II; Michell, Robert; Samara, Marilia; Hampton, Don; Jahn, Jorg-Micha
2016-01-01
A technique is presented for the periodic and systematic calibration of ground-based optical imagers. It is important to have a common system of units (Rayleighs or photon flux) for cross comparison as well as self-comparison over time. With the advancement in technology, the sensitivity of these imagers has improved so that stars can be used for more precise calibration. Background subtraction, flat fielding, star mapping, and other common techniques are combined in deriving a calibration technique appropriate for a variety of ground-based imager installations. Spectral (4278, 5577, and 8446 A ) ground-based imager data with multiple fields of view (19, 47, and 180 deg) are processed and calibrated using the techniques developed. The calibration techniques applied result in intensity measurements in agreement between different imagers using identical spectral filtering, and the intensity at each wavelength observed is within the expected range of auroral measurements. The application of these star calibration techniques, which convert raw imager counts into units of photon flux, makes it possible to do quantitative photometry. The computed photon fluxes, in units of Rayleighs, can be used for the absolute photometry between instruments or as input parameters for auroral electron transport models.
Development of evaluation technique of GMAW welding quality based on statistical analysis
NASA Astrophysics Data System (ADS)
Feng, Shengqiang; Terasaki, Hidenri; Komizo, Yuichi; Hu, Shengsun; Chen, Donggao; Ma, Zhihua
2014-11-01
Nondestructive techniques for appraising gas metal arc welding(GMAW) faults plays a very important role in on-line quality controllability and prediction of the GMAW process. On-line welding quality controllability and prediction have several disadvantages such as high cost, low efficiency, complication and greatly being affected by the environment. An enhanced, efficient evaluation technique for evaluating welding faults based on Mahalanobis distance(MD) and normal distribution is presented. In addition, a new piece of equipment, designated the weld quality tester(WQT), is developed based on the proposed evaluation technique. MD is superior to other multidimensional distances such as Euclidean distance because the covariance matrix used for calculating MD takes into account correlations in the data and scaling. The values of MD obtained from welding current and arc voltage are assumed to follow a normal distribution. The normal distribution has two parameters: the mean µ and standard deviation σ of the data. In the proposed evaluation technique used by the WQT, values of MD located in the range from zero to µ+3 σ are regarded as "good". Two experiments which involve changing the flow of shielding gas and smearing paint on the surface of the substrate are conducted in order to verify the sensitivity of the proposed evaluation technique and the feasibility of using WQT. The experimental results demonstrate the usefulness of the WQT for evaluating welding quality. The proposed technique can be applied to implement the on-line welding quality controllability and prediction, which is of great importance to design some novel equipment for weld quality detection.
Experiments on Adaptive Techniques for Host-Based Intrusion Detection
DOE Office of Scientific and Technical Information (OSTI.GOV)
DRAELOS, TIMOTHY J.; COLLINS, MICHAEL J.; DUGGAN, DAVID P.
2001-09-01
This research explores four experiments of adaptive host-based intrusion detection (ID) techniques in an attempt to develop systems that can detect novel exploits. The technique considered to have the most potential is adaptive critic designs (ACDs) because of their utilization of reinforcement learning, which allows learning exploits that are difficult to pinpoint in sensor data. Preliminary results of ID using an ACD, an Elman recurrent neural network, and a statistical anomaly detection technique demonstrate an ability to learn to distinguish between clean and exploit data. We used the Solaris Basic Security Module (BSM) as a data source and performed considerablemore » preprocessing on the raw data. A detection approach called generalized signature-based ID is recommended as a middle ground between signature-based ID, which has an inability to detect novel exploits, and anomaly detection, which detects too many events including events that are not exploits. The primary results of the ID experiments demonstrate the use of custom data for generalized signature-based intrusion detection and the ability of neural network-based systems to learn in this application environment.« less
Development of sensing techniques for weaponry health monitoring
NASA Astrophysics Data System (ADS)
Edwards, Eugene; Ruffin, Paul B.; Walker, Ebonee A.; Brantley, Christina L.
2013-04-01
Due to the costliness of destructive evaluation methods for assessing the aging and shelf-life of missile and rocket components, the identification of nondestructive evaluation methods has become increasingly important to the Army. Verifying that there is a sufficient concentration of stabilizer is a dependable indicator that the missile's double-based solid propellant is viable. The research outlined in this paper summarizes the Army Aviation and Missile Research, Development, and Engineering Center's (AMRDEC's) comparative use of nanoporous membranes, carbon nanotubes, and optical spectroscopic configured sensing techniques for detecting degradation in rocket motor propellant. The first sensing technique utilizes a gas collecting chamber consisting of nanoporous structures that trap the smaller solid propellant particles for measurement by a gas analysis device. In collaboration with NASA-Ames, sensing methods are developed that utilize functionalized single-walled carbon nanotubes as the key sensing element. The optical spectroscopic sensing method is based on a unique light collecting optical fiber system designed to detect the concentration of the propellant stabilizer. Experimental setups, laboratory results, and overall effectiveness of each technique are presented in this paper. Expectations are for the three sensing mechanisms to provide nondestructive evaluation methods that will offer cost-savings and improved weaponry health monitoring.
Current issues and future perspectives of gastric cancer screening
Hamashima, Chisato
2014-01-01
Gastric cancer remains the second leading cause of cancer death worldwide. About half of the incidence of gastric cancer is observed in East Asian countries, which show a higher mortality than other countries. The effectiveness of 3 new gastric cancer screening techniques, namely, upper gastrointestinal endoscopy, serological testing, and “screen and treat” method were extensively reviewed. Moreover, the phases of development for cancer screening were analyzed on the basis of the biomarker development road map. Several observational studies have reported the effectiveness of endoscopic screening in reducing mortality from gastric cancer. On the other hand, serologic testing has mainly been used for targeting the high-risk group for gastric cancer. To date, the effectiveness of new techniques for gastric cancer screening has remained limited. However, endoscopic screening is presently in the last trial phase of development before their introduction to population-based screening. To effectively introduce new techniques for gastric cancer screening in a community, incidence and mortality reduction from gastric cancer must be initially and thoroughly evaluated by conducting reliable studies. In addition to effectiveness evaluation, the balance of benefits and harms must be carefully assessed before introducing these new techniques for population-based screening. PMID:25320514
NASA Technical Reports Server (NTRS)
Sidney, T.; Aylott, B.; Christensen, N.; Farr, B.; Farr, W.; Feroz, F.; Gair, J.; Grover, K.; Graff, P.; Hanna, C.;
2014-01-01
The problem of reconstructing the sky position of compact binary coalescences detected via gravitational waves is a central one for future observations with the ground-based network of gravitational-wave laser interferometers, such as Advanced LIGO and Advanced Virgo. Different techniques for sky localization have been independently developed. They can be divided in two broad categories: fully coherent Bayesian techniques, which are high latency and aimed at in-depth studies of all the parameters of a source, including sky position, and "triangulation-based" techniques, which exploit the data products from the search stage of the analysis to provide an almost real-time approximation of the posterior probability density function of the sky location of a detection candidate. These techniques have previously been applied to data collected during the last science runs of gravitational-wave detectors operating in the so-called initial configuration. Here, we develop and analyze methods for assessing the self consistency of parameter estimation methods and carrying out fair comparisons between different algorithms, addressing issues of efficiency and optimality. These methods are general, and can be applied to parameter estimation problems other than sky localization. We apply these methods to two existing sky localization techniques representing the two above-mentioned categories, using a set of simulated inspiralonly signals from compact binary systems with a total mass of equal to or less than 20M solar mass and nonspinning components. We compare the relative advantages and costs of the two techniques and show that sky location uncertainties are on average a factor approx. equals 20 smaller for fully coherent techniques than for the specific variant of the triangulation-based technique used during the last science runs, at the expense of a factor approx. equals 1000 longer processing time.
NASA Astrophysics Data System (ADS)
Sidery, T.; Aylott, B.; Christensen, N.; Farr, B.; Farr, W.; Feroz, F.; Gair, J.; Grover, K.; Graff, P.; Hanna, C.; Kalogera, V.; Mandel, I.; O'Shaughnessy, R.; Pitkin, M.; Price, L.; Raymond, V.; Röver, C.; Singer, L.; van der Sluys, M.; Smith, R. J. E.; Vecchio, A.; Veitch, J.; Vitale, S.
2014-04-01
The problem of reconstructing the sky position of compact binary coalescences detected via gravitational waves is a central one for future observations with the ground-based network of gravitational-wave laser interferometers, such as Advanced LIGO and Advanced Virgo. Different techniques for sky localization have been independently developed. They can be divided in two broad categories: fully coherent Bayesian techniques, which are high latency and aimed at in-depth studies of all the parameters of a source, including sky position, and "triangulation-based" techniques, which exploit the data products from the search stage of the analysis to provide an almost real-time approximation of the posterior probability density function of the sky location of a detection candidate. These techniques have previously been applied to data collected during the last science runs of gravitational-wave detectors operating in the so-called initial configuration. Here, we develop and analyze methods for assessing the self consistency of parameter estimation methods and carrying out fair comparisons between different algorithms, addressing issues of efficiency and optimality. These methods are general, and can be applied to parameter estimation problems other than sky localization. We apply these methods to two existing sky localization techniques representing the two above-mentioned categories, using a set of simulated inspiral-only signals from compact binary systems with a total mass of ≤20M⊙ and nonspinning components. We compare the relative advantages and costs of the two techniques and show that sky location uncertainties are on average a factor ≈20 smaller for fully coherent techniques than for the specific variant of the triangulation-based technique used during the last science runs, at the expense of a factor ≈1000 longer processing time.
2014-01-01
Background Inter-professional learning has been promoted as the solution to many clinical management issues. One such issue is the correct use of asthma inhaler devices. Up to 80% of people with asthma use their inhaler device incorrectly. The implications of this are poor asthma control and quality of life. Correct inhaler technique can be taught, however these educational instructions need to be repeated if correct technique is to be maintained. It is important to maximise the opportunities to deliver this education in primary care. In light of this, it is important to explore how health care providers, in particular pharmacists and general medical practitioners, can work together in delivering inhaler technique education to patients, over time. Therefore, there is a need to develop and evaluate effective inter-professional education, which will address the need to educate patients in the correct use of their inhalers as well as equip health care professionals with skills to engage in collaborative relationships with each other. Methods This mixed methods study involves the development and evaluation of three modules of continuing education, Model 1, Model 2 and Model 3. A fourth group, Model 4, acting as a control. Model 1 consists of face-to-face continuing professional education on asthma inhaler technique, aimed at pharmacists, general medical practitioners and their practice nurses. Model 2 is an electronic online continuing education module based on Model 1 principles. Model 3 is also based on asthma inhaler technique education but employs a learning intervention targeting health care professional relationships and is based on sociocultural theory. This study took the form of a parallel group, repeated measure design. Following the completion of continuing professional education, health care professionals recruited people with asthma and followed them up for 6 months. During this period, inhaler device technique training was delivered and data on patient inhaler technique, clinical and humanistic outcomes were collected. Outcomes related to professional collaborative relationships were also measured. Discussion Challenges presented included the requirement of significant financial resources for development of study materials and limited availability of validated tools to measure health care professional collaboration over time. PMID:24708800
Bosnic-Anticevich, Sinthia Z; Stuart, Meg; Mackson, Judith; Cvetkovski, Biljana; Sainsbury, Erica; Armour, Carol; Mavritsakis, Sofia; Mendrela, Gosia; Travers-Mason, Pippa; Williamson, Margaret
2014-04-07
Inter-professional learning has been promoted as the solution to many clinical management issues. One such issue is the correct use of asthma inhaler devices. Up to 80% of people with asthma use their inhaler device incorrectly. The implications of this are poor asthma control and quality of life. Correct inhaler technique can be taught, however these educational instructions need to be repeated if correct technique is to be maintained. It is important to maximise the opportunities to deliver this education in primary care. In light of this, it is important to explore how health care providers, in particular pharmacists and general medical practitioners, can work together in delivering inhaler technique education to patients, over time. Therefore, there is a need to develop and evaluate effective inter-professional education, which will address the need to educate patients in the correct use of their inhalers as well as equip health care professionals with skills to engage in collaborative relationships with each other. This mixed methods study involves the development and evaluation of three modules of continuing education, Model 1, Model 2 and Model 3. A fourth group, Model 4, acting as a control.Model 1 consists of face-to-face continuing professional education on asthma inhaler technique, aimed at pharmacists, general medical practitioners and their practice nurses.Model 2 is an electronic online continuing education module based on Model 1 principles.Model 3 is also based on asthma inhaler technique education but employs a learning intervention targeting health care professional relationships and is based on sociocultural theory.This study took the form of a parallel group, repeated measure design. Following the completion of continuing professional education, health care professionals recruited people with asthma and followed them up for 6 months. During this period, inhaler device technique training was delivered and data on patient inhaler technique, clinical and humanistic outcomes were collected. Outcomes related to professional collaborative relationships were also measured. Challenges presented included the requirement of significant financial resources for development of study materials and limited availability of validated tools to measure health care professional collaboration over time.
NASA Astrophysics Data System (ADS)
Chandramouli, Rajarathnam; Li, Grace; Memon, Nasir D.
2002-04-01
Steganalysis techniques attempt to differentiate between stego-objects and cover-objects. In recent work we developed an explicit analytic upper bound for the steganographic capacity of LSB based steganographic techniques for a given false probability of detection. In this paper we look at adaptive steganographic techniques. Adaptive steganographic techniques take explicit steps to escape detection. We explore different techniques that can be used to adapt message embedding to the image content or to a known steganalysis technique. We investigate the advantages of adaptive steganography within an analytical framework. We also give experimental results with a state-of-the-art steganalysis technique demonstrating that adaptive embedding results in a significant number of bits embedded without detection.
Zakhia, Frédéric; de Lajudie, Philippe
2006-03-01
Taxonomy is the science that studies the relationships between organisms. It comprises classification, nomenclature, and identification. Modern bacterial taxonomy is polyphasic. This means that it is based on several molecular techniques, each one retrieving the information at different cellular levels (proteins, fatty acids, DNA...). The obtained results are combined and analysed to reach a "consensus taxonomy" of a microorganism. Until 1970, a small number of classification techniques were available for microbiologists (mainly phenotypic characterization was performed: a legume species nodulation ability for a Rhizobium, for example). With the development of techniques based on polymerase chain reaction for characterization, the bacterial taxonomy has undergone great changes. In particular, the classification of the legume nodulating bacteria has been repeatedly modified over the last 20 years. We present here a review of the currently used molecular techniques in bacterial characterization, with examples of application of these techniques for the study of the legume nodulating bacteria.
A frequency domain radar interferometric imaging (FII) technique based on high-resolution methods
NASA Astrophysics Data System (ADS)
Luce, H.; Yamamoto, M.; Fukao, S.; Helal, D.; Crochet, M.
2001-01-01
In the present work, we propose a frequency-domain interferometric imaging (FII) technique for a better knowledge of the vertical distribution of the atmospheric scatterers detected by MST radars. This is an extension of the dual frequency-domain interferometry (FDI) technique to multiple frequencies. Its objective is to reduce the ambiguity (resulting from the use of only two adjacent frequencies), inherent with the FDI technique. Different methods, commonly used in antenna array processing, are first described within the context of application to the FII technique. These methods are the Fourier-based imaging, the Capon's and the singular value decomposition method used with the MUSIC algorithm. Some preliminary simulations and tests performed on data collected with the middle and upper atmosphere (MU) radar (Shigaraki, Japan) are also presented. This work is a first step in the developments of the FII technique which seems to be very promising.
All NbN tunnel junction fabrication
NASA Technical Reports Server (NTRS)
Leduc, H. G.; Khanna, S. K.; Stern, J. A.
1987-01-01
The development of SIS tunnel junctions based on NbN for mixer applications in the submillimeter range is reported. The unique technological challenges inherent in the development of all refractory-compound superconductor-based tunnel junctions are highlighted. Current deposition and fabrication techniques are discussed, and the current status of all-NbN tunnel junctions is reported.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Kuhl, D.E.
1976-08-05
During the thirteen year duration of this contract the goal has been to develop and apply computer based analysis of radionuclide scan data so as to make available improved diagnostic information based on a knowledge of localized quantitative estimates of radionuclide concentration. Results are summarized. (CH)
NASA Astrophysics Data System (ADS)
Avitabile, P.; O'Callahan, J.
2003-07-01
Inclusion of rotational effects is critical for the accuracy of the predicted system characteristics, in almost all system modelling studies. However, experimentally derived information for the description of one or more of the components for the system will generally not have any rotational effects included in the description of the component. The lack of rotational effects has long affected the results from any system model development whether using a modal-based approach or an impedance-based approach. Several new expansion processes are described herein for the development of FRFs needed for impedance-based system models. These techniques expand experimentally derived mode shapes, residual modes from the modal parameter estimation process and FRFs directly to allow for the inclusion of the necessary rotational dof. The FRFs involving translational to rotational dofs are developed as well as the rotational to rotational dof. Examples are provided to show the use of these techniques.
DOT National Transportation Integrated Search
2012-01-31
Accelerated bridge construction (ABC) using full-depth precast deck panels is an innovative technique that brings all the benefits listed under ABC to full fruition. However, this technique needs to be evaluated and the performance of the bridge need...
A Structural and Content-Based Analysis for Web Filtering.
ERIC Educational Resources Information Center
Lee, P. Y.; Hui, S. C.; Fong, A. C. M.
2003-01-01
Presents an analysis of the distinguishing features of pornographic Web pages so that effective filtering techniques can be developed. Surveys the existing techniques for Web content filtering and describes the implementation of a Web content filtering system that uses an artificial neural network. (Author/LRW)
Developing Scenarios: Linking Environmental Scanning and Strategic Planning.
ERIC Educational Resources Information Center
Whiteley, Meredith A.; And Others
1990-01-01
The multiple scenario analysis technique for organizational planning used by multinational corporations is adaptable for colleges and universities. Arizona State University launched a futures-based planning project using the Delphi technique and cross-impact analysis to produce three alternative scenarios (stable, turbulent, and chaotic) to expand…
Comparability of a Paper-Based Language Test and a Computer-Based Language Test.
ERIC Educational Resources Information Center
Choi, Inn-Chull; Kim, Kyoung Sung; Boo, Jaeyool
2003-01-01
Utilizing the Test of English Proficiency, developed by Seoul National University (TEPS), examined comparability between the paper-based language test and the computer-based language test based on content and construct validation employing content analyses based on corpus linguistic techniques in addition to such statistical analyses as…
A review of approaches to identifying patient phenotype cohorts using electronic health records
Shivade, Chaitanya; Raghavan, Preethi; Fosler-Lussier, Eric; Embi, Peter J; Elhadad, Noemie; Johnson, Stephen B; Lai, Albert M
2014-01-01
Objective To summarize literature describing approaches aimed at automatically identifying patients with a common phenotype. Materials and methods We performed a review of studies describing systems or reporting techniques developed for identifying cohorts of patients with specific phenotypes. Every full text article published in (1) Journal of American Medical Informatics Association, (2) Journal of Biomedical Informatics, (3) Proceedings of the Annual American Medical Informatics Association Symposium, and (4) Proceedings of Clinical Research Informatics Conference within the past 3 years was assessed for inclusion in the review. Only articles using automated techniques were included. Results Ninety-seven articles met our inclusion criteria. Forty-six used natural language processing (NLP)-based techniques, 24 described rule-based systems, 41 used statistical analyses, data mining, or machine learning techniques, while 22 described hybrid systems. Nine articles described the architecture of large-scale systems developed for determining cohort eligibility of patients. Discussion We observe that there is a rise in the number of studies associated with cohort identification using electronic medical records. Statistical analyses or machine learning, followed by NLP techniques, are gaining popularity over the years in comparison with rule-based systems. Conclusions There are a variety of approaches for classifying patients into a particular phenotype. Different techniques and data sources are used, and good performance is reported on datasets at respective institutions. However, no system makes comprehensive use of electronic medical records addressing all of their known weaknesses. PMID:24201027
Application of Discrete Fracture Modeling and Upscaling Techniques to Complex Fractured Reservoirs
NASA Astrophysics Data System (ADS)
Karimi-Fard, M.; Lapene, A.; Pauget, L.
2012-12-01
During the last decade, an important effort has been made to improve data acquisition (seismic and borehole imaging) and workflow for reservoir characterization which has greatly benefited the description of fractured reservoirs. However, the geological models resulting from the interpretations need to be validated or calibrated against dynamic data. Flow modeling in fractured reservoirs remains a challenge due to the difficulty of representing mass transfers at different heterogeneity scales. The majority of the existing approaches are based on dual continuum representation where the fracture network and the matrix are represented separately and their interactions are modeled using transfer functions. These models are usually based on idealized representation of the fracture distribution which makes the integration of real data difficult. In recent years, due to increases in computer power, discrete fracture modeling techniques (DFM) are becoming popular. In these techniques the fractures are represented explicitly allowing the direct use of data. In this work we consider the DFM technique developed by Karimi-Fard et al. [1] which is based on an unstructured finite-volume discretization. The mass flux between two adjacent control-volumes is evaluated using an optimized two-point flux approximation. The result of the discretization is a list of control-volumes with the associated pore-volumes and positions, and a list of connections with the associated transmissibilities. Fracture intersections are simplified using a connectivity transformation which contributes considerably to the efficiency of the methodology. In addition, the method is designed for general purpose simulators and any connectivity based simulator can be used for flow simulations. The DFM technique is either used standalone or as part of an upscaling technique. The upscaling techniques are required for large reservoirs where the explicit representation of all fractures and faults is not possible. Karimi-Fard et al. [2] have developed an upscaling technique based on DFM representation. The original version of this technique was developed to construct a dual-porosity model from a discrete fracture description. This technique has been extended and generalized so it can be applied to a wide range of problems from reservoirs with a few or no fracture to highly fractured reservoirs. In this work, we present the application of these techniques to two three-dimensional fractured reservoirs constructed using real data. The first model contains more than 600 medium and large scale fractures. The fractures are not always connected which requires a general modeling technique. The reservoir has 50 wells (injectors and producers) and water flooding simulations are performed. The second test case is a larger reservoir with sparsely distributed faults. Single-phase simulations are performed with 5 producing wells. [1] Karimi-Fard M., Durlofsky L.J., and Aziz K. 2004. An efficient discrete-fracture model applicable for general-purpose reservoir simulators. SPE Journal, 9(2): 227-236. [2] Karimi-Fard M., Gong B., and Durlofsky L.J. 2006. Generation of coarse-scale continuum flow models from detailed fracture characterizations. Water Resources Research, 42(10): W10423.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Qi, X. Sharon, E-mail: xqi@mednet.ucla.edu; Ruan, Dan; Lee, Steve P.
2015-03-15
Purpose: To develop a practical workflow for retrospectively analyzing target and normal tissue dose–volume endpoints for various intensity modulated radiation therapy (IMRT) delivery techniques; to develop technique-specific planning goals to improve plan consistency and quality when feasible. Methods and Materials: A total of 165 consecutive head-and-neck patients from our patient registry were selected and retrospectively analyzed. All IMRT plans were generated using the same dose–volume guidelines for TomoTherapy (Tomo, Accuray), TrueBeam (TB, Varian) using fixed-field IMRT (TB-IMRT) or RAPIDARC (TB-RAPIDARC), or Siemens Oncor (Siemens-IMRT, Siemens). A MATLAB-based dose–volume extraction and analysis tool was developed to export dosimetric endpoints for eachmore » patient. With a fair stratification of patient cohort, the variation of achieved dosimetric endpoints was analyzed among different treatment techniques. Upon identification of statistically significant variations, technique-specific planning goals were derived from dynamically accumulated institutional data. Results: Retrospective analysis showed that although all techniques yielded comparable target coverage, the doses to the critical structures differed. The maximum cord doses were 34.1 ± 2.6, 42.7 ± 2.1, 43.3 ± 2.0, and 45.1 ± 1.6 Gy for Tomo, TB-IMRT, TB-RAPIDARC, and Siemens-IMRT plans, respectively. Analyses of variance showed significant differences for the maximum cord doses but no significant differences for other selected structures among the investigated IMRT delivery techniques. Subsequently, a refined technique-specific dose–volume guideline for maximum cord dose was derived at a confidence level of 95%. The dosimetric plans that failed the refined technique-specific planning goals were reoptimized according to the refined constraints. We observed better cord sparing with minimal variations for the target coverage and other organ at risk sparing for the Tomo cases, and higher parotid doses for C-arm linear accelerator–based IMRT and RAPIDARC plans. Conclusion: Patient registry–based processes allowed easy and systematic dosimetric assessment of treatment plan quality and consistency. Our analysis revealed the dependence of certain dosimetric endpoints on the treatment techniques. Technique-specific refinement of planning goals may lead to improvement in plan consistency and plan quality.« less
Space shuttle recommendations based on aircraft maintenance experience
NASA Technical Reports Server (NTRS)
Spears, J. M.; Fox, C. L.
1972-01-01
Space shuttle design recommendations based on aircraft maintenance experience are developed. The recommendations are specifically applied to the landing gear system, nondestructive inspection techniques, hydraulic system design, materials and processes, and program support.
Druzinec, Damir; Salzig, Denise; Brix, Alexander; Kraume, Matthias; Vilcinskas, Andreas; Kollewe, Christian; Czermak, Peter
2013-01-01
Due to the increasing use of insect cell based expression systems in research and industrial recombinant protein production, the development of efficient and reproducible production processes remains a challenging task. In this context, the application of online monitoring techniques is intended to ensure high and reproducible product qualities already during the early phases of process development. In the following chapter, the most common transient and stable insect cell based expression systems are briefly introduced. Novel applications of insect cell based expression systems for the production of insect derived antimicrobial peptides/proteins (AMPs) are discussed using the example of G. mellonella derived gloverin. Suitable in situ sensor techniques for insect cell culture monitoring in disposable and common bioreactor systems are outlined with respect to optical and capacitive sensor concepts. Since scale up of production processes is one of the most critical steps in process development, a conclusive overview is given about scale up aspects for industrial insect cell culture processes.
Niu, Qiang; Chi, Xiaoyi; Leu, Ming C; Ochoa, Jorge
2008-01-01
This paper describes image processing, geometric modeling and data management techniques for the development of a virtual bone surgery system. Image segmentation is used to divide CT scan data into different segments representing various regions of the bone. A region-growing algorithm is used to extract cortical bone and trabecular bone structures systematically and efficiently. Volume modeling is then used to represent the bone geometry based on the CT scan data. Material removal simulation is achieved by continuously performing Boolean subtraction of the surgical tool model from the bone model. A quadtree-based adaptive subdivision technique is developed to handle the large set of data in order to achieve the real-time simulation and visualization required for virtual bone surgery. A Marching Cubes algorithm is used to generate polygonal faces from the volumetric data. Rendering of the generated polygons is performed with the publicly available VTK (Visualization Tool Kit) software. Implementation of the developed techniques consists of developing a virtual bone-drilling software program, which allows the user to manipulate a virtual drill to make holes with the use of a PHANToM device on a bone model derived from real CT scan data.
Devices development and techniques research for space life sciences
NASA Astrophysics Data System (ADS)
Zhang, A.; Liu, B.; Zheng, C.
The development process and the status quo of the devices and techniques for space life science in China and the main research results in this field achieved by Shanghai Institute of Technical Physics SITP CAS are reviewed concisely in this paper On the base of analyzing the requirements of devices and techniques for supporting space life science experiments and researches one designment idea of developing different intelligent modules with professional function standard interface and easy to be integrated into system is put forward and the realization method of the experiment system with intelligent distributed control based on the field bus are discussed in three hierarchies Typical sensing or control function cells with certain self-determination control data management and communication abilities are designed and developed which are called Intelligent Agents Digital hardware network system which are consisted of the distributed Agents as the intelligent node is constructed with the normative opening field bus technology The multitask and real-time control application softwares are developed in the embedded RTOS circumstance which is implanted into the system hardware and space life science experiment system platform with characteristic of multitasks multi-courses professional and instant integration will be constructed
Development of lidar sensor for cloud-based measurements during convective conditions
NASA Astrophysics Data System (ADS)
Vishnu, R.; Bhavani Kumar, Y.; Rao, T. Narayana; Nair, Anish Kumar M.; Jayaraman, A.
2016-05-01
Atmospheric convection is a natural phenomena associated with heat transport. Convection is strong during daylight periods and rigorous in summer months. Severe ground heating associated with strong winds experienced during these periods. Tropics are considered as the source regions for strong convection. Formation of thunder storm clouds is common during this period. Location of cloud base and its associated dynamics is important to understand the influence of convection on the atmosphere. Lidars are sensitive to Mie scattering and are the suitable instruments for locating clouds in the atmosphere than instruments utilizing the radio frequency spectrum. Thunder storm clouds are composed of hydrometers and strongly scatter the laser light. Recently, a lidar technique was developed at National Atmospheric Research Laboratory (NARL), a Department of Space (DOS) unit, located at Gadanki near Tirupati. The lidar technique employs slant path operation and provides high resolution measurements on cloud base location in real-time. The laser based remote sensing technique allows measurement of atmosphere for every second at 7.5 m range resolution. The high resolution data permits assessment of updrafts at the cloud base. The lidar also provides real-time convective boundary layer height using aerosols as the tracers of atmospheric dynamics. The developed lidar sensor is planned for up-gradation with scanning facility to understand the cloud dynamics in the spatial direction. In this presentation, we present the lidar sensor technology and utilization of its technology for high resolution cloud base measurements during convective conditions over lidar site, Gadanki.
NASA Astrophysics Data System (ADS)
Zhao, Jin; Han-Ming, Zhang; Bin, Yan; Lei, Li; Lin-Yuan, Wang; Ai-Long, Cai
2016-03-01
Sparse-view x-ray computed tomography (CT) imaging is an interesting topic in CT field and can efficiently decrease radiation dose. Compared with spatial reconstruction, a Fourier-based algorithm has advantages in reconstruction speed and memory usage. A novel Fourier-based iterative reconstruction technique that utilizes non-uniform fast Fourier transform (NUFFT) is presented in this work along with advanced total variation (TV) regularization for a fan sparse-view CT. The proposition of a selective matrix contributes to improve reconstruction quality. The new method employs the NUFFT and its adjoin to iterate back and forth between the Fourier and image space. The performance of the proposed algorithm is demonstrated through a series of digital simulations and experimental phantom studies. Results of the proposed algorithm are compared with those of existing TV-regularized techniques based on compressed sensing method, as well as basic algebraic reconstruction technique. Compared with the existing TV-regularized techniques, the proposed Fourier-based technique significantly improves convergence rate and reduces memory allocation, respectively. Projected supported by the National High Technology Research and Development Program of China (Grant No. 2012AA011603) and the National Natural Science Foundation of China (Grant No. 61372172).
Microwave power amplifiers based on AlGaN/GaN transistors with a two-dimensional electron gas
NASA Astrophysics Data System (ADS)
Vendik, O. G.; Vendik, I. B.; Tural'chuk, P. A.; Parnes, Ya. M.; Parnes, M. D.
2016-11-01
A technique for synthesis of microwave power amplifiers based on transistors with a AlGaN/GaN heterojunction is discussed. Special focus is on the development of a technique for synthesis of transformation circuits of the power amplifier to increase efficiency with a retained high output power. The use of independent matching at the harmonic frequencies and fundamental frequency makes it possible to control the attainable efficiency in a wide frequency band along with the total suppression of harmonics beyond the operational band. Microwave power amplifiers for operation at 4 and 9 GHz have been developed and experimentally investigated.
A Minicomputer Based Scheme for Turbulence Measurements with Pulsed Doppler Ultrasound
Craig, J. I.; Saxena, Vijay; Giddens, D. P.
1979-01-01
The present paper describes the design and performance of a digital-based Doppler signal processing system that is currently being used in hemodynamics research on arteriosclerosis. The major emphasis is on the development of the digital signal processing technique and its implementation in a small but powerful minicomputer. The work reported on here is part of a larger ongoing effort that the authors are undertaking to study the structure of turbulence in blood flow and its relation to arteriosclerosis. Some of the techniques and instruments developed are felt to have a broad applicability to fluid mechanics and especially to pipe flow fluid mechanics.
A dependency-based modelling mechanism for problem solving
NASA Technical Reports Server (NTRS)
London, P.
1978-01-01
The paper develops a technique of dependency net modeling which relies on an explicit representation of justifications for beliefs held by the problem solver. Using these justifications, the modeling mechanism is able to determine the relevant lines of inference to pursue during problem solving. Three particular problem-solving difficulties which may be handled by the dependency-based technique are discussed: (1) subgoal violation detection, (2) description binding, and (3) maintaining a consistent world model.
A knowledge-based tool for multilevel decomposition of a complex design problem
NASA Technical Reports Server (NTRS)
Rogers, James L.
1989-01-01
Although much work has been done in applying artificial intelligence (AI) tools and techniques to problems in different engineering disciplines, only recently has the application of these tools begun to spread to the decomposition of complex design problems. A new tool based on AI techniques has been developed to implement a decomposition scheme suitable for multilevel optimization and display of data in an N x N matrix format.
Depletion-based techniques for super-resolution imaging of NV-diamond
NASA Astrophysics Data System (ADS)
Jaskula, Jean-Christophe; Trifonov, Alexei; Glenn, David; Walsworth, Ronald
2012-06-01
We discuss the development and application of depletion-based techniques for super-resolution imaging of NV centers in diamond: stimulated emission depletion (STED), metastable ground state depletion (GSD), and dark state depletion (DSD). NV centers in diamond do not bleach under optical excitation, are not biotoxic, and have long-lived electronic spin coherence and spin-state-dependent fluorescence. Thus NV-diamond has great potential as a fluorescent biomarker and as a magnetic biosensor.
Spectrophotometry-based detection of carbapenemase producers among Enterobacteriaceae.
Bernabeu, Sandrine; Poirel, Laurent; Nordmann, Patrice
2012-09-01
Carbapenem-hydrolyzing ß-lactamases are the most powerful ß-lactamases being able to hydrolyse almost all ß-lactams. They are mostly of the KPC, VIM, IMP, NDM, and OXA-48 type. A spectrophotometry technique based on analysis of the imipenem hydrolysis has been developed that differentiated carbapenemase- from noncarbapenemase producers. This inexpensive technique adapted to screening of carbapenemase producers may be implemented in any reference laboratory worldwide. Copyright © 2012 Elsevier Inc. All rights reserved.
NASA Technical Reports Server (NTRS)
Berendes, Todd; Sengupta, Sailes K.; Welch, Ron M.; Wielicki, Bruce A.; Navar, Murgesh
1992-01-01
A semiautomated methodology is developed for estimating cumulus cloud base heights on the basis of high spatial resolution Landsat MSS data, using various image-processing techniques to match cloud edges with their corresponding shadow edges. The cloud base height is then estimated by computing the separation distance between the corresponding generalized Hough transform reference points. The differences between the cloud base heights computed by these means and a manual verification technique are of the order of 100 m or less; accuracies of 50-70 m may soon be possible via EOS instruments.
Instantiating the art of war for effects-based operations
NASA Astrophysics Data System (ADS)
Burns, Carla L.
2002-07-01
Effects-Based Operations (EBO) is a mindset, a philosophy and an approach for planning, executing and assessing military operations for the effects they produce rather than the targets or even objectives they deal with. An EBO approach strives to provide economy of force, dynamic tasking, and reduced collateral damage. The notion of EBO is not new. Military Commanders certainly have desired effects in mind when conducting military operations. However, to date EBO has been an art of war that lacks automated techniques and tools that enable effects-based analysis and assessment. Modeling and simulation is at the heart of this challenge. The Air Force Research Laboratory (AFRL) EBO Program is developing modeling techniques and corresponding tool capabilities that can be brought to bear against the challenges presented by effects-based analysis and assessment. Effects-based course-of-action development, center of gravity/target system analysis, and wargaming capabilities are being developed and integrated to help give Commanders the information decision support required to achieve desired national security objectives. This paper presents an introduction to effects-based operations, discusses the benefits of an EBO approach, and focuses on modeling and analysis for effects-based strategy development. An overview of modeling and simulation challenges for EBO is presented, setting the stage for the detailed technical papers in the subject session.
Naresh, P; Hitesh, C; Patel, A; Kolge, T; Sharma, Archana; Mittal, K C
2013-08-01
A fourth order (LCLC) resonant converter based capacitor charging power supply (CCPS) is designed and developed for pulse power applications. Resonant converters are preferred t utilize soft switching techniques such as zero current switching (ZCS) and zero voltage switching (ZVS). An attempt has been made to overcome the disadvantages in 2nd and 3rd resonant converter topologies; hence a fourth order resonant topology is used in this paper for CCPS application. In this paper a novel fourth order LCLC based resonant converter has been explored and mathematical analysis carried out to calculate load independent constant current. This topology provides load independent constant current at switching frequency (fs) equal to resonant frequency (fr). By changing switching condition (on time and dead time) this topology has both soft switching techniques such as ZCS and ZVS for better switching action to improve the converter efficiency. This novel technique has special features such as low peak current through switches, DC blocking for transformer, utilizing transformer leakage inductance as resonant component. A prototype has been developed and tested successfully to charge a 100 μF capacitor to 200 V.
NASA Technical Reports Server (NTRS)
Walker, Carrie K.
1991-01-01
A technique has been developed for combining features of a systems architecture design and assessment tool and a software development tool. This technique reduces simulation development time and expands simulation detail. The Architecture Design and Assessment System (ADAS), developed at the Research Triangle Institute, is a set of computer-assisted engineering tools for the design and analysis of computer systems. The ADAS system is based on directed graph concepts and supports the synthesis and analysis of software algorithms mapped to candidate hardware implementations. Greater simulation detail is provided by the ADAS functional simulator. With the functional simulator, programs written in either Ada or C can be used to provide a detailed description of graph nodes. A Computer-Aided Software Engineering tool developed at the Charles Stark Draper Laboratory (CSDL CASE) automatically generates Ada or C code from engineering block diagram specifications designed with an interactive graphical interface. A technique to use the tools together has been developed, which further automates the design process.
Development of genome-based anti-virulence therapeutics to control HLB
USDA-ARS?s Scientific Manuscript database
Orthologous gene replacement technique has been developed to confirm functions of key virulence genes in 'Candidatus Liberibacters asiaticus'. These results facilitate the development of antivirulence drugs that specifically target functional domains of virulence gene products to disarm pathogenicit...
Wang, Linglan; Yan, Yuchao; Ma, Huilian; Jin, Zhonghe
2016-04-20
New developments are made in the resonant fiber optic gyro (RFOG), which is an optical sensor for the measurement of rotation rate. The digital signal processing system based on the phase modulation technique is capable of detecting the weak frequency difference induced by the Sagnac effect and suppressing the reciprocal noise in the circuit, which determines the detection sensitivity of the RFOG. A new technique based on the sinusoidal wave modulation and square wave demodulation is implemented, and the demodulation curve of the system is simulated and measured. Compared with the past technique using sinusoidal modulation and demodulation, it increases the slope of the demodulation curve by a factor of 1.56, improves the spectrum efficiency of the modulated signal, and reduces the occupancy of the field-programmable gate array resource. On the basis of this new phase modulation technique, the loop is successfully locked and achieves a short-term bias stability of 1.08°/h, which is improved by a factor of 1.47.
Design-for-Hardware-Trust Techniques, Detection Strategies and Metrics for Hardware Trojans
2015-12-14
down both rising and falling transitions. For Trojan detection , one fault , slow-‐to-‐rise or slow-‐to...in Jan. 2016. Through the course of this project we developed novel hardware Trojan detection techniques based on clock sweeping. The technique takes...algorithms to detect minor changes due to Trojan and compared them with those changes made by process variations. This technique was implemented on
Islas, Gabriela; Hernandez, Prisciliano
2017-01-01
To achieve analytical success, it is necessary to develop thorough clean-up procedures to extract analytes from the matrix. Dispersive solid phase extraction (DSPE) has been used as a pretreatment technique for the analysis of several compounds. This technique is based on the dispersion of a solid sorbent in liquid samples in the extraction isolation and clean-up of different analytes from complex matrices. DSPE has found a wide range of applications in several fields, and it is considered to be a selective, robust, and versatile technique. The applications of dispersive techniques in the analysis of veterinary drugs in different matrices involve magnetic sorbents, molecularly imprinted polymers, carbon-based nanomaterials, and the Quick, Easy, Cheap, Effective, Rugged, and Safe (QuEChERS) method. Techniques based on DSPE permit minimization of additional steps such as precipitation, centrifugation, and filtration, which decreases the manipulation of the sample. In this review, we describe the main procedures used for synthesis, characterization, and application of this pretreatment technique and how it has been applied to food analysis. PMID:29181027
Electrocardiogram signal denoising based on empirical mode decomposition technique: an overview
NASA Astrophysics Data System (ADS)
Han, G.; Lin, B.; Xu, Z.
2017-03-01
Electrocardiogram (ECG) signal is nonlinear and non-stationary weak signal which reflects whether the heart is functioning normally or abnormally. ECG signal is susceptible to various kinds of noises such as high/low frequency noises, powerline interference and baseline wander. Hence, the removal of noises from ECG signal becomes a vital link in the ECG signal processing and plays a significant role in the detection and diagnosis of heart diseases. The review will describe the recent developments of ECG signal denoising based on Empirical Mode Decomposition (EMD) technique including high frequency noise removal, powerline interference separation, baseline wander correction, the combining of EMD and Other Methods, EEMD technique. EMD technique is a quite potential and prospective but not perfect method in the application of processing nonlinear and non-stationary signal like ECG signal. The EMD combined with other algorithms is a good solution to improve the performance of noise cancellation. The pros and cons of EMD technique in ECG signal denoising are discussed in detail. Finally, the future work and challenges in ECG signal denoising based on EMD technique are clarified.
Multi-Disciplinary Techniques for Understanding Time-Varying Space-Based Imagery.
1985-05-10
problem, and I V WY" 3 discuss the impgrtage of this work to Air Force technology and to related Air Force programs. Section 1.5 provides a summary of...development of new algorithms and their realization in a hybrid optical/digital architecture. However, devices and architectures being developed in related ...and relate these representntions to object and surface contour properties of the scene. The techniques studied included Probabilistic Graph Matching
Long range science scheduling for the Hubble Space Telescope
NASA Technical Reports Server (NTRS)
Miller, Glenn; Johnston, Mark
1991-01-01
Observations with NASA's Hubble Space Telescope (HST) are scheduled with the assistance of a long-range scheduling system (SPIKE) that was developed using artificial intelligence techniques. In earlier papers, the system architecture and the constraint representation and propagation mechanisms were described. The development of high-level automated scheduling tools, including tools based on constraint satisfaction techniques and neural networks is described. The performance of these tools in scheduling HST observations is discussed.
ERIC Educational Resources Information Center
Sayre, Scott Alan
The purpose of this study was to develop and validate a computer-based system that would allow interactive video developers to integrate and manage the design components prior to production. These components of an interactive video (IVD) program include visual information in a variety of formats, audio information, and instructional techniques,…
NASA Astrophysics Data System (ADS)
Lee, Youngjin; Lee, Amy Candy; Kim, Hee-Joung
2016-09-01
Recently, significant effort has been spent on the development of photons counting detector (PCD) based on a CdTe for applications in X-ray imaging system. The motivation of developing PCDs is higher image quality. Especially, the K-edge subtraction (KES) imaging technique using a PCD is able to improve image quality and useful for increasing the contrast resolution of a target material by utilizing contrast agent. Based on above-mentioned technique, we presented an idea for an improved K-edge log-subtraction (KELS) imaging technique. The KELS imaging technique based on the PCDs can be realized by using different subtraction energy width of the energy window. In this study, the effects of the KELS imaging technique and subtraction energy width of the energy window was investigated with respect to the contrast, standard deviation, and CNR with a Monte Carlo simulation. We simulated the PCD X-ray imaging system based on a CdTe and polymethylmethacrylate (PMMA) phantom which consists of the various iodine contrast agents. To acquired KELS images, images of the phantom using above and below the iodine contrast agent K-edge absorption energy (33.2 keV) have been acquired at different energy range. According to the results, the contrast and standard deviation were decreased, when subtraction energy width of the energy window is increased. Also, the CNR using a KELS imaging technique is higher than that of the images acquired by using whole energy range. Especially, the maximum differences of CNR between whole energy range and KELS images using a 1, 2, and 3 mm diameter iodine contrast agent were acquired 11.33, 8.73, and 8.29 times, respectively. Additionally, the optimum subtraction energy width of the energy window can be acquired at 5, 4, and 3 keV for the 1, 2, and 3 mm diameter iodine contrast agent, respectively. In conclusion, we successfully established an improved KELS imaging technique and optimized subtraction energy width of the energy window, and based on our results, we recommend using this technique for high image quality.
Reyes, Camilo; Mason, Eric; Solares, C. Arturo
2014-01-01
Introduction A substantial body of literature has been devoted to the distinct characteristics and surgical options to repair the skull base. However, the skull base is an anatomically challenging location that requires a three-dimensional reconstruction approach. Furthermore, advances in endoscopic skull base surgery encompass a wide range of surgical pathology, from benign tumors to sinonasal cancer. This has resulted in the creation of wide defects that yield a new challenge in skull base reconstruction. Progress in technology and imaging has made this approach an internationally accepted method to repair these defects. Objectives Discuss historical developments and flaps available for skull base reconstruction. Data Synthesis Free grafts in skull base reconstruction are a viable option in small defects and low-flow leaks. Vascularized flaps pose a distinct advantage in large defects and high-flow leaks. When open techniques are used, free flap reconstruction techniques are often necessary to repair large entry wound defects. Conclusions Reconstruction of skull base defects requires a thorough knowledge of surgical anatomy, disease, and patient risk factors associated with high-flow cerebrospinal fluid leaks. Various reconstruction techniques are available, from free tissue grafting to vascularized flaps. Possible complications that can befall after these procedures need to be considered. Although endonasal techniques are being used with increasing frequency, open techniques are still necessary in selected cases. PMID:25992142
A comparison of time-shared vs. batch development of space software
NASA Technical Reports Server (NTRS)
Forthofer, M.
1977-01-01
In connection with a study regarding the ground support software development for the Space Shuttle, an investigation was conducted concerning the most suitable software development techniques to be employed. A time-sharing 'trial period' was used to determine whether or not time-sharing would be a cost-effective software development technique for the Ground Based Shuttle system. It was found that time-sharing substantially improved job turnaround and programmer access to the computer for the representative group of ground support programmers. Moreover, this improvement resulted in an estimated saving of over fifty programmer days during the trial period.
Olyaeemanesh, Alireza; Bavandpour, Elahe; Mobinizadeh, Mohammadreza; Ashrafinia, Mansoor; Bavandpour, Maryam; Nouhi, Mojtaba
2017-01-01
Background: Caesarean section (C-section) is the most common surgery among women worldwide, and the global rate of this surgical procedure has been continuously rising. Hence, it is significantly crucial to develop and apply highly effective and safe caesarean section techniques. In this review study, we aimed at assessing the safety and effectiveness of the Joel-Cohen-based technique and comparing the results with the transverse Pfannenstiel incision for C-section. Methods: In this study, various reliable databases such as the PubMed Central, COCHRANE, DARE, and Ovid MEDLINE were targeted. Reviews, systematic reviews, and randomized clinical trial studies comparing the Joel-Cohen-based technique and the transverse Pfannenstiel incision were selected based on the inclusion criteria. Selected studies were checked by 2 independent reviewers based on the inclusion criteria, and the quality of these studies was assessed. Then, their data were extracted and analyzed. Results: Five randomized clinical trial studies met the inclusion criteria. According to the exiting evidence, statistical results of the Joel-Cohen-based technique showed that this technique is more effective compared to the transverse Pfannenstiel incision. Metaanalysis results of the 3 outcomes were as follow: operation time (5 trials, 764 women; WMD -9.78; 95% CI:-14.49-5.07 minutes, p<0.001), blood loss (3 trials, 309 women; WMD -53.23ml; 95% –CI: 90.20-16.26 ml, p= 0.004), and post-operative hospital stay (3 trials, 453 women; WMD -.69 day; 95% CI: 1.4-0.03 day, p<0.001). Statistical results revealed a significant difference between the 2 techniques. Conclusion: According to the literature, despite having a number of side effects, the Joel-Cohen-based technique is generally more effective than the Pfannenstiel incision technique. In addition, it was recommended that the Joel-Cohen-based technique be used as a replacement for the Pfannenstiel incision technique according to the surgeons’ preferences and the patients’ conditions. PMID:29445683
DOT National Transportation Integrated Search
2017-08-01
Intelligent Compaction (IC) technique is a fast-developing technology for base and soil compaction quality control. Proof-rolling subgrades and bases using IC rollers upon completion of compaction can identify the less stiff spots and significantly i...
The emergence of top-down proteomics in clinical research
2013-01-01
Proteomic technology has advanced steadily since the development of 'soft-ionization' techniques for mass-spectrometry-based molecular identification more than two decades ago. Now, the large-scale analysis of proteins (proteomics) is a mainstay of biological research and clinical translation, with researchers seeking molecular diagnostics, as well as protein-based markers for personalized medicine. Proteomic strategies using the protease trypsin (known as bottom-up proteomics) were the first to be developed and optimized and form the dominant approach at present. However, researchers are now beginning to understand the limitations of bottom-up techniques, namely the inability to characterize and quantify intact protein molecules from a complex mixture of digested peptides. To overcome these limitations, several laboratories are taking a whole-protein-based approach, in which intact protein molecules are the analytical targets for characterization and quantification. We discuss these top-down techniques and how they have been applied to clinical research and are likely to be applied in the near future. Given the recent improvements in mass-spectrometry-based proteomics and stronger cooperation between researchers, clinicians and statisticians, both peptide-based (bottom-up) strategies and whole-protein-based (top-down) strategies are set to complement each other and help researchers and clinicians better understand and detect complex disease phenotypes. PMID:23806018
Frank N. Martin; Paul W. Tooley
2006-01-01
Molecular techniques have been developed for detection and identification of P. ramorum and other Phytophthora species that are based on the mitochondrially encoded sequences. One technique uses a Phytophthora genus specific primer to determine if a Phytophthora species is present, followed by...
Modelling and Simulation for Requirements Engineering and Options Analysis
2010-05-01
should be performed to work successfully in the domain; and process-based techniques model the processes that occur in the work domain. There is a crisp ...acad/sed/sedres/ dm /erg/cwa. DRDC Toronto CR 2010-049 39 23. Can the current technique for developing simulation models for assessments
USDA-ARS?s Scientific Manuscript database
Ambient desorption ionization techniques, such as laser desorption with electrospray ionization assistance (ELDI), direct analysis in real time (DART) and desorption electrospray ionization (DESI) have been developed as alternatives to traditional mass spectrometric-based methods. Such techniques al...
Using Word Clouds to Develop Proactive Learners
ERIC Educational Resources Information Center
Miley, Frances; Read, Andrew
2011-01-01
This article examines student responses to a technique for summarizing electronically available information based on word frequency. Students used this technique to create word clouds, using those word clouds to enhance personal and small group study. This is a qualitative study. Small focus groups were used to obtain student feedback. Feedback…
Investigation of Learners' Perceptions for Video Summarization and Recommendation
ERIC Educational Resources Information Center
Yang, Jie Chi; Chen, Sherry Y.
2012-01-01
Recently, multimedia-based learning is widespread in educational settings. A number of studies investigate how to develop effective techniques to manage a huge volume of video sources, such as summarization and recommendation. However, few studies examine how these techniques affect learners' perceptions in multimedia learning systems. This…
CAPSAS: Computer Assisted Program for the Selection of Appropriate Statistics.
ERIC Educational Resources Information Center
Shermis, Mark D.; Albert, Susan L.
A computer-assisted program has been developed for the selection of statistics or statistical techniques by both students and researchers. Based on Andrews, Klem, Davidson, O'Malley and Rodgers "A Guide for Selecting Statistical Techniques for Analyzing Social Science Data," this FORTRAN-compiled interactive computer program was…
USDA-ARS?s Scientific Manuscript database
Nondestructive methods based on fluorescence hyperspectral imaging (HSI) techniques were developed in order to detect worms on fresh-cut lettuce. The optimal wavebands for detecting worms on fresh-cut lettuce were investigated using the one-way ANOVA analysis and correlation analysis. The worm detec...
Study on development system of increasing gearbox for high-performance wind-power generator
NASA Astrophysics Data System (ADS)
Xu, Hongbin; Yan, Kejun; Zhao, Junyu
2005-12-01
Based on the analysis of the development potentiality of wind-power generator and domestic manufacture of its key parts in China, an independent development system of the Increasing Gearbox for High-performance Wind-power Generator (IGHPWG) was introduced. The main elements of the system were studied, including the procedure design, design analysis system, manufacturing technology and detecting system, and the relative important technologies were analyzed such as mixed optimal joint transmission structure of the first planetary drive with two grade parallel axle drive based on equal strength, tooth root round cutting technology before milling hard tooth surface, high-precise tooth grinding technology, heat treatment optimal technology and complex surface technique, and rig test and detection technique of IGHPWG. The development conception was advanced the data share and quality assurance system through all the elements of the development system. The increasing Gearboxes for 600KW and 1MW Wind-power Generator have been successfully developed through the application of the development system.
Combined photoacoustic and magneto-acoustic imaging.
Qu, Min; Mallidi, Srivalleesha; Mehrmohammadi, Mohammad; Ma, Li Leo; Johnston, Keith P; Sokolov, Konstantin; Emelianov, Stanislav
2009-01-01
Ultrasound is a widely used modality with excellent spatial resolution, low cost, portability, reliability and safety. In clinical practice and in the biomedical field, molecular ultrasound-based imaging techniques are desired to visualize tissue pathologies, such as cancer. In this paper, we present an advanced imaging technique - combined photoacoustic and magneto-acoustic imaging - capable of visualizing the anatomical, functional and biomechanical properties of tissues or organs. The experiments to test the combined imaging technique were performed using dual, nanoparticle-based contrast agents that exhibit the desired optical and magnetic properties. The results of our study demonstrate the feasibility of the combined photoacoustic and magneto-acoustic imaging that takes the advantages of each imaging techniques and provides high sensitivity, reliable contrast and good penetrating depth. Therefore, the developed imaging technique can be used in wide range of biomedical and clinical application.
Hanbury, Andria; Thompson, Carl; Mannion, Russell
2011-07-01
Tailored implementation strategies targeting health professionals' adoption of evidence-based recommendations are currently being developed. Research has focused on how to select an appropriate theoretical base, how to use that theoretical base to explore the local context, and how to translate theoretical constructs associated with the key factors found to influence innovation adoption into feasible and tailored implementation strategies. The reasons why an intervention is thought not to have worked are often cited as being: inappropriate choice of theoretical base; unsystematic development of the implementation strategies; and a poor evidence base to guide the process. One area of implementation research that is commonly overlooked is how to synthesize the data collected in a local context in order to identify what factors to target with the implementation strategies. This is suggested to be a critical process in the development of a theory-based intervention. The potential of multilevel modelling techniques to synthesize data collected at different hierarchical levels, for example, individual attitudes and team level variables, is discussed. Future research is needed to explore further the potential of multilevel modelling for synthesizing contextual data in implementation studies, as well as techniques for synthesizing qualitative and quantitative data.
Validation of highly reliable, real-time knowledge-based systems
NASA Technical Reports Server (NTRS)
Johnson, Sally C.
1988-01-01
Knowledge-based systems have the potential to greatly increase the capabilities of future aircraft and spacecraft and to significantly reduce support manpower needed for the space station and other space missions. However, a credible validation methodology must be developed before knowledge-based systems can be used for life- or mission-critical applications. Experience with conventional software has shown that the use of good software engineering techniques and static analysis tools can greatly reduce the time needed for testing and simulation of a system. Since exhaustive testing is infeasible, reliability must be built into the software during the design and implementation phases. Unfortunately, many of the software engineering techniques and tools used for conventional software are of little use in the development of knowledge-based systems. Therefore, research at Langley is focused on developing a set of guidelines, methods, and prototype validation tools for building highly reliable, knowledge-based systems. The use of a comprehensive methodology for building highly reliable, knowledge-based systems should significantly decrease the time needed for testing and simulation. A proven record of delivering reliable systems at the beginning of the highly visible testing and simulation phases is crucial to the acceptance of knowledge-based systems in critical applications.
NASA Technical Reports Server (NTRS)
Omura, J. K.; Simon, M. K.
1982-01-01
A theory is presented for deducing and predicting the performance of transmitter/receivers for bandwidth efficient modulations suitable for use on the linear satellite channel. The underlying principle used is the development of receiver structures based on the maximum-likelihood decision rule. The application of the performance prediction tools, e.g., channel cutoff rate and bit error probability transfer function bounds to these modulation/demodulation techniques.
Quantum technology and cryptology for information security
NASA Astrophysics Data System (ADS)
Naqvi, Syed; Riguidel, Michel
2007-04-01
Cryptology and information security are set to play a more prominent role in the near future. In this regard, quantum communication and cryptography offer new opportunities to tackle ICT security. Quantum Information Processing and Communication (QIPC) is a scientific field where new conceptual foundations and techniques are being developed. They promise to play an important role in the future of information Security. It is therefore essential to have a cross-fertilizing development between quantum technology and cryptology in order to address the security challenges of the emerging quantum era. In this article, we discuss the impact of quantum technology on the current as well as future crypto-techniques. We then analyse the assumptions on which quantum computers may operate. Then we present our vision for the distribution of security attributes using a novel form of trust based on Heisenberg's uncertainty; and, building highly secure quantum networks based on the clear transmission of single photons and/or bundles of photons able to withstand unauthorized reading as a result of secure protocols based on the observations of quantum mechanics. We argue how quantum cryptographic systems need to be developed that can take advantage of the laws of physics to provide long-term security based on solid assumptions. This requires a structured integration effort to deploy quantum technologies within the existing security infrastructure. Finally, we conclude that classical cryptographic techniques need to be redesigned and upgraded in view of the growing threat of cryptanalytic attacks posed by quantum information processing devices leading to the development of post-quantum cryptography.
NASA Technical Reports Server (NTRS)
Stephan, Amy; Erikson, Carol A.
1991-01-01
As an initial attempt to introduce expert system technology into an onboard environment, a model based diagnostic system using the TRW MARPLE software tool was integrated with prototype flight hardware and its corresponding control software. Because this experiment was designed primarily to test the effectiveness of the model based reasoning technique used, the expert system ran on a separate hardware platform, and interactions between the control software and the model based diagnostics were limited. While this project met its objective of showing that model based reasoning can effectively isolate failures in flight hardware, it also identified the need for an integrated development path for expert system and control software for onboard applications. In developing expert systems that are ready for flight, artificial intelligence techniques must be evaluated to determine whether they offer a real advantage onboard, identify which diagnostic functions should be performed by the expert systems and which are better left to the procedural software, and work closely with both the hardware and the software developers from the beginning of a project to produce a well designed and thoroughly integrated application.
The development of additive manufacturing technique for nickel-base alloys: A review
NASA Astrophysics Data System (ADS)
Zadi-Maad, Ahmad; Basuki, Arif
2018-04-01
Nickel-base alloys are an attractive alloy due to its excellent mechanical properties, a high resistance to creep deformation, corrosion, and oxidation. However, it is a hard task to control performance when casting or forging for this material. In recent years, additive manufacturing (AM) process has been implemented to replace the conventional directional solidification process for the production of nickel-base alloys. Due to its potentially lower cost and flexibility manufacturing process, AM is considered as a substitute technique for the existing. This paper provides a comprehensive review of the previous work related to the AM techniques for Ni-base alloys while highlighting current challenges and methods to solving them. The properties of conventionally manufactured Ni-base alloys are also compared with the AM fabricated alloys. The mechanical properties obtained from tension, hardness and fatigue test are included, along with discussions of the effect of post-treatment process. Recommendations for further work are also provided.
Plasmonic SERS nanochips and nanoprobes for medical diagnostics and bio-energy applications
NASA Astrophysics Data System (ADS)
Ngo, Hoan T.; Wang, Hsin-Neng; Crawford, Bridget M.; Fales, Andrew M.; Vo-Dinh, Tuan
2017-02-01
The development of rapid, easy-to-use, cost-effective, high accuracy, and high sensitive DNA detection methods for molecular diagnostics has been receiving increasing interest. Over the last five years, our laboratory has developed several chip-based DNA detection techniques including the molecular sentinel-on-chip (MSC), the multiplex MSC, and the inverse molecular sentinel-on-chip (iMS-on-Chip). In these techniques, plasmonic surface-enhanced Raman scattering (SERS) Nanowave chips were functionalized with DNA probes for single-step DNA detection. Sensing mechanisms were based on hybridization of target sequences and DNA probes, resulting in a distance change between SERS reporters and the Nanowave chip's gold surface. This distance change resulted in change in SERS intensity, thus indicating the presence and capture of the target sequences. Our techniques were single-step DNA detection techniques. Target sequences were detected by simple delivery of sample solutions onto DNA probe-functionalized Nanowave chips and SERS signals were measured after 1h - 2h incubation. Target sequence labeling or washing to remove unreacted components was not required, making the techniques simple, easy-to-use, and cost effective. The usefulness of the techniques for medical diagnostics was illustrated by the detection of genetic biomarkers for respiratory viral infection and of dengue virus 4 DNA.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Fahimian, B.
2015-06-15
Intrafraction target motion is a prominent complicating factor in the accurate targeting of radiation within the body. Methods compensating for target motion during treatment, such as gating and dynamic tumor tracking, depend on the delineation of target location as a function of time during delivery. A variety of techniques for target localization have been explored and are under active development; these include beam-level imaging of radio-opaque fiducials, fiducial-less tracking of anatomical landmarks, tracking of electromagnetic transponders, optical imaging of correlated surrogates, and volumetric imaging within treatment delivery. The Joint Imaging and Therapy Symposium will provide an overview of the techniquesmore » for real-time imaging and tracking, with special focus on emerging modes of implementation across different modalities. In particular, the symposium will explore developments in 1) Beam-level kilovoltage X-ray imaging techniques, 2) EPID-based megavoltage X-ray tracking, 3) Dynamic tracking using electromagnetic transponders, and 4) MRI-based soft-tissue tracking during radiation delivery. Learning Objectives: Understand the fundamentals of real-time imaging and tracking techniques Learn about emerging techniques in the field of real-time tracking Distinguish between the advantages and disadvantages of different tracking modalities Understand the role of real-time tracking techniques within the clinical delivery work-flow.« less
DOE Office of Scientific and Technical Information (OSTI.GOV)
Low, D.
2015-06-15
Intrafraction target motion is a prominent complicating factor in the accurate targeting of radiation within the body. Methods compensating for target motion during treatment, such as gating and dynamic tumor tracking, depend on the delineation of target location as a function of time during delivery. A variety of techniques for target localization have been explored and are under active development; these include beam-level imaging of radio-opaque fiducials, fiducial-less tracking of anatomical landmarks, tracking of electromagnetic transponders, optical imaging of correlated surrogates, and volumetric imaging within treatment delivery. The Joint Imaging and Therapy Symposium will provide an overview of the techniquesmore » for real-time imaging and tracking, with special focus on emerging modes of implementation across different modalities. In particular, the symposium will explore developments in 1) Beam-level kilovoltage X-ray imaging techniques, 2) EPID-based megavoltage X-ray tracking, 3) Dynamic tracking using electromagnetic transponders, and 4) MRI-based soft-tissue tracking during radiation delivery. Learning Objectives: Understand the fundamentals of real-time imaging and tracking techniques Learn about emerging techniques in the field of real-time tracking Distinguish between the advantages and disadvantages of different tracking modalities Understand the role of real-time tracking techniques within the clinical delivery work-flow.« less
DOE Office of Scientific and Technical Information (OSTI.GOV)
Berbeco, R.
2015-06-15
Intrafraction target motion is a prominent complicating factor in the accurate targeting of radiation within the body. Methods compensating for target motion during treatment, such as gating and dynamic tumor tracking, depend on the delineation of target location as a function of time during delivery. A variety of techniques for target localization have been explored and are under active development; these include beam-level imaging of radio-opaque fiducials, fiducial-less tracking of anatomical landmarks, tracking of electromagnetic transponders, optical imaging of correlated surrogates, and volumetric imaging within treatment delivery. The Joint Imaging and Therapy Symposium will provide an overview of the techniquesmore » for real-time imaging and tracking, with special focus on emerging modes of implementation across different modalities. In particular, the symposium will explore developments in 1) Beam-level kilovoltage X-ray imaging techniques, 2) EPID-based megavoltage X-ray tracking, 3) Dynamic tracking using electromagnetic transponders, and 4) MRI-based soft-tissue tracking during radiation delivery. Learning Objectives: Understand the fundamentals of real-time imaging and tracking techniques Learn about emerging techniques in the field of real-time tracking Distinguish between the advantages and disadvantages of different tracking modalities Understand the role of real-time tracking techniques within the clinical delivery work-flow.« less
DOE Office of Scientific and Technical Information (OSTI.GOV)
Keall, P.
2015-06-15
Intrafraction target motion is a prominent complicating factor in the accurate targeting of radiation within the body. Methods compensating for target motion during treatment, such as gating and dynamic tumor tracking, depend on the delineation of target location as a function of time during delivery. A variety of techniques for target localization have been explored and are under active development; these include beam-level imaging of radio-opaque fiducials, fiducial-less tracking of anatomical landmarks, tracking of electromagnetic transponders, optical imaging of correlated surrogates, and volumetric imaging within treatment delivery. The Joint Imaging and Therapy Symposium will provide an overview of the techniquesmore » for real-time imaging and tracking, with special focus on emerging modes of implementation across different modalities. In particular, the symposium will explore developments in 1) Beam-level kilovoltage X-ray imaging techniques, 2) EPID-based megavoltage X-ray tracking, 3) Dynamic tracking using electromagnetic transponders, and 4) MRI-based soft-tissue tracking during radiation delivery. Learning Objectives: Understand the fundamentals of real-time imaging and tracking techniques Learn about emerging techniques in the field of real-time tracking Distinguish between the advantages and disadvantages of different tracking modalities Understand the role of real-time tracking techniques within the clinical delivery work-flow.« less
MO-FG-BRD-00: Real-Time Imaging and Tracking Techniques for Intrafractional Motion Management
DOE Office of Scientific and Technical Information (OSTI.GOV)
NONE
2015-06-15
Intrafraction target motion is a prominent complicating factor in the accurate targeting of radiation within the body. Methods compensating for target motion during treatment, such as gating and dynamic tumor tracking, depend on the delineation of target location as a function of time during delivery. A variety of techniques for target localization have been explored and are under active development; these include beam-level imaging of radio-opaque fiducials, fiducial-less tracking of anatomical landmarks, tracking of electromagnetic transponders, optical imaging of correlated surrogates, and volumetric imaging within treatment delivery. The Joint Imaging and Therapy Symposium will provide an overview of the techniquesmore » for real-time imaging and tracking, with special focus on emerging modes of implementation across different modalities. In particular, the symposium will explore developments in 1) Beam-level kilovoltage X-ray imaging techniques, 2) EPID-based megavoltage X-ray tracking, 3) Dynamic tracking using electromagnetic transponders, and 4) MRI-based soft-tissue tracking during radiation delivery. Learning Objectives: Understand the fundamentals of real-time imaging and tracking techniques Learn about emerging techniques in the field of real-time tracking Distinguish between the advantages and disadvantages of different tracking modalities Understand the role of real-time tracking techniques within the clinical delivery work-flow.« less
Cooperative analysis expert situation assessment research
NASA Technical Reports Server (NTRS)
Mccown, Michael G.
1987-01-01
For the past few decades, Rome Air Development Center (RADC) has been conducting research in Artificial Intelligence (AI). When the recent advances in hardware technology made many AI techniques practical, the Intelligence and Reconnaissance Directorate of RADC initiated an applications program entitled Knowledge Based Intelligence Systems (KBIS). The goal of the program is the development of a generic Intelligent Analyst System, an open machine with the framework for intelligence analysis, natural language processing, and man-machine interface techniques, needing only the specific problem domain knowledge to be operationally useful. The development of KBIS is described.
GENIE - Generation of computational geometry-grids for internal-external flow configurations
NASA Technical Reports Server (NTRS)
Soni, B. K.
1988-01-01
Progress realized in the development of a master geometry-grid generation code GENIE is presented. The grid refinement process is enhanced by developing strategies to utilize bezier curves/surfaces and splines along with weighted transfinite interpolation technique and by formulating new forcing function for the elliptic solver based on the minimization of a non-orthogonality functional. A two step grid adaptation procedure is developed by optimally blending adaptive weightings with weighted transfinite interpolation technique. Examples of 2D-3D grids are provided to illustrate the success of these methods.
Real-Time Onboard Global Nonlinear Aerodynamic Modeling from Flight Data
NASA Technical Reports Server (NTRS)
Brandon, Jay M.; Morelli, Eugene A.
2014-01-01
Flight test and modeling techniques were developed to accurately identify global nonlinear aerodynamic models onboard an aircraft. The techniques were developed and demonstrated during piloted flight testing of an Aermacchi MB-326M Impala jet aircraft. Advanced piloting techniques and nonlinear modeling techniques based on fuzzy logic and multivariate orthogonal function methods were implemented with efficient onboard calculations and flight operations to achieve real-time maneuver monitoring and analysis, and near-real-time global nonlinear aerodynamic modeling and prediction validation testing in flight. Results demonstrated that global nonlinear aerodynamic models for a large portion of the flight envelope were identified rapidly and accurately using piloted flight test maneuvers during a single flight, with the final identified and validated models available before the aircraft landed.
Development of Control Teaching Material for Mechatronics Education Based on Experience
NASA Astrophysics Data System (ADS)
Tasaki, Takao; Watanabe, Shinichi; Shikanai, Yoshihito; Ozaki, Koichi
In this paper, we have developed a teaching material for technical high school students to understand the control technique. The material makes the students understanding the control technique through the sensibility obtained from the experience of riding the robot. We have considered the correspondence of the teaching material with the ARCS Model. Therefore, the material aims to improve the interest and the willingness to learn mechatronics and control technique by experiencing the difference of the response by the change in the control parameters. As the results of the questionnaire to the technical high school students in the class, we have verified educative effect of the teaching material which can be improved willingness of learning and interesting for mechatronics and control technique.
Further Developments of the Fringe-Imaging Skin Friction Technique
NASA Technical Reports Server (NTRS)
Zilliac, Gregory C.
1996-01-01
Various aspects and extensions of the Fringe-Imaging Skin Friction technique (FISF) have been explored through the use of several benchtop experiments and modeling. The technique has been extended to handle three-dimensional flow fields with mild shear gradients. The optical and imaging system has been refined and a PC-based application has been written that has made it possible to obtain high resolution skin friction field measurements in a reasonable period of time. The improved method was tested on a wingtip and compared with Navier-Stokes computations. Additionally, a general approach to interferogram-fringe spacing analysis has been developed that should have applications in other areas of interferometry. A detailed error analysis of the FISF technique is also included.
NASA Astrophysics Data System (ADS)
Ibrahim, Wael Refaat Anis
The present research involves the development of several fuzzy expert systems for power quality analysis and diagnosis. Intelligent systems for the prediction of abnormal system operation were also developed. The performance of all intelligent modules developed was either enhanced or completely produced through adaptive fuzzy learning techniques. Neuro-fuzzy learning is the main adaptive technique utilized. The work presents a novel approach to the interpretation of power quality from the perspective of the continuous operation of a single system. The research includes an extensive literature review pertaining to the applications of intelligent systems to power quality analysis. Basic definitions and signature events related to power quality are introduced. In addition, detailed discussions of various artificial intelligence paradigms as well as wavelet theory are included. A fuzzy-based intelligent system capable of identifying normal from abnormal operation for a given system was developed. Adaptive neuro-fuzzy learning was applied to enhance its performance. A group of fuzzy expert systems that could perform full operational diagnosis were also developed successfully. The developed systems were applied to the operational diagnosis of 3-phase induction motors and rectifier bridges. A novel approach for learning power quality waveforms and trends was developed. The technique, which is adaptive neuro fuzzy-based, learned, compressed, and stored the waveform data. The new technique was successfully tested using a wide variety of power quality signature waveforms, and using real site data. The trend-learning technique was incorporated into a fuzzy expert system that was designed to predict abnormal operation of a monitored system. The intelligent system learns and stores, in compressed format, trends leading to abnormal operation. The system then compares incoming data to the retained trends continuously. If the incoming data matches any of the learned trends, an alarm is instigated predicting the advent of system abnormal operation. The incoming data could be compared to previous trends as well as matched to trends developed through computer simulations and stored using fuzzy learning.
A scale space feature based registration technique for fusion of satellite imagery
NASA Technical Reports Server (NTRS)
Raghavan, Srini; Cromp, Robert F.; Campbell, William C.
1997-01-01
Feature based registration is one of the most reliable methods to register multi-sensor images (both active and passive imagery) since features are often more reliable than intensity or radiometric values. The only situation where a feature based approach will fail is when the scene is completely homogenous or densely textural in which case a combination of feature and intensity based methods may yield better results. In this paper, we present some preliminary results of testing our scale space feature based registration technique, a modified version of feature based method developed earlier for classification of multi-sensor imagery. The proposed approach removes the sensitivity in parameter selection experienced in the earlier version as explained later.
NASA Technical Reports Server (NTRS)
Ostroff, A. J.
1973-01-01
Some of the major difficulties associated with large orbiting astronomical telescopes are the cost of manufacturing the primary mirror to precise tolerances and the maintaining of diffraction-limited tolerances while in orbit. One successfully demonstrated approach for minimizing these problem areas is the technique of actively deforming the primary mirror by applying discrete forces to the rear of the mirror. A modal control technique, as applied to active optics, has previously been developed and analyzed. The modal control technique represents the plant to be controlled in terms of its eigenvalues and eigenfunctions which are estimated via numerical approximation techniques. The report includes an extension of previous work using the modal control technique and also describes an optimal feedback controller. The equations for both control laws are developed in state-space differential form and include such considerations as stability, controllability, and observability. These equations are general and allow the incorporation of various mode-analyzer designs; two design approaches are presented. The report also includes a technique for placing actuator and sensor locations at points on the mirror based upon the flexibility matrix of the uncontrolled or unobserved modes of the structure. The locations selected by this technique are used in the computer runs which are described. The results are based upon three different initial error distributions, two mode-analyzer designs, and both the modal and optimal control laws.
Automatic Parameter Tuning for the Morpheus Vehicle Using Particle Swarm Optimization
NASA Technical Reports Server (NTRS)
Birge, B.
2013-01-01
A high fidelity simulation using a PC based Trick framework has been developed for Johnson Space Center's Morpheus test bed flight vehicle. There is an iterative development loop of refining and testing the hardware, refining the software, comparing the software simulation to hardware performance and adjusting either or both the hardware and the simulation to extract the best performance from the hardware as well as the most realistic representation of the hardware from the software. A Particle Swarm Optimization (PSO) based technique has been developed that increases speed and accuracy of the iterative development cycle. Parameters in software can be automatically tuned to make the simulation match real world subsystem data from test flights. Special considerations for scale, linearity, discontinuities, can be all but ignored with this technique, allowing fast turnaround both for simulation tune up to match hardware changes as well as during the test and validation phase to help identify hardware issues. Software models with insufficient control authority to match hardware test data can be immediately identified and using this technique requires very little to no specialized knowledge of optimization, freeing model developers to concentrate on spacecraft engineering. Integration of the PSO into the Morpheus development cycle will be discussed as well as a case study highlighting the tool's effectiveness.
Computer assisted analysis of auroral images obtained from high altitude polar satellites
NASA Technical Reports Server (NTRS)
Samadani, Ramin; Flynn, Michael
1993-01-01
Automatic techniques that allow the extraction of physically significant parameters from auroral images were developed. This allows the processing of a much larger number of images than is currently possible with manual techniques. Our techniques were applied to diverse auroral image datasets. These results were made available to geophysicists at NASA and at universities in the form of a software system that performs the analysis. After some feedback from users, an upgraded system was transferred to NASA and to two universities. The feasibility of user-trained search and retrieval of large amounts of data using our automatically derived parameter indices was demonstrated. Techniques based on classification and regression trees (CART) were developed and applied to broaden the types of images to which the automated search and retrieval may be applied. Our techniques were tested with DE-1 auroral images.
Captive-rearing piping plovers: Developing techniques to augment wild populations
Powell, A.N.; Cuthbert, F.J.; Wemmer, L.C.; Doolittle, A.W.; Feirer, S.T.
1997-01-01
Techniques for captive-rearing and releasing piping plovers (Charadrius melodus) were developed using a surrogate species, killdeer (Charadrius vociferus). We compared captive-and parent-reared killdeer, and parent-reared piping plovers and determined that growth and behavior were similar. After surrogate trials determined that captive-rearing was feasible, we used the same methods to raise piping plover chicks from salvaged eggs. For captive-reared chick of both species, survival to fledging was higher than and behaviors similar to parent-reared chicks in the wild. Rearing techniques were fine-tuned, and ten piping plover fledglings were released to the wild. Based on our results, we developed recommendations for captive-rearing piping plovers using salvaged eggs to enhance productivity of small populations. ?? 1997 Wiley-Liss, Inc.
Captive-rearing piping plovers: developing techniques to augment wild populations
Powell, A.N.; Cuthbert, F.J.; Wemmer, L.C.; Doolittle, A.
1997-01-01
Techniques for captive-rearing and releasing piping plovers (Charadrius melodus) were developed using a surrogate species, killdeer (Charadrius vociferus). We compared captive- and parent-reared killdeer, and parent-reared piping plovers and determined that growth and behavior were similar. After surrogate trials determined that captive-rearing was feasible, we used the same methods to raise piping plover chicks from salvaged eggs. For captive-reared chick of both species, survival to fledging was higher than and behaviors similar to parent-reared chicks in the wild. Rearing techniques were fine-tuned, and ten piping plover fledglings were released to the wild. Based on our results, we developed recommendations for captive-rearing piping plovers using salvaged eggs to enhance productivity of small populations.
Genetic modification through oligonucleotide-mediated mutagenesis. A GMO regulatory challenge?
Breyer, Didier; Herman, Philippe; Brandenburger, Annick; Gheysen, Godelieve; Remaut, Erik; Soumillion, Patrice; Van Doorsselaere, Jan; Custers, René; Pauwels, Katia; Sneyers, Myriam; Reheul, Dirk
2009-01-01
In the European Union, the definition of a GMO is technology-based. This means that a novel organism will be regulated under the GMO regulatory framework only if it has been developed with the use of defined techniques. This approach is now challenged with the emergence of new techniques. In this paper, we describe regulatory and safety issues associated with the use of oligonucleotide-mediated mutagenesis to develop novel organisms. We present scientific arguments for not having organisms developed through this technique fall within the scope of the EU regulation on GMOs. We conclude that any political decision on this issue should be taken on the basis of a broad reflection at EU level, while avoiding discrepancies at international level.
Repairing Chipped Silicide Coatings on Refractory Metal Substrates
NASA Technical Reports Server (NTRS)
Youngquist, Robert
2006-01-01
The space shuttle orbiter s reaction control system (RCS) is a series of small thrusters that use hypergolic fuels to orient the orbiter in space. The RCS thrusters are constructed from a special niobium-based alloy -- the C-103. This alloy retains excellent mechanical properties from cryogenic temperature all the way up to 2,500 F (1,370 C). C-103 is susceptible to rapid oxidation at elevated temperatures. The authors have developed two methods to repair damaged R512a coatings on C-103. For the first repair technique, metal foundries, semiconductor manufacturers, and many other industries have developed and routinely use coatings that can easily be painted on metal to protect it from corrosion, including oxidation, to temperatures in excess of 2,500 F (1,370 C). This first repair technique is considered somewhat temporary. The second repair technique is based on using the native coating material of the RCS nozzles. the chipped area is ground out and a "green" R512a coating is applied to the repair area. Both repair techniques can be applied for moderate protection until the permanent laser-repair technique is available to the repair area.
Development of a noninvasive technique for the measurement of intracranial pressure
NASA Technical Reports Server (NTRS)
Ueno, T.; Shuer, L. M.; Yost, W. T.; Hargens, A. R.
1998-01-01
Intracranial pressure (ICP) dynamics are important for understanding adjustments to altered gravity. Previous flight observations document significant facial edema during exposure to microgravity, which suggests that ICP is elevated during microgravity. However, there are no experimental results obtained during space flight, primarily due to the invasiveness of currently available techniques. We have developed and refined a noninvasive technique to measure intracranial pressure noninvasively. The technique is based upon detecting skull movements of a few micrometers in association with altered intracranial pressure. We reported that the PPLL technique has enough sensitivity to detect changes in cranial distance associated with the pulsation of ICP in cadavera. In normal operations, however, we place a transducer on the scalp. Thus, we cannot rule out the possibility that the PPLL technique picks up cutaneous pulsation. The purpose of the present study was therefore to show that the PPLL technique has enough sensitivity to detect changes in cranial distance associated with cardiac cycles in vivo.
Recent patents of nanopore DNA sequencing technology: progress and challenges.
Zhou, Jianfeng; Xu, Bingqian
2010-11-01
DNA sequencing techniques witnessed fast development in the last decades, primarily driven by the Human Genome Project. Among the proposed new techniques, Nanopore was considered as a suitable candidate for the single DNA sequencing with ultrahigh speed and very low cost. Several fabrication and modification techniques have been developed to produce robust and well-defined nanopore devices. Many efforts have also been done to apply nanopore to analyze the properties of DNA molecules. By comparing with traditional sequencing techniques, nanopore has demonstrated its distinctive superiorities in main practical issues, such as sample preparation, sequencing speed, cost-effective and read-length. Although challenges still remain, recent researches in improving the capabilities of nanopore have shed a light to achieve its ultimate goal: Sequence individual DNA strand at single nucleotide level. This patent review briefly highlights recent developments and technological achievements for DNA analysis and sequencing at single molecule level, focusing on nanopore based methods.
A new technique for measuring listening and reading literacy in developing countries
NASA Astrophysics Data System (ADS)
Greene, Barbara A.; Royer, James M.; Anzalone, Stephen
1990-03-01
One problem in evaluating educational interventions in developing countries is the absence of tests that adequately reflect the culture and curriculum. The Sentence Verification Technique is a new procedure for measuring reading and listening comprehension that allows for the development of tests based on materials indigenous to a given culture. The validity of using the Sentence Verification Technique to measure reading comprehension in Grenada was evaluated in the present study. The study involved 786 students at standards 3, 4 and 5. The tests for each standard consisted of passages that varied in difficulty. The students identified as high ability students in all three standards performed better than those identified as low ability. All students performed better with easier passages. Additionally, students in higher standards performed bettter than students in lower standards on a given passage. These results supported the claim that the Sentence Verification Technique is a valid measure of reading comprehension in Grenada.
Chang, Yeong-Chan
2005-12-01
This paper addresses the problem of designing adaptive fuzzy-based (or neural network-based) robust controls for a large class of uncertain nonlinear time-varying systems. This class of systems can be perturbed by plant uncertainties, unmodeled perturbations, and external disturbances. Nonlinear H(infinity) control technique incorporated with adaptive control technique and VSC technique is employed to construct the intelligent robust stabilization controller such that an H(infinity) control is achieved. The problem of the robust tracking control design for uncertain robotic systems is employed to demonstrate the effectiveness of the developed robust stabilization control scheme. Therefore, an intelligent robust tracking controller for uncertain robotic systems in the presence of high-degree uncertainties can easily be implemented. Its solution requires only to solve a linear algebraic matrix inequality and a satisfactorily transient and asymptotical tracking performance is guaranteed. A simulation example is made to confirm the performance of the developed control algorithms.
NASA Astrophysics Data System (ADS)
Arnaud, N.; Balembois, L.; Bizouard, M. A.; Brisson, V.; Casanueva, J.; Cavalier, F.; Davier, M.; Frey, V.; Hello, P.; Huet, D.; Leroy, N.; Loriette, V.; Maksimovic, I.; Robinet, F.
2017-02-01
The second generation of Gravitational waves detectors are kilometric Michelson interferometers with additional recycling Fabry-Perot cavities on the arms and the addition of two more recycling cavities to enhance their sensitivity, with the particularity that all the mirrors are suspended. In order to control them a new technique, based on the use of auxiliary lasers, has been developed to bring the interferometer to its working point, with all the cavities on their resonance, in an adiabatic way. The implementation of this technique in Advanced Virgo is under preparation and the propagation of a stable laser through a 3-km optical fibre is one of the most problematic issues. A new technique of active phase noise cancellation based on the use of Electro Optical Modulators has been developed, and a first prototype has been successfully tested.
Demirci, Oguz; Clark, Vincent P; Calhoun, Vince D
2008-02-15
Schizophrenia is diagnosed based largely upon behavioral symptoms. Currently, no quantitative, biologically based diagnostic technique has yet been developed to identify patients with schizophrenia. Classification of individuals into patient with schizophrenia and healthy control groups based on quantitative biologically based data is of great interest to support and refine psychiatric diagnoses. We applied a novel projection pursuit technique on various components obtained with independent component analysis (ICA) of 70 subjects' fMRI activation maps obtained during an auditory oddball task. The validity of the technique was tested with a leave-one-out method and the detection performance varied between 80% and 90%. The findings suggest that the proposed data reduction algorithm is effective in classifying individuals into schizophrenia and healthy control groups and may eventually prove useful as a diagnostic tool.
ERIC Educational Resources Information Center
Earhart, James; Zamora, Irina
2015-01-01
This pilot study describes the development and initial implementation of a treatment program that uses relationship-based techniques as a basis for promoting characteristics important in learning and emotional regulation. A case example has been included as an illustration of the theoretical framework of this intervention, along with preliminary…
Computer Simulation as an Aid for Management of an Information System.
ERIC Educational Resources Information Center
Simmonds, W. H.; And Others
The aim of this study was to develop methods, based upon computer simulation, of designing information systems and illustrate the use of these methods by application to an information service. The method developed is based upon Monte Carlo and discrete event simulation techniques and is described in an earlier report - Sira report R412 Organizing…
NanoDesign: Concepts and Software for a Nanotechnology Based on Functionalized Fullerenes
NASA Technical Reports Server (NTRS)
Globus, Al; Jaffe, Richard; Chancellor, Marisa K. (Technical Monitor)
1996-01-01
Eric Drexler has proposed a hypothetical nanotechnology based on diamond and investigated the properties of such molecular systems. While attractive, diamonoid nanotechnology is not physically accessible with straightforward extensions of current laboratory techniques. We propose a nanotechnology based on functionalized fullerenes and investigate carbon nanotube based gears with teeth added via a benzyne reaction known to occur with C60. The gears are single-walled carbon nanotubes with appended coenzyme groups for teeth. Fullerenes are in widespread laboratory use and can be functionalized in many ways. Companion papers computationally demonstrate the properties of these gears (they appear to work) and the accessibility of the benzyne/nanotube reaction. This paper describes the molecular design techniques and rationale as well as the software that implements these design techniques. The software is a set of persistent C++ objects controlled by TCL command scripts. The c++/tcl interface is automatically generated by a software system called tcl_c++ developed by the author and described here. The objects keep track of different portions of the molecular machinery to allow different simulation techniques and boundary conditions to be applied as appropriate. This capability has been required to demonstrate (computationally) our gear's feasibility. A new distributed software architecture featuring a WWW universal client, CORBA distributed objects, and agent software is under consideration. The software architecture is intended to eventually enable a widely disbursed group to develop complex simulated molecular machines.
Kinetics of diffusional droplet growth in a liquid/liquid two-phase system
NASA Technical Reports Server (NTRS)
Glickman, M. E.; Fradkov, V. E.
1995-01-01
A new powerful experimental technique based on holographic observations, developed at the NASA Marshall Space Flight Center, now permits observation of small liquid droplets coarsening. This technique was developed and used for mixed-dimensional coarsening studies. Experiments were conducted on an isopycnic two-phase alloy of succinonitrile and water, annealed isothermally over a four-month period. The succinonitrile-rich droplets precipitate from a water-rich liquid matrix having a density very close to that of the droplets. The matrix and droplets, however, have different optical indices. The results of these experiments, along with the results of computer simulation based on the quasi-static diffusion approximation developed at Rensselaer are reported. These results were published recently. Copies of these papers are attached to this report.
Counsell, Serena J; Boardman, James P
2005-10-01
Preterm birth is associated with a high prevalence of neuropsychiatric impairment in childhood and adolescence, but the neural correlates underlying these disorders are not fully understood. Quantitative magnetic resonance imaging techniques have been used to investigate subtle differences in cerebral growth and development among children and adolescents born preterm or with very low birth weight. Diffusion tensor imaging and computer-assisted morphometric techniques (including voxel-based morphometry and deformation-based morphometry) have identified abnormalities in tissue microstructure and cerebral morphology among survivors of preterm birth at different ages, and some of these alterations have specific functional correlates. This chapter reviews the literature reporting differential brain development following preterm birth, with emphasis on the morphological changes that correlate with neuropsychiatric impairment.
Production of small diameter high-temperature-strength refractory metal wires
NASA Technical Reports Server (NTRS)
Petrasek, D. W.; Signorelli, R. A.; King, G. W.
1973-01-01
Special thermomechanical techniques (schedules) have been developed to produce small diameter wire from three refractory metal alloys: colombian base alloy, tantalum base alloy, and tungsten base alloy. High strengths of these wires indicate their potential for contributing increased strength to metallic composites.
A Biomechanical Modeling Guided CBCT Estimation Technique
Zhang, You; Tehrani, Joubin Nasehi; Wang, Jing
2017-01-01
Two-dimensional-to-three-dimensional (2D-3D) deformation has emerged as a new technique to estimate cone-beam computed tomography (CBCT) images. The technique is based on deforming a prior high-quality 3D CT/CBCT image to form a new CBCT image, guided by limited-view 2D projections. The accuracy of this intensity-based technique, however, is often limited in low-contrast image regions with subtle intensity differences. The solved deformation vector fields (DVFs) can also be biomechanically unrealistic. To address these problems, we have developed a biomechanical modeling guided CBCT estimation technique (Bio-CBCT-est) by combining 2D-3D deformation with finite element analysis (FEA)-based biomechanical modeling of anatomical structures. Specifically, Bio-CBCT-est first extracts the 2D-3D deformation-generated displacement vectors at the high-contrast anatomical structure boundaries. The extracted surface deformation fields are subsequently used as the boundary conditions to drive structure-based FEA to correct and fine-tune the overall deformation fields, especially those at low-contrast regions within the structure. The resulting FEA-corrected deformation fields are then fed back into 2D-3D deformation to form an iterative loop, combining the benefits of intensity-based deformation and biomechanical modeling for CBCT estimation. Using eleven lung cancer patient cases, the accuracy of the Bio-CBCT-est technique has been compared to that of the 2D-3D deformation technique and the traditional CBCT reconstruction techniques. The accuracy was evaluated in the image domain, and also in the DVF domain through clinician-tracked lung landmarks. PMID:27831866
NASA Astrophysics Data System (ADS)
Chapman, George B.; Johnson, Glenn; Burdick, Robert
1991-09-01
The CounterMeasure Association Technique (CMAT) is discussed which was developed for the Air Force, and is used to automatically recommend countermeasure and maneuver response to a pilot while he is under missile attack. The overall system is discussed, as well as several key technical components. These components include use of fuzzy sets to specify data uncertainty, use of mimic nets to train the CMAT algorithm to make the same resource optimization tradeoffs as made in a data base of library of training scenarios, and use of several data compression techniques to store the countermeasure effectiveness data base.
Detection and Sizing of Defects in Structural Components of a Nuclear Power Plant by ECT
DOE Office of Scientific and Technical Information (OSTI.GOV)
Chen Zhenmao; Miya, Kenzo
2005-04-09
In this paper, progress of ECT technique for inspection of stress corrosion cracks in a structural component of a nuclear power plant is reported. Access and scanning vehicle (robot), advanced probes for SG tube inspection, development and evaluation of new probes for welding joint, and ECT based crack sizing technique are described respectively. Based on these new techniques, it is clarified that ECT can play as a supplement of UT for the welding zone inspection. It is also proved in this work that new ECT sensors are efficient even for a stainless plate as thick as 15mm.
NASA Astrophysics Data System (ADS)
Vadolia, Gautam R.; Premjit Singh, K.
2017-04-01
Electron Beam Welding (EBW) technology is an established and widely adopted technique in nuclear research and development area. Electron beam welding was thought of as a candidate process for ITER Vacuum Vessel Fabrication. Dhruva Reactor at BARC, Mumbai and Niobium superconducting accelerator cavity at BARC has adopted the EB welding technique as a fabrication route. Study of process capability and limitations based on available literature is consolidated in this short review paper.
Nagashima, Shiori; Yoshida, Akihiro; Suzuki, Nao; Ansai, Toshihiro; Takehara, Tadamichi
2005-01-01
Genomic subtractive hybridization was used to design Prevotella nigrescens-specific primers and TaqMan probes. Based on this technique, a TaqMan real-time PCR assay was developed for quantifying four oral black-pigmented Prevotella species. The combination of real-time PCR and genomic subtractive hybridization is useful for preparing species-specific primer-probe sets for closely related species. PMID:15956428
A Fully Distributed Approach to the Design of a KBIT/SEC VHF Packet Radio Network,
1984-02-01
topological change and consequent out-modea routing data. Algorithm development has been aided by computer simulation using a finite state machine technique...development has been aided by computer simulation using a finite state machine technique to model a realistic network of up to fifty nodes. This is...use of computer based equipments in weapons systems and their associated sensors and command and control elements and the trend from voice to data
High-spatial-resolution passive microwave sounding systems
NASA Technical Reports Server (NTRS)
Staelin, D. H.; Rosenkranz, P. W.
1994-01-01
The principal contributions of this combined theoretical and experimental effort were to advance and demonstrate new and more accurate techniques for sounding atmospheric temperature, humidity, and precipitation profiles at millimeter wavelengths, and to improve the scientific basis for such soundings. Some of these techniques are being incorporated in both research and operational systems. Specific results include: (1) development of the MIT Microwave Temperature Sounder (MTS), a 118-GHz eight-channel imaging spectrometer plus a switched-frequency spectrometer near 53 GHz, for use on the NASA ER-2 high-altitude aircraft, (2) conduct of ER-2 MTS missions in multiple seasons and locations in combination with other instruments, mapping with unprecedented approximately 2-km lateral resolution atmospheric temperature and precipitation profiles, atmospheric transmittances (at both zenith and nadir), frontal systems, and hurricanes, (3) ground based 118-GHz 3-D spectral images of wavelike structure within clouds passing overhead, (4) development and analysis of approaches to ground- and space-based 5-mm wavelength sounding of the upper stratosphere and mesosphere, which supported the planning of improvements to operational weather satellites, (5) development of improved multidimensional and adaptive retrieval methods for atmospheric temperature and humidity profiles, (6) development of combined nonlinear and statistical retrieval techniques for 183-GHz humidity profile retrievals, (7) development of nonlinear statistical retrieval techniques for precipitation cell-top altitudes, and (8) numerical analyses of the impact of remote sensing data on the accuracy of numerical weather predictions; a 68-km gridded model was used to study the spectral properties of error growth.
NASA Astrophysics Data System (ADS)
Jaiswal, Neeru; Kishtawal, C. M.; Pal, P. K.
2013-02-01
India's polar orbiting satellite Oceansat-2 was launched by Indian Space Research Organisation on 23 September 2009 for applications pertaining to ocean studies and meteorology. The wind scatterometer aboard the Oceansat-2 satellite (OSCAT) covers 90 % of the global ocean within a day. In the present study, the OSCAT-derived wind fields are used to predict the genesis of tropical cyclones over the North Indian Ocean using a new technique based on data mining. The technique is based on the premise that there is some degree of similarity in low-level wind circulation among developing systems, which can be utilized to distinguish them from non-developing systems. This similarity of wind patterns has been measured quantitatively by computing the "matching index" between the given wind pattern and the wind signatures of developing systems available from the past observations. The algorithm is used to predict the tropical cyclogenesis of cyclones formed during the period 2009-11 in the North Indian Ocean. All the tropical disturbances that developed into tropical storms during the above period (2009-11), viz. PHYAN, WARD, LAILA, BANDU, PHET, GIRI, JAL, KEILA, FOUR, FIVE and THANE were predicted using the proposed method. The mean prediction lead time of the technique was 63 h. Probability of detection of the technique was 100 %, while the false alarm ratio was 2 %.
Wang, Li-Pen; Ochoa-Rodríguez, Susana; Simões, Nuno Eduardo; Onof, Christian; Maksimović, Cedo
2013-01-01
The applicability of the operational radar and raingauge networks for urban hydrology is insufficient. Radar rainfall estimates provide a good description of the spatiotemporal variability of rainfall; however, their accuracy is in general insufficient. It is therefore necessary to adjust radar measurements using raingauge data, which provide accurate point rainfall information. Several gauge-based radar rainfall adjustment techniques have been developed and mainly applied at coarser spatial and temporal scales; however, their suitability for small-scale urban hydrology is seldom explored. In this paper a review of gauge-based adjustment techniques is first provided. After that, two techniques, respectively based upon the ideas of mean bias reduction and error variance minimisation, were selected and tested using as case study an urban catchment (∼8.65 km(2)) in North-East London. The radar rainfall estimates of four historical events (2010-2012) were adjusted using in situ raingauge estimates and the adjusted rainfall fields were applied to the hydraulic model of the study area. The results show that both techniques can effectively reduce mean bias; however, the technique based upon error variance minimisation can in general better reproduce the spatial and temporal variability of rainfall, which proved to have a significant impact on the subsequent hydraulic outputs. This suggests that error variance minimisation based methods may be more appropriate for urban-scale hydrological applications.
Microfluidics for cell-based high throughput screening platforms - A review.
Du, Guansheng; Fang, Qun; den Toonder, Jaap M J
2016-01-15
In the last decades, the basic techniques of microfluidics for the study of cells such as cell culture, cell separation, and cell lysis, have been well developed. Based on cell handling techniques, microfluidics has been widely applied in the field of PCR (Polymerase Chain Reaction), immunoassays, organ-on-chip, stem cell research, and analysis and identification of circulating tumor cells. As a major step in drug discovery, high-throughput screening allows rapid analysis of thousands of chemical, biochemical, genetic or pharmacological tests in parallel. In this review, we summarize the application of microfluidics in cell-based high throughput screening. The screening methods mentioned in this paper include approaches using the perfusion flow mode, the droplet mode, and the microarray mode. We also discuss the future development of microfluidic based high throughput screening platform for drug discovery. Copyright © 2015 Elsevier B.V. All rights reserved.
Spisák, Sándor; Molnár, Béla; Galamb, Orsolya; Sipos, Ferenc; Tulassay, Zsolt
2007-08-12
The confirmation of mRNA expression studies by protein chips is of high recent interest due to the widespread application of expression arrays. In this review the advantages, technical limitations, application fields and the first results of the protein arrays is described. The bottlenecks of the increasing protein array applications are the fast decomposition of proteins, the problem with aspecific binding and the lack of amplification techniques. Today glass slide based printed, SELDI (MS) based, electrophoresis based and tissue microarray based technologies are available. The advantage of the glass slide based chips are the simplicity of their application, and relatively low cost. The SELDI based protein chip technique is applicable to minute amounts of starting material (<1 microg) but it is the most expensive one. The electrophoresis based techniques are still under intensive development. The tissue microarrays can be used for the parallel testing of the sensitivity and specificity of single antibodies on a broad range of histological specimens on a single slide. Protein chips were successfully used for serum tumor marker detection, cancer research, cell physiology studies and for the verification of mRNA expression studies. Protein chips are envisioned to be available for routine diagnostic applications if the ongoing technology development will be successful in increase in sensitivity, specificity, costs reduction and for the reduction of the necessary sample volume.
Xiao, Xia; Lei, Kin Fong; Huang, Chia-Hao
2015-01-01
Cell migration is a cellular response and results in various biological processes such as cancer metastasis, that is, the primary cause of death for cancer patients. Quantitative investigation of the correlation between cell migration and extracellular stimulation is essential for developing effective therapeutic strategies for controlling invasive cancer cells. The conventional method to determine cell migration rate based on comparison of successive images may not be an objective approach. In this work, a microfluidic chip embedded with measurement electrodes has been developed to quantitatively monitor the cell migration activity based on the impedimetric measurement technique. A no-damage wound was constructed by microfluidic phenomenon and cell migration activity under the stimulation of cytokine and an anti-cancer drug, i.e., interleukin-6 and doxorubicin, were, respectively, investigated. Impedance measurement was concurrently performed during the cell migration process. The impedance change was directly correlated to the cell migration activity; therefore, the migration rate could be calculated. In addition, a good match was found between impedance measurement and conventional imaging analysis. But the impedimetric measurement technique provides an objective and quantitative measurement. Based on our technique, cell migration rates were calculated to be 8.5, 19.1, and 34.9 μm/h under the stimulation of cytokine at concentrations of 0 (control), 5, and 10 ng/ml. This technique has high potential to be developed into a powerful analytical platform for cancer research. PMID:26180566
Flow Control Research at NASA Langley in Support of High-Lift Augmentation
NASA Technical Reports Server (NTRS)
Sellers, William L., III; Jones, Gregory S.; Moore, Mark D.
2002-01-01
The paper describes the efforts at NASA Langley to apply active and passive flow control techniques for improved high-lift systems, and advanced vehicle concepts utilizing powered high-lift techniques. The development of simplified high-lift systems utilizing active flow control is shown to provide significant weight and drag reduction benefits based on system studies. Active flow control that focuses on separation, and the development of advanced circulation control wings (CCW) utilizing unsteady excitation techniques will be discussed. The advanced CCW airfoils can provide multifunctional controls throughout the flight envelope. Computational and experimental data are shown to illustrate the benefits and issues with implementation of the technology.
A Real-Time High Performance Data Compression Technique For Space Applications
NASA Technical Reports Server (NTRS)
Yeh, Pen-Shu; Venbrux, Jack; Bhatia, Prakash; Miller, Warner H.
2000-01-01
A high performance lossy data compression technique is currently being developed for space science applications under the requirement of high-speed push-broom scanning. The technique is also error-resilient in that error propagation is contained within a few scan lines. The algorithm is based on block-transform combined with bit-plane encoding; this combination results in an embedded bit string with exactly the desirable compression rate. The lossy coder is described. The compression scheme performs well on a suite of test images typical of images from spacecraft instruments. Hardware implementations are in development; a functional chip set is expected by the end of 2001.
Report of the panel on international programs
NASA Technical Reports Server (NTRS)
Anderson, Allen Joel; Fuchs, Karl W.; Ganeka, Yasuhiro; Gaur, Vinod; Green, Andrew A.; Siegfried, W.; Lambert, Anthony; Rais, Jacub; Reighber, Christopher; Seeger, Herman
1991-01-01
The panel recommends that NASA participate and take an active role in the continuous monitoring of existing regional networks, the realization of high resolution geopotential and topographic missions, the establishment of interconnection of the reference frames as defined by different space techniques, the development and implementation of automation for all ground-to-space observing systems, calibration and validation experiments for measuring techniques and data, the establishment of international space-based networks for real-time transmission of high density space data in standardized formats, tracking and support for non-NASA missions, and the extension of state-of-the art observing and analysis techniques to developing nations.
Development and evaluation of a technique for in vivo monitoring of 60Co in human liver
NASA Astrophysics Data System (ADS)
Gomes, GH; Silva, MC; Mello, JQ; Dantas, ALA; Dantas, BM
2018-03-01
60Co is an artificial radioactive metal produced by activation of iron with neutrons. It decays by beta particles and gamma radiation and represents a risk of internal exposure of workers involved in the maintenance of nuclear power reactors. Intakes can be quantified through in vivo monitoring. This work describes the development of a technique for the quantification of 60Co in human liver. The sensitivity of the method is evaluated based on the minimum detectable effective doses. The results allow to state that the technique is suitable either for monitoring of occupational exposures or evaluation of accidental intakes.
Nonlinear ultrasonics for material state awareness
NASA Astrophysics Data System (ADS)
Jacobs, L. J.
2014-02-01
Predictive health monitoring of structural components will require the development of advanced sensing techniques capable of providing quantitative information on the damage state of structural materials. By focusing on nonlinear acoustic techniques, it is possible to measure absolute, strength based material parameters that can then be coupled with uncertainty models to enable accurate and quantitative life prediction. Starting at the material level, this review will present current research that involves a combination of sensing techniques and physics-based models to characterize damage in metallic materials. In metals, these nonlinear ultrasonic measurements can sense material state, before the formation of micro- and macro-cracks. Typically, cracks of a measurable size appear quite late in a component's total life, while the material's integrity in terms of toughness and strength gradually decreases due to the microplasticity (dislocations) and associated change in the material's microstructure. This review focuses on second harmonic generation techniques. Since these nonlinear acoustic techniques are acoustic wave based, component interrogation can be performed with bulk, surface and guided waves using the same underlying material physics; these nonlinear ultrasonic techniques provide results which are independent of the wave type used. Recent physics-based models consider the evolution of damage due to dislocations, slip bands, interstitials, and precipitates in the lattice structure, which can lead to localized damage.
Improvement in QEPAS system utilizing a second harmonic based wavelength calibration technique
NASA Astrophysics Data System (ADS)
Zhang, Qinduan; Chang, Jun; Wang, Fupeng; Wang, Zongliang; Xie, Yulei; Gong, Weihua
2018-05-01
A simple laser wavelength calibration technique, based on second harmonic signal, is demonstrated in this paper to improve the performance of quartz enhanced photoacoustic spectroscopy (QEPAS) gas sensing system, e.g. improving the signal to noise ratio (SNR), detection limit and long-term stability. Constant current, corresponding to the gas absorption line, combining f/2 frequency sinusoidal signal are used to drive the laser (constant driving mode), a software based real-time wavelength calibration technique is developed to eliminate the wavelength drift due to ambient fluctuations. Compared to conventional wavelength modulation spectroscopy (WMS), this method allows lower filtering bandwidth and averaging algorithm applied to QEPAS system, improving SNR and detection limit. In addition, the real-time wavelength calibration technique guarantees the laser output is modulated steadily at gas absorption line. Water vapor is chosen as an objective gas to evaluate its performance compared to constant driving mode and conventional WMS system. The water vapor sensor was designed insensitive to the incoherent external acoustic noise by the numerical averaging technique. As a result, the SNR increases 12.87 times in wavelength calibration technique based system compared to conventional WMS system. The new system achieved a better linear response (R2 = 0 . 9995) in concentration range from 300 to 2000 ppmv, and achieved a minimum detection limit (MDL) of 630 ppbv.
NASA Astrophysics Data System (ADS)
Jenifer, M. Annie; Jha, Madan K.
2017-05-01
Groundwater is a treasured underground resource, which plays a central role in sustainable water management. However, it being hidden and dynamic in nature, its sustainable development and management calls for precise quantification of this precious resource at an appropriate scale. This study demonstrates the efficacy of three GIS-based multi-criteria decision analysis (MCDA) techniques, viz., Analytic Hierarchy Process (AHP), Catastrophe and Entropy in evaluating groundwater potential through a case study in hard-rock aquifer systems. Using satellite imagery and relevant field data, eight thematic layers (rainfall, land slope, drainage density, soil, lineament density, geology, proximity to surface water bodies and elevation) of the factors having significant influence on groundwater occurrence were prepared. These thematic layers and their features were assigned suitable weights based on the conceptual frameworks of AHP, Catastrophe and Entropy techniques and then they were integrated in the GIS environment to generate an integrated raster layer depicting groundwater potential index of the study area. The three groundwater prospect maps thus yielded by these MCDA techniques were verified using a novel approach (concept of 'Dynamic Groundwater Potential'). The validation results revealed that the groundwater potential predicted by the AHP technique has a pronounced accuracy of 87% compared to the Catastrophe (46% accuracy) and Entropy techniques (51% accuracy). It is concluded that the AHP technique is the most reliable for the assessment of groundwater resources followed by the Entropy method. The developed groundwater potential maps can serve as a scientific guideline for the cost-effective siting of wells and the effective planning of groundwater development at a catchment or basin scale.
Computational technique for stepwise quantitative assessment of equation correctness
NASA Astrophysics Data System (ADS)
Othman, Nuru'l Izzah; Bakar, Zainab Abu
2017-04-01
Many of the computer-aided mathematics assessment systems that are available today possess the capability to implement stepwise correctness checking of a working scheme for solving equations. The computational technique for assessing the correctness of each response in the scheme mainly involves checking the mathematical equivalence and providing qualitative feedback. This paper presents a technique, known as the Stepwise Correctness Checking and Scoring (SCCS) technique that checks the correctness of each equation in terms of structural equivalence and provides quantitative feedback. The technique, which is based on the Multiset framework, adapts certain techniques from textual information retrieval involving tokenization, document modelling and similarity evaluation. The performance of the SCCS technique was tested using worked solutions on solving linear algebraic equations in one variable. 350 working schemes comprising of 1385 responses were collected using a marking engine prototype, which has been developed based on the technique. The results show that both the automated analytical scores and the automated overall scores generated by the marking engine exhibit high percent agreement, high correlation and high degree of agreement with manual scores with small average absolute and mixed errors.
Managing complex processing of medical image sequences by program supervision techniques
NASA Astrophysics Data System (ADS)
Crubezy, Monica; Aubry, Florent; Moisan, Sabine; Chameroy, Virginie; Thonnat, Monique; Di Paola, Robert
1997-05-01
Our objective is to offer clinicians wider access to evolving medical image processing (MIP) techniques, crucial to improve assessment and quantification of physiological processes, but difficult to handle for non-specialists in MIP. Based on artificial intelligence techniques, our approach consists in the development of a knowledge-based program supervision system, automating the management of MIP libraries. It comprises a library of programs, a knowledge base capturing the expertise about programs and data and a supervision engine. It selects, organizes and executes the appropriate MIP programs given a goal to achieve and a data set, with dynamic feedback based on the results obtained. It also advises users in the development of new procedures chaining MIP programs.. We have experimented the approach for an application of factor analysis of medical image sequences as a means of predicting the response of osteosarcoma to chemotherapy, with both MRI and NM dynamic image sequences. As a result our program supervision system frees clinical end-users from performing tasks outside their competence, permitting them to concentrate on clinical issues. Therefore our approach enables a better exploitation of possibilities offered by MIP and higher quality results, both in terms of robustness and reliability.
Satellite based Ocean Forecasting, the SOFT project
NASA Astrophysics Data System (ADS)
Stemmann, L.; Tintoré, J.; Moneris, S.
2003-04-01
The knowledge of future oceanic conditions would have enormous impact on human marine related areas. For such reasons, a number of international efforts are being carried out to obtain reliable and manageable ocean forecasting systems. Among the possible techniques that can be used to estimate the near future states of the ocean, an ocean forecasting system based on satellite imagery is developped through the Satelitte based Ocean ForecasTing project (SOFT). SOFT, established by the European Commission, considers the development of a forecasting system of the ocean space-time variability based on satellite data by using Artificial Intelligence techniques. This system will be merged with numerical simulation approaches, via assimilation techniques, to get a hybrid SOFT-numerical forecasting system of improved performance. The results of the project will provide efficient forecasting of sea-surface temperature structures, currents, dynamic height, and biological activity associated to chlorophyll fields. All these quantities could give valuable information on the planning and management of human activities in marine environments such as navigation, fisheries, pollution control, or coastal management. A detailed identification of present or new needs and potential end-users concerned by such an operational tool is being performed. The project would study solutions adapted to these specific needs.
Acoustic Techniques for Structural Health Monitoring
NASA Astrophysics Data System (ADS)
Frankenstein, B.; Augustin, J.; Hentschel, D.; Schubert, F.; Köhler, B.; Meyendorf, N.
2008-02-01
Future safety and maintenance strategies for industrial components and vehicles are based on combinations of monitoring systems that are permanently attached to or embedded in the structure, and periodic inspections. The latter belongs to conventional nondestructive evaluation (NDE) and can be enhanced or partially replaced by structural health monitoring systems. However, the main benefit of this technology for the future will consist of systems that can be differently designed based on improved safety philosophies, including continuous monitoring. This approach will increase the efficiency of inspection procedures at reduced inspection times. The Fraunhofer IZFP Dresden Branch has developed network nodes, miniaturized transmitter and receiver systems for active and passive acoustical techniques and sensor systems that can be attached to or embedded into components or structures. These systems have been used to demonstrate intelligent sensor networks for the monitoring of aerospace structures, railway systems, wind energy generators, piping system and other components. Material discontinuities and flaws have been detected and monitored during full scale fatigue testing. This paper will discuss opportunities and future trends in nondestructive evaluation and health monitoring based on new sensor principles and advanced microelectronics. It will outline various application examples of monitoring systems based on acoustic techniques and will indicate further needs for research and development.
NASA Technical Reports Server (NTRS)
Shen, Ji Y.; Sharpe, Lonnie, Jr.
1998-01-01
The research activity for this project is mainly to investigate the necessity and feasibility to develop a structural health monitoring system for rocket engines, and to carry out a research plan for further development of the system. More than one hundred technical papers have been searched and reviewed during the period. We concluded after this investigation that adding a new module in NASA's existing automated diagnostic system to monitor the healthy condition of rocket engine structures is a crucial task, and it's possible to develop such a system based upon the vibrational-based nondestructive damage assessment techniques. A number of such techniques have been introduced. Their advantages and disadvantages are also discussed. A global research plan has been figured out. As the first step of the overall research plan, a proposal for the next fiscal year has been submitted.
On prognostic models, artificial intelligence and censored observations.
Anand, S S; Hamilton, P W; Hughes, J G; Bell, D A
2001-03-01
The development of prognostic models for assisting medical practitioners with decision making is not a trivial task. Models need to possess a number of desirable characteristics and few, if any, current modelling approaches based on statistical or artificial intelligence can produce models that display all these characteristics. The inability of modelling techniques to provide truly useful models has led to interest in these models being purely academic in nature. This in turn has resulted in only a very small percentage of models that have been developed being deployed in practice. On the other hand, new modelling paradigms are being proposed continuously within the machine learning and statistical community and claims, often based on inadequate evaluation, being made on their superiority over traditional modelling methods. We believe that for new modelling approaches to deliver true net benefits over traditional techniques, an evaluation centric approach to their development is essential. In this paper we present such an evaluation centric approach to developing extensions to the basic k-nearest neighbour (k-NN) paradigm. We use standard statistical techniques to enhance the distance metric used and a framework based on evidence theory to obtain a prediction for the target example from the outcome of the retrieved exemplars. We refer to this new k-NN algorithm as Censored k-NN (Ck-NN). This reflects the enhancements made to k-NN that are aimed at providing a means for handling censored observations within k-NN.
Hyphenated analytical techniques for materials characterisation
NASA Astrophysics Data System (ADS)
Armstrong, Gordon; Kailas, Lekshmi
2017-09-01
This topical review will provide a survey of the current state of the art in ‘hyphenated’ techniques for characterisation of bulk materials, surface, and interfaces, whereby two or more analytical methods investigating different properties are applied simultaneously to the same sample to better characterise the sample than can be achieved by conducting separate analyses in series using different instruments. It is intended for final year undergraduates and recent graduates, who may have some background knowledge of standard analytical techniques, but are not familiar with ‘hyphenated’ techniques or hybrid instrumentation. The review will begin by defining ‘complementary’, ‘hybrid’ and ‘hyphenated’ techniques, as there is not a broad consensus among analytical scientists as to what each term means. The motivating factors driving increased development of hyphenated analytical methods will also be discussed. This introduction will conclude with a brief discussion of gas chromatography-mass spectroscopy and energy dispersive x-ray analysis in electron microscopy as two examples, in the context that combining complementary techniques for chemical analysis were among the earliest examples of hyphenated characterisation methods. The emphasis of the main review will be on techniques which are sufficiently well-established that the instrumentation is commercially available, to examine physical properties including physical, mechanical, electrical and thermal, in addition to variations in composition, rather than methods solely to identify and quantify chemical species. Therefore, the proposed topical review will address three broad categories of techniques that the reader may expect to encounter in a well-equipped materials characterisation laboratory: microscopy based techniques, scanning probe-based techniques, and thermal analysis based techniques. Examples drawn from recent literature, and a concluding case study, will be used to explain the practical issues that arise in combining different techniques. We will consider how the complementary and varied information obtained by combining these techniques may be interpreted together to better understand the sample in greater detail than that was possible before, and also how combining different techniques can simplify sample preparation and ensure reliable comparisons are made between multiple analyses on the same samples—a topic of particular importance as nanoscale technologies become more prevalent in applied and industrial research and development (R&D). The review will conclude with a brief outline of the emerging state of the art in the research laboratory, and a suggested approach to using hyphenated techniques, whether in the teaching, quality control or R&D laboratory.
Radar data smoothing filter study
NASA Technical Reports Server (NTRS)
White, J. V.
1984-01-01
The accuracy of the current Wallops Flight Facility (WFF) data smoothing techniques for a variety of radars and payloads is examined. Alternative data reduction techniques are given and recommendations are made for improving radar data processing at WFF. A data adaptive algorithm, based on Kalman filtering and smoothing techniques, is also developed for estimating payload trajectories above the atmosphere from noisy time varying radar data. This algorithm is tested and verified using radar tracking data from WFF.
Employment of adaptive learning techniques for the discrimination of acoustic emissions
NASA Astrophysics Data System (ADS)
Erkes, J. W.; McDonald, J. F.; Scarton, H. A.; Tam, K. C.; Kraft, R. P.
1983-11-01
The following aspects of this study on the discrimination of acoustic emissions (AE) were examined: (1) The analytical development and assessment of digital signal processing techniques for AE signal dereverberation, noise reduction, and source characterization; (2) The modeling and verification of some aspects of key selected techniques through a computer-based simulation; and (3) The study of signal propagation physics and their effect on received signal characteristics for relevant physical situations.
CMOS array design automation techniques. [metal oxide semiconductors
NASA Technical Reports Server (NTRS)
Ramondetta, P.; Feller, A.; Noto, R.; Lombardi, T.
1975-01-01
A low cost, quick turnaround technique for generating custom metal oxide semiconductor arrays using the standard cell approach was developed, implemented, tested and validated. Basic cell design topology and guidelines are defined based on an extensive analysis that includes circuit, layout, process, array topology and required performance considerations particularly high circuit speed.
Accommodation Strategies of College Students with Disabilities
ERIC Educational Resources Information Center
Barnard-Brak, Lucy; Lechtenberger, DeAnn; Lan, William Y.
2010-01-01
College students with disabilities develop and utilize strategies to facilitate their learning experiences due to their unique academic needs. Using a semi-structured interview technique to collect data and a technique based in grounded theory to analyze this data, the purpose of this study was to discern the meaning of disclosure for college…
Statistical Techniques for Efficient Indexing and Retrieval of Document Images
ERIC Educational Resources Information Center
Bhardwaj, Anurag
2010-01-01
We have developed statistical techniques to improve the performance of document image search systems where the intermediate step of OCR based transcription is not used. Previous research in this area has largely focused on challenges pertaining to generation of small lexicons for processing handwritten documents and enhancement of poor quality…
Alternative evaluation of innovations’ effectiveness in mechanical engineering
NASA Astrophysics Data System (ADS)
Puryaev, A. S.
2017-09-01
The aim of present work is approbation of the developed technique for assessing innovations’ effectiveness. We demonstrate an alternative assessment of innovations’ effectiveness (innovation projects) in mechanical engineering on illustrative example. It is proposed as an alternative to the traditional method technique based on the value concept and the method of “Cash flow”.
Integration of Video-Based Demonstrations to Prepare Students for the Organic Chemistry Laboratory
ERIC Educational Resources Information Center
Nadelson, Louis S.; Scaggs, Jonathan; Sheffield, Colin; McDougal, Owen M.
2015-01-01
Consistent, high-quality introductions to organic chemistry laboratory techniques effectively and efficiently support student learning in the organic chemistry laboratory. In this work, we developed and deployed a series of instructional videos to communicate core laboratory techniques and concepts. Using a quasi-experimental design, we tested the…
USDA-ARS?s Scientific Manuscript database
Objective: To prepare a new fluorescent tracer against common mycotoxins such as fumonisin B1 in order to replace 6-(4,6-Dichlorotriazinyl) aminofluorescein (6-DTAF), an expensive marker, and to develop a technique for quick detection of fumonisin B1 based on the principle of fluorescence polarizati...
A Semester-Long Project-Oriented Biochemistry Laboratory Based on "Helicobacter pylori" Urease
ERIC Educational Resources Information Center
Farnham, Kate R.; Dube, Danielle H.
2015-01-01
Here we present the development of a 13 week project-oriented biochemistry laboratory designed to introduce students to foundational biochemical techniques and then enable students to perform original research projects once they have mastered these techniques. In particular, we describe a semester-long laboratory that focuses on a biomedically…
Semantics of User Interface for Image Retrieval: Possibility Theory and Learning Techniques.
ERIC Educational Resources Information Center
Crehange, M.; And Others
1989-01-01
Discusses the need for a rich semantics for the user interface in interactive image retrieval and presents two methods for building such interfaces: possibility theory applied to fuzzy data retrieval, and a machine learning technique applied to learning the user's deep need. Prototypes developed using videodisks and knowledge-based software are…
Determining the Deacetylation Degree of Chitosan: Opportunities to Learn Instrumental Techniques
ERIC Educational Resources Information Center
Pérez-Álvarez, Leyre; Ruiz-Rubio, Leire; Vilas-Vilela, Jose Luis
2018-01-01
To enhance critical thinking and problem-solving skills, a project-based learning (PBL) approach for "Instrumental Techniques" courses in undergraduate physical chemistry was specifically developed for a pharmacy bachelor degree program. The starting point of this PBL was an open-ended question that is close to the student scientist's…
Metrics and Science Monograph Collections at the Marston Science Library, University of Florida
ERIC Educational Resources Information Center
Leonard, Michelle F.; Haas, Stephanie C.; Kisling, Vernon N.
2010-01-01
As academic libraries are increasingly supported by a matrix of database functions, the use of data mining and visualization techniques offer significant potential for future collection development and service initiatives based on quantifiable data. While data collection techniques are still not standardized and results may be skewed because of…
Make Your Own Animated Movies. Yellow Ball Workshop Film Techniques.
ERIC Educational Resources Information Center
Anderson, Yvonne
At the Yellow Ball Workshop, children and teenagers make animated films using simple art materials and camera equipment. Based on the animation techniques developed at the workshop, complete instructions for constructing backgrounds and characters and for animating the figures are provided. Setting up and using the camera, splicing film,…
App-assisted external ventricular drain insertion.
Eftekhar, Behzad
2016-09-01
The freehand technique for insertion of an external ventricular drain (EVD) is based on fixed anatomical landmarks and does not take individual variations into consideration. A patient-tailored approach based on augmented-reality techniques using devices such as smartphones can address this shortcoming. The Sina neurosurgical assist (Sina) is an Android mobile device application (app) that was designed and developed to be used as a simple intraoperative neurosurgical planning aid. It overlaps the patient's images from previously performed CT or MRI studies on the image seen through the device camera. The device is held by an assistant who aligns the images and provides information about the relative position of the target and EVD to the surgeon who is performing EVD insertion. This app can be used to provide guidance and continuous monitoring during EVD placement. The author describes the technique of Sina-assisted EVD insertion into the frontal horn of the lateral ventricle and reports on its clinical application in 5 cases as well as the results of ex vivo studies of ease of use and precision. The technique has potential for further development and use with other augmented-reality devices.
Prediction of quantitative intrathoracic fluid volume to diagnose pulmonary oedema using LabVIEW.
Urooj, Shabana; Khan, M; Ansari, A Q; Lay-Ekuakille, Aimé; Salhan, Ashok K
2012-01-01
Pulmonary oedema is a life-threatening disease that requires special attention in the area of research and clinical diagnosis. Computer-based techniques are rarely used to quantify the intrathoracic fluid volume (IFV) for diagnostic purposes. This paper discusses a software program developed to detect and diagnose pulmonary oedema using LabVIEW. The software runs on anthropometric dimensions and physiological parameters, mainly transthoracic electrical impedance (TEI). This technique is accurate and faster than existing manual techniques. The LabVIEW software was used to compute the parameters required to quantify IFV. An equation relating per cent control and IFV was obtained. The results of predicted TEI and measured TEI were compared with previously reported data to validate the developed program. It was found that the predicted values of TEI obtained from the computer-based technique were much closer to the measured values of TEI. Six new subjects were enrolled to measure and predict transthoracic impedance and hence to quantify IFV. A similar difference was also observed in the measured and predicted values of TEI for the new subjects.
Spindle speed variation technique in turning operations: Modeling and real implementation
NASA Astrophysics Data System (ADS)
Urbikain, G.; Olvera, D.; de Lacalle, L. N. López; Elías-Zúñiga, A.
2016-11-01
Chatter is still one of the most challenging problems in machining vibrations. Researchers have focused their efforts to prevent, avoid or reduce chatter vibrations by introducing more accurate predictive physical methods. Among them, the techniques based on varying the rotational speed of the spindle (or SSV, Spindle Speed Variation) have gained great relevance. However, several problems need to be addressed due to technical and practical reasons. On one hand, they can generate harmful overheating of the spindle especially at high speeds. On the other hand, the machine may be unable to perform the interpolation properly. Moreover, it is not trivial to select the most appropriate tuning parameters. This paper conducts a study of the real implementation of the SSV technique in turning systems. First, a stability model based on perturbation theory was developed for simulation purposes. Secondly, the procedure to realistically implement the technique in a conventional turning center was tested and developed. The balance between the improved stability margins and acceptable behavior of the spindle is ensured by energy consumption measurements. Mathematical model shows good agreement with experimental cutting tests.
NASA Astrophysics Data System (ADS)
Rădulescu, Dragoş; Grumezescu, Valentina; Andronescu, Ecaterina; Holban, Alina Maria; Grumezescu, Alexandru Mihai; Socol, Gabriel; Oprea, Alexandra Elena; Rădulescu, Marius; Surdu, Adrian; Trusca, Roxana; Rădulescu, Radu; Chifiriuc, Mariana Carmen; Stan, Miruna S.; Constanda, Sabrina; Dinischiotu, Anca
2016-06-01
In this study we aimed to obtain functionalized thin films based on hydroxyapatite/poly(lactic-co-glycolic acid) (HAp/PLGA) containing ceftriaxone/cefuroxime antibiotics (ATBs) deposited by Matrix Assisted Pulsed Laser Evaporation (MAPLE) technique. The prepared thin films were characterized by transmission electron microscopy (TEM), scanning electron microscopy (SEM), X-Ray diffraction (XRD), selected area electron diffraction (SAED), and infra red (IR) analysis. HAp/PLGA/ATBs thin films sustained the growth of human osteoblasts, proving their good biocompatibility. The microscopic evaluation and the culture-based quantitative assay of the E. coli biofilm development showed that the thin films inhibited the initial step of microbial attachment as well as the subsequent colonization and biofilm development on the respective surfaces. This study demonstrates that MAPLE technique could represent an appealing technique for the fabrication of antibiotics-containing polymeric implant coatings. The bioevaluation results recommend this type of surfaces for the prevention of bone implant microbial contamination and for the enhanced stimulation of the implant osseointegration process.
New Language and Old Problems in Breast Cancer Radiotherapy.
Chiricuţă, Ion Christian
2017-01-01
New developments in breast cancer radiotherapy make possible new standards in treatment recommandations based on international guidelines. Developments in radiotherapy irradiation techniques from 2D to 3D-Conformal RT and to IMRT (Intensity Modulated Arc Therapy) make possible to reduce the usual side effects on the organs at risk as: skin, lung, miocard, bone, esophagus and brahial plexus. Dispite of all these progresses acute and late side effects are present. Side effects are as old as the radiotherapy was used. New solutions are available now by improving irradiation techniques. New techniques as sentinel node procedure (SNP) or partial breast irradiation (PBRT) and immediate breast reconstruction with silicon implants (IBRIS) make necessary new considerations regarding the target volume delineations. A new language for definition of gross tumor volume (GTV), clinical target volume (CTV) based on the new diagnostic methods as PET/CT,nonaparticle MRI will have real impact on target delineation and irradiation techniques. "The new common language in breast cancer therapy" would be the first step to improve the endresults and finally the quality of life of the patients. Celsius.
Evaluation of bridge cables corrosion using acoustic emission technique
NASA Astrophysics Data System (ADS)
Li, Dongsheng; Ou, Jinping
2010-04-01
Owing to the nature of the stress, corrosion of bridge cable may result in catastrophic failure of the structure. However, using electrochemical techniques isn't fully efficient for the detection and control on line of the corrosion phenomenon. A non-destructive testing method based on acoustic emission technique monitoring bridge cable corrosion was explored. The steel strands were placed at room temperature in 5% NaCl solution. Acoustic emission (AE) characteristic parameters were recorded in the whole corrosion experiment process. Based on the plot of cumulated acoustic activity, the bridge cables corrosion included three stages. It can be clearly seen that different stages have different acoustic emission signal characteristics. The AE characteristic parameters would be increased with cables corrosion development. Finally, the bridge cables corrosion experiment with different stress state and different corrosion environment was performed. The results shows that stress magnitude only affects the bridge cable failure time, however, the AE characteristic parameters value has changed a little. It was verified that AE technique can be used to detect the bridge cable early corrosion, investigating corrosion developing trend, and in monitoring and evaluating corrosion damages.
Ultrasound Elastography: Review of Techniques and Clinical Applications
Sigrist, Rosa M.S.; Liau, Joy; Kaffas, Ahmed El; Chammas, Maria Cristina; Willmann, Juergen K.
2017-01-01
Elastography-based imaging techniques have received substantial attention in recent years for non-invasive assessment of tissue mechanical properties. These techniques take advantage of changed soft tissue elasticity in various pathologies to yield qualitative and quantitative information that can be used for diagnostic purposes. Measurements are acquired in specialized imaging modes that can detect tissue stiffness in response to an applied mechanical force (compression or shear wave). Ultrasound-based methods are of particular interest due to its many inherent advantages, such as wide availability including at the bedside and relatively low cost. Several ultrasound elastography techniques using different excitation methods have been developed. In general, these can be classified into strain imaging methods that use internal or external compression stimuli, and shear wave imaging that use ultrasound-generated traveling shear wave stimuli. While ultrasound elastography has shown promising results for non-invasive assessment of liver fibrosis, new applications in breast, thyroid, prostate, kidney and lymph node imaging are emerging. Here, we review the basic principles, foundation physics, and limitations of ultrasound elastography and summarize its current clinical use and ongoing developments in various clinical applications. PMID:28435467
Kirschvink, Joseph L.; Winklhofer, Michael; Walker, Michael M.
2010-01-01
The first demonstrations of magnetic effects on the behaviour of migratory birds and homing pigeons in laboratory and field experiments, respectively, provided evidence for the longstanding hypothesis that animals such as birds that migrate and home over long distances would benefit from possession of a magnetic sense. Subsequent identification of at least two plausible biophysical mechanisms for magnetoreception in animals, one based on biogenic magnetite and another on radical-pair biochemical reactions, led to major efforts over recent decades to test predictions of the two models, as well as efforts to understand the ultrastructure and function of the possible magnetoreceptor cells. Unfortunately, progress in understanding the magnetic sense has been challenged by: (i) the availability of a relatively small number of techniques for analysing behavioural responses to magnetic fields by animals; (ii) difficulty in achieving reproducible results using the techniques; and (iii) difficulty in development and implementation of new techniques that might bring greater experimental power. As a consequence, laboratory and field techniques used to study the magnetic sense today remain substantially unchanged, despite the huge developments in technology and instrumentation since the techniques were developed in the 1950s. New methods developed for behavioural study of the magnetic sense over the last 30 years include the use of laboratory conditioning techniques and tracking devices based on transmission of radio signals to and from satellites. Here we consider methodological developments in the study of the magnetic sense and present suggestions for increasing the reproducibility and ease of interpretation of experimental studies. We recommend that future experiments invest more effort in automating control of experiments and data capture, control of stimulation and full blinding of experiments in the rare cases where automation is impossible. We also propose new experiments to confirm whether or not animals can detect magnetic fields using the radical-pair effect together with an alternate hypothesis that may explain the dependence on light of responses by animals to magnetic field stimuli. PMID:20071390
Blagus, Rok; Lusa, Lara
2015-11-04
Prediction models are used in clinical research to develop rules that can be used to accurately predict the outcome of the patients based on some of their characteristics. They represent a valuable tool in the decision making process of clinicians and health policy makers, as they enable them to estimate the probability that patients have or will develop a disease, will respond to a treatment, or that their disease will recur. The interest devoted to prediction models in the biomedical community has been growing in the last few years. Often the data used to develop the prediction models are class-imbalanced as only few patients experience the event (and therefore belong to minority class). Prediction models developed using class-imbalanced data tend to achieve sub-optimal predictive accuracy in the minority class. This problem can be diminished by using sampling techniques aimed at balancing the class distribution. These techniques include under- and oversampling, where a fraction of the majority class samples are retained in the analysis or new samples from the minority class are generated. The correct assessment of how the prediction model is likely to perform on independent data is of crucial importance; in the absence of an independent data set, cross-validation is normally used. While the importance of correct cross-validation is well documented in the biomedical literature, the challenges posed by the joint use of sampling techniques and cross-validation have not been addressed. We show that care must be taken to ensure that cross-validation is performed correctly on sampled data, and that the risk of overestimating the predictive accuracy is greater when oversampling techniques are used. Examples based on the re-analysis of real datasets and simulation studies are provided. We identify some results from the biomedical literature where the incorrect cross-validation was performed, where we expect that the performance of oversampling techniques was heavily overestimated.
Spacecraft Maneuvering at the Sun/Earth-Moon L2 Libration Point
NASA Astrophysics Data System (ADS)
Shahid, Kamran
Spacecraft formation flying in the vicinity of the Sun/Earth-Moon libration points offers many promising possibilities for space exploration. The concept of formation flying involves the distribution of the functionality of a single spacecraft among several smaller, cooperative spacecraft. The libration points are locations relative to two large orbiting bodies where a third body with relatively small mass can remain stationary relative to the two larger bodies. The most significant perturbation experienced by a spacecraft at the libration point is effect of solar radiation pressure. This thesis presents the development of nonlinear control techniques for maneuvering control at the Sun-Earth/Moon L2 libration point. A new thruster based formation control technique is presented. We also consider a leader/follower formation architecture, and examine the station keeping control of the leader spacecraft and the formation control of the follower spacecraft using solar radiation pressure. Reference trajectories of the leader spacecraft, halo and Lissajous orbits, are determined using a numerical technique in order to take into account all major gravitational perturbations. The nonlinear controllers are developed based on Lyapunov analysis, including non-adaptive and adaptive designs. Thruster based and solar radiation pressure based control laws for spacecraft maneuvering at the Sun-Earth/Moon libration point are developed. Higher order sliding mode control is utilized to address the non-affine structure of the solar sail control inputs. The reduced input solar radiation pressure problem is properly addressed as an underactuated control problem. The development of adaptive control for solar sail equipped spacecraft is an innovation and represents and advancement in solar sailing control technology. Controller performance is evaluated in a high fidelity ephemeris model to reflect a realistic simulated space environment. The numerical results demonstrate the effectiveness of the proposed control techniques for spacecraft maneuvering using solar radiation pressure at the L2 libration point. Stationkeeping accuracies of 50m and formation maintenance accuracies of less than 1m are possible using solar radiation pressure at a sub-L2 libration point. The benefits of these control techniques include increasing libration point mission lifetimes and doubling payload mass fractions as compared to conventional propulsion methods.
Lightning Initiation Forecasting: An Operational Dual-Polarimetric Radar Technique
NASA Technical Reports Server (NTRS)
Woodard, Crystal J.; Carey, L. D.; Petersen, W. A.; Roeder, W. P.
2011-01-01
The objective of this NASA MSFC and NOAA CSTAR funded study is to develop and test operational forecast algorithms for the prediction of lightning initiation utilizing the C-band dual-polarimetric radar, UAHuntsville's Advanced Radar for Meteorological and Operational Research (ARMOR). Although there is a rich research history of radar signatures associated with lightning initiation, few studies have utilized dual-polarimetric radar signatures (e.g., Z(sub dr) columns) and capabilities (e.g., fuzzy-logic particle identification [PID] of precipitation ice) in an operational algorithm for first flash forecasting. The specific goal of this study is to develop and test polarimetric techniques that enhance the performance of current operational radar reflectivity based first flash algorithms. Improving lightning watch and warning performance will positively impact personnel safety in both work and leisure environments. Advanced warnings can provide space shuttle launch managers time to respond appropriately to secure equipment and personnel, while they can also provide appropriate warnings for spectators and players of leisure sporting events to seek safe shelter. Through the analysis of eight case dates, consisting of 35 pulse-type thunderstorms and 20 non-thunderstorm case studies, lightning initiation forecast techniques were developed and tested. The hypothesis is that the additional dual-polarimetric information could potentially reduce false alarms while maintaining high probability of detection and increasing lead-time for the prediction of the first lightning flash relative to reflectivity-only based techniques. To test the hypothesis, various physically-based techniques using polarimetric variables and/or PID categories, which are strongly correlated to initial storm electrification (e.g., large precipitation ice production via drop freezing), were benchmarked against the operational reflectivity-only based approaches to find the best compromise between forecast skill and lead-time. Forecast skill is determined by statistical analysis of probability of detection (POD), false alarm ratio (FAR), Operational Utility Index (OUI), and critical success index (CSI).
NASA Astrophysics Data System (ADS)
Conerty, Michelle D.; Castracane, James; Cacace, Anthony T.; Parnes, Steven M.; Gardner, Glendon M.; Miller, Mitchell B.
1995-05-01
Electronic Speckle Pattern Interferometry (ESPI) is a nondestructive optical evaluation technique that is capable of determining surface and subsurface integrity through the quantitative evaluation of static or vibratory motion. By utilizing state of the art developments in the areas of lasers, fiber optics and solid state detector technology, this technique has become applicable in medical research and diagnostics. Based on initial support from NIDCD and continued support from InterScience, Inc., we have been developing a range of instruments for improved diagnostic evaluation in otolaryngological applications based on the technique of ESPI. These compact fiber optic instruments are capable of making real time interferometric measurements of the target tissue. Ongoing development of image post- processing software is currently capable of extracting the desired quantitative results from the acquired interferometric images. The goal of the research is to develop a fully automated system in which the image processing and quantification will be performed in hardware in near real-time. Subsurface details of both the tympanic membrane and vocal cord dynamics could speed the diagnosis of otosclerosis, laryngeal tumors, and aid in the evaluation of surgical procedures.
A photogrammetric technique for generation of an accurate multispectral optical flow dataset
NASA Astrophysics Data System (ADS)
Kniaz, V. V.
2017-06-01
A presence of an accurate dataset is the key requirement for a successful development of an optical flow estimation algorithm. A large number of freely available optical flow datasets were developed in recent years and gave rise for many powerful algorithms. However most of the datasets include only images captured in the visible spectrum. This paper is focused on the creation of a multispectral optical flow dataset with an accurate ground truth. The generation of an accurate ground truth optical flow is a rather complex problem, as no device for error-free optical flow measurement was developed to date. Existing methods for ground truth optical flow estimation are based on hidden textures, 3D modelling or laser scanning. Such techniques are either work only with a synthetic optical flow or provide a sparse ground truth optical flow. In this paper a new photogrammetric method for generation of an accurate ground truth optical flow is proposed. The method combines the benefits of the accuracy and density of a synthetic optical flow datasets with the flexibility of laser scanning based techniques. A multispectral dataset including various image sequences was generated using the developed method. The dataset is freely available on the accompanying web site.
Review of the Potential of the Ni/Cu Plating Technique for Crystalline Silicon Solar Cells
Rehman, Atteq ur; Lee, Soo Hong
2014-01-01
Developing a better method for the metallization of silicon solar cells is integral part of realizing superior efficiency. Currently, contact realization using screen printing is the leading technology in the silicon based photovoltaic industry, as it is simple and fast. However, the problem with metallization of this kind is that it has a lower aspect ratio and higher contact resistance, which limits solar cell efficiency. The mounting cost of silver pastes and decreasing silicon wafer thicknesses encourages silicon solar cell manufacturers to develop fresh metallization techniques involving a lower quantity of silver usage and not relying pressing process of screen printing. In recent times nickel/copper (Ni/Cu) based metal plating has emerged as a metallization method that may solve these issues. This paper offers a detailed review and understanding of a Ni/Cu based plating technique for silicon solar cells. The formation of a Ni seed layer by adopting various deposition techniques and a Cu conducting layer using a light induced plating (LIP) process are appraised. Unlike screen-printed metallization, a step involving patterning is crucial for opening the masking layer. Consequently, experimental procedures involving patterning methods are also explicated. Lastly, the issues of adhesion, back ground plating, process complexity and reliability for industrial applications are also addressed. PMID:28788516
The assessment of sympathetic activity using iPPG based inter-limb coherence measurements
NASA Astrophysics Data System (ADS)
Tsoy, Maria O.; Rogatina, Kristina V.; Stiukhina, Elena S.; Postnov, Dmitry E.
2017-04-01
Photoplethysmography is an optical technique that can be used to detect blood volume changes and to measure important physiological parameters. This is low cost and non-invasive technique. However, one has to apply sensor directly to the skin. In this regard, the development on remote mothods receives the growing attention, such as imaging photoplethysmography (iPPG). Note, most of public-available iPPG systems are based on smartphone-embedded cameras, and thus have a sample frequency about 30-60 frames per second, which is enough for heart rate measurements, but may be too low for some more advanced usages of this technique. In our work, we describe the attempt to use smartphone-based iPPG technique aimed to measure the tiny mismatch in RR interval data series recorded from left and right arms. We use the transmission mode iPPG, in which the light transmitted through the medium of finger is detected by a web-camera opposite the LED source. The computational scheme by processing and analysis of the received signal was implemented using MATLAB language (MathWork Inc. in the United States). We believe that further development of our approach may lead to fast and low cost method to access the state of the sympathetic nervous system.
Analytical techniques for mechanistic characterization of EUV photoresists
NASA Astrophysics Data System (ADS)
Grzeskowiak, Steven; Narasimhan, Amrit; Murphy, Michael; Ackerman, Christian; Kaminsky, Jake; Brainard, Robert L.; Denbeaux, Greg
2017-03-01
Extreme ultraviolet (EUV, 13.5 nm) lithography is the prospective technology for high volume manufacturing by the microelectronics industry. Significant strides towards achieving adequate EUV source power and availability have been made recently, but a limited rate of improvement in photoresist performance still delays the implementation of EUV. Many fundamental questions remain to be answered about the exposure mechanisms of even the relatively well understood chemically amplified EUV photoresists. Moreover, several groups around the world are developing revolutionary metal-based resists whose EUV exposure mechanisms are even less understood. Here, we describe several evaluation techniques to help elucidate mechanistic details of EUV exposure mechanisms of chemically amplified and metal-based resists. EUV absorption coefficients are determined experimentally by measuring the transmission through a resist coated on a silicon nitride membrane. Photochemistry can be evaluated by monitoring small outgassing reaction products to provide insight into photoacid generator or metal-based resist reactivity. Spectroscopic techniques such as thin-film Fourier transform infrared (FTIR) spectroscopy can measure the chemical state of a photoresist system pre- and post-EUV exposure. Additionally, electrolysis can be used to study the interaction between photoresist components and low energy electrons. Collectively, these techniques improve our current understanding of photomechanisms for several EUV photoresist systems, which is needed to develop new, better performing materials needed for high volume manufacturing.
Improving EEG-Based Motor Imagery Classification for Real-Time Applications Using the QSA Method.
Batres-Mendoza, Patricia; Ibarra-Manzano, Mario A; Guerra-Hernandez, Erick I; Almanza-Ojeda, Dora L; Montoro-Sanjose, Carlos R; Romero-Troncoso, Rene J; Rostro-Gonzalez, Horacio
2017-01-01
We present an improvement to the quaternion-based signal analysis (QSA) technique to extract electroencephalography (EEG) signal features with a view to developing real-time applications, particularly in motor imagery (IM) cognitive processes. The proposed methodology (iQSA, improved QSA) extracts features such as the average, variance, homogeneity, and contrast of EEG signals related to motor imagery in a more efficient manner (i.e., by reducing the number of samples needed to classify the signal and improving the classification percentage) compared to the original QSA technique. Specifically, we can sample the signal in variable time periods (from 0.5 s to 3 s, in half-a-second intervals) to determine the relationship between the number of samples and their effectiveness in classifying signals. In addition, to strengthen the classification process a number of boosting-technique-based decision trees were implemented. The results show an 82.30% accuracy rate for 0.5 s samples and 73.16% for 3 s samples. This is a significant improvement compared to the original QSA technique that offered results from 33.31% to 40.82% without sampling window and from 33.44% to 41.07% with sampling window, respectively. We can thus conclude that iQSA is better suited to develop real-time applications.
Review of the Potential of the Ni/Cu Plating Technique for Crystalline Silicon Solar Cells.
Rehman, Atteq Ur; Lee, Soo Hong
2014-02-18
Developing a better method for the metallization of silicon solar cells is integral part of realizing superior efficiency. Currently, contact realization using screen printing is the leading technology in the silicon based photovoltaic industry, as it is simple and fast. However, the problem with metallization of this kind is that it has a lower aspect ratio and higher contact resistance, which limits solar cell efficiency. The mounting cost of silver pastes and decreasing silicon wafer thicknesses encourages silicon solar cell manufacturers to develop fresh metallization techniques involving a lower quantity of silver usage and not relying pressing process of screen printing. In recent times nickel/copper (Ni/Cu) based metal plating has emerged as a metallization method that may solve these issues. This paper offers a detailed review and understanding of a Ni/Cu based plating technique for silicon solar cells. The formation of a Ni seed layer by adopting various deposition techniques and a Cu conducting layer using a light induced plating (LIP) process are appraised. Unlike screen-printed metallization, a step involving patterning is crucial for opening the masking layer. Consequently, experimental procedures involving patterning methods are also explicated. Lastly, the issues of adhesion, back ground plating, process complexity and reliability for industrial applications are also addressed.
Improving EEG-Based Motor Imagery Classification for Real-Time Applications Using the QSA Method
Batres-Mendoza, Patricia; Guerra-Hernandez, Erick I.; Almanza-Ojeda, Dora L.; Montoro-Sanjose, Carlos R.
2017-01-01
We present an improvement to the quaternion-based signal analysis (QSA) technique to extract electroencephalography (EEG) signal features with a view to developing real-time applications, particularly in motor imagery (IM) cognitive processes. The proposed methodology (iQSA, improved QSA) extracts features such as the average, variance, homogeneity, and contrast of EEG signals related to motor imagery in a more efficient manner (i.e., by reducing the number of samples needed to classify the signal and improving the classification percentage) compared to the original QSA technique. Specifically, we can sample the signal in variable time periods (from 0.5 s to 3 s, in half-a-second intervals) to determine the relationship between the number of samples and their effectiveness in classifying signals. In addition, to strengthen the classification process a number of boosting-technique-based decision trees were implemented. The results show an 82.30% accuracy rate for 0.5 s samples and 73.16% for 3 s samples. This is a significant improvement compared to the original QSA technique that offered results from 33.31% to 40.82% without sampling window and from 33.44% to 41.07% with sampling window, respectively. We can thus conclude that iQSA is better suited to develop real-time applications. PMID:29348744
A framework for graph-based synthesis, analysis, and visualization of HPC cluster job data.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Mayo, Jackson R.; Kegelmeyer, W. Philip, Jr.; Wong, Matthew H.
The monitoring and system analysis of high performance computing (HPC) clusters is of increasing importance to the HPC community. Analysis of HPC job data can be used to characterize system usage and diagnose and examine failure modes and their effects. This analysis is not straightforward, however, due to the complex relationships that exist between jobs. These relationships are based on a number of factors, including shared compute nodes between jobs, proximity of jobs in time, etc. Graph-based techniques represent an approach that is particularly well suited to this problem, and provide an effective technique for discovering important relationships in jobmore » queuing and execution data. The efficacy of these techniques is rooted in the use of a semantic graph as a knowledge representation tool. In a semantic graph job data, represented in a combination of numerical and textual forms, can be flexibly processed into edges, with corresponding weights, expressing relationships between jobs, nodes, users, and other relevant entities. This graph-based representation permits formal manipulation by a number of analysis algorithms. This report presents a methodology and software implementation that leverages semantic graph-based techniques for the system-level monitoring and analysis of HPC clusters based on job queuing and execution data. Ontology development and graph synthesis is discussed with respect to the domain of HPC job data. The framework developed automates the synthesis of graphs from a database of job information. It also provides a front end, enabling visualization of the synthesized graphs. Additionally, an analysis engine is incorporated that provides performance analysis, graph-based clustering, and failure prediction capabilities for HPC systems.« less
Eksin, Ece; Zor, Erhan; Erdem, Arzum; Bingol, Haluk
2017-06-15
Recently, the low-cost effective biosensing systems based on advanced nanomaterials have received a key attention for development of novel assays for rapid and sequence-specific nucleic acid detection. The electrochemical biosensor based on reduced graphene oxide (rGO) modified disposable pencil graphite electrodes (PGEs) were developed herein for electrochemical monitoring of DNA, and also for monitoring of biointeraction occurred between anticancer drug, Daunorubicin (DNR), and DNA. First, rGO was synthesized chemically and characterized by using UV-Vis, TGA, FT-IR, Raman Spectroscopy and SEM techniques. Then, the quantity of rGO assembling onto the surface of PGE by passive adsorption was optimized. The electrochemical behavior of rGO-PGEs was examined by cyclic voltammetry (CV). rGO-PGEs were then utilized for electrochemical monitoring of surface-confined interaction between DNR and DNA using differential pulse voltammetry (DPV) technique. Additionally, voltammetric results were complemented with electrochemical impedance spectroscopy (EIS) technique. Electrochemical monitoring of DNR and DNA was resulted with satisfying detection limits 0.55µM and 2.71µg/mL, respectively. Copyright © 2017 Elsevier B.V. All rights reserved.
Demirel, Gokhan; Babur, Esra
2014-05-21
Given their simplicity and functionality, paper-based microfluidic systems are considered to be ideal and promising bioassay platforms for use in less developed countries or in point-of-care services. Although a series of innovative techniques have recently been demonstrated for the fabrication of such platforms, development of simple, inexpensive and versatile new strategies are still needed in order to reach their full potential. In this communication, we describe a simple yet facile approach to fabricate paper-based sensor platforms with a desired design through a vapor-phase polymer deposition technique. We also show that the fabricated platforms could be readily employed for the detection of various biological target molecules including glucose, protein, ALP, ALT, and uric acid. The limit of detection for each target molecule was calculated to be 25 mg dL(-1) for glucose, 1.04 g L(-1) for protein, 7.81 unit per L for ALP, 1.6 nmol L(-1) for ALT, and 0.13 mmol L(-1) for uric acid.
Williamson, J; Ranyard, R; Cuthbert, L
2000-05-01
This study is an evaluation of a process tracing method developed for naturalistic decisions, in this case a consumer choice task. The method is based on Huber et al.'s (1997) Active Information Search (AIS) technique, but develops it by providing spoken rather than written answers to respondents' questions, and by including think aloud instructions. The technique is used within a conversation-based situation, rather than the respondent thinking aloud 'into an empty space', as is conventionally the case in think aloud techniques. The method results in a concurrent verbal protocol as respondents make their decisions, and a retrospective report in the form of a post-decision summary. The method was found to be virtually non-reactive in relation to think aloud, although the variable of Preliminary Attribute Elicitation showed some evidence of reactivity. This was a methodological evaluation, and as such the data reported are essentially descriptive. Nevertheless, the data obtained indicate that the method is capable of producing information about decision processes which could have theoretical importance in terms of evaluating models of decision-making.
Application of AI techniques to infer vegetation characteristics from directional reflectance(s)
NASA Technical Reports Server (NTRS)
Kimes, D. S.; Smith, J. A.; Harrison, P. A.; Harrison, P. R.
1994-01-01
Traditionally, the remote sensing community has relied totally on spectral knowledge to extract vegetation characteristics. However, there are other knowledge bases (KB's) that can be used to significantly improve the accuracy and robustness of inference techniques. Using AI (artificial intelligence) techniques a KB system (VEG) was developed that integrates input spectral measurements with diverse KB's. These KB's consist of data sets of directional reflectance measurements, knowledge from literature, and knowledge from experts which are combined into an intelligent and efficient system for making vegetation inferences. VEG accepts spectral data of an unknown target as input, determines the best techniques for inferring the desired vegetation characteristic(s), applies the techniques to the target data, and provides a rigorous estimate of the accuracy of the inference. VEG was developed to: infer spectral hemispherical reflectance from any combination of nadir and/or off-nadir view angles; infer percent ground cover from any combination of nadir and/or off-nadir view angles; infer unknown view angle(s) from known view angle(s) (known as view angle extension); and discriminate between user defined vegetation classes using spectral and directional reflectance relationships developed from an automated learning algorithm. The errors for these techniques were generally very good ranging between 2 to 15% (proportional root mean square). The system is designed to aid scientists in developing, testing, and applying new inference techniques using directional reflectance data.
The study on network security based on software engineering
NASA Astrophysics Data System (ADS)
Jia, Shande; Ao, Qian
2012-04-01
Developing a SP is a sensitive task because the SP itself can lead to security weaknesses if it is not conform to the security properties. Hence, appropriate techniques are necessary to overcome such problems. These techniques must accompany the policy throughout its deployment phases. The main contribution of this paper is then, the proposition of three of these activities: validation, test and multi-SP conflict management. Our techniques are inspired by the well established techniques of the software engineering for which we have found some similarities with the security domain.
Intelligent Traffic Quantification System
NASA Astrophysics Data System (ADS)
Mohanty, Anita; Bhanja, Urmila; Mahapatra, Sudipta
2017-08-01
Currently, city traffic monitoring and controlling is a big issue in almost all cities worldwide. Vehicular ad-hoc Network (VANET) technique is an efficient tool to minimize this problem. Usually, different types of on board sensors are installed in vehicles to generate messages characterized by different vehicle parameters. In this work, an intelligent system based on fuzzy clustering technique is developed to reduce the number of individual messages by extracting important features from the messages of a vehicle. Therefore, the proposed fuzzy clustering technique reduces the traffic load of the network. The technique also reduces congestion and quantifies congestion.
Gene therapy progress and prospects: magnetic nanoparticle-based gene delivery.
Dobson, J
2006-02-01
The recent emphasis on the development of non-viral transfection agents for gene delivery has led to new physics and chemistry-based techniques, which take advantage of charge interactions and energetic processes. One of these techniques which shows much promise for both in vitro and in vivo transfection involves the use of biocompatible magnetic nanoparticles for gene delivery. In these systems, therapeutic or reporter genes are attached to magnetic nanoparticles, which are then focused to the target site/cells via high-field/high-gradient magnets. The technique promotes rapid transfection and, as more recent work indicates, excellent overall transfection levels as well. The advantages and difficulties associated with magnetic nanoparticle-based transfection will be discussed as will the underlying physical principles, recent studies and potential future applications.
Narumi, Ryohei; Tomonaga, Takeshi
2016-01-01
Mass spectrometry-based phosphoproteomics is an indispensible technique used in the discovery and quantification of phosphorylation events on proteins in biological samples. The application of this technique to tissue samples is especially useful for the discovery of biomarkers as well as biological studies. We herein describe the application of a large-scale phosphoproteome analysis and SRM/MRM-based quantitation to develop a strategy for the systematic discovery and validation of biomarkers using tissue samples.
Securing information display by use of visual cryptography.
Yamamoto, Hirotsugu; Hayasaki, Yoshio; Nishida, Nobuo
2003-09-01
We propose a secure display technique based on visual cryptography. The proposed technique ensures the security of visual information. The display employs a decoding mask based on visual cryptography. Without the decoding mask, the displayed information cannot be viewed. The viewing zone is limited by the decoding mask so that only one person can view the information. We have developed a set of encryption codes to maintain the designed viewing zone and have demonstrated a display that provides a limited viewing zone.
Upper atmosphere research: Reaction rate and optical measurements
NASA Technical Reports Server (NTRS)
Stief, L. J.; Allen, J. E., Jr.; Nava, D. F.; Payne, W. A., Jr.
1990-01-01
The objective is to provide photochemical, kinetic, and spectroscopic information necessary for photochemical models of the Earth's upper atmosphere and to examine reactions or reactants not presently in the models to either confirm the correctness of their exclusion or provide evidence to justify future inclusion in the models. New initiatives are being taken in technique development (many of them laser based) and in the application of established techniques to address gaps in the photochemical/kinetic data base, as well as to provide increasingly reliable information.
Multi-Step Deep Reactive Ion Etching Fabrication Process for Silicon-Based Terahertz Components
NASA Technical Reports Server (NTRS)
Reck, Theodore (Inventor); Perez, Jose Vicente Siles (Inventor); Lee, Choonsup (Inventor); Cooper, Ken B. (Inventor); Jung-Kubiak, Cecile (Inventor); Mehdi, Imran (Inventor); Chattopadhyay, Goutam (Inventor); Lin, Robert H. (Inventor); Peralta, Alejandro (Inventor)
2016-01-01
A multi-step silicon etching process has been developed to fabricate silicon-based terahertz (THz) waveguide components. This technique provides precise dimensional control across multiple etch depths with batch processing capabilities. Nonlinear and passive components such as mixers and multipliers waveguides, hybrids, OMTs and twists have been fabricated and integrated into a small silicon package. This fabrication technique enables a wafer-stacking architecture to provide ultra-compact multi-pixel receiver front-ends in the THz range.
NASA Astrophysics Data System (ADS)
Ge, Wangyao
Thin film deposition techniques are indispensable to the development of modern technologies as thin film based optical coatings, optoelectronic devices, sensors, and biological implants are the building blocks of many complicated technologies, and their performance heavily depends on the applied deposition technique. Particularly, the emergence of novel solution-processed materials, such as soft organic molecules, inorganic compounds and colloidal nanoparticles, facilitates the development of flexible and printed electronics that are inexpensive, light weight, green and smart, and these thin film devices represent future trends for new technologies. One appealing feature of solution-processed materials is that they can be deposited into thin films using solution-processed deposition techniques that are straightforward, inexpensive, high throughput and advantageous to industrialize thin film based devices. However, solution-processed techniques rely on wet deposition, which has limitations in certain applications, such as multi-layered film deposition of similar materials and blended film deposition of dissimilar materials. These limitations cannot be addressed by traditional, vacuum-based deposition techniques because these dry approaches are often too energetic and can degrade soft materials, such as polymers, such that the performance of resulting thin film based devices is compromised. The work presented in this dissertation explores a novel thin film deposition technique, namely emulsion-based, resonant infrared, matrix-assisted pulsed laser evaporation (RIR-MAPLE), which combines characteristics of wet and dry deposition techniques for solution-processed materials. Previous studies have demonstrated the feasibility of emulsion-based RIR-MAPLE to deposit uniform and continuous organic, nanoparticle and blended films, as well as hetero-structures that otherwise are difficult to achieve. However, fundamental understanding of the growth mechanisms that govern emulsion-based RIR-MAPLE is still missing, which increases the difficulty of using rational design to improve the performance of initial RIR-MAPLE devices that have been demonstrated. As a result, it is important to study the fundamentals of emulsion-based RIR-MAPLE in order to provide insight into the long-term prospects for this thin film deposition technique. This dissertation explores the fundamental deposition mechanisms of emulsion-based RIR-MAPLE by considering the effects of the emulsion target composition (namely, the primary solvent, secondary solvent, and surfactant) on the properties of deposited polymer films. The study of primary solvent effects on hydrophobic polymer deposition helps identify the unique method of film formation for emulsion-based RIR-MAPLE, which can be described as cluster-by-cluster deposition of emulsified particles that yields two levels of ordering (i.e., within the clusters and among the clusters). The generality of this film formation mechanism is tested by applying the lessons learned to hydrophilic polymer deposition. Based on these studies, the deposition design rules to achieve smooth polymer films, which are important for different device applications, are identified according to the properties of the polymer. After discussion of the fundamental deposition mechanisms, three applications of emulsion-based RIR-MAPLE, namely thin film deposition of organic solar cells, polymer/nanoparticle hybrid solar cells, and antimicrobial/fouling-release multifunctional films, are studied. The work on organic solar cells identifies the ideal deposition mode for blended films with nanoscale domain sizes, as well as demonstrates the relationships among emulsion target composition, film properties, and corresponding device performance. The studies of polymer/nanoparticle hybrid solar cells demonstrate precise control of colloidal nanoparticle deposition, in which the integrity of nanoparticles is maintained and a distinct film morphology is achieved when co-deposited with polymers. Finally, the application of antimicrobial and fouling-release multifunctional films demonstrates the importance of blended film deposition with nanoscale phase separation, a key feature to achieving reusable bio-films that can kill bacteria when illuminated with ultraviolet light. Thus, this dissertation provides great insight to the fundamentals of emulsion-based RIR-MAPLE, serves as a valuable reference for future development, and paves the pathway for wider adoption of this unique thin film deposition technique, especially for organic solar cells.
Developing a Virtual Physics World
ERIC Educational Resources Information Center
Wegener, Margaret; McIntyre, Timothy J.; McGrath, Dominic; Savage, Craig M.; Williamson, Michael
2012-01-01
In this article, the successful implementation of a development cycle for a physics teaching package based on game-like virtual reality software is reported. The cycle involved several iterations of evaluating students' use of the package followed by instructional and software development. The evaluation used a variety of techniques, including…
A probability-based approach for assessment of roadway safety hardware.
DOT National Transportation Integrated Search
2017-03-14
This report presents a general probability-based approach for assessment of roadway safety hardware (RSH). It was achieved using a reliability : analysis method and computational techniques. With the development of high-fidelity finite element (FE) m...
Electronic nanobiosensors based on two-dimensional materials
NASA Astrophysics Data System (ADS)
Ping, Jinglei
Atomically-thick two-dimensional (2D) nanomaterials have tremendous potential to be applied as transduction elements in biosensors and bioelectronics. We developed scalable methods for synthesis and large-area transfer of two-dimensional nanomaterials, particularly graphene and metal dichalcogenides (so called ``MX2'' materials). We also developed versatile fabrication methods for large arrays of field-effect transistors (FETs) and micro-electrodes with these nanomaterials based on either conventional photolithography or innovative approaches that minimize contamination of the 2D layer. By functionalizing the FETs with a computationally redesigned water-soluble mu-opioid receptor, we created selective and sensitive biosensors suitable for detection of the drug target naltrexone and the neuropeptide enkephalin at pg/mL concentrations. We also constructed DNA-functionalized biosensors and nano-particle decorated biosensors by applying related bio-nano integration techniques. Our methodology paves the way for multiplexed nanosensor arrays with all-electronic readout suitable for inexpensive point-of-care diagnostics, drug-development and biomedical research. With graphene field-effect transistors, we investigated the graphene/solution interface and developed a quantitative model for the effect of ionic screening on the graphene carrier density based on theories of the electric double layer. Finally, we have developed a technique for measuring low-level Faradaic charge-transfer current (fA) across the graphene/solution interface via real-time charge monitoring of graphene microelectrodes in ionic solution. This technique enables the development of flexible and transparent pH sensors that are promising for in vivo applications. The author acknowledges the support from the Defense Advanced Research Projects Agency (DARPA) and the U. S. Army Research Office under Grant Number W911NF1010093.
Ultrasonic sensor based defect detection and characterisation of ceramics.
Kesharaju, Manasa; Nagarajah, Romesh; Zhang, Tonzhua; Crouch, Ian
2014-01-01
Ceramic tiles, used in body armour systems, are currently inspected visually offline using an X-ray technique that is both time consuming and very expensive. The aim of this research is to develop a methodology to detect, locate and classify various manufacturing defects in Reaction Sintered Silicon Carbide (RSSC) ceramic tiles, using an ultrasonic sensing technique. Defects such as free silicon, un-sintered silicon carbide material and conventional porosity are often difficult to detect using conventional X-radiography. An alternative inspection system was developed to detect defects in ceramic components using an Artificial Neural Network (ANN) based signal processing technique. The inspection methodology proposed focuses on pre-processing of signals, de-noising, wavelet decomposition, feature extraction and post-processing of the signals for classification purposes. This research contributes to developing an on-line inspection system that would be far more cost effective than present methods and, moreover, assist manufacturers in checking the location of high density areas, defects and enable real time quality control, including the implementation of accept/reject criteria. Copyright © 2013 Elsevier B.V. All rights reserved.
A high-frequency warm shallow water acoustic communications channel model and measurements.
Chitre, Mandar
2007-11-01
Underwater acoustic communication is a core enabling technology with applications in ocean monitoring using remote sensors and autonomous underwater vehicles. One of the more challenging underwater acoustic communication channels is the medium-range very shallow warm-water channel, common in tropical coastal regions. This channel exhibits two key features-extensive time-varying multipath and high levels of non-Gaussian ambient noise due to snapping shrimp-both of which limit the performance of traditional communication techniques. A good understanding of the communications channel is key to the design of communication systems. It aids in the development of signal processing techniques as well as in the testing of the techniques via simulation. In this article, a physics-based channel model for the very shallow warm-water acoustic channel at high frequencies is developed, which are of interest to medium-range communication system developers. The model is based on ray acoustics and includes time-varying statistical effects as well as non-Gaussian ambient noise statistics observed during channel studies. The model is calibrated and its accuracy validated using measurements made at sea.
NASA Technical Reports Server (NTRS)
Lambert, Winifred C.
2000-01-01
This report describes the outcome of Phase 1 of the AMU's Improved Anvil Forecasting task. Forecasters in the 45th Weather Squadron and the Spaceflight Meteorology Group have found that anvil forecasting is a difficult task when predicting LCC and FR violations. The purpose of this task is to determine the technical feasibility of creating an anvil-forecasting tool. Work on this study was separated into three steps: literature search, forecaster discussions, and determination of technical feasibility. The literature search revealed no existing anvil-forecasting techniques. However, there appears to be growing interest in anvils in recent years. If this interest continues to grow, more information will be available to aid in developing a reliable anvil-forecasting tool. The forecaster discussion step revealed an array of methods on how better forecasting techniques could be developed. The forecasters have ideas based on sound meteorological principles and personal experience in forecasting and analyzing anvils. Based on the information gathered in the discussions with the forecasters, the conclusion of this report is that it is technically feasible at this time to develop an anvil forecasting technique that will significantly contribute to the confidence in anvil forecasts.
Application of additive laser technologies in the gas turbine blades design process
NASA Astrophysics Data System (ADS)
Shevchenko, I. V.; Rogalev, A. N.; Osipov, S. K.; Bychkov, N. M.; Komarov, I. I.
2017-11-01
An emergence of modern innovative technologies requires delivering new and modernization existing design and production processes. It is especially relevant for designing the high-temperature turbines of gas turbine engines, development of which is characterized by a transition to higher parameters of working medium in order to improve their efficient performance. A design technique for gas turbine blades based on predictive verification of thermal and hydraulic models of their cooling systems by testing of a blade prototype fabricated using the selective laser melting technology was presented in this article. Technique was proven at the time of development of the first stage blade cooling system for the high-pressure turbine. An experimental procedure for verification of a thermal model of the blades with convective cooling systems based on the comparison of heat-flux density obtained from the numerical simulation data and results of tests in a liquid-metal thermostat was developed. The techniques makes it possible to obtain an experimentally tested blade version and to exclude its experimental adjustment after the start of mass production.